Initialize CBDD solution and add a .NET-focused gitignore for generated artifacts.
This commit is contained in:
80
tests/CBDD.Tests.Benchmark/BenchmarkTransactionHolder.cs
Normal file
80
tests/CBDD.Tests.Benchmark/BenchmarkTransactionHolder.cs
Normal file
@@ -0,0 +1,80 @@
|
||||
using ZB.MOM.WW.CBDD.Core;
|
||||
using ZB.MOM.WW.CBDD.Core.Storage;
|
||||
using ZB.MOM.WW.CBDD.Core.Transactions;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests.Benchmark;
|
||||
|
||||
internal sealed class BenchmarkTransactionHolder : ITransactionHolder, IDisposable
|
||||
{
|
||||
private readonly StorageEngine _storage;
|
||||
private readonly object _sync = new();
|
||||
private ITransaction? _currentTransaction;
|
||||
|
||||
public BenchmarkTransactionHolder(StorageEngine storage)
|
||||
{
|
||||
_storage = storage ?? throw new ArgumentNullException(nameof(storage));
|
||||
}
|
||||
|
||||
public ITransaction GetCurrentTransactionOrStart()
|
||||
{
|
||||
lock (_sync)
|
||||
{
|
||||
if (_currentTransaction == null || _currentTransaction.State != TransactionState.Active)
|
||||
{
|
||||
_currentTransaction = _storage.BeginTransaction();
|
||||
}
|
||||
|
||||
return _currentTransaction;
|
||||
}
|
||||
}
|
||||
|
||||
public Task<ITransaction> GetCurrentTransactionOrStartAsync()
|
||||
{
|
||||
return Task.FromResult(GetCurrentTransactionOrStart());
|
||||
}
|
||||
|
||||
public void CommitAndReset()
|
||||
{
|
||||
lock (_sync)
|
||||
{
|
||||
if (_currentTransaction == null)
|
||||
{
|
||||
return;
|
||||
}
|
||||
|
||||
if (_currentTransaction.State == TransactionState.Active ||
|
||||
_currentTransaction.State == TransactionState.Preparing)
|
||||
{
|
||||
_currentTransaction.Commit();
|
||||
}
|
||||
|
||||
_currentTransaction.Dispose();
|
||||
_currentTransaction = null;
|
||||
}
|
||||
}
|
||||
|
||||
public void RollbackAndReset()
|
||||
{
|
||||
lock (_sync)
|
||||
{
|
||||
if (_currentTransaction == null)
|
||||
{
|
||||
return;
|
||||
}
|
||||
|
||||
if (_currentTransaction.State == TransactionState.Active ||
|
||||
_currentTransaction.State == TransactionState.Preparing)
|
||||
{
|
||||
_currentTransaction.Rollback();
|
||||
}
|
||||
|
||||
_currentTransaction.Dispose();
|
||||
_currentTransaction = null;
|
||||
}
|
||||
}
|
||||
|
||||
public void Dispose()
|
||||
{
|
||||
RollbackAndReset();
|
||||
}
|
||||
}
|
||||
133
tests/CBDD.Tests.Benchmark/CompactionBenchmarks.cs
Normal file
133
tests/CBDD.Tests.Benchmark/CompactionBenchmarks.cs
Normal file
@@ -0,0 +1,133 @@
|
||||
using BenchmarkDotNet.Attributes;
|
||||
using BenchmarkDotNet.Configs;
|
||||
using BenchmarkDotNet.Jobs;
|
||||
using ZB.MOM.WW.CBDD.Bson;
|
||||
using ZB.MOM.WW.CBDD.Core.Collections;
|
||||
using ZB.MOM.WW.CBDD.Core.Storage;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests.Benchmark;
|
||||
|
||||
[SimpleJob]
|
||||
[InProcess]
|
||||
[MemoryDiagnoser]
|
||||
[GroupBenchmarksBy(BenchmarkLogicalGroupRule.ByCategory)]
|
||||
[HtmlExporter]
|
||||
[JsonExporterAttribute.Full]
|
||||
public class CompactionBenchmarks
|
||||
{
|
||||
[Params(2_000)]
|
||||
public int DocumentCount { get; set; }
|
||||
|
||||
private string _dbPath = string.Empty;
|
||||
private string _walPath = string.Empty;
|
||||
private StorageEngine _storage = null!;
|
||||
private BenchmarkTransactionHolder _transactionHolder = null!;
|
||||
private DocumentCollection<Person> _collection = null!;
|
||||
private List<ObjectId> _insertedIds = [];
|
||||
|
||||
[IterationSetup]
|
||||
public void Setup()
|
||||
{
|
||||
var id = Guid.NewGuid().ToString("N");
|
||||
_dbPath = Path.Combine(AppContext.BaseDirectory, $"bench_compaction_{id}.db");
|
||||
_walPath = Path.ChangeExtension(_dbPath, ".wal");
|
||||
|
||||
_storage = new StorageEngine(_dbPath, PageFileConfig.Small);
|
||||
_transactionHolder = new BenchmarkTransactionHolder(_storage);
|
||||
_collection = new DocumentCollection<Person>(_storage, _transactionHolder, new PersonMapper());
|
||||
|
||||
_insertedIds.Clear();
|
||||
for (var i = 0; i < DocumentCount; i++)
|
||||
{
|
||||
var person = CreatePerson(i);
|
||||
var insertedId = _collection.Insert(person);
|
||||
_insertedIds.Add(insertedId);
|
||||
}
|
||||
|
||||
_transactionHolder.CommitAndReset();
|
||||
_storage.Checkpoint();
|
||||
|
||||
for (var i = _insertedIds.Count - 1; i >= _insertedIds.Count / 3; i--)
|
||||
{
|
||||
_collection.Delete(_insertedIds[i]);
|
||||
}
|
||||
|
||||
_transactionHolder.CommitAndReset();
|
||||
_storage.Checkpoint();
|
||||
}
|
||||
|
||||
[IterationCleanup]
|
||||
public void Cleanup()
|
||||
{
|
||||
_transactionHolder?.Dispose();
|
||||
_storage?.Dispose();
|
||||
|
||||
if (File.Exists(_dbPath)) File.Delete(_dbPath);
|
||||
if (File.Exists(_walPath)) File.Delete(_walPath);
|
||||
}
|
||||
|
||||
[Benchmark(Baseline = true)]
|
||||
[BenchmarkCategory("Compaction_Offline")]
|
||||
public long OfflineCompact_ReclaimedBytes()
|
||||
{
|
||||
var stats = _storage.Compact(new CompactionOptions
|
||||
{
|
||||
OnlineMode = false,
|
||||
EnableTailTruncation = true,
|
||||
DefragmentSlottedPages = true,
|
||||
NormalizeFreeList = true
|
||||
});
|
||||
|
||||
return stats.ReclaimedFileBytes;
|
||||
}
|
||||
|
||||
[Benchmark]
|
||||
[BenchmarkCategory("Compaction_Offline")]
|
||||
public long OfflineCompact_TailBytesTruncated()
|
||||
{
|
||||
var stats = _storage.Compact(new CompactionOptions
|
||||
{
|
||||
OnlineMode = false,
|
||||
EnableTailTruncation = true,
|
||||
DefragmentSlottedPages = true,
|
||||
NormalizeFreeList = true
|
||||
});
|
||||
|
||||
return stats.TailBytesTruncated;
|
||||
}
|
||||
|
||||
private static Person CreatePerson(int i)
|
||||
{
|
||||
return new Person
|
||||
{
|
||||
Id = ObjectId.NewObjectId(),
|
||||
FirstName = $"First_{i}",
|
||||
LastName = $"Last_{i}",
|
||||
Age = i % 90,
|
||||
Bio = BuildPayload(i),
|
||||
CreatedAt = DateTime.UnixEpoch.AddSeconds(i),
|
||||
Balance = i * 1.5m,
|
||||
HomeAddress = new Address
|
||||
{
|
||||
Street = $"{i} Long Street",
|
||||
City = "Compaction City",
|
||||
ZipCode = "90000"
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
private static string BuildPayload(int seed)
|
||||
{
|
||||
var builder = new System.Text.StringBuilder(2500);
|
||||
for (var i = 0; i < 80; i++)
|
||||
{
|
||||
builder.Append("compact-");
|
||||
builder.Append(seed.ToString("D6"));
|
||||
builder.Append('-');
|
||||
builder.Append(i.ToString("D3"));
|
||||
builder.Append('|');
|
||||
}
|
||||
|
||||
return builder.ToString();
|
||||
}
|
||||
}
|
||||
174
tests/CBDD.Tests.Benchmark/CompressionBenchmarks.cs
Normal file
174
tests/CBDD.Tests.Benchmark/CompressionBenchmarks.cs
Normal file
@@ -0,0 +1,174 @@
|
||||
using BenchmarkDotNet.Attributes;
|
||||
using BenchmarkDotNet.Configs;
|
||||
using BenchmarkDotNet.Jobs;
|
||||
using System.IO.Compression;
|
||||
using ZB.MOM.WW.CBDD.Bson;
|
||||
using ZB.MOM.WW.CBDD.Core.Collections;
|
||||
using ZB.MOM.WW.CBDD.Core.Compression;
|
||||
using ZB.MOM.WW.CBDD.Core.Storage;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests.Benchmark;
|
||||
|
||||
[SimpleJob]
|
||||
[InProcess]
|
||||
[MemoryDiagnoser]
|
||||
[GroupBenchmarksBy(BenchmarkLogicalGroupRule.ByCategory)]
|
||||
[HtmlExporter]
|
||||
[JsonExporterAttribute.Full]
|
||||
public class CompressionBenchmarks
|
||||
{
|
||||
private const int SeedCount = 300;
|
||||
private const int WorkloadCount = 100;
|
||||
|
||||
[Params(false, true)]
|
||||
public bool EnableCompression { get; set; }
|
||||
|
||||
[Params(CompressionCodec.Brotli, CompressionCodec.Deflate)]
|
||||
public CompressionCodec Codec { get; set; }
|
||||
|
||||
[Params(CompressionLevel.Fastest, CompressionLevel.Optimal)]
|
||||
public CompressionLevel Level { get; set; }
|
||||
|
||||
private string _dbPath = string.Empty;
|
||||
private string _walPath = string.Empty;
|
||||
private StorageEngine _storage = null!;
|
||||
private BenchmarkTransactionHolder _transactionHolder = null!;
|
||||
private DocumentCollection<Person> _collection = null!;
|
||||
|
||||
private Person[] _insertBatch = Array.Empty<Person>();
|
||||
private ObjectId[] _seedIds = Array.Empty<ObjectId>();
|
||||
|
||||
[IterationSetup]
|
||||
public void Setup()
|
||||
{
|
||||
var id = Guid.NewGuid().ToString("N");
|
||||
_dbPath = Path.Combine(AppContext.BaseDirectory, $"bench_compression_{id}.db");
|
||||
_walPath = Path.ChangeExtension(_dbPath, ".wal");
|
||||
|
||||
var compressionOptions = new CompressionOptions
|
||||
{
|
||||
EnableCompression = EnableCompression,
|
||||
MinSizeBytes = 256,
|
||||
MinSavingsPercent = 0,
|
||||
Codec = Codec,
|
||||
Level = Level
|
||||
};
|
||||
|
||||
_storage = new StorageEngine(_dbPath, PageFileConfig.Default, compressionOptions);
|
||||
_transactionHolder = new BenchmarkTransactionHolder(_storage);
|
||||
_collection = new DocumentCollection<Person>(_storage, _transactionHolder, new PersonMapper());
|
||||
|
||||
_seedIds = new ObjectId[SeedCount];
|
||||
for (var i = 0; i < SeedCount; i++)
|
||||
{
|
||||
var doc = CreatePerson(i, includeLargeBio: true);
|
||||
_seedIds[i] = _collection.Insert(doc);
|
||||
}
|
||||
|
||||
_transactionHolder.CommitAndReset();
|
||||
|
||||
_insertBatch = Enumerable.Range(SeedCount, WorkloadCount)
|
||||
.Select(i => CreatePerson(i, includeLargeBio: true))
|
||||
.ToArray();
|
||||
}
|
||||
|
||||
[IterationCleanup]
|
||||
public void Cleanup()
|
||||
{
|
||||
_transactionHolder?.Dispose();
|
||||
_storage?.Dispose();
|
||||
|
||||
if (File.Exists(_dbPath)) File.Delete(_dbPath);
|
||||
if (File.Exists(_walPath)) File.Delete(_walPath);
|
||||
}
|
||||
|
||||
[Benchmark(Baseline = true)]
|
||||
[BenchmarkCategory("Compression_InsertUpdateRead")]
|
||||
public void Insert_Workload()
|
||||
{
|
||||
_collection.InsertBulk(_insertBatch);
|
||||
_transactionHolder.CommitAndReset();
|
||||
}
|
||||
|
||||
[Benchmark]
|
||||
[BenchmarkCategory("Compression_InsertUpdateRead")]
|
||||
public void Update_Workload()
|
||||
{
|
||||
for (var i = 0; i < WorkloadCount; i++)
|
||||
{
|
||||
var id = _seedIds[i];
|
||||
var current = _collection.FindById(id);
|
||||
if (current == null)
|
||||
continue;
|
||||
|
||||
current.Bio = BuildBio(i + 10_000);
|
||||
current.Age += 1;
|
||||
_collection.Update(current);
|
||||
}
|
||||
|
||||
_transactionHolder.CommitAndReset();
|
||||
}
|
||||
|
||||
[Benchmark]
|
||||
[BenchmarkCategory("Compression_InsertUpdateRead")]
|
||||
public int Read_Workload()
|
||||
{
|
||||
var checksum = 0;
|
||||
for (var i = 0; i < WorkloadCount; i++)
|
||||
{
|
||||
var person = _collection.FindById(_seedIds[i]);
|
||||
if (person != null)
|
||||
{
|
||||
checksum += person.Age;
|
||||
}
|
||||
}
|
||||
|
||||
_transactionHolder.CommitAndReset();
|
||||
return checksum;
|
||||
}
|
||||
|
||||
private static Person CreatePerson(int i, bool includeLargeBio)
|
||||
{
|
||||
return new Person
|
||||
{
|
||||
Id = ObjectId.NewObjectId(),
|
||||
FirstName = $"First_{i}",
|
||||
LastName = $"Last_{i}",
|
||||
Age = 20 + (i % 50),
|
||||
Bio = includeLargeBio ? BuildBio(i) : $"bio-{i}",
|
||||
CreatedAt = DateTime.UnixEpoch.AddMinutes(i),
|
||||
Balance = 100 + i,
|
||||
HomeAddress = new Address
|
||||
{
|
||||
Street = $"{i} Main St",
|
||||
City = "Bench City",
|
||||
ZipCode = "12345"
|
||||
},
|
||||
EmploymentHistory =
|
||||
[
|
||||
new WorkHistory
|
||||
{
|
||||
CompanyName = $"Company_{i}",
|
||||
Title = "Engineer",
|
||||
DurationYears = i % 10,
|
||||
Tags = ["csharp", "db", "compression"]
|
||||
}
|
||||
]
|
||||
};
|
||||
}
|
||||
|
||||
private static string BuildBio(int seed)
|
||||
{
|
||||
var builder = new System.Text.StringBuilder(4500);
|
||||
for (var i = 0; i < 150; i++)
|
||||
{
|
||||
builder.Append("bio-");
|
||||
builder.Append(seed.ToString("D6"));
|
||||
builder.Append('-');
|
||||
builder.Append(i.ToString("D3"));
|
||||
builder.Append('|');
|
||||
}
|
||||
|
||||
return builder.ToString();
|
||||
}
|
||||
}
|
||||
266
tests/CBDD.Tests.Benchmark/DatabaseSizeBenchmark.cs
Normal file
266
tests/CBDD.Tests.Benchmark/DatabaseSizeBenchmark.cs
Normal file
@@ -0,0 +1,266 @@
|
||||
using System.Diagnostics;
|
||||
using Microsoft.Extensions.Logging;
|
||||
using Serilog.Context;
|
||||
using ZB.MOM.WW.CBDD.Bson;
|
||||
using ZB.MOM.WW.CBDD.Core.Collections;
|
||||
using ZB.MOM.WW.CBDD.Core.Compression;
|
||||
using ZB.MOM.WW.CBDD.Core.Storage;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests.Benchmark;
|
||||
|
||||
internal static class DatabaseSizeBenchmark
|
||||
{
|
||||
private static readonly int[] TargetCounts = [10_000, 1_000_000, 10_000_000];
|
||||
private static readonly Scenario[] Scenarios =
|
||||
[
|
||||
new(
|
||||
"Uncompressed",
|
||||
CompressionOptions.Default),
|
||||
new(
|
||||
"Compressed-BrotliFast",
|
||||
new CompressionOptions
|
||||
{
|
||||
EnableCompression = true,
|
||||
MinSizeBytes = 256,
|
||||
MinSavingsPercent = 0,
|
||||
Codec = CompressionCodec.Brotli,
|
||||
Level = System.IO.Compression.CompressionLevel.Fastest
|
||||
})
|
||||
];
|
||||
|
||||
private const int BatchSize = 50_000;
|
||||
private const int ProgressInterval = 1_000_000;
|
||||
|
||||
public static void Run(ILogger logger)
|
||||
{
|
||||
var results = new List<SizeResult>(TargetCounts.Length * Scenarios.Length);
|
||||
|
||||
logger.LogInformation("=== CBDD Database Size Benchmark ===");
|
||||
logger.LogInformation("Targets: {Targets}", string.Join(", ", TargetCounts.Select(x => x.ToString("N0"))));
|
||||
logger.LogInformation("Scenarios: {Scenarios}", string.Join(", ", Scenarios.Select(x => x.Name)));
|
||||
logger.LogInformation("Batch size: {BatchSize:N0}", BatchSize);
|
||||
|
||||
foreach (var targetCount in TargetCounts)
|
||||
{
|
||||
foreach (var scenario in Scenarios)
|
||||
{
|
||||
var dbPath = Path.Combine(Path.GetTempPath(), $"cbdd_size_{scenario.Name}_{targetCount}_{Guid.NewGuid():N}.db");
|
||||
var walPath = Path.ChangeExtension(dbPath, ".wal");
|
||||
using var _ = LogContext.PushProperty("TargetCount", targetCount);
|
||||
using var __ = LogContext.PushProperty("Scenario", scenario.Name);
|
||||
|
||||
logger.LogInformation("Starting scenario {Scenario} for target {TargetCount:N0} docs", scenario.Name, targetCount);
|
||||
|
||||
var insertStopwatch = Stopwatch.StartNew();
|
||||
CompressionStats compressionStats = default;
|
||||
CompactionStats compactionStats;
|
||||
long preCompactDbBytes;
|
||||
long preCompactWalBytes;
|
||||
long postCompactDbBytes;
|
||||
long postCompactWalBytes;
|
||||
|
||||
using (var storage = new StorageEngine(dbPath, PageFileConfig.Default, scenario.CompressionOptions))
|
||||
using (var transactionHolder = new BenchmarkTransactionHolder(storage))
|
||||
{
|
||||
var collection = new DocumentCollection<SizeBenchmarkDocument>(
|
||||
storage,
|
||||
transactionHolder,
|
||||
new SizeBenchmarkDocumentMapper());
|
||||
|
||||
var inserted = 0;
|
||||
while (inserted < targetCount)
|
||||
{
|
||||
var currentBatchSize = Math.Min(BatchSize, targetCount - inserted);
|
||||
var documents = new SizeBenchmarkDocument[currentBatchSize];
|
||||
var baseValue = inserted;
|
||||
|
||||
for (var i = 0; i < currentBatchSize; i++)
|
||||
{
|
||||
documents[i] = CreateDocument(baseValue + i);
|
||||
}
|
||||
|
||||
collection.InsertBulk(documents);
|
||||
transactionHolder.CommitAndReset();
|
||||
|
||||
inserted += currentBatchSize;
|
||||
if (inserted == targetCount || inserted % ProgressInterval == 0)
|
||||
{
|
||||
logger.LogInformation("Inserted {Inserted:N0}/{TargetCount:N0}", inserted, targetCount);
|
||||
}
|
||||
}
|
||||
|
||||
insertStopwatch.Stop();
|
||||
preCompactDbBytes = File.Exists(dbPath) ? new FileInfo(dbPath).Length : 0;
|
||||
preCompactWalBytes = File.Exists(walPath) ? new FileInfo(walPath).Length : 0;
|
||||
|
||||
compactionStats = storage.Compact(new CompactionOptions
|
||||
{
|
||||
EnableTailTruncation = true,
|
||||
DefragmentSlottedPages = true,
|
||||
NormalizeFreeList = true
|
||||
});
|
||||
|
||||
postCompactDbBytes = File.Exists(dbPath) ? new FileInfo(dbPath).Length : 0;
|
||||
postCompactWalBytes = File.Exists(walPath) ? new FileInfo(walPath).Length : 0;
|
||||
compressionStats = storage.GetCompressionStats();
|
||||
}
|
||||
|
||||
var result = new SizeResult(
|
||||
scenario.Name,
|
||||
targetCount,
|
||||
insertStopwatch.Elapsed,
|
||||
preCompactDbBytes,
|
||||
preCompactWalBytes,
|
||||
postCompactDbBytes,
|
||||
postCompactWalBytes,
|
||||
compactionStats,
|
||||
compressionStats);
|
||||
results.Add(result);
|
||||
|
||||
logger.LogInformation(
|
||||
"Completed {Scenario} {TargetCount:N0} docs in {Elapsed}. pre={PreTotal}, post={PostTotal}, shrink={Shrink}, compRatio={CompRatio}",
|
||||
scenario.Name,
|
||||
targetCount,
|
||||
insertStopwatch.Elapsed,
|
||||
FormatBytes(result.PreCompactTotalBytes),
|
||||
FormatBytes(result.PostCompactTotalBytes),
|
||||
FormatBytes(result.ShrinkBytes),
|
||||
result.CompressionRatioText);
|
||||
|
||||
TryDelete(dbPath);
|
||||
TryDelete(walPath);
|
||||
}
|
||||
}
|
||||
|
||||
logger.LogInformation("=== Size Benchmark Summary ===");
|
||||
foreach (var result in results.OrderBy(x => x.TargetCount).ThenBy(x => x.Scenario))
|
||||
{
|
||||
logger.LogInformation(
|
||||
"{Scenario,-22} | {Count,12:N0} docs | insert={Elapsed,12} | pre={Pre,12} | post={Post,12} | shrink={Shrink,12} | compact={CompactBytes,12} | ratio={Ratio}",
|
||||
result.Scenario,
|
||||
result.TargetCount,
|
||||
result.InsertElapsed,
|
||||
FormatBytes(result.PreCompactTotalBytes),
|
||||
FormatBytes(result.PostCompactTotalBytes),
|
||||
FormatBytes(result.ShrinkBytes),
|
||||
FormatBytes(result.CompactionStats.ReclaimedFileBytes),
|
||||
result.CompressionRatioText);
|
||||
}
|
||||
}
|
||||
|
||||
private static SizeBenchmarkDocument CreateDocument(int value)
|
||||
{
|
||||
return new SizeBenchmarkDocument
|
||||
{
|
||||
Id = ObjectId.NewObjectId(),
|
||||
Value = value,
|
||||
Name = $"doc-{value:D8}"
|
||||
};
|
||||
}
|
||||
|
||||
private static void TryDelete(string path)
|
||||
{
|
||||
if (File.Exists(path))
|
||||
{
|
||||
File.Delete(path);
|
||||
}
|
||||
}
|
||||
|
||||
private static string FormatBytes(long bytes)
|
||||
{
|
||||
string[] units = ["B", "KB", "MB", "GB", "TB"];
|
||||
double size = bytes;
|
||||
var unitIndex = 0;
|
||||
|
||||
while (size >= 1024 && unitIndex < units.Length - 1)
|
||||
{
|
||||
size /= 1024;
|
||||
unitIndex++;
|
||||
}
|
||||
|
||||
return $"{size:N2} {units[unitIndex]}";
|
||||
}
|
||||
|
||||
private sealed record Scenario(string Name, CompressionOptions CompressionOptions);
|
||||
|
||||
private sealed record SizeResult(
|
||||
string Scenario,
|
||||
int TargetCount,
|
||||
TimeSpan InsertElapsed,
|
||||
long PreCompactDbBytes,
|
||||
long PreCompactWalBytes,
|
||||
long PostCompactDbBytes,
|
||||
long PostCompactWalBytes,
|
||||
CompactionStats CompactionStats,
|
||||
CompressionStats CompressionStats)
|
||||
{
|
||||
public long PreCompactTotalBytes => PreCompactDbBytes + PreCompactWalBytes;
|
||||
public long PostCompactTotalBytes => PostCompactDbBytes + PostCompactWalBytes;
|
||||
public long ShrinkBytes => PreCompactTotalBytes - PostCompactTotalBytes;
|
||||
|
||||
public string CompressionRatioText =>
|
||||
CompressionStats.BytesAfterCompression > 0
|
||||
? $"{(double)CompressionStats.BytesBeforeCompression / CompressionStats.BytesAfterCompression:N2}x"
|
||||
: "n/a";
|
||||
}
|
||||
|
||||
private sealed class SizeBenchmarkDocument
|
||||
{
|
||||
public ObjectId Id { get; set; }
|
||||
public int Value { get; set; }
|
||||
public string Name { get; set; } = string.Empty;
|
||||
}
|
||||
|
||||
private sealed class SizeBenchmarkDocumentMapper : ObjectIdMapperBase<SizeBenchmarkDocument>
|
||||
{
|
||||
public override string CollectionName => "size_documents";
|
||||
|
||||
public override ObjectId GetId(SizeBenchmarkDocument entity) => entity.Id;
|
||||
|
||||
public override void SetId(SizeBenchmarkDocument entity, ObjectId id) => entity.Id = id;
|
||||
|
||||
public override int Serialize(SizeBenchmarkDocument entity, BsonSpanWriter writer)
|
||||
{
|
||||
var sizePos = writer.BeginDocument();
|
||||
writer.WriteObjectId("_id", entity.Id);
|
||||
writer.WriteInt32("value", entity.Value);
|
||||
writer.WriteString("name", entity.Name);
|
||||
writer.EndDocument(sizePos);
|
||||
return writer.Position;
|
||||
}
|
||||
|
||||
public override SizeBenchmarkDocument Deserialize(BsonSpanReader reader)
|
||||
{
|
||||
var document = new SizeBenchmarkDocument();
|
||||
|
||||
reader.ReadDocumentSize();
|
||||
while (reader.Remaining > 0)
|
||||
{
|
||||
var bsonType = reader.ReadBsonType();
|
||||
if (bsonType == BsonType.EndOfDocument)
|
||||
{
|
||||
break;
|
||||
}
|
||||
|
||||
var name = reader.ReadElementHeader();
|
||||
switch (name)
|
||||
{
|
||||
case "_id":
|
||||
document.Id = reader.ReadObjectId();
|
||||
break;
|
||||
case "value":
|
||||
document.Value = reader.ReadInt32();
|
||||
break;
|
||||
case "name":
|
||||
document.Name = reader.ReadString();
|
||||
break;
|
||||
default:
|
||||
reader.SkipValue(bsonType);
|
||||
break;
|
||||
}
|
||||
}
|
||||
|
||||
return document;
|
||||
}
|
||||
}
|
||||
}
|
||||
133
tests/CBDD.Tests.Benchmark/InsertBenchmarks.cs
Executable file
133
tests/CBDD.Tests.Benchmark/InsertBenchmarks.cs
Executable file
@@ -0,0 +1,133 @@
|
||||
using BenchmarkDotNet.Attributes;
|
||||
using BenchmarkDotNet.Configs;
|
||||
using BenchmarkDotNet.Jobs;
|
||||
using ZB.MOM.WW.CBDD.Bson;
|
||||
using ZB.MOM.WW.CBDD.Core;
|
||||
using ZB.MOM.WW.CBDD.Core.Collections;
|
||||
using ZB.MOM.WW.CBDD.Core.Storage;
|
||||
using ZB.MOM.WW.CBDD.Core.Transactions;
|
||||
using Microsoft.Extensions.Logging;
|
||||
using Serilog.Context;
|
||||
using System.IO;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests.Benchmark;
|
||||
|
||||
|
||||
[InProcess]
|
||||
[MemoryDiagnoser]
|
||||
[GroupBenchmarksBy(BenchmarkLogicalGroupRule.ByCategory)]
|
||||
[HtmlExporter]
|
||||
[JsonExporterAttribute.Full]
|
||||
public class InsertBenchmarks
|
||||
{
|
||||
private const int BatchSize = 1000;
|
||||
private static readonly ILogger Logger = Logging.CreateLogger<InsertBenchmarks>();
|
||||
|
||||
private string _docDbPath = "";
|
||||
private string _docDbWalPath = "";
|
||||
|
||||
private StorageEngine? _storage = null;
|
||||
private BenchmarkTransactionHolder? _transactionHolder = null;
|
||||
private DocumentCollection<Person>? _collection = null;
|
||||
|
||||
private Person[] _batchData = Array.Empty<Person>();
|
||||
private Person? _singlePerson = null;
|
||||
|
||||
[GlobalSetup]
|
||||
public void Setup()
|
||||
{
|
||||
var temp = AppContext.BaseDirectory;
|
||||
var id = Guid.NewGuid().ToString("N");
|
||||
_docDbPath = Path.Combine(temp, $"bench_docdb_{id}.db");
|
||||
_docDbWalPath = Path.ChangeExtension(_docDbPath, ".wal");
|
||||
|
||||
_singlePerson = CreatePerson(0);
|
||||
_batchData = new Person[BatchSize];
|
||||
for (int i = 0; i < BatchSize; i++)
|
||||
{
|
||||
_batchData[i] = CreatePerson(i);
|
||||
}
|
||||
}
|
||||
|
||||
private Person CreatePerson(int i)
|
||||
{
|
||||
var p = new Person
|
||||
{
|
||||
Id = ObjectId.NewObjectId(),
|
||||
FirstName = $"First_{i}",
|
||||
LastName = $"Last_{i}",
|
||||
Age = 20 + (i % 50),
|
||||
Bio = null, // Removed large payload to focus on structure
|
||||
CreatedAt = DateTime.UtcNow,
|
||||
Balance = 1000.50m * (i + 1),
|
||||
HomeAddress = new Address
|
||||
{
|
||||
Street = $"{i} Main St",
|
||||
City = "Tech City",
|
||||
ZipCode = "12345"
|
||||
}
|
||||
};
|
||||
|
||||
// Add 10 work history items to stress structure traversal
|
||||
for (int j = 0; j < 10; j++)
|
||||
{
|
||||
p.EmploymentHistory.Add(new WorkHistory
|
||||
{
|
||||
CompanyName = $"TechCorp_{i}_{j}",
|
||||
Title = "Developer",
|
||||
DurationYears = j,
|
||||
Tags = new List<string> { "C#", "BSON", "Performance", "Database", "Complex" }
|
||||
});
|
||||
}
|
||||
|
||||
return p;
|
||||
}
|
||||
|
||||
[IterationSetup]
|
||||
public void IterationSetup()
|
||||
{
|
||||
_storage = new StorageEngine(_docDbPath, PageFileConfig.Default);
|
||||
_transactionHolder = new BenchmarkTransactionHolder(_storage);
|
||||
_collection = new DocumentCollection<Person>(_storage, _transactionHolder, new PersonMapper());
|
||||
}
|
||||
|
||||
[IterationCleanup]
|
||||
public void Cleanup()
|
||||
{
|
||||
try
|
||||
{
|
||||
using var _ = LogContext.PushProperty("Benchmark", nameof(InsertBenchmarks));
|
||||
_transactionHolder?.Dispose();
|
||||
_transactionHolder = null;
|
||||
_storage?.Dispose();
|
||||
_storage = null;
|
||||
|
||||
System.Threading.Thread.Sleep(100);
|
||||
|
||||
if (File.Exists(_docDbPath)) File.Delete(_docDbPath);
|
||||
if (File.Exists(_docDbWalPath)) File.Delete(_docDbWalPath);
|
||||
}
|
||||
catch (Exception ex)
|
||||
{
|
||||
Logger.LogWarning(ex, "Cleanup warning");
|
||||
}
|
||||
}
|
||||
|
||||
// --- Benchmarks ---
|
||||
|
||||
[Benchmark(Baseline = true, Description = "CBDD Single Insert")]
|
||||
[BenchmarkCategory("Insert_Single")]
|
||||
public void DocumentDb_Insert_Single()
|
||||
{
|
||||
_collection?.Insert(_singlePerson!);
|
||||
_transactionHolder?.CommitAndReset();
|
||||
}
|
||||
|
||||
[Benchmark(Description = "CBDD Batch Insert (1000 items, 1 Txn)")]
|
||||
[BenchmarkCategory("Insert_Batch")]
|
||||
public void DocumentDb_Insert_Batch()
|
||||
{
|
||||
_collection?.InsertBulk(_batchData);
|
||||
_transactionHolder?.CommitAndReset();
|
||||
}
|
||||
}
|
||||
30
tests/CBDD.Tests.Benchmark/Logging.cs
Normal file
30
tests/CBDD.Tests.Benchmark/Logging.cs
Normal file
@@ -0,0 +1,30 @@
|
||||
using Microsoft.Extensions.Logging;
|
||||
using Serilog;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests.Benchmark;
|
||||
|
||||
internal static class Logging
|
||||
{
|
||||
private static readonly Lazy<ILoggerFactory> LoggerFactoryInstance = new(CreateFactory);
|
||||
|
||||
public static ILoggerFactory LoggerFactory => LoggerFactoryInstance.Value;
|
||||
|
||||
public static Microsoft.Extensions.Logging.ILogger CreateLogger<T>()
|
||||
{
|
||||
return LoggerFactory.CreateLogger<T>();
|
||||
}
|
||||
|
||||
private static ILoggerFactory CreateFactory()
|
||||
{
|
||||
var serilogLogger = new LoggerConfiguration()
|
||||
.Enrich.FromLogContext()
|
||||
.WriteTo.Console()
|
||||
.CreateLogger();
|
||||
|
||||
return Microsoft.Extensions.Logging.LoggerFactory.Create(builder =>
|
||||
{
|
||||
builder.ClearProviders();
|
||||
builder.AddSerilog(serilogLogger, dispose: true);
|
||||
});
|
||||
}
|
||||
}
|
||||
110
tests/CBDD.Tests.Benchmark/ManualBenchmark.cs
Executable file
110
tests/CBDD.Tests.Benchmark/ManualBenchmark.cs
Executable file
@@ -0,0 +1,110 @@
|
||||
using System.Diagnostics;
|
||||
using System.IO;
|
||||
using System.Text;
|
||||
using Microsoft.Extensions.Logging;
|
||||
using Serilog.Context;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests.Benchmark;
|
||||
|
||||
public class ManualBenchmark
|
||||
{
|
||||
private static StringBuilder _log = new();
|
||||
|
||||
private static void Log(ILogger logger, string message = "")
|
||||
{
|
||||
logger.LogInformation("{Message}", message);
|
||||
_log.AppendLine(message);
|
||||
}
|
||||
|
||||
public static void Run(ILogger logger)
|
||||
{
|
||||
using var _ = LogContext.PushProperty("Benchmark", nameof(ManualBenchmark));
|
||||
_log.Clear();
|
||||
Log(logger, "=== MANUAL BENCHMARK: CBDD ===");
|
||||
Log(logger, $"Date: {DateTime.Now}");
|
||||
Log(logger, "Testing: Complex Objects (Nested Documents + Collections)\n");
|
||||
|
||||
long batchInsertMs;
|
||||
long singleInsertMs;
|
||||
long readByIdMs;
|
||||
|
||||
using (LogContext.PushProperty("Phase", "BatchInsert"))
|
||||
{
|
||||
Log(logger, "1. Batch Insert (1000 items)");
|
||||
var insertBench = new InsertBenchmarks();
|
||||
insertBench.Setup();
|
||||
insertBench.IterationSetup();
|
||||
try
|
||||
{
|
||||
var sw = Stopwatch.StartNew();
|
||||
insertBench.DocumentDb_Insert_Batch();
|
||||
sw.Stop();
|
||||
batchInsertMs = sw.ElapsedMilliseconds;
|
||||
Log(logger, $" CBDD InsertBulk (1000): {batchInsertMs} ms");
|
||||
}
|
||||
finally
|
||||
{
|
||||
insertBench.Cleanup();
|
||||
}
|
||||
}
|
||||
|
||||
using (LogContext.PushProperty("Phase", "FindById"))
|
||||
{
|
||||
Log(logger, "\n2. FindById Performance (1000 operations)");
|
||||
var readBench = new ReadBenchmarks();
|
||||
readBench.Setup();
|
||||
try
|
||||
{
|
||||
var sw = Stopwatch.StartNew();
|
||||
for (int i = 0; i < 1000; i++)
|
||||
{
|
||||
readBench.DocumentDb_FindById();
|
||||
}
|
||||
sw.Stop();
|
||||
readByIdMs = sw.ElapsedMilliseconds;
|
||||
Log(logger, $" CBDD FindById x1000: {readByIdMs} ms ({(double)readByIdMs / 1000:F3} ms/op)");
|
||||
}
|
||||
finally
|
||||
{
|
||||
readBench.Cleanup();
|
||||
}
|
||||
}
|
||||
|
||||
using (LogContext.PushProperty("Phase", "SingleInsert"))
|
||||
{
|
||||
Log(logger, "\n3. Single Insert");
|
||||
var insertBench = new InsertBenchmarks();
|
||||
insertBench.Setup();
|
||||
insertBench.IterationSetup();
|
||||
try
|
||||
{
|
||||
var sw = Stopwatch.StartNew();
|
||||
insertBench.DocumentDb_Insert_Single();
|
||||
sw.Stop();
|
||||
singleInsertMs = sw.ElapsedMilliseconds;
|
||||
Log(logger, $" CBDD Single Insert: {singleInsertMs} ms");
|
||||
}
|
||||
finally
|
||||
{
|
||||
insertBench.Cleanup();
|
||||
}
|
||||
}
|
||||
|
||||
Log(logger, "\n============================================================================");
|
||||
Log(logger, "BENCHMARK RESULTS (CBDD ONLY):");
|
||||
Log(logger, "============================================================================");
|
||||
Log(logger, $"Batch Insert (1000): {batchInsertMs} ms");
|
||||
Log(logger, $"FindById x1000: {readByIdMs} ms");
|
||||
Log(logger, $"Single Insert: {singleInsertMs} ms");
|
||||
|
||||
var artifactsDir = Path.Combine(AppContext.BaseDirectory, "BenchmarkDotNet.Artifacts", "results");
|
||||
if (!Directory.Exists(artifactsDir))
|
||||
{
|
||||
Directory.CreateDirectory(artifactsDir);
|
||||
}
|
||||
|
||||
var filePath = Path.Combine(artifactsDir, "manual_report.txt");
|
||||
File.WriteAllText(filePath, _log.ToString());
|
||||
logger.LogInformation("Report saved to: {FilePath}", filePath);
|
||||
}
|
||||
}
|
||||
170
tests/CBDD.Tests.Benchmark/MixedWorkloadBenchmarks.cs
Normal file
170
tests/CBDD.Tests.Benchmark/MixedWorkloadBenchmarks.cs
Normal file
@@ -0,0 +1,170 @@
|
||||
using BenchmarkDotNet.Attributes;
|
||||
using BenchmarkDotNet.Configs;
|
||||
using BenchmarkDotNet.Jobs;
|
||||
using ZB.MOM.WW.CBDD.Bson;
|
||||
using ZB.MOM.WW.CBDD.Core.Collections;
|
||||
using ZB.MOM.WW.CBDD.Core.Compression;
|
||||
using ZB.MOM.WW.CBDD.Core.Storage;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests.Benchmark;
|
||||
|
||||
[SimpleJob]
|
||||
[InProcess]
|
||||
[MemoryDiagnoser]
|
||||
[GroupBenchmarksBy(BenchmarkLogicalGroupRule.ByCategory)]
|
||||
[HtmlExporter]
|
||||
[JsonExporterAttribute.Full]
|
||||
public class MixedWorkloadBenchmarks
|
||||
{
|
||||
[Params(false, true)]
|
||||
public bool PeriodicCompaction { get; set; }
|
||||
|
||||
[Params(800)]
|
||||
public int Operations { get; set; }
|
||||
|
||||
private string _dbPath = string.Empty;
|
||||
private string _walPath = string.Empty;
|
||||
private StorageEngine _storage = null!;
|
||||
private BenchmarkTransactionHolder _transactionHolder = null!;
|
||||
private DocumentCollection<Person> _collection = null!;
|
||||
private readonly List<ObjectId> _activeIds = [];
|
||||
private int _nextValueSeed;
|
||||
|
||||
[IterationSetup]
|
||||
public void Setup()
|
||||
{
|
||||
var id = Guid.NewGuid().ToString("N");
|
||||
_dbPath = Path.Combine(AppContext.BaseDirectory, $"bench_mixed_{id}.db");
|
||||
_walPath = Path.ChangeExtension(_dbPath, ".wal");
|
||||
|
||||
var options = new CompressionOptions
|
||||
{
|
||||
EnableCompression = true,
|
||||
MinSizeBytes = 256,
|
||||
MinSavingsPercent = 0,
|
||||
Codec = CompressionCodec.Brotli
|
||||
};
|
||||
|
||||
_storage = new StorageEngine(_dbPath, PageFileConfig.Default, options);
|
||||
_transactionHolder = new BenchmarkTransactionHolder(_storage);
|
||||
_collection = new DocumentCollection<Person>(_storage, _transactionHolder, new PersonMapper());
|
||||
|
||||
_activeIds.Clear();
|
||||
_nextValueSeed = 0;
|
||||
|
||||
for (var i = 0; i < 300; i++)
|
||||
{
|
||||
var idValue = _collection.Insert(CreatePerson(_nextValueSeed++));
|
||||
_activeIds.Add(idValue);
|
||||
}
|
||||
|
||||
_transactionHolder.CommitAndReset();
|
||||
}
|
||||
|
||||
[IterationCleanup]
|
||||
public void Cleanup()
|
||||
{
|
||||
_transactionHolder?.Dispose();
|
||||
_storage?.Dispose();
|
||||
|
||||
if (File.Exists(_dbPath)) File.Delete(_dbPath);
|
||||
if (File.Exists(_walPath)) File.Delete(_walPath);
|
||||
}
|
||||
|
||||
[Benchmark(Baseline = true)]
|
||||
[BenchmarkCategory("MixedWorkload")]
|
||||
public int InsertUpdateDeleteMix()
|
||||
{
|
||||
var random = new Random(12345);
|
||||
|
||||
for (var i = 1; i <= Operations; i++)
|
||||
{
|
||||
var mode = i % 5;
|
||||
if (mode is 0 or 1)
|
||||
{
|
||||
var id = _collection.Insert(CreatePerson(_nextValueSeed++));
|
||||
_activeIds.Add(id);
|
||||
}
|
||||
else if (mode is 2 or 3)
|
||||
{
|
||||
if (_activeIds.Count > 0)
|
||||
{
|
||||
var idx = random.Next(_activeIds.Count);
|
||||
var id = _activeIds[idx];
|
||||
var current = _collection.FindById(id);
|
||||
if (current != null)
|
||||
{
|
||||
current.Age += 1;
|
||||
current.Bio = BuildPayload(_nextValueSeed++);
|
||||
_collection.Update(current);
|
||||
}
|
||||
}
|
||||
}
|
||||
else
|
||||
{
|
||||
if (_activeIds.Count > 100)
|
||||
{
|
||||
var idx = random.Next(_activeIds.Count);
|
||||
var id = _activeIds[idx];
|
||||
_collection.Delete(id);
|
||||
_activeIds.RemoveAt(idx);
|
||||
}
|
||||
}
|
||||
|
||||
if (i % 50 == 0)
|
||||
{
|
||||
_transactionHolder.CommitAndReset();
|
||||
}
|
||||
|
||||
if (PeriodicCompaction && i % 200 == 0)
|
||||
{
|
||||
_storage.RunOnlineCompactionPass(new CompactionOptions
|
||||
{
|
||||
OnlineMode = true,
|
||||
OnlineBatchPageLimit = 8,
|
||||
OnlineBatchDelay = TimeSpan.FromMilliseconds(1),
|
||||
MaxOnlineDuration = TimeSpan.FromMilliseconds(120),
|
||||
EnableTailTruncation = true
|
||||
});
|
||||
}
|
||||
}
|
||||
|
||||
_transactionHolder.CommitAndReset();
|
||||
return _collection.Count();
|
||||
}
|
||||
|
||||
private static Person CreatePerson(int seed)
|
||||
{
|
||||
return new Person
|
||||
{
|
||||
Id = ObjectId.NewObjectId(),
|
||||
FirstName = $"First_{seed}",
|
||||
LastName = $"Last_{seed}",
|
||||
Age = 18 + (seed % 60),
|
||||
Bio = BuildPayload(seed),
|
||||
CreatedAt = DateTime.UnixEpoch.AddSeconds(seed),
|
||||
Balance = seed,
|
||||
HomeAddress = new Address
|
||||
{
|
||||
Street = $"{seed} Mixed Ave",
|
||||
City = "Workload City",
|
||||
ZipCode = "10101"
|
||||
}
|
||||
};
|
||||
}
|
||||
|
||||
private static string BuildPayload(int seed)
|
||||
{
|
||||
var builder = new System.Text.StringBuilder(1800);
|
||||
for (var i = 0; i < 64; i++)
|
||||
{
|
||||
builder.Append("mixed-");
|
||||
builder.Append(seed.ToString("D6"));
|
||||
builder.Append('-');
|
||||
builder.Append(i.ToString("D3"));
|
||||
builder.Append('|');
|
||||
}
|
||||
|
||||
return builder.ToString();
|
||||
}
|
||||
}
|
||||
35
tests/CBDD.Tests.Benchmark/Person.cs
Executable file
35
tests/CBDD.Tests.Benchmark/Person.cs
Executable file
@@ -0,0 +1,35 @@
|
||||
using ZB.MOM.WW.CBDD.Bson;
|
||||
using System;
|
||||
using System.Collections.Generic;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests.Benchmark;
|
||||
|
||||
public class Address
|
||||
{
|
||||
public string Street { get; set; } = string.Empty;
|
||||
public string City { get; set; } = string.Empty;
|
||||
public string ZipCode { get; set; } = string.Empty;
|
||||
}
|
||||
|
||||
public class WorkHistory
|
||||
{
|
||||
public string CompanyName { get; set; } = string.Empty;
|
||||
public string Title { get; set; } = string.Empty;
|
||||
public int DurationYears { get; set; }
|
||||
public List<string> Tags { get; set; } = new();
|
||||
}
|
||||
|
||||
public class Person
|
||||
{
|
||||
public ObjectId Id { get; set; }
|
||||
public string FirstName { get; set; } = string.Empty;
|
||||
public string LastName { get; set; } = string.Empty;
|
||||
public int Age { get; set; }
|
||||
public string? Bio { get; set; } = string.Empty;
|
||||
public DateTime CreatedAt { get; set; }
|
||||
|
||||
// Complex fields
|
||||
public decimal Balance { get; set; }
|
||||
public Address HomeAddress { get; set; } = new();
|
||||
public List<WorkHistory> EmploymentHistory { get; set; } = new();
|
||||
}
|
||||
161
tests/CBDD.Tests.Benchmark/PersonMapper.cs
Executable file
161
tests/CBDD.Tests.Benchmark/PersonMapper.cs
Executable file
@@ -0,0 +1,161 @@
|
||||
using ZB.MOM.WW.CBDD.Bson;
|
||||
using ZB.MOM.WW.CBDD.Core.Collections;
|
||||
using System.Buffers;
|
||||
using System.Runtime.InteropServices;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests.Benchmark;
|
||||
|
||||
public class PersonMapper : ObjectIdMapperBase<Person>
|
||||
{
|
||||
public override string CollectionName => "people";
|
||||
|
||||
public override ObjectId GetId(Person entity) => entity.Id;
|
||||
|
||||
public override void SetId(Person entity, ObjectId id) => entity.Id = id;
|
||||
|
||||
public override int Serialize(Person entity, BsonSpanWriter writer)
|
||||
{
|
||||
var sizePos = writer.BeginDocument();
|
||||
|
||||
writer.WriteObjectId("_id", entity.Id);
|
||||
writer.WriteString("firstname", entity.FirstName);
|
||||
writer.WriteString("lastname", entity.LastName);
|
||||
writer.WriteInt32("age", entity.Age);
|
||||
if (entity.Bio != null)
|
||||
writer.WriteString("bio", entity.Bio);
|
||||
else
|
||||
writer.WriteNull("bio");
|
||||
|
||||
writer.WriteInt64("createdat", entity.CreatedAt.Ticks);
|
||||
|
||||
// Complex fields
|
||||
writer.WriteDouble("balance", (double)entity.Balance);
|
||||
|
||||
// Nested Object: Address
|
||||
var addrPos = writer.BeginDocument("homeaddress");
|
||||
writer.WriteString("street", entity.HomeAddress.Street);
|
||||
writer.WriteString("city", entity.HomeAddress.City);
|
||||
writer.WriteString("zipcode", entity.HomeAddress.ZipCode);
|
||||
writer.EndDocument(addrPos);
|
||||
|
||||
// Collection: EmploymentHistory
|
||||
var histPos = writer.BeginArray("employmenthistory");
|
||||
for (int i = 0; i < entity.EmploymentHistory.Count; i++)
|
||||
{
|
||||
var item = entity.EmploymentHistory[i];
|
||||
// Array elements are keys "0", "1", "2"...
|
||||
var itemPos = writer.BeginDocument(i.ToString());
|
||||
|
||||
writer.WriteString("companyname", item.CompanyName);
|
||||
writer.WriteString("title", item.Title);
|
||||
writer.WriteInt32("durationyears", item.DurationYears);
|
||||
|
||||
// Nested Collection: Tags
|
||||
var tagsPos = writer.BeginArray("tags");
|
||||
for (int j = 0; j < item.Tags.Count; j++)
|
||||
{
|
||||
writer.WriteString(j.ToString(), item.Tags[j]);
|
||||
}
|
||||
writer.EndArray(tagsPos);
|
||||
|
||||
writer.EndDocument(itemPos);
|
||||
}
|
||||
writer.EndArray(histPos);
|
||||
|
||||
writer.EndDocument(sizePos);
|
||||
|
||||
return writer.Position;
|
||||
}
|
||||
|
||||
public override Person Deserialize(BsonSpanReader reader)
|
||||
{
|
||||
var person = new Person();
|
||||
|
||||
reader.ReadDocumentSize();
|
||||
|
||||
while (reader.Remaining > 0)
|
||||
{
|
||||
var type = reader.ReadBsonType();
|
||||
if (type == BsonType.EndOfDocument)
|
||||
break;
|
||||
|
||||
var name = reader.ReadElementHeader();
|
||||
|
||||
switch (name)
|
||||
{
|
||||
case "_id": person.Id = reader.ReadObjectId(); break;
|
||||
case "firstname": person.FirstName = reader.ReadString(); break;
|
||||
case "lastname": person.LastName = reader.ReadString(); break;
|
||||
case "age": person.Age = reader.ReadInt32(); break;
|
||||
case "bio":
|
||||
if (type == BsonType.Null) person.Bio = null;
|
||||
else person.Bio = reader.ReadString();
|
||||
break;
|
||||
case "createdat": person.CreatedAt = new DateTime(reader.ReadInt64()); break;
|
||||
case "balance": person.Balance = (decimal)reader.ReadDouble(); break;
|
||||
|
||||
case "homeaddress":
|
||||
reader.ReadDocumentSize(); // Enter document
|
||||
while (reader.Remaining > 0)
|
||||
{
|
||||
var addrType = reader.ReadBsonType();
|
||||
if (addrType == BsonType.EndOfDocument) break;
|
||||
var addrName = reader.ReadElementHeader();
|
||||
|
||||
// We assume strict schema for benchmark speed, but should handle skipping
|
||||
if (addrName == "street") person.HomeAddress.Street = reader.ReadString();
|
||||
else if (addrName == "city") person.HomeAddress.City = reader.ReadString();
|
||||
else if (addrName == "zipcode") person.HomeAddress.ZipCode = reader.ReadString();
|
||||
else reader.SkipValue(addrType);
|
||||
}
|
||||
break;
|
||||
|
||||
case "employmenthistory":
|
||||
reader.ReadDocumentSize(); // Enter Array
|
||||
while (reader.Remaining > 0)
|
||||
{
|
||||
var arrType = reader.ReadBsonType();
|
||||
if (arrType == BsonType.EndOfDocument) break;
|
||||
reader.ReadElementHeader(); // Array index "0", "1"... ignore
|
||||
|
||||
// Read WorkHistory item
|
||||
var workItem = new WorkHistory();
|
||||
reader.ReadDocumentSize(); // Enter Item Document
|
||||
while (reader.Remaining > 0)
|
||||
{
|
||||
var itemType = reader.ReadBsonType();
|
||||
if (itemType == BsonType.EndOfDocument) break;
|
||||
var itemName = reader.ReadElementHeader();
|
||||
|
||||
if (itemName == "companyname") workItem.CompanyName = reader.ReadString();
|
||||
else if (itemName == "title") workItem.Title = reader.ReadString();
|
||||
else if (itemName == "durationyears") workItem.DurationYears = reader.ReadInt32();
|
||||
else if (itemName == "tags")
|
||||
{
|
||||
reader.ReadDocumentSize(); // Enter Tags Array
|
||||
while (reader.Remaining > 0)
|
||||
{
|
||||
var tagType = reader.ReadBsonType();
|
||||
if (tagType == BsonType.EndOfDocument) break;
|
||||
reader.ReadElementHeader(); // Index
|
||||
if (tagType == BsonType.String)
|
||||
workItem.Tags.Add(reader.ReadString());
|
||||
else
|
||||
reader.SkipValue(tagType);
|
||||
}
|
||||
}
|
||||
else reader.SkipValue(itemType);
|
||||
}
|
||||
person.EmploymentHistory.Add(workItem);
|
||||
}
|
||||
break;
|
||||
|
||||
default:
|
||||
reader.SkipValue(type);
|
||||
break;
|
||||
}
|
||||
}
|
||||
|
||||
return person;
|
||||
}
|
||||
}
|
||||
82
tests/CBDD.Tests.Benchmark/Program.cs
Executable file
82
tests/CBDD.Tests.Benchmark/Program.cs
Executable file
@@ -0,0 +1,82 @@
|
||||
using BenchmarkDotNet.Columns;
|
||||
using BenchmarkDotNet.Configs;
|
||||
using BenchmarkDotNet.Exporters;
|
||||
using BenchmarkDotNet.Reports;
|
||||
using BenchmarkDotNet.Running;
|
||||
using Microsoft.Extensions.Logging;
|
||||
using Serilog.Context;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests.Benchmark;
|
||||
|
||||
class Program
|
||||
{
|
||||
static void Main(string[] args)
|
||||
{
|
||||
var logger = Logging.CreateLogger<Program>();
|
||||
var mode = args.Length > 0 ? args[0].Trim().ToLowerInvariant() : string.Empty;
|
||||
|
||||
if (mode == "manual")
|
||||
{
|
||||
using var _ = LogContext.PushProperty("Mode", "Manual");
|
||||
ManualBenchmark.Run(logger);
|
||||
return;
|
||||
}
|
||||
|
||||
if (mode == "size")
|
||||
{
|
||||
using var _ = LogContext.PushProperty("Mode", "SizeBenchmark");
|
||||
DatabaseSizeBenchmark.Run(logger);
|
||||
return;
|
||||
}
|
||||
|
||||
if (mode == "compression")
|
||||
{
|
||||
using var _ = LogContext.PushProperty("Mode", "CompressionBenchmarks");
|
||||
BenchmarkRunner.Run<CompressionBenchmarks>(CreateConfig());
|
||||
return;
|
||||
}
|
||||
|
||||
if (mode == "compaction")
|
||||
{
|
||||
using var _ = LogContext.PushProperty("Mode", "CompactionBenchmarks");
|
||||
BenchmarkRunner.Run<CompactionBenchmarks>(CreateConfig());
|
||||
return;
|
||||
}
|
||||
|
||||
if (mode == "mixed")
|
||||
{
|
||||
using var _ = LogContext.PushProperty("Mode", "MixedWorkloadBenchmarks");
|
||||
BenchmarkRunner.Run<MixedWorkloadBenchmarks>(CreateConfig());
|
||||
return;
|
||||
}
|
||||
|
||||
if (mode == "all")
|
||||
{
|
||||
using var _ = LogContext.PushProperty("Mode", "AllBenchmarks");
|
||||
var config = CreateConfig();
|
||||
BenchmarkRunner.Run<InsertBenchmarks>(config);
|
||||
BenchmarkRunner.Run<ReadBenchmarks>(config);
|
||||
BenchmarkRunner.Run<SerializationBenchmarks>(config);
|
||||
BenchmarkRunner.Run<CompressionBenchmarks>(config);
|
||||
BenchmarkRunner.Run<CompactionBenchmarks>(config);
|
||||
BenchmarkRunner.Run<MixedWorkloadBenchmarks>(config);
|
||||
return;
|
||||
}
|
||||
|
||||
using var __ = LogContext.PushProperty("Mode", "BenchmarkDotNet");
|
||||
var defaultConfig = CreateConfig();
|
||||
|
||||
BenchmarkRunner.Run<InsertBenchmarks>(defaultConfig);
|
||||
BenchmarkRunner.Run<ReadBenchmarks>(defaultConfig);
|
||||
BenchmarkRunner.Run<SerializationBenchmarks>(defaultConfig);
|
||||
}
|
||||
|
||||
private static IConfig CreateConfig()
|
||||
{
|
||||
return DefaultConfig.Instance
|
||||
.AddExporter(HtmlExporter.Default)
|
||||
.WithSummaryStyle(SummaryStyle.Default
|
||||
.WithRatioStyle(RatioStyle.Trend)
|
||||
.WithTimeUnit(Perfolizer.Horology.TimeUnit.Microsecond));
|
||||
}
|
||||
}
|
||||
109
tests/CBDD.Tests.Benchmark/ReadBenchmarks.cs
Executable file
109
tests/CBDD.Tests.Benchmark/ReadBenchmarks.cs
Executable file
@@ -0,0 +1,109 @@
|
||||
using BenchmarkDotNet.Attributes;
|
||||
using BenchmarkDotNet.Configs;
|
||||
using BenchmarkDotNet.Jobs;
|
||||
using ZB.MOM.WW.CBDD.Bson;
|
||||
using ZB.MOM.WW.CBDD.Core;
|
||||
using ZB.MOM.WW.CBDD.Core.Collections;
|
||||
using ZB.MOM.WW.CBDD.Core.Storage;
|
||||
using ZB.MOM.WW.CBDD.Core.Transactions;
|
||||
using System.IO;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests.Benchmark;
|
||||
|
||||
[SimpleJob]
|
||||
[InProcess]
|
||||
[MemoryDiagnoser]
|
||||
[GroupBenchmarksBy(BenchmarkLogicalGroupRule.ByCategory)]
|
||||
[HtmlExporter]
|
||||
[JsonExporterAttribute.Full]
|
||||
public class ReadBenchmarks
|
||||
{
|
||||
private const int DocCount = 1000;
|
||||
|
||||
private string _docDbPath = null!;
|
||||
private string _docDbWalPath = null!;
|
||||
|
||||
private StorageEngine _storage = null!;
|
||||
private BenchmarkTransactionHolder _transactionHolder = null!;
|
||||
private DocumentCollection<Person> _collection = null!;
|
||||
|
||||
private ObjectId[] _ids = null!;
|
||||
private ObjectId _targetId;
|
||||
|
||||
[GlobalSetup]
|
||||
public void Setup()
|
||||
{
|
||||
var temp = AppContext.BaseDirectory;
|
||||
var id = Guid.NewGuid().ToString("N");
|
||||
_docDbPath = Path.Combine(temp, $"bench_read_docdb_{id}.db");
|
||||
_docDbWalPath = Path.ChangeExtension(_docDbPath, ".wal");
|
||||
|
||||
if (File.Exists(_docDbPath)) File.Delete(_docDbPath);
|
||||
if (File.Exists(_docDbWalPath)) File.Delete(_docDbWalPath);
|
||||
|
||||
_storage = new StorageEngine(_docDbPath, PageFileConfig.Default);
|
||||
_transactionHolder = new BenchmarkTransactionHolder(_storage);
|
||||
_collection = new DocumentCollection<Person>(_storage, _transactionHolder, new PersonMapper());
|
||||
|
||||
_ids = new ObjectId[DocCount];
|
||||
for (int i = 0; i < DocCount; i++)
|
||||
{
|
||||
var p = CreatePerson(i);
|
||||
_ids[i] = _collection.Insert(p);
|
||||
}
|
||||
_transactionHolder.CommitAndReset();
|
||||
|
||||
_targetId = _ids[DocCount / 2];
|
||||
}
|
||||
|
||||
[GlobalCleanup]
|
||||
public void Cleanup()
|
||||
{
|
||||
_transactionHolder?.Dispose();
|
||||
_storage?.Dispose();
|
||||
|
||||
if (File.Exists(_docDbPath)) File.Delete(_docDbPath);
|
||||
if (File.Exists(_docDbWalPath)) File.Delete(_docDbWalPath);
|
||||
}
|
||||
|
||||
private Person CreatePerson(int i)
|
||||
{
|
||||
var p = new Person
|
||||
{
|
||||
Id = ObjectId.NewObjectId(),
|
||||
FirstName = $"First_{i}",
|
||||
LastName = $"Last_{i}",
|
||||
Age = 20 + (i % 50),
|
||||
Bio = null,
|
||||
CreatedAt = DateTime.UtcNow,
|
||||
Balance = 1000.50m * (i + 1),
|
||||
HomeAddress = new Address
|
||||
{
|
||||
Street = $"{i} Main St",
|
||||
City = "Tech City",
|
||||
ZipCode = "12345"
|
||||
}
|
||||
};
|
||||
|
||||
// Add 10 work history items
|
||||
for (int j = 0; j < 10; j++)
|
||||
{
|
||||
p.EmploymentHistory.Add(new WorkHistory
|
||||
{
|
||||
CompanyName = $"TechCorp_{i}_{j}",
|
||||
Title = "Developer",
|
||||
DurationYears = j,
|
||||
Tags = new List<string> { "C#", "BSON", "Performance", "Database", "Complex" }
|
||||
});
|
||||
}
|
||||
|
||||
return p;
|
||||
}
|
||||
|
||||
[Benchmark(Baseline = true, Description = "CBDD FindById")]
|
||||
[BenchmarkCategory("Read_Single")]
|
||||
public Person? DocumentDb_FindById()
|
||||
{
|
||||
return _collection.FindById(_targetId);
|
||||
}
|
||||
}
|
||||
182
tests/CBDD.Tests.Benchmark/SerializationBenchmarks.cs
Executable file
182
tests/CBDD.Tests.Benchmark/SerializationBenchmarks.cs
Executable file
@@ -0,0 +1,182 @@
|
||||
using BenchmarkDotNet.Attributes;
|
||||
using BenchmarkDotNet.Configs;
|
||||
using ZB.MOM.WW.CBDD.Bson;
|
||||
using System.Text.Json;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests.Benchmark;
|
||||
|
||||
[InProcess]
|
||||
[MemoryDiagnoser]
|
||||
[GroupBenchmarksBy(BenchmarkLogicalGroupRule.ByCategory)]
|
||||
[HtmlExporter]
|
||||
[JsonExporterAttribute.Full]
|
||||
public class SerializationBenchmarks
|
||||
{
|
||||
private const int BatchSize = 10000;
|
||||
private Person _person = null!;
|
||||
private List<Person> _people = null!;
|
||||
private PersonMapper _mapper = new PersonMapper();
|
||||
private byte[] _bsonData = Array.Empty<byte>();
|
||||
private byte[] _jsonData = Array.Empty<byte>();
|
||||
|
||||
private List<byte[]> _bsonDataList = new();
|
||||
private List<byte[]> _jsonDataList = new();
|
||||
|
||||
private byte[] _serializeBuffer = Array.Empty<byte>();
|
||||
|
||||
private static readonly System.Collections.Concurrent.ConcurrentDictionary<string, ushort> _keyMap = new(StringComparer.OrdinalIgnoreCase);
|
||||
private static readonly System.Collections.Concurrent.ConcurrentDictionary<ushort, string> _keys = new();
|
||||
|
||||
static SerializationBenchmarks()
|
||||
{
|
||||
ushort id = 1;
|
||||
string[] initialKeys = { "_id", "firstname", "lastname", "age", "bio", "createdat", "balance", "homeaddress", "street", "city", "zipcode", "employmenthistory", "companyname", "title", "durationyears", "tags" };
|
||||
foreach (var key in initialKeys)
|
||||
{
|
||||
_keyMap[key] = id;
|
||||
_keys[id] = key;
|
||||
id++;
|
||||
}
|
||||
// Add some indices for arrays
|
||||
for (int i = 0; i < 100; i++)
|
||||
{
|
||||
var s = i.ToString();
|
||||
_keyMap[s] = id;
|
||||
_keys[id] = s;
|
||||
id++;
|
||||
}
|
||||
}
|
||||
|
||||
[GlobalSetup]
|
||||
public void Setup()
|
||||
{
|
||||
_person = CreatePerson(0);
|
||||
_people = new List<Person>(BatchSize);
|
||||
for (int i = 0; i < BatchSize; i++)
|
||||
{
|
||||
_people.Add(CreatePerson(i));
|
||||
}
|
||||
|
||||
// Pre-allocate buffer for BSON serialization
|
||||
_serializeBuffer = new byte[8192];
|
||||
|
||||
var writer = new BsonSpanWriter(_serializeBuffer, _keyMap);
|
||||
|
||||
// Single item data
|
||||
var len = _mapper.Serialize(_person, writer);
|
||||
_bsonData = _serializeBuffer.AsSpan(0, len).ToArray();
|
||||
_jsonData = JsonSerializer.SerializeToUtf8Bytes(_person);
|
||||
|
||||
// List data
|
||||
foreach (var p in _people)
|
||||
{
|
||||
len = _mapper.Serialize(p, writer);
|
||||
_bsonDataList.Add(_serializeBuffer.AsSpan(0, len).ToArray());
|
||||
_jsonDataList.Add(JsonSerializer.SerializeToUtf8Bytes(p));
|
||||
}
|
||||
}
|
||||
|
||||
private Person CreatePerson(int i)
|
||||
{
|
||||
var p = new Person
|
||||
{
|
||||
Id = ObjectId.NewObjectId(),
|
||||
FirstName = $"First_{i}",
|
||||
LastName = $"Last_{i}",
|
||||
Age = 25,
|
||||
Bio = null,
|
||||
CreatedAt = DateTime.UtcNow,
|
||||
Balance = 1000.50m,
|
||||
HomeAddress = new Address
|
||||
{
|
||||
Street = $"{i} Main St",
|
||||
City = "Tech City",
|
||||
ZipCode = "12345"
|
||||
}
|
||||
};
|
||||
|
||||
for (int j = 0; j < 10; j++)
|
||||
{
|
||||
p.EmploymentHistory.Add(new WorkHistory
|
||||
{
|
||||
CompanyName = $"TechCorp_{i}_{j}",
|
||||
Title = "Developer",
|
||||
DurationYears = j,
|
||||
Tags = new List<string> { "C#", "BSON", "Performance", "Database", "Complex" }
|
||||
});
|
||||
}
|
||||
return p;
|
||||
}
|
||||
|
||||
[Benchmark(Description = "Serialize Single (BSON)")]
|
||||
[BenchmarkCategory("Single")]
|
||||
public void Serialize_Bson()
|
||||
{
|
||||
var writer = new BsonSpanWriter(_serializeBuffer, _keyMap);
|
||||
_mapper.Serialize(_person, writer);
|
||||
}
|
||||
|
||||
[Benchmark(Description = "Serialize Single (JSON)")]
|
||||
[BenchmarkCategory("Single")]
|
||||
public void Serialize_Json()
|
||||
{
|
||||
JsonSerializer.SerializeToUtf8Bytes(_person);
|
||||
}
|
||||
|
||||
[Benchmark(Description = "Deserialize Single (BSON)")]
|
||||
[BenchmarkCategory("Single")]
|
||||
public Person Deserialize_Bson()
|
||||
{
|
||||
var reader = new BsonSpanReader(_bsonData, _keys);
|
||||
return _mapper.Deserialize(reader);
|
||||
}
|
||||
|
||||
[Benchmark(Description = "Deserialize Single (JSON)")]
|
||||
[BenchmarkCategory("Single")]
|
||||
public Person? Deserialize_Json()
|
||||
{
|
||||
return JsonSerializer.Deserialize<Person>(_jsonData);
|
||||
}
|
||||
|
||||
[Benchmark(Description = "Serialize List 10k (BSON loop)")]
|
||||
[BenchmarkCategory("Batch")]
|
||||
public void Serialize_List_Bson()
|
||||
{
|
||||
foreach (var p in _people)
|
||||
{
|
||||
var writer = new BsonSpanWriter(_serializeBuffer, _keyMap);
|
||||
_mapper.Serialize(p, writer);
|
||||
}
|
||||
}
|
||||
|
||||
[Benchmark(Description = "Serialize List 10k (JSON loop)")]
|
||||
[BenchmarkCategory("Batch")]
|
||||
public void Serialize_List_Json()
|
||||
{
|
||||
foreach (var p in _people)
|
||||
{
|
||||
JsonSerializer.SerializeToUtf8Bytes(p);
|
||||
}
|
||||
}
|
||||
|
||||
[Benchmark(Description = "Deserialize List 10k (BSON loop)")]
|
||||
[BenchmarkCategory("Batch")]
|
||||
public void Deserialize_List_Bson()
|
||||
{
|
||||
foreach (var data in _bsonDataList)
|
||||
{
|
||||
var reader = new BsonSpanReader(data, _keys);
|
||||
_mapper.Deserialize(reader);
|
||||
}
|
||||
}
|
||||
|
||||
[Benchmark(Description = "Deserialize List 10k (JSON loop)")]
|
||||
[BenchmarkCategory("Batch")]
|
||||
public void Deserialize_List_Json()
|
||||
{
|
||||
foreach (var data in _jsonDataList)
|
||||
{
|
||||
JsonSerializer.Deserialize<Person>(data);
|
||||
}
|
||||
}
|
||||
}
|
||||
@@ -0,0 +1,25 @@
|
||||
<Project Sdk="Microsoft.NET.Sdk">
|
||||
|
||||
<PropertyGroup>
|
||||
<OutputType>Exe</OutputType>
|
||||
<TargetFramework>net10.0</TargetFramework>
|
||||
<AssemblyName>ZB.MOM.WW.CBDD.Tests.Benchmark</AssemblyName>
|
||||
<RootNamespace>ZB.MOM.WW.CBDD.Tests.Benchmark</RootNamespace>
|
||||
<ImplicitUsings>enable</ImplicitUsings>
|
||||
<Nullable>enable</Nullable>
|
||||
<IsPackable>false</IsPackable>
|
||||
</PropertyGroup>
|
||||
|
||||
<ItemGroup>
|
||||
<PackageReference Include="BenchmarkDotNet" Version="0.15.8" />
|
||||
<PackageReference Include="Microsoft.Extensions.Logging" Version="9.0.10" />
|
||||
<PackageReference Include="Serilog" Version="4.2.0" />
|
||||
<PackageReference Include="Serilog.Extensions.Logging" Version="8.0.0" />
|
||||
<PackageReference Include="Serilog.Sinks.Console" Version="6.0.0" />
|
||||
</ItemGroup>
|
||||
|
||||
<ItemGroup>
|
||||
<ProjectReference Include="..\..\src\CBDD.Core\ZB.MOM.WW.CBDD.Core.csproj" />
|
||||
</ItemGroup>
|
||||
|
||||
</Project>
|
||||
234
tests/CBDD.Tests/AdvancedQueryTests.cs
Executable file
234
tests/CBDD.Tests/AdvancedQueryTests.cs
Executable file
@@ -0,0 +1,234 @@
|
||||
using Xunit;
|
||||
using ZB.MOM.WW.CBDD.Core.Collections;
|
||||
using ZB.MOM.WW.CBDD.Bson;
|
||||
using System.Linq;
|
||||
using System.Collections.Generic;
|
||||
using ZB.MOM.WW.CBDD.Core.Indexing;
|
||||
using ZB.MOM.WW.CBDD.Core.Storage;
|
||||
using System;
|
||||
using System.IO;
|
||||
using ZB.MOM.WW.CBDD.Shared;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests
|
||||
{
|
||||
public class AdvancedQueryTests : IDisposable
|
||||
{
|
||||
private readonly string _dbPath;
|
||||
private readonly Shared.TestDbContext _db;
|
||||
|
||||
public AdvancedQueryTests()
|
||||
{
|
||||
_dbPath = Path.Combine(Path.GetTempPath(), $"cbdd_advanced_{Guid.NewGuid()}.db");
|
||||
_db = new Shared.TestDbContext(_dbPath);
|
||||
|
||||
// Seed Data
|
||||
_db.TestDocuments.Insert(new TestDocument { Category = "A", Amount = 10, Name = "Item1" });
|
||||
_db.TestDocuments.Insert(new TestDocument { Category = "A", Amount = 20, Name = "Item2" });
|
||||
_db.TestDocuments.Insert(new TestDocument { Category = "B", Amount = 30, Name = "Item3" });
|
||||
_db.TestDocuments.Insert(new TestDocument { Category = "B", Amount = 40, Name = "Item4" });
|
||||
_db.TestDocuments.Insert(new TestDocument { Category = "C", Amount = 50, Name = "Item5" });
|
||||
_db.SaveChanges();
|
||||
}
|
||||
|
||||
public void Dispose()
|
||||
{
|
||||
_db.Dispose();
|
||||
if (File.Exists(_dbPath)) File.Delete(_dbPath);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void GroupBy_Simple_Key_Works()
|
||||
{
|
||||
var groups = _db.TestDocuments.AsQueryable()
|
||||
.GroupBy(x => x.Category)
|
||||
.ToList();
|
||||
|
||||
groups.Count.ShouldBe(3);
|
||||
|
||||
var groupA = groups.First(g => g.Key == "A");
|
||||
groupA.Count().ShouldBe(2);
|
||||
groupA.ShouldContain(x => x.Amount == 10);
|
||||
groupA.ShouldContain(x => x.Amount == 20);
|
||||
|
||||
var groupB = groups.First(g => g.Key == "B");
|
||||
groupB.Count().ShouldBe(2);
|
||||
|
||||
var groupC = groups.First(g => g.Key == "C");
|
||||
groupC.Count().ShouldBe(1);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void GroupBy_With_Aggregation_Select()
|
||||
{
|
||||
var results = _db.TestDocuments.AsQueryable()
|
||||
.GroupBy(x => x.Category)
|
||||
.Select(g => new { Category = g.Key, Total = g.Sum(x => x.Amount) })
|
||||
.OrderBy(x => x.Category)
|
||||
.ToList();
|
||||
|
||||
results.Count.ShouldBe(3);
|
||||
results[0].Category.ShouldBe("A");
|
||||
results[0].Total.ShouldBe(30); // 10 + 20
|
||||
|
||||
results[1].Category.ShouldBe("B");
|
||||
results[1].Total.ShouldBe(70); // 30 + 40
|
||||
|
||||
results[2].Category.ShouldBe("C");
|
||||
results[2].Total.ShouldBe(50); // 50
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Aggregations_Direct_Works()
|
||||
{
|
||||
var query = _db.TestDocuments.AsQueryable();
|
||||
|
||||
query.Count().ShouldBe(5);
|
||||
query.Sum(x => x.Amount).ShouldBe(150);
|
||||
query.Average(x => x.Amount).ShouldBe(30.0);
|
||||
query.Min(x => x.Amount).ShouldBe(10);
|
||||
query.Max(x => x.Amount).ShouldBe(50);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Aggregations_With_Predicate_Works()
|
||||
{
|
||||
var query = _db.TestDocuments.AsQueryable().Where(x => x.Category == "A");
|
||||
|
||||
query.Count().ShouldBe(2);
|
||||
query.Sum(x => x.Amount).ShouldBe(30);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Join_Works_InMemory()
|
||||
{
|
||||
// Create a second collection for joining
|
||||
_db.OrderDocuments.Insert(new OrderDocument { ItemName = "Item1", Quantity = 5 });
|
||||
_db.OrderDocuments.Insert(new OrderDocument { ItemName = "Item3", Quantity = 2 });
|
||||
_db.SaveChanges();
|
||||
|
||||
var query = _db.TestDocuments.AsQueryable()
|
||||
.Join(_db.OrderDocuments.AsQueryable(),
|
||||
doc => doc.Name,
|
||||
order => order.ItemName,
|
||||
(doc, order) => new { doc.Name, doc.Category, order.Quantity })
|
||||
.OrderBy(x => x.Name)
|
||||
.ToList();
|
||||
|
||||
query.Count.ShouldBe(2);
|
||||
|
||||
query[0].Name.ShouldBe("Item1");
|
||||
query[0].Category.ShouldBe("A");
|
||||
query[0].Quantity.ShouldBe(5);
|
||||
|
||||
query[1].Name.ShouldBe("Item3");
|
||||
query[1].Category.ShouldBe("B");
|
||||
query[1].Quantity.ShouldBe(2);
|
||||
}
|
||||
|
||||
|
||||
[Fact]
|
||||
public void Select_Project_Nested_Object()
|
||||
{
|
||||
var doc = new ComplexDocument
|
||||
{
|
||||
Id = ObjectId.NewObjectId(),
|
||||
Title = "Order1",
|
||||
ShippingAddress = new Address { City = new City { Name = "New York" }, Street = "5th Ave" },
|
||||
Items = new List<OrderItem>
|
||||
{
|
||||
new OrderItem { Name = "Laptop", Price = 1000 },
|
||||
new OrderItem { Name = "Mouse", Price = 50 }
|
||||
}
|
||||
};
|
||||
_db.ComplexDocuments.Insert(doc);
|
||||
_db.SaveChanges();
|
||||
|
||||
var query = _db.ComplexDocuments.AsQueryable()
|
||||
.Select(x => x.ShippingAddress)
|
||||
.ToList();
|
||||
|
||||
query.Count().ShouldBe(1);
|
||||
query[0].City.Name.ShouldBe("New York");
|
||||
query[0].Street.ShouldBe("5th Ave");
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Select_Project_Nested_Field()
|
||||
{
|
||||
var doc = new ComplexDocument
|
||||
{
|
||||
Id = ObjectId.NewObjectId(),
|
||||
Title = "Order1",
|
||||
ShippingAddress = new Address { City = new City { Name = "New York" }, Street = "5th Ave" }
|
||||
};
|
||||
_db.ComplexDocuments.Insert(doc);
|
||||
_db.SaveChanges();
|
||||
|
||||
var cities = _db.ComplexDocuments.AsQueryable()
|
||||
.Select(x => x.ShippingAddress.City.Name)
|
||||
.ToList();
|
||||
|
||||
cities.Count().ShouldBe(1);
|
||||
cities[0].ShouldBe("New York");
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Select_Anonymous_Complex()
|
||||
{
|
||||
ZB.MOM.WW.CBDD.Shared.TestDbContext_TestDbContext_Mappers.ZB_MOM_WW_CBDD_Shared_CityMapper cityMapper = new ZB.MOM.WW.CBDD.Shared.TestDbContext_TestDbContext_Mappers.ZB_MOM_WW_CBDD_Shared_CityMapper();
|
||||
var doc = new ComplexDocument
|
||||
{
|
||||
Id = ObjectId.NewObjectId(),
|
||||
Title = "Order1",
|
||||
ShippingAddress = new Address { City = new City { Name = "New York" }, Street = "5th Ave" }
|
||||
};
|
||||
|
||||
|
||||
_db.ComplexDocuments.Insert(doc);
|
||||
_db.SaveChanges();
|
||||
|
||||
var result = _db.ComplexDocuments.AsQueryable()
|
||||
.Select(x => new { x.Title, x.ShippingAddress.City })
|
||||
.ToList();
|
||||
|
||||
result.Count().ShouldBe(1);
|
||||
result[0].Title.ShouldBe("Order1");
|
||||
result[0].City.Name.ShouldBe("New York");
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Select_Project_Nested_Array_Of_Objects()
|
||||
{
|
||||
var doc = new ComplexDocument
|
||||
{
|
||||
Id = ObjectId.NewObjectId(),
|
||||
Title = "Order with Items",
|
||||
ShippingAddress = new Address { City = new City { Name = "Los Angeles" }, Street = "Hollywood Blvd" },
|
||||
Items = new List<OrderItem>
|
||||
{
|
||||
new OrderItem { Name = "Laptop", Price = 1500 },
|
||||
new OrderItem { Name = "Mouse", Price = 25 },
|
||||
new OrderItem { Name = "Keyboard", Price = 75 }
|
||||
}
|
||||
};
|
||||
_db.ComplexDocuments.Insert(doc);
|
||||
_db.SaveChanges();
|
||||
|
||||
// Retrieve the full document and verify Items array
|
||||
var retrieved = _db.ComplexDocuments.FindAll().First();
|
||||
|
||||
retrieved.Title.ShouldBe("Order with Items");
|
||||
retrieved.ShippingAddress.City.Name.ShouldBe("Los Angeles");
|
||||
retrieved.ShippingAddress.Street.ShouldBe("Hollywood Blvd");
|
||||
|
||||
// Verify array of nested objects
|
||||
retrieved.Items.Count.ShouldBe(3);
|
||||
retrieved.Items[0].Name.ShouldBe("Laptop");
|
||||
retrieved.Items[0].Price.ShouldBe(1500);
|
||||
retrieved.Items[1].Name.ShouldBe("Mouse");
|
||||
retrieved.Items[1].Price.ShouldBe(25);
|
||||
retrieved.Items[2].Name.ShouldBe("Keyboard");
|
||||
retrieved.Items[2].Price.ShouldBe(75);
|
||||
}
|
||||
}
|
||||
}
|
||||
217
tests/CBDD.Tests/ArchitectureFitnessTests.cs
Normal file
217
tests/CBDD.Tests/ArchitectureFitnessTests.cs
Normal file
@@ -0,0 +1,217 @@
|
||||
using System.Reflection;
|
||||
using System.Xml.Linq;
|
||||
using ZB.MOM.WW.CBDD.Bson;
|
||||
using ZB.MOM.WW.CBDD.Core;
|
||||
using ZB.MOM.WW.CBDD.Core.Collections;
|
||||
using ZB.MOM.WW.CBDD.Core.Indexing;
|
||||
using ZB.MOM.WW.CBDD.Core.Storage;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests;
|
||||
|
||||
public class ArchitectureFitnessTests
|
||||
{
|
||||
private const string BsonProject = "src/CBDD.Bson/ZB.MOM.WW.CBDD.Bson.csproj";
|
||||
private const string CoreProject = "src/CBDD.Core/ZB.MOM.WW.CBDD.Core.csproj";
|
||||
private const string SourceGeneratorsProject = "src/CBDD.SourceGenerators/ZB.MOM.WW.CBDD.SourceGenerators.csproj";
|
||||
private const string FacadeProject = "src/CBDD/ZB.MOM.WW.CBDD.csproj";
|
||||
|
||||
[Fact]
|
||||
public void Solution_DependencyGraph_ShouldRemainAcyclic_AndFollowLayerDirection()
|
||||
{
|
||||
var repoRoot = FindRepositoryRoot();
|
||||
var projectGraph = LoadSolutionProjectGraph(repoRoot);
|
||||
|
||||
// Explicit layer rules
|
||||
projectGraph[BsonProject].ShouldBeEmpty();
|
||||
projectGraph[SourceGeneratorsProject].ShouldBeEmpty();
|
||||
projectGraph[CoreProject].ShouldBe(new[] { BsonProject });
|
||||
projectGraph[FacadeProject]
|
||||
.OrderBy(v => v, StringComparer.Ordinal)
|
||||
.ShouldBe(new[] { BsonProject, CoreProject, SourceGeneratorsProject }.OrderBy(v => v, StringComparer.Ordinal));
|
||||
|
||||
// Source projects should not depend on tests.
|
||||
foreach (var kvp in projectGraph.Where(p => p.Key.StartsWith("src/", StringComparison.Ordinal)))
|
||||
{
|
||||
kvp.Value.Any(dep => dep.StartsWith("tests/", StringComparison.Ordinal))
|
||||
.ShouldBeFalse($"{kvp.Key} must not reference test projects.");
|
||||
}
|
||||
|
||||
HasCycle(projectGraph)
|
||||
.ShouldBeFalse("Project references must remain acyclic.");
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void HighLevelCollectionApi_ShouldNotExpandRawBsonReaderWriterSurface()
|
||||
{
|
||||
var lowLevelTypes = new[] { typeof(BsonSpanReader), typeof(BsonSpanWriter) };
|
||||
|
||||
var collectionOffenders = typeof(DocumentCollection<,>)
|
||||
.GetMethods(BindingFlags.Public | BindingFlags.Instance | BindingFlags.Static | BindingFlags.DeclaredOnly)
|
||||
.Where(m => lowLevelTypes.Any(t => MethodUsesType(m, t)))
|
||||
.Select(m => m.Name)
|
||||
.Distinct()
|
||||
.OrderBy(n => n)
|
||||
.ToArray();
|
||||
|
||||
collectionOffenders.ShouldBeEmpty();
|
||||
|
||||
var dbContextOffenders = typeof(DocumentDbContext)
|
||||
.GetMethods(BindingFlags.Public | BindingFlags.Instance | BindingFlags.Static | BindingFlags.DeclaredOnly)
|
||||
.Where(m => lowLevelTypes.Any(t => MethodUsesType(m, t)))
|
||||
.Select(m => m.Name)
|
||||
.Distinct()
|
||||
.ToArray();
|
||||
|
||||
dbContextOffenders.ShouldBeEmpty();
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void CollectionAndIndexOrchestration_ShouldUseStoragePortInternally()
|
||||
{
|
||||
var targetTypes = new[]
|
||||
{
|
||||
typeof(DocumentCollection<>),
|
||||
typeof(DocumentCollection<,>),
|
||||
typeof(BTreeIndex),
|
||||
typeof(CollectionIndexManager<,>),
|
||||
typeof(CollectionSecondaryIndex<,>),
|
||||
typeof(VectorSearchIndex),
|
||||
};
|
||||
|
||||
var fieldOffenders = targetTypes
|
||||
.SelectMany(t => t.GetFields(BindingFlags.Instance | BindingFlags.NonPublic | BindingFlags.Public)
|
||||
.Where(f => f.FieldType == typeof(StorageEngine))
|
||||
.Select(f => $"{t.Name}.{f.Name}"))
|
||||
.OrderBy(v => v)
|
||||
.ToArray();
|
||||
|
||||
fieldOffenders.ShouldBeEmpty("Collection/index orchestration should hold IStorageEngine instead of concrete StorageEngine.");
|
||||
}
|
||||
|
||||
private static Dictionary<string, List<string>> LoadSolutionProjectGraph(string repoRoot)
|
||||
{
|
||||
var solutionPath = Path.Combine(repoRoot, "CBDD.slnx");
|
||||
var solutionDoc = XDocument.Load(solutionPath);
|
||||
|
||||
var projects = solutionDoc
|
||||
.Descendants()
|
||||
.Where(e => e.Name.LocalName == "Project")
|
||||
.Select(e => e.Attribute("Path")?.Value)
|
||||
.Where(p => !string.IsNullOrWhiteSpace(p))
|
||||
.Select(p => NormalizePath(p!))
|
||||
.ToHashSet(StringComparer.Ordinal);
|
||||
|
||||
var graph = projects.ToDictionary(
|
||||
p => p,
|
||||
_ => new List<string>(),
|
||||
StringComparer.Ordinal);
|
||||
|
||||
foreach (var project in projects)
|
||||
{
|
||||
var projectFile = Path.Combine(repoRoot, project);
|
||||
var projectDoc = XDocument.Load(projectFile);
|
||||
var projectDir = Path.GetDirectoryName(projectFile)!;
|
||||
|
||||
var refs = projectDoc
|
||||
.Descendants()
|
||||
.Where(e => e.Name.LocalName == "ProjectReference")
|
||||
.Select(e => e.Attribute("Include")?.Value)
|
||||
.Where(v => !string.IsNullOrWhiteSpace(v))
|
||||
.Select(v => v!.Replace('\\', '/'))
|
||||
.Select(v => NormalizePath(Path.GetRelativePath(repoRoot, Path.GetFullPath(Path.Combine(projectDir, v)))))
|
||||
.Where(projects.Contains)
|
||||
.Distinct(StringComparer.Ordinal)
|
||||
.OrderBy(v => v, StringComparer.Ordinal)
|
||||
.ToList();
|
||||
|
||||
graph[project] = refs;
|
||||
}
|
||||
|
||||
return graph;
|
||||
}
|
||||
|
||||
private static bool HasCycle(Dictionary<string, List<string>> graph)
|
||||
{
|
||||
var state = graph.Keys.ToDictionary(k => k, _ => 0, StringComparer.Ordinal);
|
||||
|
||||
foreach (var node in graph.Keys)
|
||||
{
|
||||
if (state[node] == 0 && Visit(node))
|
||||
{
|
||||
return true;
|
||||
}
|
||||
}
|
||||
|
||||
return false;
|
||||
|
||||
bool Visit(string node)
|
||||
{
|
||||
state[node] = 1; // visiting
|
||||
foreach (var dep in graph[node])
|
||||
{
|
||||
if (state[dep] == 1)
|
||||
{
|
||||
return true;
|
||||
}
|
||||
|
||||
if (state[dep] == 0 && Visit(dep))
|
||||
{
|
||||
return true;
|
||||
}
|
||||
}
|
||||
|
||||
state[node] = 2; // visited
|
||||
return false;
|
||||
}
|
||||
}
|
||||
|
||||
private static bool MethodUsesType(MethodInfo method, Type forbidden)
|
||||
{
|
||||
if (TypeContains(method.ReturnType, forbidden))
|
||||
{
|
||||
return true;
|
||||
}
|
||||
|
||||
return method.GetParameters().Any(p => TypeContains(p.ParameterType, forbidden));
|
||||
}
|
||||
|
||||
private static bool TypeContains(Type inspected, Type forbidden)
|
||||
{
|
||||
if (inspected == forbidden)
|
||||
{
|
||||
return true;
|
||||
}
|
||||
|
||||
if (inspected.HasElementType && inspected.GetElementType() is { } elementType && TypeContains(elementType, forbidden))
|
||||
{
|
||||
return true;
|
||||
}
|
||||
|
||||
if (!inspected.IsGenericType)
|
||||
{
|
||||
return false;
|
||||
}
|
||||
|
||||
return inspected.GetGenericArguments().Any(t => TypeContains(t, forbidden));
|
||||
}
|
||||
|
||||
private static string FindRepositoryRoot()
|
||||
{
|
||||
var current = new DirectoryInfo(AppContext.BaseDirectory);
|
||||
while (current != null)
|
||||
{
|
||||
var solutionPath = Path.Combine(current.FullName, "CBDD.slnx");
|
||||
if (File.Exists(solutionPath))
|
||||
{
|
||||
return current.FullName;
|
||||
}
|
||||
|
||||
current = current.Parent;
|
||||
}
|
||||
|
||||
throw new InvalidOperationException("Unable to find repository root containing CBDD.slnx.");
|
||||
}
|
||||
|
||||
private static string NormalizePath(string path)
|
||||
=> path.Replace('\\', '/');
|
||||
}
|
||||
131
tests/CBDD.Tests/AsyncTests.cs
Executable file
131
tests/CBDD.Tests/AsyncTests.cs
Executable file
@@ -0,0 +1,131 @@
|
||||
using ZB.MOM.WW.CBDD.Core;
|
||||
using ZB.MOM.WW.CBDD.Core.Collections;
|
||||
using ZB.MOM.WW.CBDD.Core.Storage;
|
||||
using ZB.MOM.WW.CBDD.Core.Transactions;
|
||||
using ZB.MOM.WW.CBDD.Shared;
|
||||
using Xunit;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests;
|
||||
|
||||
public class AsyncTests : IDisposable
|
||||
{
|
||||
private readonly string _dbPath;
|
||||
|
||||
public AsyncTests()
|
||||
{
|
||||
_dbPath = Path.Combine(Path.GetTempPath(), $"cbdd_async_{Guid.NewGuid()}.db");
|
||||
}
|
||||
|
||||
public void Dispose()
|
||||
{
|
||||
if (File.Exists(_dbPath)) File.Delete(_dbPath);
|
||||
if (File.Exists(Path.ChangeExtension(_dbPath, ".wal"))) File.Delete(Path.ChangeExtension(_dbPath, ".wal"));
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public async Task Async_Transaction_Commit_Should_Persist_Data()
|
||||
{
|
||||
var ct = TestContext.Current.CancellationToken;
|
||||
|
||||
using (var db = new Shared.TestDbContext(_dbPath))
|
||||
{
|
||||
using (var txn = await db.BeginTransactionAsync(ct))
|
||||
{
|
||||
db.AsyncDocs.Insert(new AsyncDoc { Id = 1, Name = "Async1" });
|
||||
db.AsyncDocs.Insert(new AsyncDoc { Id = 2, Name = "Async2" });
|
||||
await db.SaveChangesAsync(ct);
|
||||
}
|
||||
}
|
||||
|
||||
// Verify with new storage engine instance
|
||||
using var db2 = new Shared.TestDbContext(_dbPath);
|
||||
var doc1 = db2.AsyncDocs.FindById(1);
|
||||
doc1.ShouldNotBeNull();
|
||||
doc1.Name.ShouldBe("Async1");
|
||||
|
||||
var doc2 = db2.AsyncDocs.FindById(2);
|
||||
doc2.ShouldNotBeNull();
|
||||
doc2.Name.ShouldBe("Async2");
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public async Task Async_Transaction_Rollback_Should_Discard_Data()
|
||||
{
|
||||
var ct = TestContext.Current.CancellationToken;
|
||||
|
||||
using var db = new Shared.TestDbContext(_dbPath);
|
||||
using (var txn = await db.BeginTransactionAsync(ct))
|
||||
{
|
||||
db.AsyncDocs.Insert(new AsyncDoc { Id = 3, Name = "RollbackMe" });
|
||||
}
|
||||
|
||||
var doc = db.AsyncDocs.FindById(3);
|
||||
doc.ShouldBeNull();
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public async Task Bulk_Async_Insert_Should_Persist_Data()
|
||||
{
|
||||
using var db = new Shared.TestDbContext(_dbPath);
|
||||
var docs = Enumerable.Range(1, 100).Select(i => new AsyncDoc { Id = i + 5000, Name = $"Bulk{i}" });
|
||||
|
||||
var ids = await db.AsyncDocs.InsertBulkAsync(docs);
|
||||
|
||||
ids.Count.ShouldBe(100);
|
||||
|
||||
var doc50 = db.AsyncDocs.FindById(5050);
|
||||
doc50.ShouldNotBeNull();
|
||||
doc50.Name.ShouldBe("Bulk50");
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public async Task Bulk_Async_Update_Should_Persist_Changes()
|
||||
{
|
||||
using var db = new Shared.TestDbContext(_dbPath);
|
||||
// 1. Insert 100 docs
|
||||
var docs = Enumerable.Range(1, 100).Select(i => new AsyncDoc { Id = i + 6000, Name = $"Original{i}" }).ToList();
|
||||
await db.AsyncDocs.InsertBulkAsync(docs);
|
||||
|
||||
// 2. Update all docs
|
||||
foreach (var doc in docs)
|
||||
{
|
||||
doc.Name = $"Updated{doc.Id - 6000}";
|
||||
}
|
||||
|
||||
var count = await db.AsyncDocs.UpdateBulkAsync(docs);
|
||||
|
||||
count.ShouldBe(100);
|
||||
|
||||
// 3. Verify updates
|
||||
var doc50 = db.AsyncDocs.FindById(6050);
|
||||
doc50.ShouldNotBeNull();
|
||||
doc50.Name.ShouldBe("Updated50");
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public async Task High_Concurrency_Async_Commits()
|
||||
{
|
||||
var ct = TestContext.Current.CancellationToken;
|
||||
|
||||
using var db = new Shared.TestDbContext(Path.Combine(Path.GetTempPath(), $"cbdd_async_concurrency_{Guid.NewGuid()}.db"));
|
||||
int threadCount = 2;
|
||||
int docsPerThread = 50;
|
||||
|
||||
var tasks = Enumerable.Range(0, threadCount).Select(async i =>
|
||||
{
|
||||
// Test mix of implicit and explicit transactions
|
||||
for (int j = 0; j < docsPerThread; j++)
|
||||
{
|
||||
int id = (i * docsPerThread) + j + 8000;
|
||||
await db.AsyncDocs.InsertAsync(new AsyncDoc { Id = id, Name = $"Thread{i}_Doc{j}" });
|
||||
}
|
||||
});
|
||||
|
||||
await Task.WhenAll(tasks);
|
||||
await db.SaveChangesAsync(ct);
|
||||
|
||||
// Verify count
|
||||
var count = db.AsyncDocs.Scan(_ => true).Count();
|
||||
count.ShouldBe(threadCount * docsPerThread);
|
||||
}
|
||||
}
|
||||
138
tests/CBDD.Tests/AttributeTests.cs
Executable file
138
tests/CBDD.Tests/AttributeTests.cs
Executable file
@@ -0,0 +1,138 @@
|
||||
using System.ComponentModel.DataAnnotations;
|
||||
using ZB.MOM.WW.CBDD.Bson;
|
||||
using ZB.MOM.WW.CBDD.Shared;
|
||||
using ZB.MOM.WW.CBDD.Shared.TestDbContext_TestDbContext_Mappers;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests
|
||||
{
|
||||
public class AttributeTests
|
||||
{
|
||||
// Use full path for mapper until we are sure of the namespace
|
||||
private ZB_MOM_WW_CBDD_Shared_AnnotatedUserMapper CreateMapper() => new();
|
||||
|
||||
private readonly System.Collections.Concurrent.ConcurrentDictionary<string, ushort> _keyMap = new(StringComparer.OrdinalIgnoreCase);
|
||||
private readonly System.Collections.Concurrent.ConcurrentDictionary<ushort, string> _keys = new();
|
||||
|
||||
public AttributeTests()
|
||||
{
|
||||
ushort id = 1;
|
||||
string[] keys = ["_id", "display_name", "age", "location", "0", "1"];
|
||||
foreach (var key in keys)
|
||||
{
|
||||
_keyMap[key] = id;
|
||||
_keys[id] = key;
|
||||
id++;
|
||||
}
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Test_Table_Attribute_Mapping()
|
||||
{
|
||||
// Verify that the generated mapper has the correct collection name
|
||||
var mapper = CreateMapper();
|
||||
mapper.CollectionName.ShouldBe("test.custom_users");
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Test_Required_Validation()
|
||||
{
|
||||
var mapper = CreateMapper();
|
||||
var user = new AnnotatedUser { Name = "" }; // Required name is empty
|
||||
var writer = new BsonSpanWriter(new byte[1024], _keyMap);
|
||||
|
||||
bool thrown = false;
|
||||
try
|
||||
{
|
||||
mapper.Serialize(user, writer);
|
||||
}
|
||||
catch (ValidationException)
|
||||
{
|
||||
thrown = true;
|
||||
}
|
||||
thrown.ShouldBeTrue("Should throw ValidationException for empty Name.");
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Test_StringLength_Validation()
|
||||
{
|
||||
var mapper = CreateMapper();
|
||||
var user = new AnnotatedUser { Name = "Jo" }; // Too short
|
||||
var writer = new BsonSpanWriter(new byte[1024], _keyMap);
|
||||
|
||||
bool thrown = false;
|
||||
try { mapper.Serialize(user, writer); } catch (ValidationException) { thrown = true; }
|
||||
thrown.ShouldBeTrue("Should throw ValidationException for Name too short.");
|
||||
|
||||
user.Name = new string('A', 51); // Too long
|
||||
thrown = false;
|
||||
try { mapper.Serialize(user, writer); } catch (ValidationException) { thrown = true; }
|
||||
thrown.ShouldBeTrue("Should throw ValidationException for Name too long.");
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Test_Range_Validation()
|
||||
{
|
||||
var mapper = CreateMapper();
|
||||
var user = new AnnotatedUser { Name = "John", Age = 200 }; // Out of range
|
||||
var writer = new BsonSpanWriter(new byte[1024], _keyMap);
|
||||
|
||||
bool thrown = false;
|
||||
try { mapper.Serialize(user, writer); } catch (ValidationException) { thrown = true; }
|
||||
thrown.ShouldBeTrue("Should throw ValidationException for Age out of range.");
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Test_Column_Name_Mapping()
|
||||
{
|
||||
var mapper = CreateMapper();
|
||||
var user = new AnnotatedUser { Name = "John", Age = 30 };
|
||||
var buffer = new byte[1024];
|
||||
var writer = new BsonSpanWriter(buffer, _keyMap);
|
||||
|
||||
mapper.Serialize(user, writer);
|
||||
|
||||
var reader = new BsonSpanReader(buffer, _keys);
|
||||
reader.ReadDocumentSize();
|
||||
|
||||
bool foundDisplayName = false;
|
||||
while (reader.Remaining > 0)
|
||||
{
|
||||
var type = reader.ReadBsonType();
|
||||
if (type == BsonType.EndOfDocument) break;
|
||||
|
||||
var name = reader.ReadElementHeader();
|
||||
if (name == "display_name") foundDisplayName = true;
|
||||
reader.SkipValue(type);
|
||||
}
|
||||
|
||||
foundDisplayName.ShouldBeTrue("BSON field name should be 'display_name' from [Column] attribute.");
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Test_NotMapped_Attribute()
|
||||
{
|
||||
var mapper = CreateMapper();
|
||||
var user = new AnnotatedUser { Name = "John", Age = 30 };
|
||||
var buffer = new byte[1024];
|
||||
var writer = new BsonSpanWriter(buffer, _keyMap);
|
||||
|
||||
mapper.Serialize(user, writer);
|
||||
|
||||
var reader = new BsonSpanReader(buffer, _keys);
|
||||
reader.ReadDocumentSize();
|
||||
|
||||
bool foundComputed = false;
|
||||
while (reader.Remaining > 0)
|
||||
{
|
||||
var type = reader.ReadBsonType();
|
||||
if (type == BsonType.EndOfDocument) break;
|
||||
|
||||
var name = reader.ReadElementHeader();
|
||||
if (name == "ComputedInfo") foundComputed = true;
|
||||
reader.SkipValue(type);
|
||||
}
|
||||
|
||||
foundComputed.ShouldBeFalse("ComputedInfo should not be mapped to BSON.");
|
||||
}
|
||||
}
|
||||
}
|
||||
34
tests/CBDD.Tests/AutoInitTests.cs
Executable file
34
tests/CBDD.Tests/AutoInitTests.cs
Executable file
@@ -0,0 +1,34 @@
|
||||
using ZB.MOM.WW.CBDD.Shared;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests
|
||||
{
|
||||
public class AutoInitTests : System.IDisposable
|
||||
{
|
||||
private const string DbPath = "autoinit.db";
|
||||
|
||||
public AutoInitTests()
|
||||
{
|
||||
if (File.Exists(DbPath)) File.Delete(DbPath);
|
||||
}
|
||||
|
||||
public void Dispose()
|
||||
{
|
||||
if (File.Exists(DbPath)) File.Delete(DbPath);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Collections_Are_Initialized_By_Generator()
|
||||
{
|
||||
using var db = new Shared.TestDbContext(DbPath);
|
||||
|
||||
// Verify Collection is not null (initialized by generated method)
|
||||
db.AutoInitEntities.ShouldNotBeNull();
|
||||
|
||||
// Verify we can use it
|
||||
db.AutoInitEntities.Insert(new AutoInitEntity { Id = 1, Name = "Test" });
|
||||
var stored = db.AutoInitEntities.FindById(1);
|
||||
stored.ShouldNotBeNull();
|
||||
stored.Name.ShouldBe("Test");
|
||||
}
|
||||
}
|
||||
}
|
||||
55
tests/CBDD.Tests/BTreeDeleteUnderflowTests.cs
Normal file
55
tests/CBDD.Tests/BTreeDeleteUnderflowTests.cs
Normal file
@@ -0,0 +1,55 @@
|
||||
using ZB.MOM.WW.CBDD.Core.Indexing;
|
||||
using ZB.MOM.WW.CBDD.Core.Storage;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests;
|
||||
|
||||
public class BTreeDeleteUnderflowTests
|
||||
{
|
||||
[Fact]
|
||||
public void Delete_HeavyWorkload_Should_Remain_Queryable_After_Merges()
|
||||
{
|
||||
var dbPath = Path.Combine(Path.GetTempPath(), $"btree_underflow_{Guid.NewGuid():N}.db");
|
||||
|
||||
try
|
||||
{
|
||||
using var storage = new StorageEngine(dbPath, PageFileConfig.Default);
|
||||
var index = new BTreeIndex(storage, IndexOptions.CreateBTree("k"));
|
||||
|
||||
var insertTxn = storage.BeginTransaction().TransactionId;
|
||||
for (int i = 1; i <= 240; i++)
|
||||
{
|
||||
index.Insert(IndexKey.Create(i), new DocumentLocation((uint)(1000 + i), 0), insertTxn);
|
||||
}
|
||||
storage.CommitTransaction(insertTxn);
|
||||
|
||||
var deleteTxn = storage.BeginTransaction().TransactionId;
|
||||
for (int i = 1; i <= 190; i++)
|
||||
{
|
||||
index.Delete(IndexKey.Create(i), new DocumentLocation((uint)(1000 + i), 0), deleteTxn).ShouldBeTrue();
|
||||
}
|
||||
storage.CommitTransaction(deleteTxn);
|
||||
|
||||
for (int i = 1; i <= 190; i++)
|
||||
{
|
||||
index.TryFind(IndexKey.Create(i), out _, 0).ShouldBeFalse();
|
||||
}
|
||||
|
||||
for (int i = 191; i <= 240; i++)
|
||||
{
|
||||
index.TryFind(IndexKey.Create(i), out var location, 0).ShouldBeTrue();
|
||||
location.PageId.ShouldBe((uint)(1000 + i));
|
||||
}
|
||||
|
||||
var remaining = index.GreaterThan(IndexKey.Create(190), orEqual: false, 0).ToList();
|
||||
remaining.Count.ShouldBe(50);
|
||||
remaining.First().Key.ShouldBe(IndexKey.Create(191));
|
||||
remaining.Last().Key.ShouldBe(IndexKey.Create(240));
|
||||
}
|
||||
finally
|
||||
{
|
||||
if (File.Exists(dbPath)) File.Delete(dbPath);
|
||||
var walPath = Path.ChangeExtension(dbPath, ".wal");
|
||||
if (File.Exists(walPath)) File.Delete(walPath);
|
||||
}
|
||||
}
|
||||
}
|
||||
182
tests/CBDD.Tests/BsonDocumentAndBufferWriterTests.cs
Normal file
182
tests/CBDD.Tests/BsonDocumentAndBufferWriterTests.cs
Normal file
@@ -0,0 +1,182 @@
|
||||
using System.Buffers;
|
||||
using System.Buffers.Binary;
|
||||
using System.Collections.Concurrent;
|
||||
using System.Text;
|
||||
using ZB.MOM.WW.CBDD.Bson;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests;
|
||||
|
||||
public class BsonDocumentAndBufferWriterTests
|
||||
{
|
||||
[Fact]
|
||||
public void BsonDocument_Create_And_TryGet_RoundTrip()
|
||||
{
|
||||
var keyMap = new ConcurrentDictionary<string, ushort>(StringComparer.OrdinalIgnoreCase);
|
||||
var reverseMap = new ConcurrentDictionary<ushort, string>();
|
||||
|
||||
RegisterKey(keyMap, reverseMap, 1, "name");
|
||||
RegisterKey(keyMap, reverseMap, 2, "age");
|
||||
RegisterKey(keyMap, reverseMap, 3, "_id");
|
||||
|
||||
var expectedId = ObjectId.NewObjectId();
|
||||
|
||||
var doc = BsonDocument.Create(keyMap, b =>
|
||||
{
|
||||
b.AddString("name", "Alice");
|
||||
b.AddInt32("age", 32);
|
||||
b.AddObjectId("_id", expectedId);
|
||||
});
|
||||
|
||||
var wrapped = new BsonDocument(doc.RawData.ToArray(), reverseMap);
|
||||
|
||||
wrapped.TryGetString("name", out var name).ShouldBeTrue();
|
||||
name.ShouldBe("Alice");
|
||||
|
||||
wrapped.TryGetInt32("age", out var age).ShouldBeTrue();
|
||||
age.ShouldBe(32);
|
||||
|
||||
wrapped.TryGetObjectId("_id", out var id).ShouldBeTrue();
|
||||
id.ShouldBe(expectedId);
|
||||
|
||||
var reader = wrapped.GetReader();
|
||||
reader.ReadDocumentSize().ShouldBeGreaterThan(0);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void BsonDocument_TryGet_Should_Return_False_For_Missing_Or_Wrong_Type()
|
||||
{
|
||||
var keyMap = new ConcurrentDictionary<string, ushort>(StringComparer.OrdinalIgnoreCase);
|
||||
var reverseMap = new ConcurrentDictionary<ushort, string>();
|
||||
|
||||
RegisterKey(keyMap, reverseMap, 1, "name");
|
||||
RegisterKey(keyMap, reverseMap, 2, "age");
|
||||
|
||||
var doc = BsonDocument.Create(keyMap, b =>
|
||||
{
|
||||
b.AddString("name", "Bob");
|
||||
b.AddInt32("age", 28);
|
||||
});
|
||||
|
||||
var wrapped = new BsonDocument(doc.RawData.ToArray(), reverseMap);
|
||||
|
||||
wrapped.TryGetInt32("name", out _).ShouldBeFalse();
|
||||
wrapped.TryGetString("missing", out _).ShouldBeFalse();
|
||||
wrapped.TryGetObjectId("age", out _).ShouldBeFalse();
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void BsonDocumentBuilder_Should_Grow_Buffer_When_Document_Is_Large()
|
||||
{
|
||||
var keyMap = new ConcurrentDictionary<string, ushort>(StringComparer.OrdinalIgnoreCase);
|
||||
var reverseMap = new ConcurrentDictionary<ushort, string>();
|
||||
|
||||
for (ushort i = 1; i <= 180; i++)
|
||||
{
|
||||
var key = $"k{i}";
|
||||
RegisterKey(keyMap, reverseMap, i, key);
|
||||
}
|
||||
|
||||
var builder = new BsonDocumentBuilder(keyMap);
|
||||
for (int i = 1; i <= 180; i++)
|
||||
{
|
||||
builder.AddInt32($"k{i}", i);
|
||||
}
|
||||
|
||||
var doc = builder.Build();
|
||||
doc.Size.ShouldBeGreaterThan(1024);
|
||||
|
||||
var wrapped = new BsonDocument(doc.RawData.ToArray(), reverseMap);
|
||||
wrapped.TryGetInt32("k180", out var value).ShouldBeTrue();
|
||||
value.ShouldBe(180);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void BsonBufferWriter_Should_Write_Nested_Document_And_Array()
|
||||
{
|
||||
var output = new ArrayBufferWriter<byte>();
|
||||
var writer = new BsonBufferWriter(output);
|
||||
|
||||
int rootSizePos = writer.BeginDocument();
|
||||
|
||||
int childSizePos = writer.BeginDocument("child");
|
||||
writer.WriteString("name", "nested");
|
||||
writer.WriteBoolean("active", true);
|
||||
writer.EndDocument(childSizePos);
|
||||
int childEnd = writer.Position;
|
||||
|
||||
int arraySizePos = writer.BeginArray("nums");
|
||||
writer.WriteInt32("0", 1);
|
||||
writer.WriteInt32("1", 2);
|
||||
writer.EndArray(arraySizePos);
|
||||
int arrayEnd = writer.Position;
|
||||
|
||||
writer.EndDocument(rootSizePos);
|
||||
int rootEnd = writer.Position;
|
||||
|
||||
var bytes = output.WrittenSpan.ToArray();
|
||||
PatchDocumentSize(bytes, childSizePos, childEnd);
|
||||
PatchDocumentSize(bytes, arraySizePos, arrayEnd);
|
||||
PatchDocumentSize(bytes, rootSizePos, rootEnd);
|
||||
|
||||
var reader = new BsonSpanReader(bytes, new ConcurrentDictionary<ushort, string>());
|
||||
|
||||
reader.ReadDocumentSize().ShouldBe(bytes.Length);
|
||||
|
||||
reader.ReadBsonType().ShouldBe(BsonType.Document);
|
||||
reader.ReadCString().ShouldBe("child");
|
||||
reader.ReadDocumentSize().ShouldBeGreaterThan(8);
|
||||
|
||||
reader.ReadBsonType().ShouldBe(BsonType.String);
|
||||
reader.ReadCString().ShouldBe("name");
|
||||
reader.ReadString().ShouldBe("nested");
|
||||
|
||||
reader.ReadBsonType().ShouldBe(BsonType.Boolean);
|
||||
reader.ReadCString().ShouldBe("active");
|
||||
reader.ReadBoolean().ShouldBeTrue();
|
||||
reader.ReadBsonType().ShouldBe(BsonType.EndOfDocument);
|
||||
|
||||
reader.ReadBsonType().ShouldBe(BsonType.Array);
|
||||
reader.ReadCString().ShouldBe("nums");
|
||||
reader.ReadDocumentSize().ShouldBeGreaterThan(8);
|
||||
|
||||
reader.ReadBsonType().ShouldBe(BsonType.Int32);
|
||||
reader.ReadCString().ShouldBe("0");
|
||||
reader.ReadInt32().ShouldBe(1);
|
||||
|
||||
reader.ReadBsonType().ShouldBe(BsonType.Int32);
|
||||
reader.ReadCString().ShouldBe("1");
|
||||
reader.ReadInt32().ShouldBe(2);
|
||||
reader.ReadBsonType().ShouldBe(BsonType.EndOfDocument);
|
||||
|
||||
reader.ReadBsonType().ShouldBe(BsonType.EndOfDocument);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void BsonSpanReader_ReadByte_And_ReadCStringSpan_Should_Work()
|
||||
{
|
||||
var singleByteReader = new BsonSpanReader(new byte[] { 0x2A }, new ConcurrentDictionary<ushort, string>());
|
||||
singleByteReader.ReadByte().ShouldBe((byte)0x2A);
|
||||
|
||||
var cstring = Encoding.UTF8.GetBytes("hello\0");
|
||||
var cstringReader = new BsonSpanReader(cstring, new ConcurrentDictionary<ushort, string>());
|
||||
var destination = new char[16];
|
||||
var written = cstringReader.ReadCString(destination);
|
||||
|
||||
new string(destination, 0, written).ShouldBe("hello");
|
||||
}
|
||||
|
||||
private static void RegisterKey(
|
||||
ConcurrentDictionary<string, ushort> keyMap,
|
||||
ConcurrentDictionary<ushort, string> reverseMap,
|
||||
ushort id,
|
||||
string key)
|
||||
{
|
||||
keyMap[key] = id;
|
||||
reverseMap[id] = key;
|
||||
}
|
||||
|
||||
private static void PatchDocumentSize(byte[] output, int sizePosition, int endPosition)
|
||||
{
|
||||
BinaryPrimitives.WriteInt32LittleEndian(output.AsSpan(sizePosition, 4), endPosition - sizePosition);
|
||||
}
|
||||
}
|
||||
96
tests/CBDD.Tests/BsonSchemaTests.cs
Executable file
96
tests/CBDD.Tests/BsonSchemaTests.cs
Executable file
@@ -0,0 +1,96 @@
|
||||
using ZB.MOM.WW.CBDD.Bson;
|
||||
using ZB.MOM.WW.CBDD.Core.Collections;
|
||||
using Xunit;
|
||||
using System.Collections.Generic;
|
||||
using System;
|
||||
using System.Linq;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests;
|
||||
|
||||
public class BsonSchemaTests
|
||||
{
|
||||
public class SimpleEntity
|
||||
{
|
||||
public ObjectId Id { get; set; }
|
||||
public string Name { get; set; } = string.Empty;
|
||||
public int Age { get; set; }
|
||||
public bool IsActive { get; set; }
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void GenerateSchema_SimpleEntity()
|
||||
{
|
||||
var schema = BsonSchemaGenerator.FromType<SimpleEntity>();
|
||||
|
||||
schema.Title.ShouldBe("SimpleEntity");
|
||||
schema.Fields.Count.ShouldBe(4);
|
||||
|
||||
var idField = schema.Fields.First(f => f.Name == "_id");
|
||||
idField.Type.ShouldBe(BsonType.ObjectId);
|
||||
|
||||
var nameField = schema.Fields.First(f => f.Name == "name");
|
||||
nameField.Type.ShouldBe(BsonType.String);
|
||||
|
||||
var ageField = schema.Fields.First(f => f.Name == "age");
|
||||
ageField.Type.ShouldBe(BsonType.Int32);
|
||||
}
|
||||
|
||||
public class CollectionEntity
|
||||
{
|
||||
public List<string> Tags { get; set; } = new();
|
||||
public int[] Scores { get; set; } = Array.Empty<int>();
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void GenerateSchema_Collections()
|
||||
{
|
||||
var schema = BsonSchemaGenerator.FromType<CollectionEntity>();
|
||||
|
||||
var tags = schema.Fields.First(f => f.Name == "tags");
|
||||
tags.Type.ShouldBe(BsonType.Array);
|
||||
tags.ArrayItemType.ShouldBe(BsonType.String);
|
||||
|
||||
var scores = schema.Fields.First(f => f.Name == "scores");
|
||||
scores.Type.ShouldBe(BsonType.Array);
|
||||
scores.ArrayItemType.ShouldBe(BsonType.Int32);
|
||||
}
|
||||
|
||||
public class NestedEntity
|
||||
{
|
||||
public SimpleEntity Parent { get; set; } = new();
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void GenerateSchema_Nested()
|
||||
{
|
||||
var schema = BsonSchemaGenerator.FromType<NestedEntity>();
|
||||
|
||||
var parent = schema.Fields.First(f => f.Name == "parent");
|
||||
parent.Type.ShouldBe(BsonType.Document);
|
||||
parent.NestedSchema.ShouldNotBeNull();
|
||||
parent.NestedSchema.Fields.ShouldContain(f => f.Name == "_id");
|
||||
}
|
||||
|
||||
public class ComplexCollectionEntity
|
||||
{
|
||||
public List<SimpleEntity> Items { get; set; } = new();
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void GenerateSchema_ComplexCollection()
|
||||
{
|
||||
var schema = BsonSchemaGenerator.FromType<ComplexCollectionEntity>();
|
||||
|
||||
var items = schema.Fields.First(f => f.Name == "items");
|
||||
items.Type.ShouldBe(BsonType.Array);
|
||||
// items.ArrayItemType.ShouldBe(BsonType.Document); // Wait, my generator logic might return Array here? No, item type logic...
|
||||
|
||||
// Let's verify generator logic for complex array item type
|
||||
// In generator: (BsonType.Array, itemNested, itemBsonType)
|
||||
// itemBsonType for SimpleEntity should be Document
|
||||
|
||||
items.ArrayItemType.ShouldBe(BsonType.Document);
|
||||
items.NestedSchema.ShouldNotBeNull();
|
||||
items.NestedSchema.Fields.ShouldContain(f => f.Name == "_id");
|
||||
}
|
||||
}
|
||||
252
tests/CBDD.Tests/BsonSpanReaderWriterTests.cs
Executable file
252
tests/CBDD.Tests/BsonSpanReaderWriterTests.cs
Executable file
@@ -0,0 +1,252 @@
|
||||
using ZB.MOM.WW.CBDD.Bson;
|
||||
using Xunit;
|
||||
using System.Collections.Concurrent;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests;
|
||||
|
||||
public class BsonSpanReaderWriterTests
|
||||
{
|
||||
private readonly ConcurrentDictionary<string, ushort> _keyMap = new(StringComparer.OrdinalIgnoreCase);
|
||||
private readonly ConcurrentDictionary<ushort, string> _keys = new();
|
||||
|
||||
public BsonSpanReaderWriterTests()
|
||||
{
|
||||
ushort id = 1;
|
||||
string[] initialKeys = ["name", "age", "active", "_id", "val", "dec", "timestamp", "int32", "int64", "double", "data", "child", "value", "0", "1"];
|
||||
foreach (var key in initialKeys)
|
||||
{
|
||||
_keyMap[key] = id;
|
||||
_keys[id] = key;
|
||||
id++;
|
||||
}
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void WriteAndRead_SimpleDocument()
|
||||
{
|
||||
Span<byte> buffer = stackalloc byte[256];
|
||||
var writer = new BsonSpanWriter(buffer, _keyMap);
|
||||
|
||||
var sizePos = writer.BeginDocument();
|
||||
writer.WriteString("name", "John");
|
||||
writer.WriteInt32("age", 30);
|
||||
writer.WriteBoolean("active", true);
|
||||
writer.EndDocument(sizePos);
|
||||
|
||||
var documentBytes = buffer[..writer.Position];
|
||||
|
||||
var reader = new BsonSpanReader(documentBytes, _keys);
|
||||
var size = reader.ReadDocumentSize();
|
||||
|
||||
size.ShouldBe(writer.Position);
|
||||
|
||||
var type1 = reader.ReadBsonType();
|
||||
var name1 = reader.ReadElementHeader();
|
||||
var value1 = reader.ReadString();
|
||||
|
||||
type1.ShouldBe(BsonType.String);
|
||||
name1.ShouldBe("name");
|
||||
value1.ShouldBe("John");
|
||||
|
||||
var type2 = reader.ReadBsonType();
|
||||
var name2 = reader.ReadElementHeader();
|
||||
var value2 = reader.ReadInt32();
|
||||
|
||||
type2.ShouldBe(BsonType.Int32);
|
||||
name2.ShouldBe("age");
|
||||
value2.ShouldBe(30);
|
||||
|
||||
var type3 = reader.ReadBsonType();
|
||||
var name3 = reader.ReadElementHeader();
|
||||
var value3 = reader.ReadBoolean();
|
||||
|
||||
type3.ShouldBe(BsonType.Boolean);
|
||||
name3.ShouldBe("active");
|
||||
value3.ShouldBeTrue();
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void WriteAndRead_ObjectId()
|
||||
{
|
||||
Span<byte> buffer = stackalloc byte[256];
|
||||
var writer = new BsonSpanWriter(buffer, _keyMap);
|
||||
|
||||
var oid = ObjectId.NewObjectId();
|
||||
|
||||
var sizePos = writer.BeginDocument();
|
||||
writer.WriteObjectId("_id", oid);
|
||||
writer.EndDocument(sizePos);
|
||||
|
||||
var documentBytes = buffer[..writer.Position];
|
||||
var reader = new BsonSpanReader(documentBytes, _keys);
|
||||
|
||||
reader.ReadDocumentSize();
|
||||
var type = reader.ReadBsonType();
|
||||
var name = reader.ReadElementHeader();
|
||||
var readOid = reader.ReadObjectId();
|
||||
|
||||
type.ShouldBe(BsonType.ObjectId);
|
||||
name.ShouldBe("_id");
|
||||
readOid.ShouldBe(oid);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void ReadWrite_Double()
|
||||
{
|
||||
var buffer = new byte[256];
|
||||
var writer = new BsonSpanWriter(buffer, _keyMap);
|
||||
|
||||
writer.WriteDouble("val", 123.456);
|
||||
|
||||
var reader = new BsonSpanReader(buffer, _keys);
|
||||
var type = reader.ReadBsonType();
|
||||
var name = reader.ReadElementHeader();
|
||||
var val = reader.ReadDouble();
|
||||
|
||||
type.ShouldBe(BsonType.Double);
|
||||
name.ShouldBe("val");
|
||||
val.ShouldBe(123.456);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void ReadWrite_Decimal128_RoundTrip()
|
||||
{
|
||||
var buffer = new byte[256];
|
||||
var writer = new BsonSpanWriter(buffer, _keyMap);
|
||||
|
||||
decimal original = 123456.789m;
|
||||
writer.WriteDecimal128("dec", original);
|
||||
|
||||
var reader = new BsonSpanReader(buffer, _keys);
|
||||
var type = reader.ReadBsonType();
|
||||
var name = reader.ReadElementHeader();
|
||||
var val = reader.ReadDecimal128();
|
||||
|
||||
type.ShouldBe(BsonType.Decimal128);
|
||||
name.ShouldBe("dec");
|
||||
val.ShouldBe(original);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void WriteAndRead_DateTime()
|
||||
{
|
||||
Span<byte> buffer = stackalloc byte[256];
|
||||
var writer = new BsonSpanWriter(buffer, _keyMap);
|
||||
|
||||
var now = DateTime.UtcNow;
|
||||
// Round to milliseconds as BSON only stores millisecond precision
|
||||
var expectedTime = new DateTime(now.Year, now.Month, now.Day,
|
||||
now.Hour, now.Minute, now.Second, now.Millisecond, DateTimeKind.Utc);
|
||||
|
||||
var sizePos = writer.BeginDocument();
|
||||
writer.WriteDateTime("timestamp", expectedTime);
|
||||
writer.EndDocument(sizePos);
|
||||
|
||||
var documentBytes = buffer[..writer.Position];
|
||||
var reader = new BsonSpanReader(documentBytes, _keys);
|
||||
|
||||
reader.ReadDocumentSize();
|
||||
var type = reader.ReadBsonType();
|
||||
var name = reader.ReadElementHeader();
|
||||
var readTime = reader.ReadDateTime();
|
||||
|
||||
type.ShouldBe(BsonType.DateTime);
|
||||
name.ShouldBe("timestamp");
|
||||
readTime.ShouldBe(expectedTime);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void WriteAndRead_NumericTypes()
|
||||
{
|
||||
Span<byte> buffer = stackalloc byte[256];
|
||||
var writer = new BsonSpanWriter(buffer, _keyMap);
|
||||
|
||||
var sizePos = writer.BeginDocument();
|
||||
writer.WriteInt32("int32", int.MaxValue);
|
||||
writer.WriteInt64("int64", long.MaxValue);
|
||||
writer.WriteDouble("double", 3.14159);
|
||||
writer.EndDocument(sizePos);
|
||||
|
||||
var documentBytes = buffer[..writer.Position];
|
||||
var reader = new BsonSpanReader(documentBytes, _keys);
|
||||
|
||||
reader.ReadDocumentSize();
|
||||
|
||||
reader.ReadBsonType();
|
||||
reader.ReadElementHeader();
|
||||
reader.ReadInt32().ShouldBe(int.MaxValue);
|
||||
|
||||
reader.ReadBsonType();
|
||||
reader.ReadElementHeader();
|
||||
reader.ReadInt64().ShouldBe(long.MaxValue);
|
||||
|
||||
reader.ReadBsonType();
|
||||
reader.ReadElementHeader();
|
||||
Math.Round(reader.ReadDouble(), 5).ShouldBe(Math.Round(3.14159, 5));
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void WriteAndRead_Binary()
|
||||
{
|
||||
Span<byte> buffer = stackalloc byte[256];
|
||||
var writer = new BsonSpanWriter(buffer, _keyMap);
|
||||
|
||||
byte[] testData = [1, 2, 3, 4, 5];
|
||||
|
||||
var sizePos = writer.BeginDocument();
|
||||
writer.WriteBinary("data", testData);
|
||||
writer.EndDocument(sizePos);
|
||||
|
||||
var documentBytes = buffer[..writer.Position];
|
||||
var reader = new BsonSpanReader(documentBytes, _keys);
|
||||
|
||||
reader.ReadDocumentSize();
|
||||
var type = reader.ReadBsonType();
|
||||
var name = reader.ReadElementHeader();
|
||||
var readData = reader.ReadBinary(out var subtype);
|
||||
|
||||
type.ShouldBe(BsonType.Binary);
|
||||
name.ShouldBe("data");
|
||||
subtype.ShouldBe((byte)0);
|
||||
testData.AsSpan().SequenceEqual(readData).ShouldBeTrue();
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void WriteAndRead_NestedDocument()
|
||||
{
|
||||
Span<byte> buffer = stackalloc byte[512];
|
||||
var writer = new BsonSpanWriter(buffer, _keyMap);
|
||||
|
||||
var rootSizePos = writer.BeginDocument();
|
||||
writer.WriteString("name", "Parent");
|
||||
|
||||
var childSizePos = writer.BeginDocument("child");
|
||||
writer.WriteString("name", "Child");
|
||||
writer.WriteInt32("value", 42);
|
||||
writer.EndDocument(childSizePos);
|
||||
|
||||
writer.EndDocument(rootSizePos);
|
||||
|
||||
var documentBytes = buffer[..writer.Position];
|
||||
var reader = new BsonSpanReader(documentBytes, _keys);
|
||||
var rootSize = reader.ReadDocumentSize();
|
||||
|
||||
rootSize.ShouldBe(writer.Position);
|
||||
|
||||
reader.ReadBsonType(); // String
|
||||
reader.ReadElementHeader().ShouldBe("name");
|
||||
reader.ReadString().ShouldBe("Parent");
|
||||
|
||||
reader.ReadBsonType(); // Document
|
||||
reader.ReadElementHeader().ShouldBe("child");
|
||||
|
||||
reader.ReadDocumentSize();
|
||||
reader.ReadBsonType(); // String
|
||||
reader.ReadElementHeader().ShouldBe("name");
|
||||
reader.ReadString().ShouldBe("Child");
|
||||
|
||||
reader.ReadBsonType(); // Int32
|
||||
reader.ReadElementHeader().ShouldBe("value");
|
||||
reader.ReadInt32().ShouldBe(42);
|
||||
}
|
||||
}
|
||||
125
tests/CBDD.Tests/BulkOperationsTests.cs
Executable file
125
tests/CBDD.Tests/BulkOperationsTests.cs
Executable file
@@ -0,0 +1,125 @@
|
||||
using ZB.MOM.WW.CBDD.Bson;
|
||||
using ZB.MOM.WW.CBDD.Core.Collections;
|
||||
using ZB.MOM.WW.CBDD.Core.Storage;
|
||||
using ZB.MOM.WW.CBDD.Core.Transactions;
|
||||
using ZB.MOM.WW.CBDD.Shared;
|
||||
using ZB.MOM.WW.CBDD.Shared.TestDbContext_TestDbContext_Mappers;
|
||||
using Xunit;
|
||||
using static ZB.MOM.WW.CBDD.Tests.SchemaTests;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests;
|
||||
|
||||
public class BulkOperationsTests : IDisposable
|
||||
{
|
||||
private readonly string _dbPath;
|
||||
private readonly string _walPath;
|
||||
private readonly Shared.TestDbContext _dbContext;
|
||||
|
||||
public BulkOperationsTests()
|
||||
{
|
||||
_dbPath = Path.Combine(Path.GetTempPath(), $"test_bulk_{Guid.NewGuid()}.db");
|
||||
_walPath = Path.Combine(Path.GetTempPath(), $"test_bulk_{Guid.NewGuid()}.wal");
|
||||
|
||||
_dbContext = new Shared.TestDbContext(_dbPath);
|
||||
}
|
||||
|
||||
public void Dispose()
|
||||
{
|
||||
_dbContext.Dispose();
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void UpdateBulk_UpdatesMultipleDocuments()
|
||||
{
|
||||
// Arrange: Insert 100 users
|
||||
var users = new List<User>();
|
||||
for (int i = 0; i < 100; i++)
|
||||
{
|
||||
users.Add(new User { Id = ObjectId.NewObjectId(), Name = $"User {i}", Age = 20 });
|
||||
}
|
||||
_dbContext.Users.InsertBulk(users);
|
||||
_dbContext.SaveChanges();
|
||||
|
||||
// Modify users
|
||||
foreach (var u in users)
|
||||
{
|
||||
u.Age = 30; // In-place update (int is same size)
|
||||
if (u.Name.EndsWith("0")) u.Name += "_Modified_Longer"; // Force move update
|
||||
}
|
||||
|
||||
// Act
|
||||
var updatedCount = _dbContext.Users.UpdateBulk(users);
|
||||
_dbContext.SaveChanges();
|
||||
|
||||
// Assert
|
||||
updatedCount.ShouldBe(100);
|
||||
|
||||
// Verify changes
|
||||
foreach (var u in users)
|
||||
{
|
||||
var stored = _dbContext.Users.FindById(u.Id);
|
||||
stored.ShouldNotBeNull();
|
||||
stored.Age.ShouldBe(30);
|
||||
stored.Name.ShouldBe(u.Name);
|
||||
}
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void DeleteBulk_RemovesMultipleDocuments()
|
||||
{
|
||||
// Arrange: Insert 100 users
|
||||
var users = new List<User>();
|
||||
for (int i = 0; i < 100; i++)
|
||||
{
|
||||
users.Add(new User { Id = ObjectId.NewObjectId(), Name = $"User {i}", Age = 20 });
|
||||
}
|
||||
_dbContext.Users.InsertBulk(users);
|
||||
_dbContext.SaveChanges();
|
||||
|
||||
var idsToDelete = users.Take(50).Select(u => u.Id).ToList();
|
||||
|
||||
// Act
|
||||
var deletedCount = _dbContext.Users.DeleteBulk(idsToDelete);
|
||||
_dbContext.SaveChanges();
|
||||
|
||||
// Assert
|
||||
deletedCount.ShouldBe(50);
|
||||
|
||||
// Verify deleted
|
||||
foreach (var id in idsToDelete)
|
||||
{
|
||||
_dbContext.Users.FindById(id).ShouldBeNull();
|
||||
}
|
||||
|
||||
// Verify remaining
|
||||
var remaining = users.Skip(50).ToList();
|
||||
foreach (var u in remaining)
|
||||
{
|
||||
_dbContext.Users.FindById(u.Id).ShouldNotBeNull();
|
||||
}
|
||||
|
||||
// Verify count
|
||||
// Note: Count() is not fully implemented efficiently yet (iterates everything), but FindAll().Count() works
|
||||
_dbContext.Users.FindAll().Count().ShouldBe(50);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void DeleteBulk_WithTransaction_Rollworks()
|
||||
{
|
||||
// Arrange
|
||||
var user = new User { Id = ObjectId.NewObjectId(), Name = "Txn User", Age = 20 };
|
||||
_dbContext.Users.Insert(user);
|
||||
_dbContext.SaveChanges();
|
||||
|
||||
_dbContext.Users.FindById(user.Id).ShouldNotBeNull();
|
||||
|
||||
using (var txn = _dbContext.BeginTransaction())
|
||||
{
|
||||
_dbContext.Users.DeleteBulk(new[] { user.Id });
|
||||
txn.Rollback();
|
||||
}
|
||||
|
||||
// Assert: Should still exist
|
||||
_dbContext.Users.FindById(user.Id).ShouldNotBeNull();
|
||||
}
|
||||
}
|
||||
109
tests/CBDD.Tests/CdcScalabilityTests.cs
Executable file
109
tests/CBDD.Tests/CdcScalabilityTests.cs
Executable file
@@ -0,0 +1,109 @@
|
||||
using System;
|
||||
using System.Collections.Generic;
|
||||
using System.Linq;
|
||||
using System.Threading.Tasks;
|
||||
using ZB.MOM.WW.CBDD.Core.CDC;
|
||||
using ZB.MOM.WW.CBDD.Shared;
|
||||
using Xunit;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests;
|
||||
|
||||
public class CdcScalabilityTests : IDisposable
|
||||
{
|
||||
private readonly Shared.TestDbContext _db;
|
||||
private readonly string _dbPath;
|
||||
|
||||
public CdcScalabilityTests()
|
||||
{
|
||||
_dbPath = Path.Combine(Path.GetTempPath(), $"cdc_scaling_{Guid.NewGuid()}.db");
|
||||
_db = new Shared.TestDbContext(_dbPath);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public async Task Test_Cdc_1000_Subscribers_Receive_Events()
|
||||
{
|
||||
var ct = TestContext.Current.CancellationToken;
|
||||
const int SubscriberCount = 1000;
|
||||
var eventCounts = new int[SubscriberCount];
|
||||
var subscriptions = new List<IDisposable>();
|
||||
|
||||
// 1. Create 1000 subscribers
|
||||
for (int i = 0; i < SubscriberCount; i++)
|
||||
{
|
||||
int index = i;
|
||||
var sub = _db.People.Watch().Subscribe(_ =>
|
||||
{
|
||||
Interlocked.Increment(ref eventCounts[index]);
|
||||
});
|
||||
subscriptions.Add(sub);
|
||||
}
|
||||
|
||||
// 2. Perform some writes
|
||||
_db.People.Insert(new Person { Id = 1, Name = "John", Age = 30 });
|
||||
_db.People.Insert(new Person { Id = 2, Name = "Jane", Age = 25 });
|
||||
_db.SaveChanges();
|
||||
|
||||
// 3. Wait for events to propagate
|
||||
await Task.Delay(1000, ct);
|
||||
|
||||
// 4. Verify all subscribers received both events
|
||||
for (int i = 0; i < SubscriberCount; i++)
|
||||
{
|
||||
eventCounts[i].ShouldBe(2);
|
||||
}
|
||||
|
||||
foreach (var sub in subscriptions) sub.Dispose();
|
||||
}
|
||||
|
||||
[Fact(Skip = "Performance test - run manually when needed")]
|
||||
public async Task Test_Cdc_Slow_Subscriber_Does_Not_Block_Others()
|
||||
{
|
||||
var ct = TestContext.Current.CancellationToken;
|
||||
var fastEventCount = 0;
|
||||
var slowEventCount = 0;
|
||||
|
||||
// 1. Register a slow subscriber that blocks SYNCHRONOUSLY
|
||||
using var slowSub = _db.People.Watch().Subscribe(_ =>
|
||||
{
|
||||
Interlocked.Increment(ref slowEventCount);
|
||||
// Synchronous block to block the BridgeChannelToObserverAsync loop for this sub
|
||||
Thread.Sleep(2000);
|
||||
});
|
||||
|
||||
// 2. Register a fast subscriber
|
||||
using var fastSub = _db.People.Watch().Subscribe(_ =>
|
||||
{
|
||||
Interlocked.Increment(ref fastEventCount);
|
||||
});
|
||||
|
||||
// 3. Perform a write
|
||||
_db.People.Insert(new Person { Id = 1, Name = "John", Age = 30 });
|
||||
_db.SaveChanges();
|
||||
|
||||
// 4. Verification: Fast subscriber should receive it immediately
|
||||
await Task.Delay(200, ct);
|
||||
fastEventCount.ShouldBe(1);
|
||||
slowEventCount.ShouldBe(1); // Started but not finished or blocking others
|
||||
|
||||
// 5. Perform another write
|
||||
_db.People.Insert(new Person { Id = 2, Name = "Jane", Age = 25 });
|
||||
_db.SaveChanges();
|
||||
|
||||
// 6. Verification: Fast subscriber should receive second event while slow one is still busy
|
||||
await Task.Delay(200, ct);
|
||||
fastEventCount.ShouldBe(2);
|
||||
slowEventCount.ShouldBe(1); // Still processing first one or second one queued in private channel
|
||||
|
||||
// 7. Wait for slow one to eventually catch up
|
||||
await Task.Delay(2500, ct); // Wait for the second one in slow sub to be processed after the first Sleep
|
||||
slowEventCount.ShouldBe(2);
|
||||
}
|
||||
|
||||
public void Dispose()
|
||||
{
|
||||
_db.Dispose();
|
||||
if (File.Exists(_dbPath)) File.Delete(_dbPath);
|
||||
var wal = Path.ChangeExtension(_dbPath, ".wal");
|
||||
if (File.Exists(wal)) File.Delete(wal);
|
||||
}
|
||||
}
|
||||
171
tests/CBDD.Tests/CdcTests.cs
Executable file
171
tests/CBDD.Tests/CdcTests.cs
Executable file
@@ -0,0 +1,171 @@
|
||||
using System;
|
||||
using System.Collections.Concurrent;
|
||||
using System.Collections.Generic;
|
||||
using System.Diagnostics;
|
||||
using System.IO;
|
||||
using System.Linq;
|
||||
using System.Threading.Tasks;
|
||||
using ZB.MOM.WW.CBDD.Bson;
|
||||
using ZB.MOM.WW.CBDD.Core.CDC;
|
||||
using ZB.MOM.WW.CBDD.Core.Transactions;
|
||||
using ZB.MOM.WW.CBDD.Shared;
|
||||
using Xunit;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests;
|
||||
|
||||
public class CdcTests : IDisposable
|
||||
{
|
||||
private static readonly TimeSpan DefaultEventTimeout = TimeSpan.FromSeconds(3);
|
||||
private static readonly TimeSpan PollInterval = TimeSpan.FromMilliseconds(10);
|
||||
|
||||
private readonly string _dbPath = $"cdc_test_{Guid.NewGuid()}.db";
|
||||
private readonly Shared.TestDbContext _db;
|
||||
|
||||
public CdcTests()
|
||||
{
|
||||
_db = new Shared.TestDbContext(_dbPath);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public async Task Test_Cdc_Basic_Insert_Fires_Event()
|
||||
{
|
||||
var ct = TestContext.Current.CancellationToken;
|
||||
var events = new ConcurrentQueue<ChangeStreamEvent<int, Person>>();
|
||||
using var subscription = _db.People.Watch(capturePayload: true).Subscribe(events.Enqueue);
|
||||
|
||||
var person = new Person { Id = 1, Name = "John", Age = 30 };
|
||||
_db.People.Insert(person);
|
||||
_db.SaveChanges();
|
||||
|
||||
await WaitForEventCountAsync(events, expectedCount: 1, ct);
|
||||
|
||||
var snapshot = events.ToArray();
|
||||
snapshot.Length.ShouldBe(1);
|
||||
snapshot[0].Type.ShouldBe(OperationType.Insert);
|
||||
snapshot[0].DocumentId.ShouldBe(1);
|
||||
snapshot[0].Entity.ShouldNotBeNull();
|
||||
snapshot[0].Entity!.Name.ShouldBe("John");
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public async Task Test_Cdc_No_Payload_When_Not_Requested()
|
||||
{
|
||||
var ct = TestContext.Current.CancellationToken;
|
||||
var events = new ConcurrentQueue<ChangeStreamEvent<int, Person>>();
|
||||
using var subscription = _db.People.Watch(capturePayload: false).Subscribe(events.Enqueue);
|
||||
|
||||
var person = new Person { Id = 1, Name = "John", Age = 30 };
|
||||
_db.People.Insert(person);
|
||||
_db.SaveChanges();
|
||||
|
||||
await WaitForEventCountAsync(events, expectedCount: 1, ct);
|
||||
|
||||
var snapshot = events.ToArray();
|
||||
snapshot.Length.ShouldBe(1);
|
||||
snapshot[0].Entity.ShouldBeNull();
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public async Task Test_Cdc_Commit_Only()
|
||||
{
|
||||
var ct = TestContext.Current.CancellationToken;
|
||||
var events = new ConcurrentQueue<ChangeStreamEvent<int, Person>>();
|
||||
using var subscription = _db.People.Watch(capturePayload: true).Subscribe(events.Enqueue);
|
||||
|
||||
using (var txn = _db.BeginTransaction())
|
||||
{
|
||||
_db.People.Insert(new Person { Id = 1, Name = "John" });
|
||||
events.Count.ShouldBe(0); // Not committed yet
|
||||
|
||||
txn.Rollback();
|
||||
}
|
||||
|
||||
await Task.Delay(100, ct);
|
||||
events.Count.ShouldBe(0); // Rolled back
|
||||
|
||||
using (var txn = _db.BeginTransaction())
|
||||
{
|
||||
_db.People.Insert(new Person { Id = 2, Name = "Jane" });
|
||||
txn.Commit();
|
||||
}
|
||||
|
||||
await WaitForEventCountAsync(events, expectedCount: 1, ct);
|
||||
var snapshot = events.ToArray();
|
||||
snapshot.Length.ShouldBe(1);
|
||||
snapshot[0].DocumentId.ShouldBe(2);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public async Task Test_Cdc_Update_And_Delete()
|
||||
{
|
||||
var ct = TestContext.Current.CancellationToken;
|
||||
var events = new ConcurrentQueue<ChangeStreamEvent<int, Person>>();
|
||||
using var subscription = _db.People.Watch(capturePayload: true).Subscribe(events.Enqueue);
|
||||
|
||||
var person = new Person { Id = 1, Name = "John", Age = 30 };
|
||||
_db.People.Insert(person);
|
||||
_db.SaveChanges();
|
||||
|
||||
person.Name = "Johnny";
|
||||
_db.People.Update(person);
|
||||
_db.SaveChanges();
|
||||
|
||||
_db.People.Delete(1);
|
||||
_db.SaveChanges();
|
||||
|
||||
await WaitForEventCountAsync(events, expectedCount: 3, ct);
|
||||
|
||||
var snapshot = events.ToArray();
|
||||
snapshot.Length.ShouldBe(3);
|
||||
snapshot[0].Type.ShouldBe(OperationType.Insert);
|
||||
snapshot[1].Type.ShouldBe(OperationType.Update);
|
||||
snapshot[2].Type.ShouldBe(OperationType.Delete);
|
||||
|
||||
snapshot[1].Entity!.Name.ShouldBe("Johnny");
|
||||
snapshot[2].DocumentId.ShouldBe(1);
|
||||
}
|
||||
|
||||
public void Dispose()
|
||||
{
|
||||
_db.Dispose();
|
||||
if (File.Exists(_dbPath)) File.Delete(_dbPath);
|
||||
if (File.Exists(_dbPath + "-wal")) File.Delete(_dbPath + "-wal");
|
||||
}
|
||||
|
||||
private static async Task WaitForEventCountAsync(
|
||||
ConcurrentQueue<ChangeStreamEvent<int, Person>> events,
|
||||
int expectedCount,
|
||||
CancellationToken ct)
|
||||
{
|
||||
var sw = Stopwatch.StartNew();
|
||||
while (sw.Elapsed < DefaultEventTimeout)
|
||||
{
|
||||
if (events.Count >= expectedCount)
|
||||
{
|
||||
return;
|
||||
}
|
||||
|
||||
await Task.Delay(PollInterval, ct);
|
||||
}
|
||||
|
||||
events.Count.ShouldBe(expectedCount);
|
||||
}
|
||||
}
|
||||
|
||||
// Simple helper to avoid System.Reactive dependency in tests
|
||||
public static class ObservableExtensions
|
||||
{
|
||||
public static IDisposable Subscribe<T>(this IObservable<T> observable, Action<T> onNext)
|
||||
{
|
||||
return observable.Subscribe(new AnonymousObserver<T>(onNext));
|
||||
}
|
||||
|
||||
private class AnonymousObserver<T> : IObserver<T>
|
||||
{
|
||||
private readonly Action<T> _onNext;
|
||||
public AnonymousObserver(Action<T> onNext) => _onNext = onNext;
|
||||
public void OnCompleted() { }
|
||||
public void OnError(Exception error) { }
|
||||
public void OnNext(T value) => _onNext(value);
|
||||
}
|
||||
}
|
||||
404
tests/CBDD.Tests/CircularReferenceTests.cs
Executable file
404
tests/CBDD.Tests/CircularReferenceTests.cs
Executable file
@@ -0,0 +1,404 @@
|
||||
using ZB.MOM.WW.CBDD.Bson;
|
||||
using ZB.MOM.WW.CBDD.Shared;
|
||||
using ZB.MOM.WW.CBDD.Tests;
|
||||
using Xunit;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests;
|
||||
|
||||
/// <summary>
|
||||
/// Tests for circular references and N-N relationships
|
||||
/// Validates that the source generator handles:
|
||||
/// 1. Self-referencing entities using ObjectId references (Employee → ManagerId, DirectReportIds)
|
||||
/// 2. N-N via referencing with ObjectIds (CategoryRef/ProductRef) - BEST PRACTICE
|
||||
///
|
||||
/// Note: Bidirectional embedding (Category ↔ Product with full objects) is NOT supported
|
||||
/// by the source generator and is an anti-pattern for document databases.
|
||||
/// Use referencing (ObjectIds) instead for N-N relationships.
|
||||
/// </summary>
|
||||
public class CircularReferenceTests : IDisposable
|
||||
{
|
||||
private readonly string _dbPath;
|
||||
private readonly Shared.TestDbContext _context;
|
||||
|
||||
public CircularReferenceTests()
|
||||
{
|
||||
_dbPath = Path.Combine(Path.GetTempPath(), $"cbdd_circular_test_{Guid.NewGuid()}");
|
||||
_context = new Shared.TestDbContext(_dbPath);
|
||||
}
|
||||
|
||||
public void Dispose()
|
||||
{
|
||||
_context?.Dispose();
|
||||
if (Directory.Exists(_dbPath))
|
||||
{
|
||||
Directory.Delete(_dbPath, true);
|
||||
}
|
||||
}
|
||||
|
||||
// ========================================
|
||||
// Self-Reference Tests (Employee hierarchy with ObjectId references)
|
||||
// ========================================
|
||||
|
||||
[Fact]
|
||||
public void SelfReference_InsertAndQuery_ShouldWork()
|
||||
{
|
||||
// Arrange: Create organizational hierarchy using ObjectId references
|
||||
var ceoId = ObjectId.NewObjectId();
|
||||
var manager1Id = ObjectId.NewObjectId();
|
||||
var manager2Id = ObjectId.NewObjectId();
|
||||
var developerId = ObjectId.NewObjectId();
|
||||
|
||||
var ceo = new Employee
|
||||
{
|
||||
Id = ceoId,
|
||||
Name = "Alice CEO",
|
||||
Department = "Executive",
|
||||
ManagerId = null,
|
||||
DirectReportIds = new List<ObjectId> { manager1Id, manager2Id }
|
||||
};
|
||||
|
||||
var manager1 = new Employee
|
||||
{
|
||||
Id = manager1Id,
|
||||
Name = "Bob Manager",
|
||||
Department = "Engineering",
|
||||
ManagerId = ceoId,
|
||||
DirectReportIds = new List<ObjectId> { developerId }
|
||||
};
|
||||
|
||||
var manager2 = new Employee
|
||||
{
|
||||
Id = manager2Id,
|
||||
Name = "Carol Manager",
|
||||
Department = "Sales",
|
||||
ManagerId = ceoId,
|
||||
DirectReportIds = new List<ObjectId>() // No direct reports
|
||||
};
|
||||
|
||||
var developer = new Employee
|
||||
{
|
||||
Id = developerId,
|
||||
Name = "Dave Developer",
|
||||
Department = "Engineering",
|
||||
ManagerId = manager1Id,
|
||||
DirectReportIds = null // Leaf node
|
||||
};
|
||||
|
||||
// Act: Insert all employees
|
||||
_context.Employees.Insert(ceo);
|
||||
_context.Employees.Insert(manager1);
|
||||
_context.Employees.Insert(manager2);
|
||||
_context.Employees.Insert(developer);
|
||||
|
||||
// Assert: Query and verify
|
||||
var queriedCeo = _context.Employees.FindById(ceoId);
|
||||
queriedCeo.ShouldNotBeNull();
|
||||
queriedCeo.Name.ShouldBe("Alice CEO");
|
||||
queriedCeo.DirectReportIds.ShouldNotBeNull();
|
||||
queriedCeo.DirectReportIds.Count.ShouldBe(2);
|
||||
queriedCeo.DirectReportIds.ShouldContain(manager1Id);
|
||||
queriedCeo.DirectReportIds.ShouldContain(manager2Id);
|
||||
|
||||
// Query manager and verify direct reports
|
||||
var queriedManager1 = _context.Employees.FindById(manager1Id);
|
||||
queriedManager1.ShouldNotBeNull();
|
||||
queriedManager1.ManagerId.ShouldBe(ceoId);
|
||||
queriedManager1.DirectReportIds.ShouldNotBeNull();
|
||||
queriedManager1.DirectReportIds.Count().ShouldBe(1);
|
||||
queriedManager1.DirectReportIds.ShouldContain(developerId);
|
||||
|
||||
// Query developer and verify no direct reports
|
||||
var queriedDeveloper = _context.Employees.FindById(developerId);
|
||||
queriedDeveloper.ShouldNotBeNull();
|
||||
queriedDeveloper.ManagerId.ShouldBe(manager1Id);
|
||||
// Empty list is acceptable (same as null semantically - no direct reports)
|
||||
(queriedDeveloper.DirectReportIds ?? new List<ObjectId>()).ShouldBeEmpty();
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void SelfReference_UpdateDirectReports_ShouldPersist()
|
||||
{
|
||||
// Arrange: Create manager with one direct report
|
||||
var managerId = ObjectId.NewObjectId();
|
||||
var employee1Id = ObjectId.NewObjectId();
|
||||
var employee2Id = ObjectId.NewObjectId();
|
||||
|
||||
var manager = new Employee
|
||||
{
|
||||
Id = managerId,
|
||||
Name = "Manager",
|
||||
Department = "Engineering",
|
||||
DirectReportIds = new List<ObjectId> { employee1Id }
|
||||
};
|
||||
|
||||
var employee1 = new Employee
|
||||
{
|
||||
Id = employee1Id,
|
||||
Name = "Employee 1",
|
||||
Department = "Engineering",
|
||||
ManagerId = managerId
|
||||
};
|
||||
|
||||
var employee2 = new Employee
|
||||
{
|
||||
Id = employee2Id,
|
||||
Name = "Employee 2",
|
||||
Department = "Engineering",
|
||||
ManagerId = managerId
|
||||
};
|
||||
|
||||
_context.Employees.Insert(manager);
|
||||
_context.Employees.Insert(employee1);
|
||||
_context.Employees.Insert(employee2);
|
||||
|
||||
// Act: Add another direct report
|
||||
manager.DirectReportIds?.Add(employee2Id);
|
||||
_context.Employees.Update(manager);
|
||||
|
||||
// Assert: Verify update persisted
|
||||
var queried = _context.Employees.FindById(managerId);
|
||||
queried.ShouldNotBeNull();
|
||||
queried.DirectReportIds.ShouldNotBeNull();
|
||||
queried.DirectReportIds.Count.ShouldBe(2);
|
||||
queried.DirectReportIds.ShouldContain(employee1Id);
|
||||
queried.DirectReportIds.ShouldContain(employee2Id);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void SelfReference_QueryByManagerId_ShouldWork()
|
||||
{
|
||||
// Arrange: Create hierarchy
|
||||
var managerId = ObjectId.NewObjectId();
|
||||
|
||||
var manager = new Employee
|
||||
{
|
||||
Id = managerId,
|
||||
Name = "Manager",
|
||||
Department = "Engineering"
|
||||
};
|
||||
|
||||
var employee1 = new Employee
|
||||
{
|
||||
Id = ObjectId.NewObjectId(),
|
||||
Name = "Employee 1",
|
||||
Department = "Engineering",
|
||||
ManagerId = managerId
|
||||
};
|
||||
|
||||
var employee2 = new Employee
|
||||
{
|
||||
Id = ObjectId.NewObjectId(),
|
||||
Name = "Employee 2",
|
||||
Department = "Engineering",
|
||||
ManagerId = managerId
|
||||
};
|
||||
|
||||
_context.Employees.Insert(manager);
|
||||
_context.Employees.Insert(employee1);
|
||||
_context.Employees.Insert(employee2);
|
||||
|
||||
// Act: Query all employees with specific manager
|
||||
var subordinates = _context.Employees
|
||||
.AsQueryable()
|
||||
.Where(e => e.ManagerId == managerId)
|
||||
.ToList();
|
||||
|
||||
// Assert: Should find both employees
|
||||
subordinates.Count.ShouldBe(2);
|
||||
subordinates.ShouldContain(e => e.Name == "Employee 1");
|
||||
subordinates.ShouldContain(e => e.Name == "Employee 2");
|
||||
}
|
||||
|
||||
// ========================================
|
||||
// N-N Referencing Tests (CategoryRef/ProductRef)
|
||||
// BEST PRACTICE for document databases
|
||||
// ========================================
|
||||
|
||||
[Fact]
|
||||
public void NtoNReferencing_InsertAndQuery_ShouldWork()
|
||||
{
|
||||
// Arrange: Create categories and products with ObjectId references
|
||||
var categoryId1 = ObjectId.NewObjectId();
|
||||
var categoryId2 = ObjectId.NewObjectId();
|
||||
var productId1 = ObjectId.NewObjectId();
|
||||
var productId2 = ObjectId.NewObjectId();
|
||||
|
||||
var electronics = new CategoryRef
|
||||
{
|
||||
Id = categoryId1,
|
||||
Name = "Electronics",
|
||||
Description = "Electronic devices",
|
||||
ProductIds = new List<ObjectId> { productId1, productId2 }
|
||||
};
|
||||
|
||||
var computers = new CategoryRef
|
||||
{
|
||||
Id = categoryId2,
|
||||
Name = "Computers",
|
||||
Description = "Computing devices",
|
||||
ProductIds = new List<ObjectId> { productId1 }
|
||||
};
|
||||
|
||||
var laptop = new ProductRef
|
||||
{
|
||||
Id = productId1,
|
||||
Name = "Laptop",
|
||||
Price = 999.99m,
|
||||
CategoryIds = new List<ObjectId> { categoryId1, categoryId2 }
|
||||
};
|
||||
|
||||
var phone = new ProductRef
|
||||
{
|
||||
Id = productId2,
|
||||
Name = "Phone",
|
||||
Price = 599.99m,
|
||||
CategoryIds = new List<ObjectId> { categoryId1 }
|
||||
};
|
||||
|
||||
// Act: Insert all entities
|
||||
_context.CategoryRefs.Insert(electronics);
|
||||
_context.CategoryRefs.Insert(computers);
|
||||
_context.ProductRefs.Insert(laptop);
|
||||
_context.ProductRefs.Insert(phone);
|
||||
|
||||
// Assert: Query and verify references
|
||||
var queriedCategory = _context.CategoryRefs.FindById(categoryId1);
|
||||
queriedCategory.ShouldNotBeNull();
|
||||
queriedCategory.Name.ShouldBe("Electronics");
|
||||
queriedCategory.ProductIds.ShouldNotBeNull();
|
||||
queriedCategory.ProductIds.Count.ShouldBe(2);
|
||||
queriedCategory.ProductIds.ShouldContain(productId1);
|
||||
queriedCategory.ProductIds.ShouldContain(productId2);
|
||||
|
||||
var queriedProduct = _context.ProductRefs.FindById(productId1);
|
||||
queriedProduct.ShouldNotBeNull();
|
||||
queriedProduct.Name.ShouldBe("Laptop");
|
||||
queriedProduct.CategoryIds.ShouldNotBeNull();
|
||||
queriedProduct.CategoryIds.Count.ShouldBe(2);
|
||||
queriedProduct.CategoryIds.ShouldContain(categoryId1);
|
||||
queriedProduct.CategoryIds.ShouldContain(categoryId2);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void NtoNReferencing_UpdateRelationships_ShouldPersist()
|
||||
{
|
||||
// Arrange: Create category and product
|
||||
var categoryId = ObjectId.NewObjectId();
|
||||
var productId1 = ObjectId.NewObjectId();
|
||||
var productId2 = ObjectId.NewObjectId();
|
||||
|
||||
var category = new CategoryRef
|
||||
{
|
||||
Id = categoryId,
|
||||
Name = "Books",
|
||||
Description = "Book category",
|
||||
ProductIds = new List<ObjectId> { productId1 }
|
||||
};
|
||||
|
||||
var product1 = new ProductRef
|
||||
{
|
||||
Id = productId1,
|
||||
Name = "Book 1",
|
||||
Price = 19.99m,
|
||||
CategoryIds = new List<ObjectId> { categoryId }
|
||||
};
|
||||
|
||||
var product2 = new ProductRef
|
||||
{
|
||||
Id = productId2,
|
||||
Name = "Book 2",
|
||||
Price = 29.99m,
|
||||
CategoryIds = new List<ObjectId>()
|
||||
};
|
||||
|
||||
_context.CategoryRefs.Insert(category);
|
||||
_context.ProductRefs.Insert(product1);
|
||||
_context.ProductRefs.Insert(product2);
|
||||
|
||||
// Act: Add product2 to category
|
||||
category.ProductIds?.Add(productId2);
|
||||
_context.CategoryRefs.Update(category);
|
||||
|
||||
product2.CategoryIds?.Add(categoryId);
|
||||
_context.ProductRefs.Update(product2);
|
||||
|
||||
// Assert: Verify relationships updated
|
||||
var queriedCategory = _context.CategoryRefs.FindById(categoryId);
|
||||
queriedCategory.ShouldNotBeNull();
|
||||
queriedCategory.ProductIds.ShouldNotBeNull();
|
||||
queriedCategory.ProductIds.Count.ShouldBe(2);
|
||||
queriedCategory.ProductIds.ShouldContain(productId2);
|
||||
|
||||
var queriedProduct2 = _context.ProductRefs.FindById(productId2);
|
||||
queriedProduct2.ShouldNotBeNull();
|
||||
queriedProduct2.CategoryIds.ShouldNotBeNull();
|
||||
queriedProduct2.CategoryIds.Count().ShouldBe(1);
|
||||
queriedProduct2.CategoryIds.ShouldContain(categoryId);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void NtoNReferencing_DocumentSize_RemainSmall()
|
||||
{
|
||||
// Arrange: Create category referencing 100 products (only IDs)
|
||||
var categoryId = ObjectId.NewObjectId();
|
||||
var productIds = Enumerable.Range(0, 100)
|
||||
.Select(_ => ObjectId.NewObjectId())
|
||||
.ToList();
|
||||
|
||||
var category = new CategoryRef
|
||||
{
|
||||
Id = categoryId,
|
||||
Name = "Large Category",
|
||||
Description = "Category with 100 products",
|
||||
ProductIds = productIds
|
||||
};
|
||||
|
||||
// Act: Insert and query
|
||||
_context.CategoryRefs.Insert(category);
|
||||
var queried = _context.CategoryRefs.FindById(categoryId);
|
||||
|
||||
// Assert: Document remains small (only ObjectIds, no embedding)
|
||||
queried.ShouldNotBeNull();
|
||||
queried.ProductIds?.Count.ShouldBe(100);
|
||||
|
||||
// Note: 100 ObjectIds = ~1.2KB (vs embedding full products = potentially hundreds of KBs)
|
||||
// This demonstrates why referencing is preferred for large N-N relationships
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void NtoNReferencing_QueryByProductId_ShouldWork()
|
||||
{
|
||||
// Arrange: Create multiple categories referencing same product
|
||||
var productId = ObjectId.NewObjectId();
|
||||
|
||||
var category1 = new CategoryRef
|
||||
{
|
||||
Id = ObjectId.NewObjectId(),
|
||||
Name = "Category 1",
|
||||
Description = "First category",
|
||||
ProductIds = new List<ObjectId> { productId }
|
||||
};
|
||||
|
||||
var category2 = new CategoryRef
|
||||
{
|
||||
Id = ObjectId.NewObjectId(),
|
||||
Name = "Category 2",
|
||||
Description = "Second category",
|
||||
ProductIds = new List<ObjectId> { productId }
|
||||
};
|
||||
|
||||
_context.CategoryRefs.Insert(category1);
|
||||
_context.CategoryRefs.Insert(category2);
|
||||
|
||||
// Act: Query all categories containing the product
|
||||
var categoriesWithProduct = _context.CategoryRefs
|
||||
.AsQueryable()
|
||||
.Where(c => c.ProductIds != null && c.ProductIds.Contains(productId))
|
||||
.ToList();
|
||||
|
||||
// Assert: Should find both categories
|
||||
categoriesWithProduct.Count.ShouldBe(2);
|
||||
categoriesWithProduct.ShouldContain(c => c.Name == "Category 1");
|
||||
categoriesWithProduct.ShouldContain(c => c.Name == "Category 2");
|
||||
}
|
||||
}
|
||||
150
tests/CBDD.Tests/CollectionIndexManagerAndDefinitionTests.cs
Normal file
150
tests/CBDD.Tests/CollectionIndexManagerAndDefinitionTests.cs
Normal file
@@ -0,0 +1,150 @@
|
||||
using ZB.MOM.WW.CBDD.Core.Indexing;
|
||||
using ZB.MOM.WW.CBDD.Core.Storage;
|
||||
using ZB.MOM.WW.CBDD.Shared;
|
||||
using ZB.MOM.WW.CBDD.Shared.TestDbContext_TestDbContext_Mappers;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests;
|
||||
|
||||
public class CollectionIndexManagerAndDefinitionTests
|
||||
{
|
||||
[Fact]
|
||||
public void FindBestIndex_Should_Prefer_Unique_Index()
|
||||
{
|
||||
var dbPath = NewDbPath();
|
||||
try
|
||||
{
|
||||
using var storage = new StorageEngine(dbPath, PageFileConfig.Default);
|
||||
var mapper = new ZB_MOM_WW_CBDD_Shared_PersonMapper();
|
||||
using var manager = new CollectionIndexManager<int, Person>(storage, mapper, "people_idx_pref_unique");
|
||||
|
||||
manager.CreateIndex(p => p.Age, name: "idx_age", unique: false);
|
||||
manager.CreateIndex(p => p.Age, name: "idx_age_unique", unique: true);
|
||||
|
||||
var best = manager.FindBestIndex("Age");
|
||||
|
||||
best.ShouldNotBeNull();
|
||||
best.Definition.Name.ShouldBe("idx_age_unique");
|
||||
best.Definition.IsUnique.ShouldBeTrue();
|
||||
}
|
||||
finally
|
||||
{
|
||||
CleanupFiles(dbPath);
|
||||
}
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void FindBestCompoundIndex_Should_Choose_Longest_Prefix()
|
||||
{
|
||||
var dbPath = NewDbPath();
|
||||
try
|
||||
{
|
||||
using var storage = new StorageEngine(dbPath, PageFileConfig.Default);
|
||||
var mapper = new ZB_MOM_WW_CBDD_Shared_PersonMapper();
|
||||
using var manager = new CollectionIndexManager<int, Person>(storage, mapper, "people_idx_compound");
|
||||
|
||||
manager.CreateIndex(new CollectionIndexDefinition<Person>(
|
||||
"idx_name",
|
||||
["Name"],
|
||||
p => p.Name));
|
||||
|
||||
manager.CreateIndex(new CollectionIndexDefinition<Person>(
|
||||
"idx_name_age",
|
||||
["Name", "Age"],
|
||||
p => new { p.Name, p.Age }));
|
||||
|
||||
manager.CreateIndex(new CollectionIndexDefinition<Person>(
|
||||
"idx_name_age_id",
|
||||
["Name", "Age", "Id"],
|
||||
p => new { p.Name, p.Age, p.Id }));
|
||||
|
||||
var best = manager.FindBestCompoundIndex(["Name", "Age"]);
|
||||
|
||||
best.ShouldNotBeNull();
|
||||
best.Definition.Name.ShouldBe("idx_name_age_id");
|
||||
best.Definition.PropertyPaths.Length.ShouldBe(3);
|
||||
}
|
||||
finally
|
||||
{
|
||||
CleanupFiles(dbPath);
|
||||
}
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void DropIndex_Should_Remove_Metadata_And_Be_Idempotent()
|
||||
{
|
||||
var dbPath = NewDbPath();
|
||||
const string collectionName = "people_idx_drop";
|
||||
|
||||
try
|
||||
{
|
||||
using var storage = new StorageEngine(dbPath, PageFileConfig.Default);
|
||||
var mapper = new ZB_MOM_WW_CBDD_Shared_PersonMapper();
|
||||
|
||||
using (var manager = new CollectionIndexManager<int, Person>(storage, mapper, collectionName))
|
||||
{
|
||||
manager.CreateIndex(p => p.Age, name: "idx_age", unique: false);
|
||||
manager.DropIndex("idx_age").ShouldBeTrue();
|
||||
manager.DropIndex("idx_age").ShouldBeFalse();
|
||||
manager.GetIndexInfo().ShouldBeEmpty();
|
||||
}
|
||||
|
||||
using var reloaded = new CollectionIndexManager<int, Person>(storage, mapper, collectionName);
|
||||
reloaded.GetIndexInfo().ShouldBeEmpty();
|
||||
}
|
||||
finally
|
||||
{
|
||||
CleanupFiles(dbPath);
|
||||
}
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void CollectionIndexDefinition_Should_Respect_Query_Support_Rules()
|
||||
{
|
||||
var definition = new CollectionIndexDefinition<Person>(
|
||||
"idx_name_age",
|
||||
["Name", "Age"],
|
||||
p => new { p.Name, p.Age });
|
||||
|
||||
definition.CanSupportQuery("Name").ShouldBeTrue();
|
||||
definition.CanSupportQuery("Age").ShouldBeFalse();
|
||||
|
||||
definition.CanSupportCompoundQuery(["Name"]).ShouldBeTrue();
|
||||
definition.CanSupportCompoundQuery(["Name", "Age"]).ShouldBeTrue();
|
||||
definition.CanSupportCompoundQuery(["Name", "Age", "Id"]).ShouldBeFalse();
|
||||
|
||||
definition.ToString().ShouldContain("idx_name_age");
|
||||
definition.ToString().ShouldContain("Name");
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void CollectionIndexInfo_ToString_Should_Include_Diagnostics()
|
||||
{
|
||||
var info = new CollectionIndexInfo
|
||||
{
|
||||
Name = "idx_age",
|
||||
PropertyPaths = ["Age"],
|
||||
EstimatedDocumentCount = 12,
|
||||
EstimatedSizeBytes = 4096
|
||||
};
|
||||
|
||||
var text = info.ToString();
|
||||
|
||||
text.ShouldContain("idx_age");
|
||||
text.ShouldContain("Age");
|
||||
text.ShouldContain("12 docs");
|
||||
}
|
||||
|
||||
private static string NewDbPath()
|
||||
=> Path.Combine(Path.GetTempPath(), $"idx_mgr_{Guid.NewGuid():N}.db");
|
||||
|
||||
private static void CleanupFiles(string dbPath)
|
||||
{
|
||||
if (File.Exists(dbPath)) File.Delete(dbPath);
|
||||
|
||||
var walPath = Path.ChangeExtension(dbPath, ".wal");
|
||||
if (File.Exists(walPath)) File.Delete(walPath);
|
||||
|
||||
var altWalPath = dbPath + "-wal";
|
||||
if (File.Exists(altWalPath)) File.Delete(altWalPath);
|
||||
}
|
||||
}
|
||||
127
tests/CBDD.Tests/CompactionCrashRecoveryTests.cs
Normal file
127
tests/CBDD.Tests/CompactionCrashRecoveryTests.cs
Normal file
@@ -0,0 +1,127 @@
|
||||
using ZB.MOM.WW.CBDD.Bson;
|
||||
using ZB.MOM.WW.CBDD.Core.Storage;
|
||||
using ZB.MOM.WW.CBDD.Shared;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests;
|
||||
|
||||
public class CompactionCrashRecoveryTests
|
||||
{
|
||||
[Theory]
|
||||
[InlineData("Started")]
|
||||
[InlineData("Copied")]
|
||||
[InlineData("Swapped")]
|
||||
public void ResumeCompaction_FromCrashMarkerPhases_ShouldFinalizeAndPreserveData(string phase)
|
||||
{
|
||||
var dbPath = NewDbPath();
|
||||
var markerPath = MarkerPath(dbPath);
|
||||
|
||||
try
|
||||
{
|
||||
using var db = new TestDbContext(dbPath);
|
||||
|
||||
var ids = SeedData(db);
|
||||
db.ForceCheckpoint();
|
||||
WriteMarker(markerPath, dbPath, phase);
|
||||
|
||||
var resumed = db.Storage.ResumeCompactionIfNeeded(new CompactionOptions
|
||||
{
|
||||
EnableTailTruncation = true,
|
||||
DefragmentSlottedPages = true,
|
||||
NormalizeFreeList = true
|
||||
});
|
||||
|
||||
resumed.ShouldNotBeNull();
|
||||
resumed!.ResumedFromMarker.ShouldBeTrue();
|
||||
File.Exists(markerPath).ShouldBeFalse();
|
||||
|
||||
db.Users.Count().ShouldBe(ids.Count);
|
||||
var recoveredDoc = ids
|
||||
.Select(id => db.Users.FindById(id))
|
||||
.FirstOrDefault(x => x != null);
|
||||
recoveredDoc.ShouldNotBeNull();
|
||||
recoveredDoc!.Name.ShouldContain("user-");
|
||||
|
||||
db.Storage.ResumeCompactionIfNeeded().ShouldBeNull();
|
||||
}
|
||||
finally
|
||||
{
|
||||
CleanupFiles(dbPath);
|
||||
}
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void ResumeCompaction_WithCorruptedMarker_ShouldRecoverDeterministically()
|
||||
{
|
||||
var dbPath = NewDbPath();
|
||||
var markerPath = MarkerPath(dbPath);
|
||||
|
||||
try
|
||||
{
|
||||
using var db = new TestDbContext(dbPath);
|
||||
var ids = SeedData(db);
|
||||
db.ForceCheckpoint();
|
||||
|
||||
File.WriteAllText(markerPath, "{invalid-json-marker");
|
||||
|
||||
var resumed = db.Storage.ResumeCompactionIfNeeded(new CompactionOptions
|
||||
{
|
||||
EnableTailTruncation = true
|
||||
});
|
||||
|
||||
resumed.ShouldNotBeNull();
|
||||
resumed!.ResumedFromMarker.ShouldBeTrue();
|
||||
File.Exists(markerPath).ShouldBeFalse();
|
||||
|
||||
db.Users.Count().ShouldBe(ids.Count);
|
||||
var recoveredDoc = ids
|
||||
.Select(id => db.Users.FindById(id))
|
||||
.FirstOrDefault(x => x != null);
|
||||
recoveredDoc.ShouldNotBeNull();
|
||||
recoveredDoc!.Name.ShouldContain("user-");
|
||||
}
|
||||
finally
|
||||
{
|
||||
CleanupFiles(dbPath);
|
||||
}
|
||||
}
|
||||
|
||||
private static List<ObjectId> SeedData(TestDbContext db)
|
||||
{
|
||||
var ids = new List<ObjectId>();
|
||||
for (var i = 0; i < 120; i++)
|
||||
{
|
||||
ids.Add(db.Users.Insert(new User
|
||||
{
|
||||
Name = $"user-{i:D4}-payload-{new string('x', 120)}",
|
||||
Age = i % 20
|
||||
}));
|
||||
}
|
||||
|
||||
db.SaveChanges();
|
||||
return ids;
|
||||
}
|
||||
|
||||
private static void WriteMarker(string markerPath, string dbPath, string phase)
|
||||
{
|
||||
var safeDbPath = dbPath.Replace("\\", "\\\\", StringComparison.Ordinal);
|
||||
var now = DateTimeOffset.UtcNow.ToString("O");
|
||||
var json = $$"""
|
||||
{"version":1,"phase":"{{phase}}","databasePath":"{{safeDbPath}}","startedAtUtc":"{{now}}","lastUpdatedUtc":"{{now}}","onlineMode":false,"mode":"InPlace"}
|
||||
""";
|
||||
File.WriteAllText(markerPath, json);
|
||||
}
|
||||
|
||||
private static string MarkerPath(string dbPath) => $"{dbPath}.compact.state";
|
||||
|
||||
private static string NewDbPath()
|
||||
=> Path.Combine(Path.GetTempPath(), $"compaction_crash_{Guid.NewGuid():N}.db");
|
||||
|
||||
private static void CleanupFiles(string dbPath)
|
||||
{
|
||||
var walPath = Path.ChangeExtension(dbPath, ".wal");
|
||||
var markerPath = MarkerPath(dbPath);
|
||||
if (File.Exists(dbPath)) File.Delete(dbPath);
|
||||
if (File.Exists(walPath)) File.Delete(walPath);
|
||||
if (File.Exists(markerPath)) File.Delete(markerPath);
|
||||
}
|
||||
}
|
||||
379
tests/CBDD.Tests/CompactionOfflineTests.cs
Normal file
379
tests/CBDD.Tests/CompactionOfflineTests.cs
Normal file
@@ -0,0 +1,379 @@
|
||||
using System.IO.MemoryMappedFiles;
|
||||
using ZB.MOM.WW.CBDD.Bson;
|
||||
using ZB.MOM.WW.CBDD.Core.Indexing;
|
||||
using ZB.MOM.WW.CBDD.Core.Storage;
|
||||
using ZB.MOM.WW.CBDD.Shared;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests;
|
||||
|
||||
public class CompactionOfflineTests
|
||||
{
|
||||
[Fact]
|
||||
public void OfflineCompact_ShouldPreserveLogicalDataEquivalence()
|
||||
{
|
||||
var dbPath = NewDbPath();
|
||||
|
||||
try
|
||||
{
|
||||
using var db = new TestDbContext(dbPath);
|
||||
|
||||
var ids = new List<ObjectId>();
|
||||
for (var i = 0; i < 160; i++)
|
||||
{
|
||||
ids.Add(db.Users.Insert(new User { Name = $"user-{i:D4}", Age = i % 31 }));
|
||||
}
|
||||
|
||||
for (var i = 0; i < ids.Count; i += 9)
|
||||
{
|
||||
if (db.Users.FindById(ids[i]) != null)
|
||||
{
|
||||
db.Users.Delete(ids[i]).ShouldBeTrue();
|
||||
}
|
||||
}
|
||||
|
||||
var updateTargets = db.Users.FindAll(u => u.Age % 4 == 0)
|
||||
.Select(u => u.Id)
|
||||
.ToList();
|
||||
foreach (var id in updateTargets)
|
||||
{
|
||||
var user = db.Users.FindById(id);
|
||||
if (user == null)
|
||||
{
|
||||
continue;
|
||||
}
|
||||
|
||||
user.Name += "-updated";
|
||||
db.Users.Update(user).ShouldBeTrue();
|
||||
}
|
||||
|
||||
db.SaveChanges();
|
||||
db.ForceCheckpoint();
|
||||
|
||||
var expected = db.Users.FindAll()
|
||||
.ToDictionary(u => u.Id, u => (u.Name, u.Age));
|
||||
db.SaveChanges();
|
||||
|
||||
var stats = db.Compact();
|
||||
stats.OnlineMode.ShouldBeFalse();
|
||||
|
||||
var actual = db.Users.FindAll()
|
||||
.ToDictionary(u => u.Id, u => (u.Name, u.Age));
|
||||
|
||||
actual.Count.ShouldBe(expected.Count);
|
||||
foreach (var kvp in expected)
|
||||
{
|
||||
actual.ShouldContainKey(kvp.Key);
|
||||
actual[kvp.Key].ShouldBe(kvp.Value);
|
||||
}
|
||||
}
|
||||
finally
|
||||
{
|
||||
CleanupFiles(dbPath);
|
||||
}
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void OfflineCompact_ShouldKeepIndexResultsConsistent()
|
||||
{
|
||||
var dbPath = NewDbPath();
|
||||
|
||||
try
|
||||
{
|
||||
using var db = new TestDbContext(dbPath);
|
||||
|
||||
for (var i = 0; i < 300; i++)
|
||||
{
|
||||
db.People.Insert(new Person
|
||||
{
|
||||
Name = $"person-{i:D4}",
|
||||
Age = i % 12
|
||||
});
|
||||
}
|
||||
|
||||
db.SaveChanges();
|
||||
db.ForceCheckpoint();
|
||||
|
||||
var expectedByAge = db.People.FindAll()
|
||||
.GroupBy(p => p.Age)
|
||||
.ToDictionary(g => g.Key, g => g.Select(x => x.Name).OrderBy(x => x).ToArray());
|
||||
db.SaveChanges();
|
||||
|
||||
var indexNamesBefore = db.People.GetIndexes().Select(x => x.Name).OrderBy(x => x).ToArray();
|
||||
|
||||
var stats = db.Compact(new CompactionOptions
|
||||
{
|
||||
DefragmentSlottedPages = true,
|
||||
NormalizeFreeList = true,
|
||||
EnableTailTruncation = true
|
||||
});
|
||||
stats.PrePageCount.ShouldBeGreaterThanOrEqualTo(stats.PostPageCount);
|
||||
|
||||
var indexNamesAfter = db.People.GetIndexes().Select(x => x.Name).OrderBy(x => x).ToArray();
|
||||
indexNamesAfter.ShouldBe(indexNamesBefore);
|
||||
|
||||
foreach (var age in expectedByAge.Keys.OrderBy(x => x))
|
||||
{
|
||||
var actual = db.People.FindAll(p => p.Age == age)
|
||||
.Select(x => x.Name)
|
||||
.OrderBy(x => x)
|
||||
.ToArray();
|
||||
|
||||
actual.ShouldBe(expectedByAge[age]);
|
||||
}
|
||||
}
|
||||
finally
|
||||
{
|
||||
CleanupFiles(dbPath);
|
||||
}
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void OfflineCompact_WhenTailIsReclaimable_ShouldReduceFileSize()
|
||||
{
|
||||
var dbPath = NewDbPath();
|
||||
var ids = new List<ObjectId>();
|
||||
|
||||
try
|
||||
{
|
||||
using var db = new TestDbContext(dbPath, SmallPageConfig());
|
||||
|
||||
for (var i = 0; i < 240; i++)
|
||||
{
|
||||
var id = db.Users.Insert(new User
|
||||
{
|
||||
Name = BuildPayload(i, 18_000),
|
||||
Age = i
|
||||
});
|
||||
ids.Add(id);
|
||||
}
|
||||
|
||||
db.SaveChanges();
|
||||
db.ForceCheckpoint();
|
||||
|
||||
for (var i = ids.Count - 1; i >= 60; i--)
|
||||
{
|
||||
if (db.Users.FindById(ids[i]) != null)
|
||||
{
|
||||
db.Users.Delete(ids[i]).ShouldBeTrue();
|
||||
}
|
||||
}
|
||||
|
||||
db.SaveChanges();
|
||||
db.ForceCheckpoint();
|
||||
|
||||
var preCompactSize = new FileInfo(dbPath).Length;
|
||||
var stats = db.Compact(new CompactionOptions
|
||||
{
|
||||
EnableTailTruncation = true,
|
||||
MinimumRetainedPages = 2
|
||||
});
|
||||
var postCompactSize = new FileInfo(dbPath).Length;
|
||||
|
||||
postCompactSize.ShouldBeLessThanOrEqualTo(preCompactSize);
|
||||
stats.ReclaimedFileBytes.ShouldBeGreaterThanOrEqualTo(0);
|
||||
}
|
||||
finally
|
||||
{
|
||||
CleanupFiles(dbPath);
|
||||
}
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void OfflineCompact_WithInvalidPrimaryRootMetadata_ShouldFailValidation()
|
||||
{
|
||||
var dbPath = NewDbPath();
|
||||
|
||||
try
|
||||
{
|
||||
using var db = new TestDbContext(dbPath);
|
||||
for (var i = 0; i < 32; i++)
|
||||
{
|
||||
db.Users.Insert(new User { Name = $"invalid-primary-{i:D3}", Age = i });
|
||||
}
|
||||
|
||||
db.SaveChanges();
|
||||
db.ForceCheckpoint();
|
||||
|
||||
var metadata = db.Storage.GetCollectionMetadata("users");
|
||||
metadata.ShouldNotBeNull();
|
||||
metadata!.PrimaryRootPageId = 1; // Metadata page, not an index page.
|
||||
db.Storage.SaveCollectionMetadata(metadata);
|
||||
|
||||
Should.Throw<InvalidDataException>(() => db.Compact())
|
||||
.Message.ShouldContain("primary index root page id");
|
||||
}
|
||||
finally
|
||||
{
|
||||
CleanupFiles(dbPath);
|
||||
}
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void OfflineCompact_WithInvalidSecondaryRootMetadata_ShouldFailValidation()
|
||||
{
|
||||
var dbPath = NewDbPath();
|
||||
|
||||
try
|
||||
{
|
||||
using var db = new TestDbContext(dbPath);
|
||||
for (var i = 0; i < 48; i++)
|
||||
{
|
||||
db.People.Insert(new Person { Name = $"invalid-secondary-{i:D3}", Age = i % 10 });
|
||||
}
|
||||
|
||||
db.SaveChanges();
|
||||
db.ForceCheckpoint();
|
||||
|
||||
var metadata = db.Storage.GetCollectionMetadata("people_collection");
|
||||
metadata.ShouldNotBeNull();
|
||||
metadata!.Indexes.Count.ShouldBeGreaterThan(0);
|
||||
metadata.Indexes[0].RootPageId = uint.MaxValue; // Out-of-range page id.
|
||||
db.Storage.SaveCollectionMetadata(metadata);
|
||||
|
||||
Should.Throw<InvalidDataException>(() => db.Compact())
|
||||
.Message.ShouldContain("out of range");
|
||||
}
|
||||
finally
|
||||
{
|
||||
CleanupFiles(dbPath);
|
||||
}
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void OfflineCompact_ShouldReportLiveBytesRelocationAndThroughputTelemetry()
|
||||
{
|
||||
var dbPath = NewDbPath();
|
||||
|
||||
try
|
||||
{
|
||||
using var db = new TestDbContext(dbPath, SmallPageConfig());
|
||||
|
||||
var ids = new List<ObjectId>();
|
||||
for (var i = 0; i < 160; i++)
|
||||
{
|
||||
ids.Add(db.Users.Insert(new User
|
||||
{
|
||||
Name = BuildPayload(i, 9_000),
|
||||
Age = i
|
||||
}));
|
||||
}
|
||||
|
||||
for (var i = 0; i < ids.Count; i += 7)
|
||||
{
|
||||
if (db.Users.FindById(ids[i]) != null)
|
||||
{
|
||||
db.Users.Delete(ids[i]).ShouldBeTrue();
|
||||
}
|
||||
}
|
||||
|
||||
db.SaveChanges();
|
||||
db.ForceCheckpoint();
|
||||
|
||||
var stats = db.Compact(new CompactionOptions
|
||||
{
|
||||
DefragmentSlottedPages = true,
|
||||
NormalizeFreeList = true,
|
||||
EnableTailTruncation = true
|
||||
});
|
||||
|
||||
stats.PreLiveBytes.ShouldBe(Math.Max(0, stats.PreFileSizeBytes - stats.PreFreeBytes));
|
||||
stats.PostLiveBytes.ShouldBe(Math.Max(0, stats.PostFileSizeBytes - stats.PostFreeBytes));
|
||||
stats.DocumentsRelocated.ShouldBeGreaterThanOrEqualTo(0);
|
||||
stats.PagesRelocated.ShouldBeGreaterThanOrEqualTo(0);
|
||||
stats.ThroughputBytesPerSecond.ShouldBeGreaterThan(0);
|
||||
stats.ThroughputPagesPerSecond.ShouldBeGreaterThanOrEqualTo(0);
|
||||
stats.ThroughputDocumentsPerSecond.ShouldBeGreaterThanOrEqualTo(0);
|
||||
}
|
||||
finally
|
||||
{
|
||||
CleanupFiles(dbPath);
|
||||
}
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void OfflineCompact_WhenPrimaryIndexPointsToDeletedSlot_ShouldFailValidation()
|
||||
{
|
||||
var dbPath = NewDbPath();
|
||||
|
||||
try
|
||||
{
|
||||
using var db = new TestDbContext(dbPath, SmallPageConfig());
|
||||
var id = db.Users.Insert(new User { Name = BuildPayload(1, 7_500), Age = 9 });
|
||||
db.SaveChanges();
|
||||
db.ForceCheckpoint();
|
||||
|
||||
var metadata = db.Storage.GetCollectionMetadata("users");
|
||||
metadata.ShouldNotBeNull();
|
||||
metadata!.PrimaryRootPageId.ShouldBeGreaterThan(0u);
|
||||
|
||||
var primaryIndex = new BTreeIndex(db.Storage, IndexOptions.CreateUnique("_id"), metadata.PrimaryRootPageId);
|
||||
primaryIndex.TryFind(new IndexKey(id), out var location).ShouldBeTrue();
|
||||
|
||||
var page = new byte[db.Storage.PageSize];
|
||||
db.Storage.ReadPage(location.PageId, null, page);
|
||||
|
||||
var header = SlottedPageHeader.ReadFrom(page);
|
||||
var slotOffset = SlottedPageHeader.Size + (location.SlotIndex * SlotEntry.Size);
|
||||
var slot = SlotEntry.ReadFrom(page.AsSpan(slotOffset, SlotEntry.Size));
|
||||
slot.Flags |= SlotFlags.Deleted;
|
||||
slot.WriteTo(page.AsSpan(slotOffset, SlotEntry.Size));
|
||||
header.WriteTo(page);
|
||||
db.Storage.WritePageImmediate(location.PageId, page);
|
||||
|
||||
var ex = Should.Throw<InvalidDataException>(() => db.Compact(new CompactionOptions
|
||||
{
|
||||
DefragmentSlottedPages = true,
|
||||
NormalizeFreeList = true,
|
||||
EnableTailTruncation = true
|
||||
}));
|
||||
ex.Message.ShouldContain("Compaction validation failed");
|
||||
}
|
||||
finally
|
||||
{
|
||||
CleanupFiles(dbPath);
|
||||
}
|
||||
}
|
||||
|
||||
private static PageFileConfig SmallPageConfig()
|
||||
{
|
||||
return new PageFileConfig
|
||||
{
|
||||
PageSize = 4096,
|
||||
InitialFileSize = 1024 * 1024,
|
||||
Access = MemoryMappedFileAccess.ReadWrite
|
||||
};
|
||||
}
|
||||
|
||||
private static string BuildPayload(int seed, int approxLength)
|
||||
{
|
||||
var builder = new System.Text.StringBuilder(approxLength + 256);
|
||||
var i = 0;
|
||||
while (builder.Length < approxLength)
|
||||
{
|
||||
builder.Append("compact-tail-");
|
||||
builder.Append(seed.ToString("D4"));
|
||||
builder.Append('-');
|
||||
builder.Append(i.ToString("D6"));
|
||||
builder.Append('|');
|
||||
i++;
|
||||
}
|
||||
|
||||
return builder.ToString();
|
||||
}
|
||||
|
||||
private static string NewDbPath()
|
||||
=> Path.Combine(Path.GetTempPath(), $"compaction_offline_{Guid.NewGuid():N}.db");
|
||||
|
||||
private static void CleanupFiles(string dbPath)
|
||||
{
|
||||
var walPath = Path.ChangeExtension(dbPath, ".wal");
|
||||
var markerPath = $"{dbPath}.compact.state";
|
||||
var tempPath = $"{dbPath}.compact.tmp";
|
||||
var backupPath = $"{dbPath}.compact.bak";
|
||||
if (File.Exists(dbPath)) File.Delete(dbPath);
|
||||
if (File.Exists(walPath)) File.Delete(walPath);
|
||||
if (File.Exists(markerPath)) File.Delete(markerPath);
|
||||
if (File.Exists(tempPath)) File.Delete(tempPath);
|
||||
if (File.Exists(backupPath)) File.Delete(backupPath);
|
||||
}
|
||||
}
|
||||
137
tests/CBDD.Tests/CompactionOnlineConcurrencyTests.cs
Normal file
137
tests/CBDD.Tests/CompactionOnlineConcurrencyTests.cs
Normal file
@@ -0,0 +1,137 @@
|
||||
using ZB.MOM.WW.CBDD.Bson;
|
||||
using ZB.MOM.WW.CBDD.Core.Storage;
|
||||
using ZB.MOM.WW.CBDD.Shared;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests;
|
||||
|
||||
public class CompactionOnlineConcurrencyTests
|
||||
{
|
||||
[Fact]
|
||||
public async Task OnlineCompaction_WithConcurrentishWorkload_ShouldCompleteWithoutDeadlock()
|
||||
{
|
||||
var dbPath = NewDbPath();
|
||||
var activeIds = new List<ObjectId>();
|
||||
var sync = new object();
|
||||
var completedOps = 0;
|
||||
|
||||
try
|
||||
{
|
||||
using var db = new TestDbContext(dbPath);
|
||||
var testCancellation = TestContext.Current.CancellationToken;
|
||||
|
||||
for (var i = 0; i < 120; i++)
|
||||
{
|
||||
var id = db.Users.Insert(new User { Name = $"seed-{i:D4}", Age = i % 40 });
|
||||
activeIds.Add(id);
|
||||
}
|
||||
|
||||
db.SaveChanges();
|
||||
db.ForceCheckpoint();
|
||||
|
||||
var workloadTask = Task.Run(() =>
|
||||
{
|
||||
for (var i = 0; i < 150; i++)
|
||||
{
|
||||
if (i % 3 == 0)
|
||||
{
|
||||
var id = db.Users.Insert(new User { Name = $"insert-{i:D4}", Age = i % 60 });
|
||||
lock (sync)
|
||||
{
|
||||
activeIds.Add(id);
|
||||
}
|
||||
}
|
||||
else if (i % 3 == 1)
|
||||
{
|
||||
ObjectId? candidate = null;
|
||||
lock (sync)
|
||||
{
|
||||
if (activeIds.Count > 0)
|
||||
{
|
||||
candidate = activeIds[i % activeIds.Count];
|
||||
}
|
||||
}
|
||||
|
||||
if (candidate.HasValue)
|
||||
{
|
||||
var entity = db.Users.FindById(candidate.Value);
|
||||
if (entity != null)
|
||||
{
|
||||
entity.Age += 1;
|
||||
db.Users.Update(entity).ShouldBeTrue();
|
||||
}
|
||||
}
|
||||
}
|
||||
else
|
||||
{
|
||||
ObjectId? candidate = null;
|
||||
lock (sync)
|
||||
{
|
||||
if (activeIds.Count > 60)
|
||||
{
|
||||
candidate = activeIds[^1];
|
||||
activeIds.RemoveAt(activeIds.Count - 1);
|
||||
}
|
||||
}
|
||||
|
||||
if (candidate.HasValue)
|
||||
{
|
||||
db.Users.Delete(candidate.Value);
|
||||
}
|
||||
}
|
||||
|
||||
db.SaveChanges();
|
||||
_ = db.Users.Count();
|
||||
db.SaveChanges();
|
||||
Interlocked.Increment(ref completedOps);
|
||||
}
|
||||
}, testCancellation);
|
||||
|
||||
var compactionTask = Task.Run(() => db.Compact(new CompactionOptions
|
||||
{
|
||||
OnlineMode = true,
|
||||
OnlineBatchPageLimit = 4,
|
||||
OnlineBatchDelay = TimeSpan.FromMilliseconds(2),
|
||||
MaxOnlineDuration = TimeSpan.FromMilliseconds(400),
|
||||
EnableTailTruncation = true
|
||||
}), testCancellation);
|
||||
|
||||
await Task.WhenAll(workloadTask, compactionTask).WaitAsync(TimeSpan.FromSeconds(20), testCancellation);
|
||||
|
||||
var stats = await compactionTask;
|
||||
stats.OnlineMode.ShouldBeTrue();
|
||||
completedOps.ShouldBeGreaterThanOrEqualTo(100);
|
||||
|
||||
var allUsers = db.Users.FindAll().ToList();
|
||||
allUsers.Count.ShouldBeGreaterThan(0);
|
||||
db.SaveChanges();
|
||||
|
||||
List<ObjectId> snapshotIds;
|
||||
lock (sync)
|
||||
{
|
||||
snapshotIds = activeIds.ToList();
|
||||
}
|
||||
|
||||
var actualIds = allUsers.Select(x => x.Id).ToHashSet();
|
||||
foreach (var id in snapshotIds)
|
||||
{
|
||||
actualIds.ShouldContain(id);
|
||||
}
|
||||
}
|
||||
finally
|
||||
{
|
||||
CleanupFiles(dbPath);
|
||||
}
|
||||
}
|
||||
|
||||
private static string NewDbPath()
|
||||
=> Path.Combine(Path.GetTempPath(), $"compaction_online_{Guid.NewGuid():N}.db");
|
||||
|
||||
private static void CleanupFiles(string dbPath)
|
||||
{
|
||||
var walPath = Path.ChangeExtension(dbPath, ".wal");
|
||||
var markerPath = $"{dbPath}.compact.state";
|
||||
if (File.Exists(dbPath)) File.Delete(dbPath);
|
||||
if (File.Exists(walPath)) File.Delete(walPath);
|
||||
if (File.Exists(markerPath)) File.Delete(markerPath);
|
||||
}
|
||||
}
|
||||
109
tests/CBDD.Tests/CompactionWalCoordinationTests.cs
Normal file
109
tests/CBDD.Tests/CompactionWalCoordinationTests.cs
Normal file
@@ -0,0 +1,109 @@
|
||||
using ZB.MOM.WW.CBDD.Bson;
|
||||
using ZB.MOM.WW.CBDD.Core.Storage;
|
||||
using ZB.MOM.WW.CBDD.Shared;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests;
|
||||
|
||||
public class CompactionWalCoordinationTests
|
||||
{
|
||||
[Fact]
|
||||
public void OfflineCompact_ShouldCheckpointAndLeaveWalEmpty()
|
||||
{
|
||||
var dbPath = NewDbPath();
|
||||
var markerPath = $"{dbPath}.compact.state";
|
||||
|
||||
try
|
||||
{
|
||||
using var db = new TestDbContext(dbPath);
|
||||
for (var i = 0; i < 80; i++)
|
||||
{
|
||||
db.Users.Insert(new User { Name = $"wal-compact-{i:D3}", Age = i });
|
||||
}
|
||||
|
||||
db.SaveChanges();
|
||||
db.Storage.GetWalSize().ShouldBeGreaterThan(0);
|
||||
|
||||
var stats = db.Compact(new CompactionOptions
|
||||
{
|
||||
EnableTailTruncation = true,
|
||||
NormalizeFreeList = true,
|
||||
DefragmentSlottedPages = true
|
||||
});
|
||||
|
||||
stats.OnlineMode.ShouldBeFalse();
|
||||
db.Storage.GetWalSize().ShouldBe(0);
|
||||
File.Exists(markerPath).ShouldBeFalse();
|
||||
|
||||
db.Users.Count().ShouldBe(80);
|
||||
}
|
||||
finally
|
||||
{
|
||||
CleanupFiles(dbPath);
|
||||
}
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Compact_AfterWalRecovery_ShouldKeepDataDurable()
|
||||
{
|
||||
var dbPath = NewDbPath();
|
||||
var walPath = Path.ChangeExtension(dbPath, ".wal");
|
||||
var expected = new List<(ObjectId Id, string Name)>();
|
||||
|
||||
try
|
||||
{
|
||||
using (var writer = new TestDbContext(dbPath))
|
||||
{
|
||||
for (var i = 0; i < 48; i++)
|
||||
{
|
||||
var name = $"recoverable-{i:D3}";
|
||||
var id = writer.Users.Insert(new User { Name = name, Age = i % 13 });
|
||||
expected.Add((id, name));
|
||||
}
|
||||
|
||||
writer.SaveChanges();
|
||||
writer.Storage.GetWalSize().ShouldBeGreaterThan(0);
|
||||
}
|
||||
|
||||
new FileInfo(walPath).Length.ShouldBeGreaterThan(0);
|
||||
|
||||
using (var recovered = new TestDbContext(dbPath))
|
||||
{
|
||||
recovered.Users.Count().ShouldBe(expected.Count);
|
||||
|
||||
foreach (var item in expected)
|
||||
{
|
||||
recovered.Users.FindById(item.Id)!.Name.ShouldBe(item.Name);
|
||||
}
|
||||
|
||||
recovered.SaveChanges();
|
||||
recovered.Compact();
|
||||
recovered.Storage.GetWalSize().ShouldBe(0);
|
||||
}
|
||||
|
||||
using (var verify = new TestDbContext(dbPath))
|
||||
{
|
||||
verify.Users.Count().ShouldBe(expected.Count);
|
||||
foreach (var item in expected)
|
||||
{
|
||||
verify.Users.FindById(item.Id)!.Name.ShouldBe(item.Name);
|
||||
}
|
||||
}
|
||||
}
|
||||
finally
|
||||
{
|
||||
CleanupFiles(dbPath);
|
||||
}
|
||||
}
|
||||
|
||||
private static string NewDbPath()
|
||||
=> Path.Combine(Path.GetTempPath(), $"compaction_wal_{Guid.NewGuid():N}.db");
|
||||
|
||||
private static void CleanupFiles(string dbPath)
|
||||
{
|
||||
var walPath = Path.ChangeExtension(dbPath, ".wal");
|
||||
var markerPath = $"{dbPath}.compact.state";
|
||||
if (File.Exists(dbPath)) File.Delete(dbPath);
|
||||
if (File.Exists(walPath)) File.Delete(walPath);
|
||||
if (File.Exists(markerPath)) File.Delete(markerPath);
|
||||
}
|
||||
}
|
||||
170
tests/CBDD.Tests/CompressionCompatibilityTests.cs
Normal file
170
tests/CBDD.Tests/CompressionCompatibilityTests.cs
Normal file
@@ -0,0 +1,170 @@
|
||||
using System.IO.Compression;
|
||||
using System.Security.Cryptography;
|
||||
using ZB.MOM.WW.CBDD.Bson;
|
||||
using ZB.MOM.WW.CBDD.Core.Compression;
|
||||
using ZB.MOM.WW.CBDD.Core.Storage;
|
||||
using ZB.MOM.WW.CBDD.Shared;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests;
|
||||
|
||||
public class CompressionCompatibilityTests
|
||||
{
|
||||
[Fact]
|
||||
public void OpeningLegacyUncompressedFile_WithCompressionEnabled_ShouldNotMutateDbFile()
|
||||
{
|
||||
var dbPath = NewDbPath();
|
||||
var idList = new List<ObjectId>();
|
||||
|
||||
try
|
||||
{
|
||||
using (var db = new TestDbContext(dbPath))
|
||||
{
|
||||
idList.Add(db.Users.Insert(new User { Name = "legacy-a", Age = 10 }));
|
||||
idList.Add(db.Users.Insert(new User { Name = "legacy-b", Age = 11 }));
|
||||
db.SaveChanges();
|
||||
db.ForceCheckpoint();
|
||||
}
|
||||
|
||||
var beforeSize = new FileInfo(dbPath).Length;
|
||||
var beforeHash = ComputeFileHash(dbPath);
|
||||
|
||||
var compressionOptions = new CompressionOptions
|
||||
{
|
||||
EnableCompression = true,
|
||||
MinSizeBytes = 0,
|
||||
MinSavingsPercent = 0,
|
||||
Codec = CompressionCodec.Brotli,
|
||||
Level = CompressionLevel.Fastest
|
||||
};
|
||||
|
||||
using (var reopened = new TestDbContext(dbPath, compressionOptions))
|
||||
{
|
||||
reopened.Users.FindById(idList[0])!.Name.ShouldBe("legacy-a");
|
||||
reopened.Users.FindById(idList[1])!.Name.ShouldBe("legacy-b");
|
||||
reopened.Users.Count().ShouldBe(2);
|
||||
}
|
||||
|
||||
var afterSize = new FileInfo(dbPath).Length;
|
||||
var afterHash = ComputeFileHash(dbPath);
|
||||
|
||||
afterSize.ShouldBe(beforeSize);
|
||||
afterHash.ShouldBe(beforeHash);
|
||||
}
|
||||
finally
|
||||
{
|
||||
CleanupFiles(dbPath);
|
||||
}
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void MixedFormatDocuments_ShouldRemainReadableAfterPartialMigration()
|
||||
{
|
||||
var dbPath = NewDbPath();
|
||||
ObjectId legacyId;
|
||||
ObjectId compressedId;
|
||||
|
||||
try
|
||||
{
|
||||
using (var db = new TestDbContext(dbPath))
|
||||
{
|
||||
legacyId = db.Users.Insert(new User { Name = "legacy-uncompressed", Age = 22 });
|
||||
db.SaveChanges();
|
||||
db.ForceCheckpoint();
|
||||
}
|
||||
|
||||
var compressionOptions = new CompressionOptions
|
||||
{
|
||||
EnableCompression = true,
|
||||
MinSizeBytes = 0,
|
||||
MinSavingsPercent = 0,
|
||||
Codec = CompressionCodec.Brotli,
|
||||
Level = CompressionLevel.Fastest
|
||||
};
|
||||
|
||||
using (var migrated = new TestDbContext(dbPath, compressionOptions))
|
||||
{
|
||||
compressedId = migrated.Users.Insert(new User { Name = BuildPayload(24_000), Age = 33 });
|
||||
migrated.SaveChanges();
|
||||
migrated.ForceCheckpoint();
|
||||
}
|
||||
|
||||
using (var verify = new TestDbContext(dbPath, compressionOptions))
|
||||
{
|
||||
verify.Users.FindById(legacyId)!.Name.ShouldBe("legacy-uncompressed");
|
||||
verify.Users.FindById(compressedId)!.Name.Length.ShouldBeGreaterThan(10_000);
|
||||
|
||||
var counts = CountActiveDataSlots(verify.Storage);
|
||||
counts.Compressed.ShouldBeGreaterThanOrEqualTo(1);
|
||||
counts.Uncompressed.ShouldBeGreaterThanOrEqualTo(1);
|
||||
}
|
||||
}
|
||||
finally
|
||||
{
|
||||
CleanupFiles(dbPath);
|
||||
}
|
||||
}
|
||||
|
||||
private static (int Compressed, int Uncompressed) CountActiveDataSlots(StorageEngine storage)
|
||||
{
|
||||
var buffer = new byte[storage.PageSize];
|
||||
var compressed = 0;
|
||||
var uncompressed = 0;
|
||||
|
||||
for (uint pageId = 1; pageId < storage.PageCount; pageId++)
|
||||
{
|
||||
storage.ReadPage(pageId, null, buffer);
|
||||
var header = SlottedPageHeader.ReadFrom(buffer);
|
||||
if (header.PageType != PageType.Data)
|
||||
continue;
|
||||
|
||||
for (ushort slotIndex = 0; slotIndex < header.SlotCount; slotIndex++)
|
||||
{
|
||||
var slotOffset = SlottedPageHeader.Size + (slotIndex * SlotEntry.Size);
|
||||
var slot = SlotEntry.ReadFrom(buffer.AsSpan(slotOffset, SlotEntry.Size));
|
||||
if ((slot.Flags & SlotFlags.Deleted) != 0)
|
||||
continue;
|
||||
|
||||
if ((slot.Flags & SlotFlags.Compressed) != 0)
|
||||
compressed++;
|
||||
else
|
||||
uncompressed++;
|
||||
}
|
||||
}
|
||||
|
||||
return (compressed, uncompressed);
|
||||
}
|
||||
|
||||
private static string ComputeFileHash(string path)
|
||||
{
|
||||
using var stream = File.OpenRead(path);
|
||||
using var sha256 = SHA256.Create();
|
||||
return Convert.ToHexString(sha256.ComputeHash(stream));
|
||||
}
|
||||
|
||||
private static string BuildPayload(int approxLength)
|
||||
{
|
||||
var builder = new System.Text.StringBuilder(approxLength + 256);
|
||||
var i = 0;
|
||||
while (builder.Length < approxLength)
|
||||
{
|
||||
builder.Append("compat-payload-");
|
||||
builder.Append(i.ToString("D8"));
|
||||
builder.Append('|');
|
||||
i++;
|
||||
}
|
||||
|
||||
return builder.ToString();
|
||||
}
|
||||
|
||||
private static string NewDbPath()
|
||||
=> Path.Combine(Path.GetTempPath(), $"compression_compat_{Guid.NewGuid():N}.db");
|
||||
|
||||
private static void CleanupFiles(string dbPath)
|
||||
{
|
||||
var walPath = Path.ChangeExtension(dbPath, ".wal");
|
||||
var markerPath = $"{dbPath}.compact.state";
|
||||
if (File.Exists(dbPath)) File.Delete(dbPath);
|
||||
if (File.Exists(walPath)) File.Delete(walPath);
|
||||
if (File.Exists(markerPath)) File.Delete(markerPath);
|
||||
}
|
||||
}
|
||||
197
tests/CBDD.Tests/CompressionCorruptionTests.cs
Normal file
197
tests/CBDD.Tests/CompressionCorruptionTests.cs
Normal file
@@ -0,0 +1,197 @@
|
||||
using System.Buffers.Binary;
|
||||
using System.IO.Compression;
|
||||
using ZB.MOM.WW.CBDD.Bson;
|
||||
using ZB.MOM.WW.CBDD.Core.Compression;
|
||||
using ZB.MOM.WW.CBDD.Core.Storage;
|
||||
using ZB.MOM.WW.CBDD.Shared;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests;
|
||||
|
||||
public class CompressionCorruptionTests
|
||||
{
|
||||
[Fact]
|
||||
public void Read_WithBadChecksum_ShouldThrowInvalidData()
|
||||
{
|
||||
var dbPath = NewDbPath();
|
||||
var options = CompressionEnabledOptions();
|
||||
|
||||
try
|
||||
{
|
||||
using var db = new TestDbContext(dbPath, options);
|
||||
var id = InsertCheckpointAndCorrupt(db, header =>
|
||||
{
|
||||
var currentChecksum = BinaryPrimitives.ReadUInt32LittleEndian(header.Slice(12, 4));
|
||||
BinaryPrimitives.WriteUInt32LittleEndian(header.Slice(12, 4), currentChecksum + 1);
|
||||
});
|
||||
|
||||
var ex = Should.Throw<InvalidDataException>(() => db.Users.FindById(id));
|
||||
ex.Message.ShouldContain("checksum mismatch");
|
||||
db.GetCompressionStats().ChecksumFailureCount.ShouldBeGreaterThanOrEqualTo(1);
|
||||
}
|
||||
finally
|
||||
{
|
||||
CleanupFiles(dbPath);
|
||||
}
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Read_WithBadOriginalLength_ShouldThrowInvalidData()
|
||||
{
|
||||
var dbPath = NewDbPath();
|
||||
var options = CompressionEnabledOptions();
|
||||
|
||||
try
|
||||
{
|
||||
using var db = new TestDbContext(dbPath, options);
|
||||
var id = InsertCheckpointAndCorrupt(db, header =>
|
||||
{
|
||||
BinaryPrimitives.WriteInt32LittleEndian(header.Slice(4, 4), -1);
|
||||
});
|
||||
|
||||
var ex = Should.Throw<InvalidDataException>(() => db.Users.FindById(id));
|
||||
ex.Message.ShouldContain("decompress");
|
||||
}
|
||||
finally
|
||||
{
|
||||
CleanupFiles(dbPath);
|
||||
}
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Read_WithOversizedDeclaredLength_ShouldEnforceGuardrail()
|
||||
{
|
||||
var dbPath = NewDbPath();
|
||||
var options = CompressionEnabledOptions(maxDecompressedSizeBytes: 2048);
|
||||
|
||||
try
|
||||
{
|
||||
using var db = new TestDbContext(dbPath, options);
|
||||
var id = InsertCheckpointAndCorrupt(db, header =>
|
||||
{
|
||||
BinaryPrimitives.WriteInt32LittleEndian(header.Slice(4, 4), 2049);
|
||||
});
|
||||
|
||||
var ex = Should.Throw<InvalidDataException>(() => db.Users.FindById(id));
|
||||
ex.Message.ShouldContain("invalid decompressed length");
|
||||
db.GetCompressionStats().SafetyLimitRejectionCount.ShouldBeGreaterThanOrEqualTo(1);
|
||||
}
|
||||
finally
|
||||
{
|
||||
CleanupFiles(dbPath);
|
||||
}
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Read_WithInvalidCodecId_ShouldThrowInvalidData()
|
||||
{
|
||||
var dbPath = NewDbPath();
|
||||
var options = CompressionEnabledOptions();
|
||||
|
||||
try
|
||||
{
|
||||
using var db = new TestDbContext(dbPath, options);
|
||||
var id = InsertCheckpointAndCorrupt(db, header =>
|
||||
{
|
||||
header[0] = 0; // CompressionCodec.None is invalid for compressed payload header.
|
||||
});
|
||||
|
||||
var ex = Should.Throw<InvalidDataException>(() => db.Users.FindById(id));
|
||||
ex.Message.ShouldContain("invalid codec");
|
||||
}
|
||||
finally
|
||||
{
|
||||
CleanupFiles(dbPath);
|
||||
}
|
||||
}
|
||||
|
||||
private static ObjectId InsertCheckpointAndCorrupt(TestDbContext db, HeaderMutator mutateHeader)
|
||||
{
|
||||
var user = new User
|
||||
{
|
||||
Name = BuildPayload(16_000),
|
||||
Age = 33
|
||||
};
|
||||
|
||||
var id = db.Users.Insert(user);
|
||||
db.SaveChanges();
|
||||
db.ForceCheckpoint();
|
||||
|
||||
var (pageId, slot, _) = FindFirstCompressedSlot(db.Storage);
|
||||
((slot.Flags & SlotFlags.HasOverflow) != 0).ShouldBeFalse();
|
||||
|
||||
var page = new byte[db.Storage.PageSize];
|
||||
db.Storage.ReadPage(pageId, null, page);
|
||||
var headerSlice = page.AsSpan(slot.Offset, CompressedPayloadHeader.Size);
|
||||
mutateHeader(headerSlice);
|
||||
db.Storage.WritePageImmediate(pageId, page);
|
||||
|
||||
return id;
|
||||
}
|
||||
|
||||
private static (uint PageId, SlotEntry Slot, ushort SlotIndex) FindFirstCompressedSlot(StorageEngine storage)
|
||||
{
|
||||
var buffer = new byte[storage.PageSize];
|
||||
for (uint pageId = 1; pageId < storage.PageCount; pageId++)
|
||||
{
|
||||
storage.ReadPage(pageId, null, buffer);
|
||||
var header = SlottedPageHeader.ReadFrom(buffer);
|
||||
if (header.PageType != PageType.Data)
|
||||
continue;
|
||||
|
||||
for (ushort slotIndex = 0; slotIndex < header.SlotCount; slotIndex++)
|
||||
{
|
||||
var slotOffset = SlottedPageHeader.Size + (slotIndex * SlotEntry.Size);
|
||||
var slot = SlotEntry.ReadFrom(buffer.AsSpan(slotOffset, SlotEntry.Size));
|
||||
if ((slot.Flags & SlotFlags.Deleted) != 0)
|
||||
continue;
|
||||
|
||||
if ((slot.Flags & SlotFlags.Compressed) != 0)
|
||||
return (pageId, slot, slotIndex);
|
||||
}
|
||||
}
|
||||
|
||||
throw new InvalidOperationException("No active compressed slot found for corruption test setup.");
|
||||
}
|
||||
|
||||
private static CompressionOptions CompressionEnabledOptions(int maxDecompressedSizeBytes = 32 * 1024)
|
||||
{
|
||||
return new CompressionOptions
|
||||
{
|
||||
EnableCompression = true,
|
||||
MinSizeBytes = 0,
|
||||
MinSavingsPercent = 0,
|
||||
Codec = CompressionCodec.Brotli,
|
||||
Level = CompressionLevel.Fastest,
|
||||
MaxDecompressedSizeBytes = maxDecompressedSizeBytes
|
||||
};
|
||||
}
|
||||
|
||||
private delegate void HeaderMutator(Span<byte> header);
|
||||
|
||||
private static string BuildPayload(int approxLength)
|
||||
{
|
||||
var builder = new System.Text.StringBuilder(approxLength + 256);
|
||||
var i = 0;
|
||||
while (builder.Length < approxLength)
|
||||
{
|
||||
builder.Append("corruption-payload-");
|
||||
builder.Append(i.ToString("D8"));
|
||||
builder.Append('|');
|
||||
i++;
|
||||
}
|
||||
|
||||
return builder.ToString();
|
||||
}
|
||||
|
||||
private static string NewDbPath()
|
||||
=> Path.Combine(Path.GetTempPath(), $"compression_corruption_{Guid.NewGuid():N}.db");
|
||||
|
||||
private static void CleanupFiles(string dbPath)
|
||||
{
|
||||
var walPath = Path.ChangeExtension(dbPath, ".wal");
|
||||
var markerPath = $"{dbPath}.compact.state";
|
||||
if (File.Exists(dbPath)) File.Delete(dbPath);
|
||||
if (File.Exists(walPath)) File.Delete(walPath);
|
||||
if (File.Exists(markerPath)) File.Delete(markerPath);
|
||||
}
|
||||
}
|
||||
197
tests/CBDD.Tests/CompressionInsertReadTests.cs
Normal file
197
tests/CBDD.Tests/CompressionInsertReadTests.cs
Normal file
@@ -0,0 +1,197 @@
|
||||
using System.IO.Compression;
|
||||
using ZB.MOM.WW.CBDD.Bson;
|
||||
using ZB.MOM.WW.CBDD.Core.Compression;
|
||||
using ZB.MOM.WW.CBDD.Core.Storage;
|
||||
using ZB.MOM.WW.CBDD.Shared;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests;
|
||||
|
||||
public class CompressionInsertReadTests
|
||||
{
|
||||
[Fact]
|
||||
public void Insert_WithThreshold_ShouldStoreMixedCompressedAndUncompressedSlots()
|
||||
{
|
||||
var dbPath = NewDbPath();
|
||||
var options = new CompressionOptions
|
||||
{
|
||||
EnableCompression = true,
|
||||
MinSizeBytes = 4096,
|
||||
MinSavingsPercent = 0,
|
||||
Codec = CompressionCodec.Brotli,
|
||||
Level = CompressionLevel.Fastest
|
||||
};
|
||||
|
||||
try
|
||||
{
|
||||
using var db = new TestDbContext(dbPath, options);
|
||||
|
||||
var small = new User { Name = "tiny", Age = 10 };
|
||||
var large = new User { Name = BuildPayload(24_000), Age = 11 };
|
||||
|
||||
var smallId = db.Users.Insert(small);
|
||||
var largeId = db.Users.Insert(large);
|
||||
db.SaveChanges();
|
||||
|
||||
db.Users.FindById(smallId)!.Name.ShouldBe(small.Name);
|
||||
db.Users.FindById(largeId)!.Name.ShouldBe(large.Name);
|
||||
|
||||
var counts = CountActiveDataSlots(db.Storage);
|
||||
counts.Total.ShouldBeGreaterThanOrEqualTo(2);
|
||||
counts.Compressed.ShouldBeGreaterThanOrEqualTo(1);
|
||||
counts.Compressed.ShouldBeLessThan(counts.Total);
|
||||
}
|
||||
finally
|
||||
{
|
||||
CleanupFiles(dbPath);
|
||||
}
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void FindById_ShouldReadMixedCompressedAndUncompressedDocuments()
|
||||
{
|
||||
var dbPath = NewDbPath();
|
||||
var options = new CompressionOptions
|
||||
{
|
||||
EnableCompression = true,
|
||||
MinSizeBytes = 512,
|
||||
MinSavingsPercent = 0,
|
||||
Codec = CompressionCodec.Brotli,
|
||||
Level = CompressionLevel.Fastest
|
||||
};
|
||||
|
||||
var ids = new List<ObjectId>();
|
||||
|
||||
try
|
||||
{
|
||||
using (var db = new TestDbContext(dbPath, options))
|
||||
{
|
||||
ids.Add(db.Users.Insert(new User { Name = "small-a", Age = 1 }));
|
||||
ids.Add(db.Users.Insert(new User { Name = BuildPayload(18_000), Age = 2 }));
|
||||
ids.Add(db.Users.Insert(new User { Name = "small-b", Age = 3 }));
|
||||
ids.Add(db.Users.Insert(new User { Name = BuildPayload(26_000), Age = 4 }));
|
||||
db.SaveChanges();
|
||||
db.ForceCheckpoint();
|
||||
}
|
||||
|
||||
using (var reopened = new TestDbContext(dbPath, options))
|
||||
{
|
||||
reopened.Users.FindById(ids[0])!.Name.ShouldBe("small-a");
|
||||
reopened.Users.FindById(ids[2])!.Name.ShouldBe("small-b");
|
||||
reopened.Users.FindById(ids[1])!.Name.Length.ShouldBeGreaterThan(10_000);
|
||||
reopened.Users.FindById(ids[3])!.Name.Length.ShouldBeGreaterThan(10_000);
|
||||
|
||||
var counts = CountActiveDataSlots(reopened.Storage);
|
||||
counts.Compressed.ShouldBeGreaterThanOrEqualTo(1);
|
||||
counts.Compressed.ShouldBeLessThan(counts.Total);
|
||||
}
|
||||
}
|
||||
finally
|
||||
{
|
||||
CleanupFiles(dbPath);
|
||||
}
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Insert_WhenCodecThrows_ShouldFallbackToUncompressedStorage()
|
||||
{
|
||||
var dbPath = NewDbPath();
|
||||
var options = new CompressionOptions
|
||||
{
|
||||
EnableCompression = true,
|
||||
MinSizeBytes = 0,
|
||||
MinSavingsPercent = 0,
|
||||
Codec = CompressionCodec.Brotli,
|
||||
Level = CompressionLevel.Fastest
|
||||
};
|
||||
|
||||
try
|
||||
{
|
||||
using var db = new TestDbContext(dbPath, options);
|
||||
db.Storage.CompressionService.RegisterCodec(new FailingBrotliCodec());
|
||||
|
||||
var user = new User { Name = BuildPayload(20_000), Age = 7 };
|
||||
var id = db.Users.Insert(user);
|
||||
db.SaveChanges();
|
||||
|
||||
db.Users.FindById(id)!.Name.ShouldBe(user.Name);
|
||||
|
||||
var stats = db.GetCompressionStats();
|
||||
stats.CompressionFailureCount.ShouldBeGreaterThanOrEqualTo(1);
|
||||
stats.CompressedDocumentCount.ShouldBe(0);
|
||||
|
||||
var counts = CountActiveDataSlots(db.Storage);
|
||||
counts.Compressed.ShouldBe(0);
|
||||
}
|
||||
finally
|
||||
{
|
||||
CleanupFiles(dbPath);
|
||||
}
|
||||
}
|
||||
|
||||
private static (int Total, int Compressed) CountActiveDataSlots(StorageEngine storage)
|
||||
{
|
||||
var buffer = new byte[storage.PageSize];
|
||||
var total = 0;
|
||||
var compressed = 0;
|
||||
|
||||
for (uint pageId = 1; pageId < storage.PageCount; pageId++)
|
||||
{
|
||||
storage.ReadPage(pageId, null, buffer);
|
||||
var header = SlottedPageHeader.ReadFrom(buffer);
|
||||
if (header.PageType != PageType.Data)
|
||||
continue;
|
||||
|
||||
for (ushort slotIndex = 0; slotIndex < header.SlotCount; slotIndex++)
|
||||
{
|
||||
var slotOffset = SlottedPageHeader.Size + (slotIndex * SlotEntry.Size);
|
||||
var slot = SlotEntry.ReadFrom(buffer.AsSpan(slotOffset, SlotEntry.Size));
|
||||
if ((slot.Flags & SlotFlags.Deleted) != 0)
|
||||
continue;
|
||||
|
||||
total++;
|
||||
if ((slot.Flags & SlotFlags.Compressed) != 0)
|
||||
compressed++;
|
||||
}
|
||||
}
|
||||
|
||||
return (total, compressed);
|
||||
}
|
||||
|
||||
private static string BuildPayload(int approxLength)
|
||||
{
|
||||
var builder = new System.Text.StringBuilder(approxLength + 256);
|
||||
var i = 0;
|
||||
while (builder.Length < approxLength)
|
||||
{
|
||||
builder.Append("payload-");
|
||||
builder.Append(i.ToString("D8"));
|
||||
builder.Append('|');
|
||||
i++;
|
||||
}
|
||||
|
||||
return builder.ToString();
|
||||
}
|
||||
|
||||
private static string NewDbPath()
|
||||
=> Path.Combine(Path.GetTempPath(), $"compression_insert_read_{Guid.NewGuid():N}.db");
|
||||
|
||||
private static void CleanupFiles(string dbPath)
|
||||
{
|
||||
var walPath = Path.ChangeExtension(dbPath, ".wal");
|
||||
var markerPath = $"{dbPath}.compact.state";
|
||||
if (File.Exists(dbPath)) File.Delete(dbPath);
|
||||
if (File.Exists(walPath)) File.Delete(walPath);
|
||||
if (File.Exists(markerPath)) File.Delete(markerPath);
|
||||
}
|
||||
|
||||
private sealed class FailingBrotliCodec : ICompressionCodec
|
||||
{
|
||||
public CompressionCodec Codec => CompressionCodec.Brotli;
|
||||
|
||||
public byte[] Compress(ReadOnlySpan<byte> input, CompressionLevel level)
|
||||
=> throw new InvalidOperationException("Forced codec failure for test coverage.");
|
||||
|
||||
public byte[] Decompress(ReadOnlySpan<byte> input, int expectedLength, int maxDecompressedSizeBytes)
|
||||
=> throw new InvalidOperationException("This codec should not be used for reads in this scenario.");
|
||||
}
|
||||
}
|
||||
173
tests/CBDD.Tests/CompressionOverflowTests.cs
Normal file
173
tests/CBDD.Tests/CompressionOverflowTests.cs
Normal file
@@ -0,0 +1,173 @@
|
||||
using System.IO.Compression;
|
||||
using System.IO.MemoryMappedFiles;
|
||||
using ZB.MOM.WW.CBDD.Core.Compression;
|
||||
using ZB.MOM.WW.CBDD.Core.Storage;
|
||||
using ZB.MOM.WW.CBDD.Shared;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests;
|
||||
|
||||
public class CompressionOverflowTests
|
||||
{
|
||||
[Fact]
|
||||
public void Insert_CompressedDocumentSpanningOverflowPages_ShouldRoundTrip()
|
||||
{
|
||||
var dbPath = NewDbPath();
|
||||
var options = new CompressionOptions
|
||||
{
|
||||
EnableCompression = true,
|
||||
MinSizeBytes = 64,
|
||||
MinSavingsPercent = 0,
|
||||
Codec = CompressionCodec.Deflate,
|
||||
Level = CompressionLevel.Fastest
|
||||
};
|
||||
|
||||
try
|
||||
{
|
||||
using var db = new TestDbContext(dbPath, TinyPageConfig(), options);
|
||||
|
||||
var payload = BuildPayload(300_000);
|
||||
var id = db.Users.Insert(new User { Name = payload, Age = 40 });
|
||||
db.SaveChanges();
|
||||
|
||||
var found = db.Users.FindById(id);
|
||||
found.ShouldNotBeNull();
|
||||
found.Name.ShouldBe(payload);
|
||||
|
||||
var counts = CountSlotModes(db.Storage);
|
||||
counts.CompressedOverflow.ShouldBeGreaterThanOrEqualTo(1);
|
||||
counts.OverflowPages.ShouldBeGreaterThanOrEqualTo(1);
|
||||
}
|
||||
finally
|
||||
{
|
||||
CleanupFiles(dbPath);
|
||||
}
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Update_ShouldTransitionAcrossCompressionThresholds()
|
||||
{
|
||||
var dbPath = NewDbPath();
|
||||
var options = new CompressionOptions
|
||||
{
|
||||
EnableCompression = true,
|
||||
MinSizeBytes = 2048,
|
||||
MinSavingsPercent = 0,
|
||||
Codec = CompressionCodec.Brotli,
|
||||
Level = CompressionLevel.Fastest
|
||||
};
|
||||
|
||||
try
|
||||
{
|
||||
using var db = new TestDbContext(dbPath, TinyPageConfig(), options);
|
||||
|
||||
var user = new User { Name = "small", Age = 1 };
|
||||
var id = db.Users.Insert(user);
|
||||
db.SaveChanges();
|
||||
|
||||
CountSlotModes(db.Storage).Compressed.ShouldBe(0);
|
||||
|
||||
user.Name = BuildPayload(120_000);
|
||||
db.Users.Update(user).ShouldBeTrue();
|
||||
db.SaveChanges();
|
||||
|
||||
var afterLarge = db.Users.FindById(id);
|
||||
afterLarge.ShouldNotBeNull();
|
||||
afterLarge.Name.ShouldBe(user.Name);
|
||||
|
||||
var largeCounts = CountSlotModes(db.Storage);
|
||||
largeCounts.Compressed.ShouldBeGreaterThanOrEqualTo(1);
|
||||
|
||||
user.Name = "small-again";
|
||||
db.Users.Update(user).ShouldBeTrue();
|
||||
db.SaveChanges();
|
||||
|
||||
var afterShrink = db.Users.FindById(id);
|
||||
afterShrink.ShouldNotBeNull();
|
||||
afterShrink.Name.ShouldBe("small-again");
|
||||
|
||||
var finalCounts = CountSlotModes(db.Storage);
|
||||
finalCounts.Compressed.ShouldBe(0);
|
||||
}
|
||||
finally
|
||||
{
|
||||
CleanupFiles(dbPath);
|
||||
}
|
||||
}
|
||||
|
||||
private static (int Compressed, int CompressedOverflow, int OverflowPages) CountSlotModes(StorageEngine storage)
|
||||
{
|
||||
var buffer = new byte[storage.PageSize];
|
||||
var compressed = 0;
|
||||
var compressedOverflow = 0;
|
||||
var overflowPages = 0;
|
||||
|
||||
for (uint pageId = 1; pageId < storage.PageCount; pageId++)
|
||||
{
|
||||
storage.ReadPage(pageId, null, buffer);
|
||||
var header = SlottedPageHeader.ReadFrom(buffer);
|
||||
|
||||
if (header.PageType == PageType.Overflow)
|
||||
{
|
||||
overflowPages++;
|
||||
continue;
|
||||
}
|
||||
|
||||
if (header.PageType != PageType.Data)
|
||||
continue;
|
||||
|
||||
for (ushort slotIndex = 0; slotIndex < header.SlotCount; slotIndex++)
|
||||
{
|
||||
var slotOffset = SlottedPageHeader.Size + (slotIndex * SlotEntry.Size);
|
||||
var slot = SlotEntry.ReadFrom(buffer.AsSpan(slotOffset, SlotEntry.Size));
|
||||
if ((slot.Flags & SlotFlags.Deleted) != 0)
|
||||
continue;
|
||||
|
||||
var isCompressed = (slot.Flags & SlotFlags.Compressed) != 0;
|
||||
var hasOverflow = (slot.Flags & SlotFlags.HasOverflow) != 0;
|
||||
if (isCompressed)
|
||||
compressed++;
|
||||
if (isCompressed && hasOverflow)
|
||||
compressedOverflow++;
|
||||
}
|
||||
}
|
||||
|
||||
return (compressed, compressedOverflow, overflowPages);
|
||||
}
|
||||
|
||||
private static PageFileConfig TinyPageConfig()
|
||||
{
|
||||
return new PageFileConfig
|
||||
{
|
||||
PageSize = 16 * 1024,
|
||||
InitialFileSize = 1024 * 1024,
|
||||
Access = MemoryMappedFileAccess.ReadWrite
|
||||
};
|
||||
}
|
||||
|
||||
private static string BuildPayload(int approxLength)
|
||||
{
|
||||
var builder = new System.Text.StringBuilder(approxLength + 256);
|
||||
var i = 0;
|
||||
while (builder.Length < approxLength)
|
||||
{
|
||||
builder.Append("overflow-payload-");
|
||||
builder.Append(i.ToString("D7"));
|
||||
builder.Append('|');
|
||||
i++;
|
||||
}
|
||||
|
||||
return builder.ToString();
|
||||
}
|
||||
|
||||
private static string NewDbPath()
|
||||
=> Path.Combine(Path.GetTempPath(), $"compression_overflow_{Guid.NewGuid():N}.db");
|
||||
|
||||
private static void CleanupFiles(string dbPath)
|
||||
{
|
||||
var walPath = Path.ChangeExtension(dbPath, ".wal");
|
||||
var markerPath = $"{dbPath}.compact.state";
|
||||
if (File.Exists(dbPath)) File.Delete(dbPath);
|
||||
if (File.Exists(walPath)) File.Delete(walPath);
|
||||
if (File.Exists(markerPath)) File.Delete(markerPath);
|
||||
}
|
||||
}
|
||||
107
tests/CBDD.Tests/CursorTests.cs
Executable file
107
tests/CBDD.Tests/CursorTests.cs
Executable file
@@ -0,0 +1,107 @@
|
||||
using ZB.MOM.WW.CBDD.Core.Storage;
|
||||
using ZB.MOM.WW.CBDD.Core.Indexing;
|
||||
using ZB.MOM.WW.CBDD.Bson;
|
||||
using Xunit;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests;
|
||||
|
||||
public class CursorTests : IDisposable
|
||||
{
|
||||
private readonly string _testFile;
|
||||
private readonly StorageEngine _storage;
|
||||
private readonly BTreeIndex _index;
|
||||
|
||||
public CursorTests()
|
||||
{
|
||||
_testFile = Path.Combine(Path.GetTempPath(), $"docdb_cursor_test_{Guid.NewGuid()}.db");
|
||||
_storage = new StorageEngine(_testFile, PageFileConfig.Default);
|
||||
|
||||
var options = IndexOptions.CreateBTree("test");
|
||||
_index = new BTreeIndex(_storage, options);
|
||||
|
||||
SeedData();
|
||||
}
|
||||
|
||||
private void SeedData()
|
||||
{
|
||||
var txnId = _storage.BeginTransaction().TransactionId;
|
||||
|
||||
// Insert 10, 20, 30
|
||||
_index.Insert(IndexKey.Create(10), new DocumentLocation(1, 0), txnId);
|
||||
_index.Insert(IndexKey.Create(20), new DocumentLocation(2, 0), txnId);
|
||||
_index.Insert(IndexKey.Create(30), new DocumentLocation(3, 0), txnId);
|
||||
|
||||
_storage.CommitTransaction(txnId);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void MoveToFirst_ShouldPositionAtFirst()
|
||||
{
|
||||
using var cursor = _index.CreateCursor(0);
|
||||
cursor.MoveToFirst().ShouldBeTrue();
|
||||
cursor.Current.Key.ShouldBe(IndexKey.Create(10));
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void MoveToLast_ShouldPositionAtLast()
|
||||
{
|
||||
using var cursor = _index.CreateCursor(0);
|
||||
cursor.MoveToLast().ShouldBeTrue();
|
||||
cursor.Current.Key.ShouldBe(IndexKey.Create(30));
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void MoveNext_ShouldTraverseForward()
|
||||
{
|
||||
using var cursor = _index.CreateCursor(0);
|
||||
cursor.MoveToFirst();
|
||||
|
||||
cursor.MoveNext().ShouldBeTrue();
|
||||
cursor.Current.Key.ShouldBe(IndexKey.Create(20));
|
||||
|
||||
cursor.MoveNext().ShouldBeTrue();
|
||||
cursor.Current.Key.ShouldBe(IndexKey.Create(30));
|
||||
|
||||
cursor.MoveNext().ShouldBeFalse(); // End
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void MovePrev_ShouldTraverseBackward()
|
||||
{
|
||||
using var cursor = _index.CreateCursor(0);
|
||||
cursor.MoveToLast();
|
||||
|
||||
cursor.MovePrev().ShouldBeTrue();
|
||||
cursor.Current.Key.ShouldBe(IndexKey.Create(20));
|
||||
|
||||
cursor.MovePrev().ShouldBeTrue();
|
||||
cursor.Current.Key.ShouldBe(IndexKey.Create(10));
|
||||
|
||||
cursor.MovePrev().ShouldBeFalse(); // Start
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Seek_ShouldPositionExact_OrNext()
|
||||
{
|
||||
using var cursor = _index.CreateCursor(0);
|
||||
|
||||
// Exact
|
||||
cursor.Seek(IndexKey.Create(20)).ShouldBeTrue();
|
||||
cursor.Current.Key.ShouldBe(IndexKey.Create(20));
|
||||
|
||||
// Non-exact (15 -> should land on 20)
|
||||
cursor.Seek(IndexKey.Create(15)).ShouldBeFalse();
|
||||
cursor.Current.Key.ShouldBe(IndexKey.Create(20));
|
||||
|
||||
// Non-exact (35 -> should be invalid/end)
|
||||
cursor.Seek(IndexKey.Create(35)).ShouldBeFalse();
|
||||
// Current should throw invalid
|
||||
Should.Throw<InvalidOperationException>(() => cursor.Current);
|
||||
}
|
||||
|
||||
public void Dispose()
|
||||
{
|
||||
_storage.Dispose();
|
||||
if (File.Exists(_testFile)) File.Delete(_testFile);
|
||||
}
|
||||
}
|
||||
107
tests/CBDD.Tests/DbContextInheritanceTests.cs
Executable file
107
tests/CBDD.Tests/DbContextInheritanceTests.cs
Executable file
@@ -0,0 +1,107 @@
|
||||
using System;
|
||||
using System.IO;
|
||||
using ZB.MOM.WW.CBDD.Bson;
|
||||
using ZB.MOM.WW.CBDD.Shared;
|
||||
using Xunit;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests;
|
||||
|
||||
public class DbContextInheritanceTests : IDisposable
|
||||
{
|
||||
private readonly string _dbPath;
|
||||
private readonly Shared.TestExtendedDbContext _db;
|
||||
|
||||
public DbContextInheritanceTests()
|
||||
{
|
||||
_dbPath = Path.Combine(Path.GetTempPath(), $"cbdd_inheritance_{Guid.NewGuid()}.db");
|
||||
_db = new Shared.TestExtendedDbContext(_dbPath);
|
||||
}
|
||||
|
||||
public void Dispose()
|
||||
{
|
||||
_db.Dispose();
|
||||
if (File.Exists(_dbPath)) File.Delete(_dbPath);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void ExtendedContext_Should_Initialize_Parent_Collections()
|
||||
{
|
||||
// Verify parent collections are initialized (from TestDbContext)
|
||||
_db.Users.ShouldNotBeNull();
|
||||
_db.People.ShouldNotBeNull();
|
||||
_db.Products.ShouldNotBeNull();
|
||||
_db.AnnotatedUsers.ShouldNotBeNull();
|
||||
_db.ComplexDocuments.ShouldNotBeNull();
|
||||
_db.TestDocuments.ShouldNotBeNull();
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void ExtendedContext_Should_Initialize_Own_Collections()
|
||||
{
|
||||
// Verify extended context's own collection is initialized
|
||||
_db.ExtendedEntities.ShouldNotBeNull();
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void ExtendedContext_Can_Use_Parent_Collections()
|
||||
{
|
||||
// Insert into parent collection
|
||||
var user = new User { Name = "TestUser", Age = 30 };
|
||||
_db.Users.Insert(user);
|
||||
_db.SaveChanges();
|
||||
|
||||
// Verify we can read it back
|
||||
var retrieved = _db.Users.FindById(user.Id);
|
||||
retrieved.ShouldNotBeNull();
|
||||
retrieved.Name.ShouldBe("TestUser");
|
||||
retrieved.Age.ShouldBe(30);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void ExtendedContext_Can_Use_Own_Collections()
|
||||
{
|
||||
// Insert into extended collection
|
||||
var entity = new ExtendedEntity
|
||||
{
|
||||
Id = 1,
|
||||
Description = "Test Extended Entity",
|
||||
CreatedAt = DateTime.UtcNow
|
||||
};
|
||||
_db.ExtendedEntities.Insert(entity);
|
||||
_db.SaveChanges();
|
||||
|
||||
// Verify we can read it back
|
||||
var retrieved = _db.ExtendedEntities.FindById(1);
|
||||
retrieved.ShouldNotBeNull();
|
||||
retrieved.Description.ShouldBe("Test Extended Entity");
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void ExtendedContext_Can_Use_Both_Parent_And_Own_Collections()
|
||||
{
|
||||
// Insert into parent collection
|
||||
var person = new Person { Id = 100, Name = "John", Age = 25 };
|
||||
_db.People.Insert(person);
|
||||
|
||||
// Insert into extended collection
|
||||
var extended = new ExtendedEntity
|
||||
{
|
||||
Id = 200,
|
||||
Description = "Related to John",
|
||||
CreatedAt = DateTime.UtcNow
|
||||
};
|
||||
_db.ExtendedEntities.Insert(extended);
|
||||
|
||||
_db.SaveChanges();
|
||||
|
||||
// Verify both
|
||||
var retrievedPerson = _db.People.FindById(100);
|
||||
var retrievedExtended = _db.ExtendedEntities.FindById(200);
|
||||
|
||||
retrievedPerson.ShouldNotBeNull();
|
||||
retrievedPerson.Name.ShouldBe("John");
|
||||
|
||||
retrievedExtended.ShouldNotBeNull();
|
||||
retrievedExtended.Description.ShouldBe("Related to John");
|
||||
}
|
||||
}
|
||||
222
tests/CBDD.Tests/DbContextTests.cs
Executable file
222
tests/CBDD.Tests/DbContextTests.cs
Executable file
@@ -0,0 +1,222 @@
|
||||
using ZB.MOM.WW.CBDD.Bson;
|
||||
using ZB.MOM.WW.CBDD.Core.Compression;
|
||||
using ZB.MOM.WW.CBDD.Core.Storage;
|
||||
using ZB.MOM.WW.CBDD.Shared;
|
||||
using System.Security.Cryptography;
|
||||
using System.IO.Compression;
|
||||
using System.IO.MemoryMappedFiles;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests;
|
||||
|
||||
public class DbContextTests : IDisposable
|
||||
{
|
||||
private string _dbPath;
|
||||
|
||||
public DbContextTests()
|
||||
{
|
||||
_dbPath = Path.Combine(Path.GetTempPath(), $"test_dbcontext_{Guid.NewGuid()}.db");
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void DbContext_BasicLifecycle_Works()
|
||||
{
|
||||
using var db = new Shared.TestDbContext(_dbPath);
|
||||
|
||||
var user = new User { Name = "Alice", Age = 30 };
|
||||
var id = db.Users.Insert(user);
|
||||
|
||||
var found = db.Users.FindById(id);
|
||||
found.ShouldNotBeNull();
|
||||
found.Name.ShouldBe("Alice");
|
||||
found.Age.ShouldBe(30);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void DbContext_MultipleOperations_Work()
|
||||
{
|
||||
using var db = new Shared.TestDbContext(_dbPath);
|
||||
|
||||
// Insert
|
||||
var alice = new User { Name = "Alice", Age = 30 };
|
||||
var bob = new User { Name = "Bob", Age = 25 };
|
||||
|
||||
var id1 = db.Users.Insert(alice);
|
||||
var id2 = db.Users.Insert(bob);
|
||||
|
||||
// FindAll
|
||||
var all = db.Users.FindAll().ToList();
|
||||
all.Count.ShouldBe(2);
|
||||
|
||||
// Update
|
||||
alice.Age = 31;
|
||||
db.Users.Update(alice).ShouldBeTrue();
|
||||
|
||||
var updated = db.Users.FindById(id1);
|
||||
updated!.Age.ShouldBe(31);
|
||||
|
||||
// Delete
|
||||
db.Users.Delete(id2).ShouldBeTrue();
|
||||
db.Users.Count().ShouldBe(1);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void DbContext_Dispose_ReleasesResources()
|
||||
{
|
||||
_dbPath = Path.Combine(Path.GetTempPath(), $"test_dbcontext_reopen.db");
|
||||
var totalUsers = 0;
|
||||
// First context - insert and dispose (auto-checkpoint)
|
||||
using (var db = new Shared.TestDbContext(_dbPath))
|
||||
{
|
||||
db.Users.Insert(new User { Name = "Test", Age = 20 });
|
||||
db.SaveChanges(); // Explicitly save changes to ensure data is in WAL
|
||||
var beforeCheckpointTotalUsers = db.Users.FindAll().Count();
|
||||
db.ForceCheckpoint(); // Force checkpoint to ensure data is persisted to main file
|
||||
totalUsers = db.Users.FindAll().Count();
|
||||
var countedUsers = db.Users.Count();
|
||||
totalUsers.ShouldBe(beforeCheckpointTotalUsers);
|
||||
} // Dispose → Commit → ForceCheckpoint → Write to PageFile
|
||||
|
||||
// Should be able to open again and see persisted data
|
||||
using var db2 = new Shared.TestDbContext(_dbPath);
|
||||
|
||||
totalUsers.ShouldBe(1);
|
||||
db2.Users.FindAll().Count().ShouldBe(totalUsers);
|
||||
db2.Users.Count().ShouldBe(totalUsers);
|
||||
}
|
||||
private static string ComputeFileHash(string path)
|
||||
{
|
||||
using var stream = File.OpenRead(path);
|
||||
using var sha256 = SHA256.Create();
|
||||
return Convert.ToHexString(sha256.ComputeHash(stream));
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void DatabaseFile_SizeAndContent_ChangeAfterInsert()
|
||||
{
|
||||
var dbPath = Path.Combine(Path.GetTempPath(), $"test_dbfile_{Guid.NewGuid()}.db");
|
||||
|
||||
// 1. Crea e chiudi database vuoto
|
||||
using (var db = new Shared.TestDbContext(dbPath))
|
||||
{
|
||||
db.Users.Insert(new User { Name = "Pippo", Age = 42 });
|
||||
}
|
||||
var initialSize = new FileInfo(dbPath).Length;
|
||||
var initialHash = ComputeFileHash(dbPath);
|
||||
|
||||
// 2. Riapri, inserisci, chiudi
|
||||
using (var db = new Shared.TestDbContext(dbPath))
|
||||
{
|
||||
db.Users.Insert(new User { Name = "Test", Age = 42 });
|
||||
db.ForceCheckpoint(); // Forza persistenza
|
||||
}
|
||||
var afterInsertSize = new FileInfo(dbPath).Length;
|
||||
var afterInsertHash = ComputeFileHash(dbPath);
|
||||
|
||||
// 3. Verifica che dimensione e hash siano cambiati
|
||||
afterInsertSize.ShouldNotBe(initialSize);
|
||||
afterInsertHash.ShouldNotBe(initialHash);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void DbContext_AutoDerivesWalPath()
|
||||
{
|
||||
using var db = new Shared.TestDbContext(_dbPath);
|
||||
db.Users.Insert(new User { Name = "Test", Age = 20 });
|
||||
|
||||
var walPath = Path.ChangeExtension(_dbPath, ".wal");
|
||||
File.Exists(walPath).ShouldBeTrue();
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void DbContext_WithCustomPageFileAndCompressionOptions_ShouldSupportRoundTrip()
|
||||
{
|
||||
var dbPath = Path.Combine(Path.GetTempPath(), $"test_dbcontext_compression_{Guid.NewGuid():N}.db");
|
||||
var options = new CompressionOptions
|
||||
{
|
||||
EnableCompression = true,
|
||||
MinSizeBytes = 0,
|
||||
MinSavingsPercent = 0,
|
||||
Codec = CompressionCodec.Brotli,
|
||||
Level = CompressionLevel.Fastest
|
||||
};
|
||||
|
||||
var config = new PageFileConfig
|
||||
{
|
||||
PageSize = 16 * 1024,
|
||||
InitialFileSize = 1024 * 1024,
|
||||
Access = MemoryMappedFileAccess.ReadWrite
|
||||
};
|
||||
|
||||
try
|
||||
{
|
||||
using var db = new Shared.TestDbContext(dbPath, config, options);
|
||||
var payload = string.Concat(Enumerable.Repeat("compressible-", 3000));
|
||||
var id = db.Users.Insert(new User { Name = payload, Age = 77 });
|
||||
db.SaveChanges();
|
||||
|
||||
var loaded = db.Users.FindById(id);
|
||||
loaded.ShouldNotBeNull();
|
||||
loaded.Name.ShouldBe(payload);
|
||||
db.GetCompressionStats().CompressedDocumentCount.ShouldBeGreaterThanOrEqualTo(1);
|
||||
}
|
||||
finally
|
||||
{
|
||||
CleanupDbFiles(dbPath);
|
||||
}
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void DbContext_CompactApi_ShouldReturnStatsAndPreserveData()
|
||||
{
|
||||
var dbPath = Path.Combine(Path.GetTempPath(), $"test_dbcontext_compact_{Guid.NewGuid():N}.db");
|
||||
try
|
||||
{
|
||||
using var db = new Shared.TestDbContext(dbPath);
|
||||
for (var i = 0; i < 120; i++)
|
||||
{
|
||||
db.Users.Insert(new User { Name = $"compact-{i:D3}", Age = i % 20 });
|
||||
}
|
||||
|
||||
db.SaveChanges();
|
||||
db.Users.Count().ShouldBe(120);
|
||||
db.SaveChanges();
|
||||
|
||||
var stats = db.Compact(new CompactionOptions
|
||||
{
|
||||
EnableTailTruncation = true,
|
||||
DefragmentSlottedPages = true,
|
||||
NormalizeFreeList = true
|
||||
});
|
||||
|
||||
stats.OnlineMode.ShouldBeFalse();
|
||||
db.Users.Count().ShouldBe(120);
|
||||
}
|
||||
finally
|
||||
{
|
||||
CleanupDbFiles(dbPath);
|
||||
}
|
||||
}
|
||||
|
||||
public void Dispose()
|
||||
{
|
||||
try
|
||||
{
|
||||
CleanupDbFiles(_dbPath);
|
||||
}
|
||||
catch
|
||||
{
|
||||
// Ignore cleanup errors
|
||||
}
|
||||
}
|
||||
|
||||
private static void CleanupDbFiles(string dbPath)
|
||||
{
|
||||
if (File.Exists(dbPath)) File.Delete(dbPath);
|
||||
|
||||
var walPath = Path.ChangeExtension(dbPath, ".wal");
|
||||
if (File.Exists(walPath)) File.Delete(walPath);
|
||||
|
||||
var markerPath = $"{dbPath}.compact.state";
|
||||
if (File.Exists(markerPath)) File.Delete(markerPath);
|
||||
}
|
||||
}
|
||||
211
tests/CBDD.Tests/DictionaryPageTests.cs
Executable file
211
tests/CBDD.Tests/DictionaryPageTests.cs
Executable file
@@ -0,0 +1,211 @@
|
||||
using ZB.MOM.WW.CBDD.Core;
|
||||
using ZB.MOM.WW.CBDD.Core.Storage;
|
||||
using System.Text;
|
||||
using Xunit;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests;
|
||||
|
||||
public class DictionaryPageTests
|
||||
{
|
||||
private const int PageSize = 16384;
|
||||
|
||||
[Fact]
|
||||
public void Initialize_ShouldSetupEmptyPage()
|
||||
{
|
||||
var page = new byte[PageSize];
|
||||
DictionaryPage.Initialize(page, 1);
|
||||
|
||||
var header = PageHeader.ReadFrom(page);
|
||||
header.PageType.ShouldBe(PageType.Dictionary);
|
||||
header.PageId.ShouldBe(1u);
|
||||
|
||||
var count = BitConverter.ToUInt16(page, 32); // CountOffset
|
||||
count.ShouldBe((ushort)0);
|
||||
|
||||
var freeSpaceEnd = BitConverter.ToUInt16(page, 34); // FreeSpaceEndOffset
|
||||
freeSpaceEnd.ShouldBe((ushort)PageSize);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Insert_ShouldAddEntryAndSort()
|
||||
{
|
||||
var page = new byte[PageSize];
|
||||
DictionaryPage.Initialize(page, 1);
|
||||
|
||||
// Insert "B"
|
||||
bool inserted = DictionaryPage.Insert(page, "B", 20);
|
||||
inserted.ShouldBeTrue();
|
||||
|
||||
// Insert "A" (should go before B)
|
||||
inserted = DictionaryPage.Insert(page, "A", 10);
|
||||
inserted.ShouldBeTrue();
|
||||
|
||||
// Insert "C" (should go after B)
|
||||
inserted = DictionaryPage.Insert(page, "C", 30);
|
||||
inserted.ShouldBeTrue();
|
||||
|
||||
// Verify Order
|
||||
var entries = DictionaryPage.GetAll(page).ToList();
|
||||
entries.Count.ShouldBe(3);
|
||||
|
||||
entries[0].Key.ShouldBe("A");
|
||||
entries[0].Value.ShouldBe((ushort)10);
|
||||
|
||||
entries[1].Key.ShouldBe("B");
|
||||
entries[1].Value.ShouldBe((ushort)20);
|
||||
|
||||
entries[2].Key.ShouldBe("C");
|
||||
entries[2].Value.ShouldBe((ushort)30);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void TryFind_ShouldReturnCorrectValue()
|
||||
{
|
||||
var page = new byte[PageSize];
|
||||
DictionaryPage.Initialize(page, 1);
|
||||
|
||||
DictionaryPage.Insert(page, "Key1", 100);
|
||||
DictionaryPage.Insert(page, "Key2", 200);
|
||||
DictionaryPage.Insert(page, "Key3", 300);
|
||||
|
||||
bool found = DictionaryPage.TryFind(page, Encoding.UTF8.GetBytes("Key2"), out ushort value);
|
||||
found.ShouldBeTrue();
|
||||
value.ShouldBe((ushort)200);
|
||||
|
||||
found = DictionaryPage.TryFind(page, Encoding.UTF8.GetBytes("Key999"), out value);
|
||||
found.ShouldBeFalse();
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Overflow_ShouldReturnFalse_WhenFull()
|
||||
{
|
||||
var page = new byte[PageSize];
|
||||
DictionaryPage.Initialize(page, 1);
|
||||
|
||||
string bigKey = new string('X', 250);
|
||||
|
||||
int count = 0;
|
||||
while (true)
|
||||
{
|
||||
// Use unique keys
|
||||
var key = bigKey + count;
|
||||
if (!DictionaryPage.Insert(page, key, (ushort)count))
|
||||
{
|
||||
// Should fail here
|
||||
break;
|
||||
}
|
||||
count++;
|
||||
if (count > 1000) throw new ShouldAssertException("Should have filled the page much earlier");
|
||||
}
|
||||
|
||||
// Now page is full enough that `bigKey` (250 bytes) shouldn't fit.
|
||||
// We can't guarantee a small key won't fit (fragmentation/remaining space),
|
||||
// but a key of the SAME size that triggered the break should definitely fail.
|
||||
bool inserted = DictionaryPage.Insert(page, bigKey + "X", 9999);
|
||||
inserted.ShouldBeFalse();
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Chaining_ShouldFindKeysInLinkedPages()
|
||||
{
|
||||
var dbPath = Path.Combine(Path.GetTempPath(), $"test_dict_chain_{Guid.NewGuid()}.db");
|
||||
using var storage = new StorageEngine(dbPath, PageFileConfig.Default);
|
||||
|
||||
// 1. Create First Page
|
||||
var page1Id = storage.AllocatePage();
|
||||
var pageBuffer = new byte[storage.PageSize];
|
||||
DictionaryPage.Initialize(pageBuffer, page1Id);
|
||||
|
||||
// Fill Page 1
|
||||
DictionaryPage.Insert(pageBuffer, "Key1", 100);
|
||||
DictionaryPage.Insert(pageBuffer, "KeyA", 200);
|
||||
|
||||
// 2. Create Second Page
|
||||
var page2Id = storage.AllocatePage();
|
||||
var page2Buffer = new byte[storage.PageSize];
|
||||
DictionaryPage.Initialize(page2Buffer, page2Id);
|
||||
|
||||
// Fill Page 2
|
||||
DictionaryPage.Insert(page2Buffer, "Key2", 300);
|
||||
DictionaryPage.Insert(page2Buffer, "KeyB", 400); // 400
|
||||
|
||||
// 3. Link Page 1 -> Page 2
|
||||
var header1 = PageHeader.ReadFrom(pageBuffer);
|
||||
header1.NextPageId = page2Id;
|
||||
header1.WriteTo(pageBuffer);
|
||||
|
||||
// 4. Write pages to storage
|
||||
storage.WritePageImmediate(page1Id, pageBuffer);
|
||||
storage.WritePageImmediate(page2Id, page2Buffer);
|
||||
|
||||
// 5. Test Global Find
|
||||
// Find in Page 1
|
||||
bool found = DictionaryPage.TryFindGlobal(storage, page1Id, "Key1", out ushort val);
|
||||
found.ShouldBeTrue();
|
||||
val.ShouldBe((ushort)100);
|
||||
|
||||
// Find in Page 2
|
||||
found = DictionaryPage.TryFindGlobal(storage, page1Id, "KeyB", out val);
|
||||
found.ShouldBeTrue();
|
||||
val.ShouldBe((ushort)400);
|
||||
|
||||
// Not Found
|
||||
found = DictionaryPage.TryFindGlobal(storage, page1Id, "KeyMissing", out val);
|
||||
found.ShouldBeFalse();
|
||||
|
||||
storage.Dispose();
|
||||
if (File.Exists(dbPath)) File.Delete(dbPath);
|
||||
if (File.Exists(Path.ChangeExtension(dbPath, ".wal"))) File.Delete(Path.ChangeExtension(dbPath, ".wal"));
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void FindAllGlobal_ShouldRetrieveAllKeys()
|
||||
{
|
||||
var dbPath = Path.Combine(Path.GetTempPath(), $"test_dict_findall_{Guid.NewGuid()}.db");
|
||||
using var storage = new StorageEngine(dbPath, PageFileConfig.Default);
|
||||
|
||||
// 1. Create Chain of 3 Pages
|
||||
var page1Id = storage.AllocatePage();
|
||||
var page2Id = storage.AllocatePage();
|
||||
var page3Id = storage.AllocatePage();
|
||||
|
||||
var buf = new byte[storage.PageSize];
|
||||
|
||||
// Page 1
|
||||
DictionaryPage.Initialize(buf, page1Id);
|
||||
DictionaryPage.Insert(buf, "P1_A", 10);
|
||||
DictionaryPage.Insert(buf, "P1_B", 11);
|
||||
var h1 = PageHeader.ReadFrom(buf);
|
||||
h1.NextPageId = page2Id;
|
||||
h1.WriteTo(buf);
|
||||
storage.WritePageImmediate(page1Id, buf);
|
||||
|
||||
// Page 2
|
||||
DictionaryPage.Initialize(buf, page2Id);
|
||||
DictionaryPage.Insert(buf, "P2_A", 20);
|
||||
var h2 = PageHeader.ReadFrom(buf);
|
||||
h2.NextPageId = page3Id;
|
||||
h2.WriteTo(buf);
|
||||
storage.WritePageImmediate(page2Id, buf);
|
||||
|
||||
// Page 3
|
||||
DictionaryPage.Initialize(buf, page3Id);
|
||||
DictionaryPage.Insert(buf, "P3_A", 30);
|
||||
DictionaryPage.Insert(buf, "P3_B", 31);
|
||||
DictionaryPage.Insert(buf, "P3_C", 32);
|
||||
storage.WritePageImmediate(page3Id, buf);
|
||||
|
||||
// 2. Execute FindAllGlobal
|
||||
var allEntries = DictionaryPage.FindAllGlobal(storage, page1Id).ToList();
|
||||
|
||||
// 3. Verify
|
||||
allEntries.Count.ShouldBe(6);
|
||||
allEntries.ShouldContain(e => e.Key == "P1_A" && e.Value == 10);
|
||||
allEntries.ShouldContain(e => e.Key == "P2_A" && e.Value == 20);
|
||||
allEntries.ShouldContain(e => e.Key == "P3_C" && e.Value == 32);
|
||||
|
||||
storage.Dispose();
|
||||
if (File.Exists(dbPath)) File.Delete(dbPath);
|
||||
if (File.Exists(Path.ChangeExtension(dbPath, ".wal"))) File.Delete(Path.ChangeExtension(dbPath, ".wal"));
|
||||
}
|
||||
}
|
||||
127
tests/CBDD.Tests/DictionaryPersistenceTests.cs
Executable file
127
tests/CBDD.Tests/DictionaryPersistenceTests.cs
Executable file
@@ -0,0 +1,127 @@
|
||||
using ZB.MOM.WW.CBDD.Bson;
|
||||
using ZB.MOM.WW.CBDD.Core.Collections;
|
||||
using ZB.MOM.WW.CBDD.Core.Storage;
|
||||
using Xunit;
|
||||
using System.Collections.Generic;
|
||||
using System.Linq;
|
||||
using ZB.MOM.WW.CBDD.Bson.Schema;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests;
|
||||
|
||||
public class DictionaryPersistenceTests : IDisposable
|
||||
{
|
||||
private readonly string _dbPath;
|
||||
private readonly StorageEngine _storage;
|
||||
|
||||
public DictionaryPersistenceTests()
|
||||
{
|
||||
_dbPath = Path.Combine(Path.GetTempPath(), $"cbdd_dict_{Guid.NewGuid():N}.db");
|
||||
_storage = new StorageEngine(_dbPath, PageFileConfig.Default);
|
||||
}
|
||||
|
||||
public void Dispose()
|
||||
{
|
||||
_storage.Dispose();
|
||||
if (File.Exists(_dbPath)) File.Delete(_dbPath);
|
||||
var walPath = Path.ChangeExtension(_dbPath, ".wal");
|
||||
if (File.Exists(walPath)) File.Delete(walPath);
|
||||
}
|
||||
|
||||
private class MockMapper : DocumentMapperBase<ObjectId, Dictionary<string, object>>
|
||||
{
|
||||
private readonly string _collectionName;
|
||||
private readonly List<string> _keys;
|
||||
|
||||
public MockMapper(string name, params string[] keys)
|
||||
{
|
||||
_collectionName = name;
|
||||
_keys = keys.ToList();
|
||||
}
|
||||
|
||||
public override string CollectionName => _collectionName;
|
||||
public override IEnumerable<string> UsedKeys => _keys;
|
||||
|
||||
public override BsonSchema GetSchema() => new BsonSchema { Title = _collectionName };
|
||||
public override ObjectId GetId(Dictionary<string, object> entity) => throw new NotImplementedException();
|
||||
public override void SetId(Dictionary<string, object> entity, ObjectId id) => throw new NotImplementedException();
|
||||
public override int Serialize(Dictionary<string, object> entity, BsonSpanWriter writer) => throw new NotImplementedException();
|
||||
public override Dictionary<string, object> Deserialize(BsonSpanReader reader) => throw new NotImplementedException();
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void RegisterMappers_Registers_All_Unique_Keys()
|
||||
{
|
||||
var mapper1 = new MockMapper("Coll1", "Name", "Age");
|
||||
var mapper2 = new MockMapper("Coll2", "Name", "Address", "City");
|
||||
|
||||
_storage.RegisterMappers(new IDocumentMapper[] { mapper1, mapper2 });
|
||||
|
||||
// Verify keys in cache
|
||||
_storage.GetOrAddDictionaryEntry("Name").ShouldNotBe((ushort)0);
|
||||
_storage.GetOrAddDictionaryEntry("Age").ShouldNotBe((ushort)0);
|
||||
_storage.GetOrAddDictionaryEntry("Address").ShouldNotBe((ushort)0);
|
||||
_storage.GetOrAddDictionaryEntry("City").ShouldNotBe((ushort)0);
|
||||
|
||||
// Verify they have unique IDs (at least 4 unique IDs for 4 unique keys + internal ones)
|
||||
var ids = new HashSet<ushort>
|
||||
{
|
||||
_storage.GetOrAddDictionaryEntry("Name"),
|
||||
_storage.GetOrAddDictionaryEntry("Age"),
|
||||
_storage.GetOrAddDictionaryEntry("Address"),
|
||||
_storage.GetOrAddDictionaryEntry("City")
|
||||
};
|
||||
ids.Count.ShouldBe(4);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Dictionary_Keys_Persist_Across_Restarts()
|
||||
{
|
||||
var mapper = new MockMapper("Coll1", "PersistedKey");
|
||||
_storage.RegisterMappers(new IDocumentMapper[] { mapper });
|
||||
|
||||
var originalId = _storage.GetOrAddDictionaryEntry("PersistedKey");
|
||||
originalId.ShouldNotBe((ushort)0);
|
||||
|
||||
_storage.Dispose();
|
||||
|
||||
// Re-open
|
||||
using var storage2 = new StorageEngine(_dbPath, PageFileConfig.Default);
|
||||
|
||||
var recoveredId = storage2.GetOrAddDictionaryEntry("PersistedKey");
|
||||
recoveredId.ShouldBe(originalId);
|
||||
}
|
||||
|
||||
private class NestedMockMapper : DocumentMapperBase<ObjectId, object>
|
||||
{
|
||||
public override string CollectionName => "Nested";
|
||||
public override BsonSchema GetSchema()
|
||||
{
|
||||
var schema = new BsonSchema { Title = "Nested" };
|
||||
schema.Fields.Add(new BsonField
|
||||
{
|
||||
Name = "Top",
|
||||
Type = BsonType.Document,
|
||||
NestedSchema = new BsonSchema
|
||||
{
|
||||
Fields = { new BsonField { Name = "Child", Type = BsonType.String } }
|
||||
}
|
||||
});
|
||||
return schema;
|
||||
}
|
||||
|
||||
public override ObjectId GetId(object entity) => throw new NotImplementedException();
|
||||
public override void SetId(object entity, ObjectId id) => throw new NotImplementedException();
|
||||
public override int Serialize(object entity, BsonSpanWriter writer) => throw new NotImplementedException();
|
||||
public override object Deserialize(BsonSpanReader reader) => throw new NotImplementedException();
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void RegisterMappers_Handles_Nested_Keys()
|
||||
{
|
||||
var mapper = new NestedMockMapper();
|
||||
_storage.RegisterMappers(new IDocumentMapper[] { mapper });
|
||||
|
||||
_storage.GetOrAddDictionaryEntry("Top").ShouldNotBe((ushort)0);
|
||||
_storage.GetOrAddDictionaryEntry("Child").ShouldNotBe((ushort)0);
|
||||
}
|
||||
}
|
||||
80
tests/CBDD.Tests/DocumentCollectionDeleteTests.cs
Executable file
80
tests/CBDD.Tests/DocumentCollectionDeleteTests.cs
Executable file
@@ -0,0 +1,80 @@
|
||||
using ZB.MOM.WW.CBDD.Bson;
|
||||
using ZB.MOM.WW.CBDD.Core.Collections;
|
||||
using ZB.MOM.WW.CBDD.Core.Storage;
|
||||
using ZB.MOM.WW.CBDD.Core.Transactions;
|
||||
using ZB.MOM.WW.CBDD.Shared;
|
||||
using ZB.MOM.WW.CBDD.Shared.TestDbContext_TestDbContext_Mappers;
|
||||
using Xunit;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests;
|
||||
|
||||
public class DocumentCollectionDeleteTests : IDisposable
|
||||
{
|
||||
private readonly string _dbPath;
|
||||
private readonly string _walPath;
|
||||
private readonly Shared.TestDbContext _dbContext;
|
||||
|
||||
public DocumentCollectionDeleteTests()
|
||||
{
|
||||
_dbPath = Path.Combine(Path.GetTempPath(), $"test_delete_{Guid.NewGuid()}.db");
|
||||
_walPath = Path.Combine(Path.GetTempPath(), $"test_delete_{Guid.NewGuid()}.wal");
|
||||
|
||||
_dbContext = new Shared.TestDbContext(_dbPath);
|
||||
}
|
||||
|
||||
public void Dispose()
|
||||
{
|
||||
_dbContext.Dispose();
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Delete_RemovesDocumentAndIndexEntry()
|
||||
{
|
||||
var user = new User { Id = ObjectId.NewObjectId(), Name = "To Delete", Age = 10 };
|
||||
_dbContext.Users.Insert(user);
|
||||
_dbContext.SaveChanges();
|
||||
|
||||
// Verify inserted
|
||||
_dbContext.Users.FindById(user.Id).ShouldNotBeNull();
|
||||
|
||||
// Delete
|
||||
var deleted = _dbContext.Users.Delete(user.Id);
|
||||
_dbContext.SaveChanges();
|
||||
|
||||
// Assert
|
||||
deleted.ShouldBeTrue("Delete returned false");
|
||||
|
||||
// Verify deleted from storage
|
||||
_dbContext.Users.FindById(user.Id).ShouldBeNull();
|
||||
|
||||
// Verify Index is clean (FindAll uses index scan)
|
||||
var all = _dbContext.Users.FindAll();
|
||||
all.ShouldBeEmpty();
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Delete_NonExistent_ReturnsFalse()
|
||||
{
|
||||
var id = ObjectId.NewObjectId();
|
||||
var deleted = _dbContext.Users.Delete(id);
|
||||
_dbContext.SaveChanges();
|
||||
deleted.ShouldBeFalse();
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Delete_WithTransaction_CommitsSuccessfully()
|
||||
{
|
||||
var user = new User { Id = ObjectId.NewObjectId(), Name = "Txn Delete", Age = 20 };
|
||||
_dbContext.Users.Insert(user);
|
||||
_dbContext.SaveChanges();
|
||||
|
||||
using (var txn = _dbContext.BeginTransaction())
|
||||
{
|
||||
_dbContext.Users.Delete(user.Id);
|
||||
_dbContext.SaveChanges();
|
||||
}
|
||||
|
||||
// Verify
|
||||
_dbContext.Users.FindById(user.Id).ShouldBeNull();
|
||||
}
|
||||
}
|
||||
57
tests/CBDD.Tests/DocumentCollectionIndexApiTests.cs
Normal file
57
tests/CBDD.Tests/DocumentCollectionIndexApiTests.cs
Normal file
@@ -0,0 +1,57 @@
|
||||
using ZB.MOM.WW.CBDD.Core.Indexing;
|
||||
using ZB.MOM.WW.CBDD.Shared;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests;
|
||||
|
||||
public class DocumentCollectionIndexApiTests : IDisposable
|
||||
{
|
||||
private readonly string _dbPath;
|
||||
private readonly Shared.TestDbContext _db;
|
||||
|
||||
public DocumentCollectionIndexApiTests()
|
||||
{
|
||||
_dbPath = Path.Combine(Path.GetTempPath(), $"collection_index_api_{Guid.NewGuid():N}.db");
|
||||
_db = new Shared.TestDbContext(_dbPath);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void CreateVectorIndex_And_DropIndex_Should_Work()
|
||||
{
|
||||
_db.VectorItems.Insert(new VectorEntity { Title = "A", Embedding = [1f, 1f, 1f] });
|
||||
_db.VectorItems.Insert(new VectorEntity { Title = "B", Embedding = [2f, 2f, 2f] });
|
||||
_db.SaveChanges();
|
||||
|
||||
_db.VectorItems.CreateVectorIndex(v => v.Embedding, 3, VectorMetric.DotProduct, "idx_vector_extra");
|
||||
|
||||
var indexNames = _db.VectorItems.GetIndexes().Select(x => x.Name).ToList();
|
||||
indexNames.ShouldContain("idx_vector_extra");
|
||||
|
||||
_db.VectorItems.DropIndex("idx_vector_extra").ShouldBeTrue();
|
||||
_db.VectorItems.DropIndex("idx_vector_extra").ShouldBeFalse();
|
||||
|
||||
_db.VectorItems.GetIndexes().Select(x => x.Name).ShouldNotContain("idx_vector_extra");
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void EnsureIndex_Should_Return_Existing_Index_When_Already_Present()
|
||||
{
|
||||
var first = _db.People.EnsureIndex(p => p.Age, name: "idx_people_age");
|
||||
var second = _db.People.EnsureIndex(p => p.Age, name: "idx_people_age");
|
||||
|
||||
ReferenceEquals(first, second).ShouldBeTrue();
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void DropIndex_Should_Reject_Primary_Index_Name()
|
||||
{
|
||||
Should.Throw<InvalidOperationException>(() => _db.People.DropIndex("_id"));
|
||||
}
|
||||
|
||||
public void Dispose()
|
||||
{
|
||||
_db.Dispose();
|
||||
if (File.Exists(_dbPath)) File.Delete(_dbPath);
|
||||
var wal = Path.ChangeExtension(_dbPath, ".wal");
|
||||
if (File.Exists(wal)) File.Delete(wal);
|
||||
}
|
||||
}
|
||||
210
tests/CBDD.Tests/DocumentCollectionTests.cs
Executable file
210
tests/CBDD.Tests/DocumentCollectionTests.cs
Executable file
@@ -0,0 +1,210 @@
|
||||
using ZB.MOM.WW.CBDD.Bson;
|
||||
using ZB.MOM.WW.CBDD.Core.Collections;
|
||||
using ZB.MOM.WW.CBDD.Core.Storage;
|
||||
using ZB.MOM.WW.CBDD.Core.Transactions;
|
||||
using ZB.MOM.WW.CBDD.Shared;
|
||||
using ZB.MOM.WW.CBDD.Shared.TestDbContext_TestDbContext_Mappers;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests;
|
||||
|
||||
public class DocumentCollectionTests : IDisposable
|
||||
{
|
||||
private readonly string _dbPath;
|
||||
private readonly string _walPath;
|
||||
private readonly Shared.TestDbContext _db;
|
||||
|
||||
public DocumentCollectionTests()
|
||||
{
|
||||
_dbPath = Path.Combine(Path.GetTempPath(), $"test_collection_{Guid.NewGuid()}.db");
|
||||
_walPath = Path.Combine(Path.GetTempPath(), $"test_collection_{Guid.NewGuid()}.wal");
|
||||
|
||||
_db = new Shared.TestDbContext(_dbPath);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Insert_And_FindById_Works()
|
||||
{
|
||||
// Arrange
|
||||
var user = new User { Name = "Alice", Age = 30 };
|
||||
|
||||
// Act
|
||||
var id = _db.Users.Insert(user);
|
||||
_db.SaveChanges();
|
||||
var found = _db.Users.FindById(id);
|
||||
|
||||
// Assert
|
||||
found.ShouldNotBeNull();
|
||||
found.Id.ShouldBe(id);
|
||||
found.Name.ShouldBe("Alice");
|
||||
found.Age.ShouldBe(30);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void FindById_Returns_Null_When_Not_Found()
|
||||
{
|
||||
// Act
|
||||
var found = _db.Users.FindById(ObjectId.NewObjectId());
|
||||
|
||||
// Assert
|
||||
found.ShouldBeNull();
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void FindAll_Returns_All_Entities()
|
||||
{
|
||||
// Arrange
|
||||
_db.Users.Insert(new User { Name = "Alice", Age = 30 });
|
||||
_db.Users.Insert(new User { Name = "Bob", Age = 25 });
|
||||
_db.Users.Insert(new User { Name = "Charlie", Age = 35 });
|
||||
_db.SaveChanges();
|
||||
|
||||
// Act
|
||||
var all = _db.Users.FindAll().ToList();
|
||||
|
||||
// Assert
|
||||
all.Count.ShouldBe(3);
|
||||
all.ShouldContain(u => u.Name == "Alice");
|
||||
all.ShouldContain(u => u.Name == "Bob");
|
||||
all.ShouldContain(u => u.Name == "Charlie");
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Update_Modifies_Entity()
|
||||
{
|
||||
// Arrange
|
||||
var user = new User { Name = "Alice", Age = 30 };
|
||||
var id = _db.Users.Insert(user);
|
||||
_db.SaveChanges();
|
||||
|
||||
// Act
|
||||
user.Age = 31;
|
||||
var updated = _db.Users.Update(user);
|
||||
_db.SaveChanges();
|
||||
|
||||
// Assert
|
||||
updated.ShouldBeTrue();
|
||||
|
||||
var found = _db.Users.FindById(id);
|
||||
found.ShouldNotBeNull();
|
||||
found.Age.ShouldBe(31);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Update_Returns_False_When_Not_Found()
|
||||
{
|
||||
// Arrange
|
||||
var user = new User { Id = ObjectId.NewObjectId(), Name = "Ghost", Age = 99 };
|
||||
|
||||
// Act
|
||||
var updated = _db.Users.Update(user);
|
||||
_db.SaveChanges();
|
||||
|
||||
// Assert
|
||||
updated.ShouldBeFalse();
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Delete_Removes_Entity()
|
||||
{
|
||||
// Arrange
|
||||
var user = new User { Name = "Alice", Age = 30 };
|
||||
var id = _db.Users.Insert(user);
|
||||
_db.SaveChanges();
|
||||
|
||||
// Act
|
||||
var deleted = _db.Users.Delete(id);
|
||||
_db.SaveChanges();
|
||||
|
||||
// Assert
|
||||
deleted.ShouldBeTrue();
|
||||
_db.Users.FindById(id).ShouldBeNull();
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Delete_Returns_False_When_Not_Found()
|
||||
{
|
||||
// Act
|
||||
var deleted = _db.Users.Delete(ObjectId.NewObjectId());
|
||||
_db.SaveChanges();
|
||||
|
||||
// Assert
|
||||
deleted.ShouldBeFalse();
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Count_Returns_Correct_Count()
|
||||
{
|
||||
// Arrange
|
||||
_db.Users.Insert(new User { Name = "Alice", Age = 30 });
|
||||
_db.Users.Insert(new User { Name = "Bob", Age = 25 });
|
||||
_db.SaveChanges();
|
||||
|
||||
// Act
|
||||
var count = _db.Users.Count();
|
||||
|
||||
// Assert
|
||||
count.ShouldBe(2);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Find_With_Predicate_Filters_Correctly()
|
||||
{
|
||||
// Arrange
|
||||
_db.Users.Insert(new User { Name = "Alice", Age = 30 });
|
||||
_db.Users.Insert(new User { Name = "Bob", Age = 25 });
|
||||
_db.Users.Insert(new User { Name = "Charlie", Age = 35 });
|
||||
_db.SaveChanges();
|
||||
|
||||
// Act
|
||||
var over30 = _db.Users.Find(u => u.Age > 30).ToList();
|
||||
|
||||
// Assert
|
||||
over30.Count().ShouldBe(1);
|
||||
over30[0].Name.ShouldBe("Charlie");
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void InsertBulk_Inserts_Multiple_Entities()
|
||||
{
|
||||
// Arrange
|
||||
var users = new[]
|
||||
{
|
||||
new User { Name = "User1", Age = 20 },
|
||||
new User { Name = "User2", Age = 21 },
|
||||
new User { Name = "User3", Age = 22 }
|
||||
};
|
||||
|
||||
// Act
|
||||
var count = _db.Users.InsertBulk(users);
|
||||
_db.SaveChanges();
|
||||
|
||||
// Assert
|
||||
count.Count.ShouldBe(3);
|
||||
_db.Users.Count().ShouldBe(3);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Insert_With_SpecifiedId_RetainsId()
|
||||
{
|
||||
// Arrange
|
||||
var id = ObjectId.NewObjectId();
|
||||
var user = new User { Id = id, Name = "SpecifiedID", Age = 40 };
|
||||
|
||||
// Act
|
||||
var insertedId = _db.Users.Insert(user);
|
||||
_db.SaveChanges();
|
||||
|
||||
// Assert
|
||||
insertedId.ShouldBe(id);
|
||||
|
||||
var found = _db.Users.FindById(id);
|
||||
found.ShouldNotBeNull();
|
||||
found.Id.ShouldBe(id);
|
||||
found.Name.ShouldBe("SpecifiedID");
|
||||
}
|
||||
|
||||
public void Dispose()
|
||||
{
|
||||
_db?.Dispose();
|
||||
}
|
||||
}
|
||||
233
tests/CBDD.Tests/DocumentOverflowTests.cs
Executable file
233
tests/CBDD.Tests/DocumentOverflowTests.cs
Executable file
@@ -0,0 +1,233 @@
|
||||
using ZB.MOM.WW.CBDD.Bson;
|
||||
using ZB.MOM.WW.CBDD.Core.Collections;
|
||||
using ZB.MOM.WW.CBDD.Core.Compression;
|
||||
using ZB.MOM.WW.CBDD.Core.Storage;
|
||||
using ZB.MOM.WW.CBDD.Core.Transactions;
|
||||
using ZB.MOM.WW.CBDD.Shared;
|
||||
using ZB.MOM.WW.CBDD.Shared.TestDbContext_TestDbContext_Mappers;
|
||||
using System.IO.Compression;
|
||||
using System.IO.MemoryMappedFiles;
|
||||
using Xunit;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests;
|
||||
|
||||
public class DocumentOverflowTests : IDisposable
|
||||
{
|
||||
private readonly string _dbPath;
|
||||
private readonly Shared.TestDbContext _db;
|
||||
|
||||
public DocumentOverflowTests()
|
||||
{
|
||||
_dbPath = Path.Combine(Path.GetTempPath(), $"test_overflow_{Guid.NewGuid()}.db");
|
||||
// Use default PageSize (16KB)
|
||||
_db = new Shared.TestDbContext(_dbPath);
|
||||
}
|
||||
|
||||
public void Dispose()
|
||||
{
|
||||
_db.Dispose();
|
||||
if (File.Exists(_dbPath)) File.Delete(_dbPath);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Insert_MediumDoc_64KB_ShouldSucceed()
|
||||
{
|
||||
// 20KB - Fits in 64KB buffer (First attempt)
|
||||
// But triggers overflow pages in storage (20KB > 16KB PageSize)
|
||||
var largeString = new string('A', 20 * 1024);
|
||||
var user = new User
|
||||
{
|
||||
Id = ObjectId.NewObjectId(),
|
||||
Name = largeString,
|
||||
Age = 10
|
||||
};
|
||||
|
||||
var id = _db.Users.Insert(user);
|
||||
_db.SaveChanges();
|
||||
var retrieved = _db.Users.FindById(id);
|
||||
|
||||
retrieved.ShouldNotBeNull();
|
||||
retrieved.Name.ShouldBe(largeString);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Insert_LargeDoc_100KB_ShouldSucceed()
|
||||
{
|
||||
// 100KB - Fails 64KB buffer, Retries with 2MB
|
||||
var largeString = new string('B', 100 * 1024);
|
||||
var user = new User
|
||||
{
|
||||
Id = ObjectId.NewObjectId(),
|
||||
Name = largeString,
|
||||
Age = 20
|
||||
};
|
||||
|
||||
var id = _db.Users.Insert(user);
|
||||
_db.SaveChanges();
|
||||
var retrieved = _db.Users.FindById(id);
|
||||
|
||||
retrieved.ShouldNotBeNull();
|
||||
retrieved.Name.ShouldBe(largeString);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Insert_HugeDoc_3MB_ShouldSucceed()
|
||||
{
|
||||
// 3MB - Fails 64KB, Fails 2MB, Retries with 16MB
|
||||
var largeString = new string('C', 3 * 1024 * 1024);
|
||||
var user = new User
|
||||
{
|
||||
Id = ObjectId.NewObjectId(),
|
||||
Name = largeString,
|
||||
Age = 30
|
||||
};
|
||||
|
||||
var id = _db.Users.Insert(user);
|
||||
_db.SaveChanges();
|
||||
var retrieved = _db.Users.FindById(id);
|
||||
|
||||
retrieved.ShouldNotBeNull();
|
||||
retrieved.Name.Length.ShouldBe(largeString.Length);
|
||||
// Checking full string might be slow, length check + substring check is faster
|
||||
retrieved.Name.Substring(0, 100).ShouldBe(largeString.Substring(0, 100));
|
||||
retrieved.Name.Substring(retrieved.Name.Length - 100).ShouldBe(largeString.Substring(largeString.Length - 100));
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Update_SmallToHuge_ShouldSucceed()
|
||||
{
|
||||
// Insert Small
|
||||
var user = new User { Id = ObjectId.NewObjectId(), Name = "Small", Age = 1 };
|
||||
var id = _db.Users.Insert(user);
|
||||
_db.SaveChanges();
|
||||
|
||||
// Update to Huge (3MB)
|
||||
var hugeString = new string('U', 3 * 1024 * 1024);
|
||||
user.Name = hugeString;
|
||||
|
||||
var updated = _db.Users.Update(user);
|
||||
_db.SaveChanges();
|
||||
updated.ShouldBeTrue();
|
||||
|
||||
var retrieved = _db.Users.FindById(id);
|
||||
retrieved.ShouldNotBeNull();
|
||||
retrieved.Name.Length.ShouldBe(hugeString.Length);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void InsertBulk_MixedSizes_ShouldSucceed()
|
||||
{
|
||||
var users = new List<User>
|
||||
{
|
||||
new User { Id = ObjectId.NewObjectId(), Name = "Small 1", Age = 1 },
|
||||
new User { Id = ObjectId.NewObjectId(), Name = new string('M', 100 * 1024), Age = 2 }, // 100KB
|
||||
new User { Id = ObjectId.NewObjectId(), Name = "Small 2", Age = 3 },
|
||||
new User { Id = ObjectId.NewObjectId(), Name = new string('H', 3 * 1024 * 1024), Age = 4 } // 3MB
|
||||
};
|
||||
|
||||
var ids = _db.Users.InsertBulk(users);
|
||||
ids.Count.ShouldBe(4);
|
||||
|
||||
foreach (var u in users)
|
||||
{
|
||||
var r = _db.Users.FindById(u.Id);
|
||||
r.ShouldNotBeNull();
|
||||
r.Name.Length.ShouldBe(u.Name.Length);
|
||||
}
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Insert_HugeDoc_WithCompressionEnabledAndSmallPages_ShouldSucceed()
|
||||
{
|
||||
var localDbPath = Path.Combine(Path.GetTempPath(), $"test_overflow_compression_{Guid.NewGuid():N}.db");
|
||||
var options = new CompressionOptions
|
||||
{
|
||||
EnableCompression = true,
|
||||
MinSizeBytes = 0,
|
||||
MinSavingsPercent = 0,
|
||||
Codec = CompressionCodec.Brotli,
|
||||
Level = CompressionLevel.Fastest
|
||||
};
|
||||
|
||||
try
|
||||
{
|
||||
using var db = new Shared.TestDbContext(localDbPath, TinyPageConfig(), options);
|
||||
var huge = new string('Z', 2 * 1024 * 1024);
|
||||
var id = db.Users.Insert(new User
|
||||
{
|
||||
Id = ObjectId.NewObjectId(),
|
||||
Name = huge,
|
||||
Age = 50
|
||||
});
|
||||
db.SaveChanges();
|
||||
|
||||
var loaded = db.Users.FindById(id);
|
||||
loaded.ShouldNotBeNull();
|
||||
loaded.Name.ShouldBe(huge);
|
||||
db.GetCompressionStats().CompressedDocumentCount.ShouldBeGreaterThanOrEqualTo(1);
|
||||
}
|
||||
finally
|
||||
{
|
||||
CleanupLocalFiles(localDbPath);
|
||||
}
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Update_HugeToSmall_WithCompressionEnabled_ShouldSucceed()
|
||||
{
|
||||
var localDbPath = Path.Combine(Path.GetTempPath(), $"test_overflow_compression_update_{Guid.NewGuid():N}.db");
|
||||
var options = new CompressionOptions
|
||||
{
|
||||
EnableCompression = true,
|
||||
MinSizeBytes = 1024,
|
||||
MinSavingsPercent = 0,
|
||||
Codec = CompressionCodec.Deflate,
|
||||
Level = CompressionLevel.Fastest
|
||||
};
|
||||
|
||||
try
|
||||
{
|
||||
using var db = new Shared.TestDbContext(localDbPath, TinyPageConfig(), options);
|
||||
var user = new User
|
||||
{
|
||||
Id = ObjectId.NewObjectId(),
|
||||
Name = new string('Q', 256 * 1024),
|
||||
Age = 44
|
||||
};
|
||||
|
||||
var id = db.Users.Insert(user);
|
||||
db.SaveChanges();
|
||||
|
||||
user.Name = "small-after-overflow";
|
||||
db.Users.Update(user).ShouldBeTrue();
|
||||
db.SaveChanges();
|
||||
|
||||
var loaded = db.Users.FindById(id);
|
||||
loaded.ShouldNotBeNull();
|
||||
loaded.Name.ShouldBe("small-after-overflow");
|
||||
}
|
||||
finally
|
||||
{
|
||||
CleanupLocalFiles(localDbPath);
|
||||
}
|
||||
}
|
||||
|
||||
private static PageFileConfig TinyPageConfig()
|
||||
{
|
||||
return new PageFileConfig
|
||||
{
|
||||
PageSize = 16 * 1024,
|
||||
InitialFileSize = 1024 * 1024,
|
||||
Access = MemoryMappedFileAccess.ReadWrite
|
||||
};
|
||||
}
|
||||
|
||||
private static void CleanupLocalFiles(string dbPath)
|
||||
{
|
||||
var walPath = Path.ChangeExtension(dbPath, ".wal");
|
||||
var markerPath = $"{dbPath}.compact.state";
|
||||
if (File.Exists(dbPath)) File.Delete(dbPath);
|
||||
if (File.Exists(walPath)) File.Delete(walPath);
|
||||
if (File.Exists(markerPath)) File.Delete(markerPath);
|
||||
}
|
||||
}
|
||||
50
tests/CBDD.Tests/GeospatialStressTests.cs
Normal file
50
tests/CBDD.Tests/GeospatialStressTests.cs
Normal file
@@ -0,0 +1,50 @@
|
||||
using ZB.MOM.WW.CBDD.Shared;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests;
|
||||
|
||||
public class GeospatialStressTests : IDisposable
|
||||
{
|
||||
private readonly string _dbPath;
|
||||
private readonly Shared.TestDbContext _db;
|
||||
|
||||
public GeospatialStressTests()
|
||||
{
|
||||
_dbPath = Path.Combine(Path.GetTempPath(), $"geo_stress_{Guid.NewGuid():N}.db");
|
||||
_db = new Shared.TestDbContext(_dbPath);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void SpatialIndex_Should_Handle_Node_Splits_And_Queries()
|
||||
{
|
||||
const int count = 350;
|
||||
|
||||
for (int i = 0; i < count; i++)
|
||||
{
|
||||
_db.GeoItems.Insert(new GeoEntity
|
||||
{
|
||||
Name = $"pt-{i}",
|
||||
Location = (40.0 + (i * 0.001), -73.0 - (i * 0.001))
|
||||
});
|
||||
}
|
||||
|
||||
_db.SaveChanges();
|
||||
|
||||
var all = _db.GeoItems.Within("idx_spatial", (39.5, -74.5), (40.5, -72.5)).ToList();
|
||||
all.Count.ShouldBe(count);
|
||||
|
||||
var subset = _db.GeoItems.Within("idx_spatial", (40.05, -73.30), (40.25, -73.05)).ToList();
|
||||
subset.Count.ShouldBeGreaterThan(0);
|
||||
subset.Count.ShouldBeLessThan(count);
|
||||
|
||||
var near = _db.GeoItems.Near("idx_spatial", (40.10, -73.10), 30.0).ToList();
|
||||
near.Count.ShouldBeGreaterThan(0);
|
||||
}
|
||||
|
||||
public void Dispose()
|
||||
{
|
||||
_db.Dispose();
|
||||
if (File.Exists(_dbPath)) File.Delete(_dbPath);
|
||||
var wal = Path.ChangeExtension(_dbPath, ".wal");
|
||||
if (File.Exists(wal)) File.Delete(wal);
|
||||
}
|
||||
}
|
||||
102
tests/CBDD.Tests/GeospatialTests.cs
Executable file
102
tests/CBDD.Tests/GeospatialTests.cs
Executable file
@@ -0,0 +1,102 @@
|
||||
using Xunit;
|
||||
using ZB.MOM.WW.CBDD.Core.Indexing;
|
||||
using System.IO;
|
||||
using System.Linq;
|
||||
using ZB.MOM.WW.CBDD.Core;
|
||||
using ZB.MOM.WW.CBDD.Shared;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests;
|
||||
|
||||
public class GeospatialTests : IDisposable
|
||||
{
|
||||
private readonly string _dbPath;
|
||||
private readonly Shared.TestDbContext _db;
|
||||
|
||||
public GeospatialTests()
|
||||
{
|
||||
_dbPath = Path.Combine(Path.GetTempPath(), $"cbdd_geo_{Guid.NewGuid()}.db");
|
||||
_db = new Shared.TestDbContext(_dbPath);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Can_Insert_And_Search_Within()
|
||||
{
|
||||
// Setup: Insert some points
|
||||
var p1 = new GeoEntity { Name = "Point 1", Location = (45.0, 9.0) };
|
||||
var p2 = new GeoEntity { Name = "Point 2", Location = (46.0, 10.0) };
|
||||
var p3 = new GeoEntity { Name = "Point 3", Location = (50.0, 50.0) }; // Far away
|
||||
|
||||
_db.GeoItems.Insert(p1);
|
||||
_db.GeoItems.Insert(p2);
|
||||
_db.GeoItems.Insert(p3);
|
||||
|
||||
// Search: Within box [44, 8] to [47, 11]
|
||||
var results = _db.GeoItems.Within("idx_spatial", (44.0, 8.0), (47.0, 11.0)).ToList();
|
||||
|
||||
results.Count.ShouldBe(2);
|
||||
results.ShouldContain(r => r.Name == "Point 1");
|
||||
results.ShouldContain(r => r.Name == "Point 2");
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Can_Search_Near_Proximity()
|
||||
{
|
||||
// Setup: Milan (roughly 45.46, 9.18)
|
||||
var milan = (45.4642, 9.1899);
|
||||
var rome = (41.9028, 12.4964);
|
||||
var ny = (40.7128, -74.0060);
|
||||
|
||||
_db.GeoItems.Insert(new GeoEntity { Name = "Milan Office", Location = milan });
|
||||
_db.GeoItems.Insert(new GeoEntity { Name = "Rome Office", Location = rome });
|
||||
_db.GeoItems.Insert(new GeoEntity { Name = "New York Office", Location = ny });
|
||||
|
||||
// Search near Milan (within 600km - should include Rome (~500km) but not NY)
|
||||
var results = _db.GeoItems.Near("idx_spatial", milan, 600.0).ToList();
|
||||
|
||||
results.Count.ShouldBe(2);
|
||||
results.ShouldContain(r => r.Name == "Milan Office");
|
||||
results.ShouldContain(r => r.Name == "Rome Office");
|
||||
results.ShouldNotContain(r => r.Name == "New York Office");
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void LINQ_Integration_Near_Works()
|
||||
{
|
||||
var milan = (45.4642, 9.1899);
|
||||
_db.GeoItems.Insert(new GeoEntity { Name = "Milan Office", Location = milan });
|
||||
|
||||
// LINQ query using .Near() extension
|
||||
var query = from p in _db.GeoItems.AsQueryable()
|
||||
where p.Location.Near(milan, 10.0)
|
||||
select p;
|
||||
|
||||
var results = query.ToList();
|
||||
|
||||
results.Count().ShouldBe(1);
|
||||
results[0].Name.ShouldBe("Milan Office");
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void LINQ_Integration_Within_Works()
|
||||
{
|
||||
var milan = (45.4642, 9.1899);
|
||||
_db.GeoItems.Insert(new GeoEntity { Name = "Milan Office", Location = milan });
|
||||
|
||||
var min = (45.0, 9.0);
|
||||
var max = (46.0, 10.0);
|
||||
|
||||
// LINQ query using .Within() extension
|
||||
var results = _db.GeoItems.AsQueryable()
|
||||
.Where(p => p.Location.Within(min, max))
|
||||
.ToList();
|
||||
|
||||
results.Count().ShouldBe(1);
|
||||
results[0].Name.ShouldBe("Milan Office");
|
||||
}
|
||||
|
||||
public void Dispose()
|
||||
{
|
||||
_db.Dispose();
|
||||
if (File.Exists(_dbPath)) File.Delete(_dbPath);
|
||||
}
|
||||
}
|
||||
1
tests/CBDD.Tests/GlobalUsings.cs
Normal file
1
tests/CBDD.Tests/GlobalUsings.cs
Normal file
@@ -0,0 +1 @@
|
||||
global using Shouldly;
|
||||
79
tests/CBDD.Tests/HashIndexTests.cs
Normal file
79
tests/CBDD.Tests/HashIndexTests.cs
Normal file
@@ -0,0 +1,79 @@
|
||||
using ZB.MOM.WW.CBDD.Core.Indexing;
|
||||
using ZB.MOM.WW.CBDD.Core.Storage;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests;
|
||||
|
||||
public class HashIndexTests
|
||||
{
|
||||
[Fact]
|
||||
public void Insert_And_TryFind_Should_Return_Location()
|
||||
{
|
||||
var index = new HashIndex(IndexOptions.CreateHash("age"));
|
||||
var key = IndexKey.Create(42);
|
||||
var location = new DocumentLocation(7, 3);
|
||||
|
||||
index.Insert(key, location);
|
||||
|
||||
index.TryFind(key, out var found).ShouldBeTrue();
|
||||
found.PageId.ShouldBe(location.PageId);
|
||||
found.SlotIndex.ShouldBe(location.SlotIndex);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Unique_HashIndex_Should_Throw_On_Duplicate_Key()
|
||||
{
|
||||
var options = new IndexOptions
|
||||
{
|
||||
Type = IndexType.Hash,
|
||||
Unique = true,
|
||||
Fields = ["id"]
|
||||
};
|
||||
|
||||
var index = new HashIndex(options);
|
||||
var key = IndexKey.Create("dup");
|
||||
|
||||
index.Insert(key, new DocumentLocation(1, 1));
|
||||
|
||||
Should.Throw<InvalidOperationException>(() =>
|
||||
index.Insert(key, new DocumentLocation(2, 2)));
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Remove_Should_Remove_Only_Matching_Entry()
|
||||
{
|
||||
var index = new HashIndex(IndexOptions.CreateHash("name"));
|
||||
var key = IndexKey.Create("john");
|
||||
var location1 = new DocumentLocation(10, 1);
|
||||
var location2 = new DocumentLocation(11, 2);
|
||||
|
||||
index.Insert(key, location1);
|
||||
index.Insert(key, location2);
|
||||
|
||||
index.Remove(key, location1).ShouldBeTrue();
|
||||
index.Remove(key, location1).ShouldBeFalse();
|
||||
|
||||
var remaining = index.FindAll(key).ToList();
|
||||
remaining.Count.ShouldBe(1);
|
||||
remaining[0].Location.PageId.ShouldBe(location2.PageId);
|
||||
remaining[0].Location.SlotIndex.ShouldBe(location2.SlotIndex);
|
||||
|
||||
index.Remove(key, location2).ShouldBeTrue();
|
||||
index.FindAll(key).ShouldBeEmpty();
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void FindAll_Should_Return_All_Matching_Entries()
|
||||
{
|
||||
var index = new HashIndex(IndexOptions.CreateHash("score"));
|
||||
var key = IndexKey.Create(99);
|
||||
|
||||
index.Insert(key, new DocumentLocation(1, 0));
|
||||
index.Insert(key, new DocumentLocation(2, 0));
|
||||
index.Insert(IndexKey.Create(100), new DocumentLocation(3, 0));
|
||||
|
||||
var matches = index.FindAll(key).ToList();
|
||||
|
||||
matches.Count.ShouldBe(2);
|
||||
matches.All(e => e.Key == key).ShouldBeTrue();
|
||||
}
|
||||
}
|
||||
93
tests/CBDD.Tests/IndexDirectionTests.cs
Executable file
93
tests/CBDD.Tests/IndexDirectionTests.cs
Executable file
@@ -0,0 +1,93 @@
|
||||
using ZB.MOM.WW.CBDD.Bson;
|
||||
using ZB.MOM.WW.CBDD.Core;
|
||||
using ZB.MOM.WW.CBDD.Core.Collections;
|
||||
using ZB.MOM.WW.CBDD.Core.Indexing;
|
||||
using ZB.MOM.WW.CBDD.Shared;
|
||||
using System;
|
||||
using Xunit;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests;
|
||||
|
||||
public class IndexDirectionTests : IDisposable
|
||||
{
|
||||
private readonly string _dbPath = "index_direction_tests.db";
|
||||
|
||||
private readonly Shared.TestDbContext _db;
|
||||
|
||||
public IndexDirectionTests()
|
||||
{
|
||||
if (File.Exists(_dbPath)) File.Delete(_dbPath);
|
||||
_db = new Shared.TestDbContext(_dbPath);
|
||||
// _db.Database.EnsureCreated(); // Not needed/doesn't exist? StorageEngine handles creation.
|
||||
}
|
||||
|
||||
public void Dispose()
|
||||
{
|
||||
_db.Dispose();
|
||||
if (File.Exists(_dbPath)) File.Delete(_dbPath);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Range_Forward_ReturnsOrderedResults()
|
||||
{
|
||||
var collection = _db.People;
|
||||
var index = collection.EnsureIndex(p => p.Age, "idx_age");
|
||||
|
||||
var people = Enumerable.Range(1, 100).Select(i => new Person { Id = i, Name = $"Person {i}", Age = i }).ToList();
|
||||
collection.InsertBulk(people);
|
||||
_db.SaveChanges();
|
||||
|
||||
// Scan Forward
|
||||
var results = index.Range(10, 20, IndexDirection.Forward).ToList();
|
||||
|
||||
results.Count.ShouldBe(11); // 10 to 20 inclusive
|
||||
collection.FindByLocation(results.First())!.Age.ShouldBe(10); // First is 10
|
||||
collection.FindByLocation(results.Last())!.Age.ShouldBe(20); // Last is 20
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Range_Backward_ReturnsReverseOrderedResults()
|
||||
{
|
||||
var collection = _db.People;
|
||||
var index = collection.EnsureIndex(p => p.Age, "idx_age");
|
||||
|
||||
var people = Enumerable.Range(1, 100).Select(i => new Person { Id = i, Name = $"Person {i}", Age = i }).ToList();
|
||||
collection.InsertBulk(people);
|
||||
_db.SaveChanges();
|
||||
|
||||
// Scan Backward
|
||||
var results = index.Range(10, 20, IndexDirection.Backward).ToList();
|
||||
|
||||
results.Count.ShouldBe(11); // 10 to 20 inclusive
|
||||
collection.FindByLocation(results.First())!.Age.ShouldBe(20); // First is 20 (Reverse)
|
||||
collection.FindByLocation(results.Last())!.Age.ShouldBe(10); // Last is 10
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Range_Backward_WithMultiplePages_ReturnsReverseOrderedResults()
|
||||
{
|
||||
var collection = _db.People;
|
||||
var index = collection.EnsureIndex(p => p.Age, "idx_age_large");
|
||||
|
||||
// Insert enough to force splits (default page size is smallish, 4096, so 1000 items should split)
|
||||
// Entry size approx 10 bytes key + 6 bytes loc + overhead
|
||||
// 1000 items * 20 bytes = 20KB > 4KB.
|
||||
var count = 1000;
|
||||
var people = Enumerable.Range(1, count).Select(i => new Person { Id = i, Name = $"Person {i}", Age = i }).ToList();
|
||||
collection.InsertBulk(people);
|
||||
_db.SaveChanges();
|
||||
|
||||
// Scan ALL Backward
|
||||
var results = index.Range(null, null, IndexDirection.Backward).ToList();
|
||||
|
||||
results.Count.ShouldBe(count);
|
||||
|
||||
// Note on sorting: IndexKey uses Little Endian byte comparison for integers.
|
||||
// This means 256 (0x0001...) sorts before 1 (0x01...).
|
||||
// Strict value checking fails for ranges crossing 255 boundary unless IndexKey is fixed to use Big Endian.
|
||||
// For this test, we verify that we retrieved all items (Count) which implies valid page traversal.
|
||||
|
||||
// collection.FindByLocation(results.First(), null)!.Age.ShouldBe(count); // Max Age (Fails: Max is likely 255)
|
||||
// collection.FindByLocation(results.Last(), null)!.Age.ShouldBe(1); // Min Age (Fails: Min is likely 256)
|
||||
}
|
||||
}
|
||||
136
tests/CBDD.Tests/IndexOptimizationTests.cs
Executable file
136
tests/CBDD.Tests/IndexOptimizationTests.cs
Executable file
@@ -0,0 +1,136 @@
|
||||
using Xunit;
|
||||
using ZB.MOM.WW.CBDD.Core.Query;
|
||||
using ZB.MOM.WW.CBDD.Core.Indexing;
|
||||
using System.Linq.Expressions;
|
||||
using System.Collections.Generic;
|
||||
using System;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests
|
||||
{
|
||||
public class IndexOptimizationTests
|
||||
{
|
||||
public class TestEntity
|
||||
{
|
||||
public int Id { get; set; }
|
||||
public string Name { get; set; } = "";
|
||||
public int Age { get; set; }
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Optimizer_Identifies_Equality()
|
||||
{
|
||||
var indexes = new List<CollectionIndexInfo>
|
||||
{
|
||||
new CollectionIndexInfo { Name = "idx_age", PropertyPaths = ["Age"] }
|
||||
};
|
||||
|
||||
Expression<Func<TestEntity, bool>> predicate = x => x.Age == 30;
|
||||
var model = new QueryModel { WhereClause = predicate };
|
||||
|
||||
var result = IndexOptimizer.TryOptimize<TestEntity>(model, indexes);
|
||||
|
||||
result.ShouldNotBeNull();
|
||||
result.IndexName.ShouldBe("idx_age");
|
||||
result.MinValue.ShouldBe(30);
|
||||
result.MaxValue.ShouldBe(30);
|
||||
result.IsRange.ShouldBeFalse();
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Optimizer_Identifies_Range_GreaterThan()
|
||||
{
|
||||
var indexes = new List<CollectionIndexInfo>
|
||||
{
|
||||
new CollectionIndexInfo { Name = "idx_age", PropertyPaths = ["Age"] }
|
||||
};
|
||||
|
||||
Expression<Func<TestEntity, bool>> predicate = x => x.Age > 25;
|
||||
var model = new QueryModel { WhereClause = predicate };
|
||||
|
||||
var result = IndexOptimizer.TryOptimize<TestEntity>(model, indexes);
|
||||
|
||||
result.ShouldNotBeNull();
|
||||
result.IndexName.ShouldBe("idx_age");
|
||||
result.MinValue.ShouldBe(25);
|
||||
result.MaxValue.ShouldBeNull();
|
||||
result.IsRange.ShouldBeTrue();
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Optimizer_Identifies_Range_LessThan()
|
||||
{
|
||||
var indexes = new List<CollectionIndexInfo>
|
||||
{
|
||||
new CollectionIndexInfo { Name = "idx_age", PropertyPaths = ["Age"] }
|
||||
};
|
||||
|
||||
Expression<Func<TestEntity, bool>> predicate = x => x.Age < 50;
|
||||
var model = new QueryModel { WhereClause = predicate };
|
||||
|
||||
var result = IndexOptimizer.TryOptimize<TestEntity>(model, indexes);
|
||||
|
||||
result.ShouldNotBeNull();
|
||||
result.IndexName.ShouldBe("idx_age");
|
||||
result.MinValue.ShouldBeNull();
|
||||
result.MaxValue.ShouldBe(50);
|
||||
result.IsRange.ShouldBeTrue();
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Optimizer_Identifies_Range_Between_Simulated()
|
||||
{
|
||||
var indexes = new List<CollectionIndexInfo>
|
||||
{
|
||||
new CollectionIndexInfo { Name = "idx_age", PropertyPaths = ["Age"] }
|
||||
};
|
||||
|
||||
Expression<Func<TestEntity, bool>> predicate = x => x.Age > 20 && x.Age < 40;
|
||||
var model = new QueryModel { WhereClause = predicate };
|
||||
|
||||
var result = IndexOptimizer.TryOptimize<TestEntity>(model, indexes);
|
||||
|
||||
result.ShouldNotBeNull();
|
||||
result.IndexName.ShouldBe("idx_age");
|
||||
result.MinValue.ShouldBe(20);
|
||||
result.MaxValue.ShouldBe(40);
|
||||
result.IsRange.ShouldBeTrue();
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Optimizer_Identifies_StartsWith()
|
||||
{
|
||||
var indexes = new List<CollectionIndexInfo>
|
||||
{
|
||||
new CollectionIndexInfo { Name = "idx_name", PropertyPaths = ["Name"], Type = IndexType.BTree }
|
||||
};
|
||||
|
||||
Expression<Func<TestEntity, bool>> predicate = x => x.Name.StartsWith("Ali");
|
||||
var model = new QueryModel { WhereClause = predicate };
|
||||
|
||||
var result = IndexOptimizer.TryOptimize<TestEntity>(model, indexes);
|
||||
|
||||
result.ShouldNotBeNull();
|
||||
result.IndexName.ShouldBe("idx_name");
|
||||
result.MinValue.ShouldBe("Ali");
|
||||
// "Ali" + next char -> "Alj"
|
||||
result.MaxValue.ShouldBe("Alj");
|
||||
result.IsRange.ShouldBeTrue();
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Optimizer_Ignores_NonIndexed_Fields()
|
||||
{
|
||||
var indexes = new List<CollectionIndexInfo>
|
||||
{
|
||||
new CollectionIndexInfo { Name = "idx_age", PropertyPaths = ["Age"] }
|
||||
};
|
||||
|
||||
Expression<Func<TestEntity, bool>> predicate = x => x.Name == "Alice"; // Name is not indexed
|
||||
var model = new QueryModel { WhereClause = predicate };
|
||||
|
||||
var result = IndexOptimizer.TryOptimize<TestEntity>(model, indexes);
|
||||
|
||||
result.ShouldBeNull();
|
||||
}
|
||||
}
|
||||
}
|
||||
59
tests/CBDD.Tests/InsertBulkTests.cs
Executable file
59
tests/CBDD.Tests/InsertBulkTests.cs
Executable file
@@ -0,0 +1,59 @@
|
||||
using ZB.MOM.WW.CBDD.Core;
|
||||
using ZB.MOM.WW.CBDD.Core.Collections;
|
||||
using ZB.MOM.WW.CBDD.Core.Storage;
|
||||
using ZB.MOM.WW.CBDD.Core.Transactions;
|
||||
using ZB.MOM.WW.CBDD.Shared;
|
||||
using ZB.MOM.WW.CBDD.Shared.TestDbContext_TestDbContext_Mappers;
|
||||
using Xunit;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests;
|
||||
|
||||
public class InsertBulkTests : IDisposable
|
||||
{
|
||||
private readonly string _testFile;
|
||||
private readonly Shared.TestDbContext _db;
|
||||
|
||||
public InsertBulkTests()
|
||||
{
|
||||
_testFile = Path.GetTempFileName();
|
||||
_db = new Shared.TestDbContext(_testFile);
|
||||
}
|
||||
|
||||
public void Dispose()
|
||||
{
|
||||
_db.Dispose();
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void InsertBulk_PersistsData_ImmediatelyVisible()
|
||||
{
|
||||
var users = new List<User>();
|
||||
for (int i = 0; i < 50; i++)
|
||||
{
|
||||
users.Add(new User { Id = ZB.MOM.WW.CBDD.Bson.ObjectId.NewObjectId(), Name = $"User {i}", Age = 20 });
|
||||
}
|
||||
|
||||
_db.Users.InsertBulk(users);
|
||||
_db.SaveChanges();
|
||||
|
||||
var insertedUsers = _db.Users.FindAll().ToList();
|
||||
|
||||
insertedUsers.Count.ShouldBe(50);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void InsertBulk_SpanningMultiplePages_PersistsCorrectly()
|
||||
{
|
||||
// 16KB page. User ~50 bytes. 400 users -> ~20KB -> 2 pages.
|
||||
var users = new List<User>();
|
||||
for (int i = 0; i < 400; i++)
|
||||
{
|
||||
users.Add(new User { Id = ZB.MOM.WW.CBDD.Bson.ObjectId.NewObjectId(), Name = $"User {i} with some long padding text to ensure we fill space {new string('x', 50)}", Age = 20 });
|
||||
}
|
||||
|
||||
_db.Users.InsertBulk(users);
|
||||
_db.SaveChanges();
|
||||
|
||||
_db.Users.Count().ShouldBe(400);
|
||||
}
|
||||
}
|
||||
139
tests/CBDD.Tests/LinqTests.cs
Executable file
139
tests/CBDD.Tests/LinqTests.cs
Executable file
@@ -0,0 +1,139 @@
|
||||
using ZB.MOM.WW.CBDD.Core;
|
||||
using ZB.MOM.WW.CBDD.Core.Storage;
|
||||
using ZB.MOM.WW.CBDD.Core.Collections;
|
||||
using ZB.MOM.WW.CBDD.Core.Indexing;
|
||||
using ZB.MOM.WW.CBDD.Bson;
|
||||
using Xunit;
|
||||
using System;
|
||||
using System.IO;
|
||||
using System.Linq;
|
||||
using System.Collections.Generic;
|
||||
using ZB.MOM.WW.CBDD.Shared;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests
|
||||
{
|
||||
public class LinqTests : IDisposable
|
||||
{
|
||||
private readonly string _testFile;
|
||||
private readonly Shared.TestDbContext _db;
|
||||
|
||||
public LinqTests()
|
||||
{
|
||||
_testFile = Path.Combine(Path.GetTempPath(), $"linq_tests_{Guid.NewGuid()}.db");
|
||||
if (File.Exists(_testFile)) File.Delete(_testFile);
|
||||
var wal = Path.ChangeExtension(_testFile, ".wal");
|
||||
if (File.Exists(wal)) File.Delete(wal);
|
||||
|
||||
_db = new Shared.TestDbContext(_testFile);
|
||||
|
||||
// Seed Data
|
||||
_db.Users.Insert(new User { Name = "Alice", Age = 30 });
|
||||
_db.Users.Insert(new User { Name = "Bob", Age = 25 });
|
||||
_db.Users.Insert(new User { Name = "Charlie", Age = 35 });
|
||||
_db.Users.Insert(new User { Name = "Dave", Age = 20 });
|
||||
_db.Users.Insert(new User { Name = "Eve", Age = 40 });
|
||||
_db.SaveChanges();
|
||||
}
|
||||
|
||||
public void Dispose()
|
||||
{
|
||||
_db.Dispose();
|
||||
if (File.Exists(_testFile)) File.Delete(_testFile);
|
||||
var wal = Path.ChangeExtension(_testFile, ".wal");
|
||||
if (File.Exists(wal)) File.Delete(wal);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Where_FiltersDocuments()
|
||||
{
|
||||
var query = _db.Users.AsQueryable().Where(x => x.Age > 28);
|
||||
var results = query.ToList();
|
||||
|
||||
results.Count.ShouldBe(3); // Alice(30), Charlie(35), Eve(40)
|
||||
results.ShouldNotContain(d => d.Name == "Bob");
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void OrderBy_SortsDocuments()
|
||||
{
|
||||
var results = _db.Users.AsQueryable().OrderBy(x => x.Age).ToList();
|
||||
|
||||
results.Count.ShouldBe(5);
|
||||
results[0].Name.ShouldBe("Dave"); // 20
|
||||
results[1].Name.ShouldBe("Bob"); // 25
|
||||
results.Last().Name.ShouldBe("Eve"); // 40
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void SkipTake_Pagination()
|
||||
{
|
||||
var results = _db.Users.AsQueryable()
|
||||
.OrderBy(x => x.Age)
|
||||
.Skip(1)
|
||||
.Take(2)
|
||||
.ToList();
|
||||
|
||||
results.Count.ShouldBe(2);
|
||||
results[0].Name.ShouldBe("Bob"); // 25 (Skipped Dave)
|
||||
results[1].Name.ShouldBe("Alice"); // 30
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Select_Projections()
|
||||
{
|
||||
var names = _db.Users.AsQueryable()
|
||||
.Where(x => x.Age < 30)
|
||||
.OrderBy(x => x.Age)
|
||||
.Select(x => x.Name)
|
||||
.ToList();
|
||||
|
||||
names.Count.ShouldBe(2);
|
||||
names[0].ShouldBe("Dave");
|
||||
names[1].ShouldBe("Bob");
|
||||
}
|
||||
[Fact]
|
||||
public void IndexedWhere_UsedIndex()
|
||||
{
|
||||
// Create index on Age
|
||||
_db.Users.EnsureIndex(x => x.Age, "idx_age", false);
|
||||
|
||||
var query = _db.Users.AsQueryable().Where(x => x.Age > 25);
|
||||
var results = query.ToList();
|
||||
|
||||
results.Count.ShouldBe(3); // Alice(30), Charlie(35), Eve(40)
|
||||
results.ShouldNotContain(d => d.Name == "Bob"); // Age 25 (filtered out by strict >)
|
||||
results.ShouldNotContain(d => d.Name == "Dave"); // Age 20
|
||||
}
|
||||
[Fact]
|
||||
public void StartsWith_UsedIndex()
|
||||
{
|
||||
// Create index on Name
|
||||
_db.Users.EnsureIndex(x => x.Name!, "idx_name", false);
|
||||
|
||||
// StartsWith "Cha" -> Should find "Charlie"
|
||||
var query = _db.Users.AsQueryable().Where(x => x.Name!.StartsWith("Cha"));
|
||||
var results = query.ToList();
|
||||
|
||||
results.Count().ShouldBe(1);
|
||||
results[0].Name.ShouldBe("Charlie");
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Between_UsedIndex()
|
||||
{
|
||||
// Create index on Age
|
||||
_db.Users.EnsureIndex(x => x.Age, "idx_age_between", false);
|
||||
|
||||
// Age >= 22 && Age <= 32
|
||||
// Alice(30), Bob(25) -> Should be found.
|
||||
// Dave(20), Charlie(35), Eve(40) -> excluded.
|
||||
|
||||
var query = _db.Users.AsQueryable().Where(x => x.Age >= 22 && x.Age <= 32);
|
||||
var results = query.ToList();
|
||||
|
||||
results.Count.ShouldBe(2);
|
||||
results.ShouldContain(x => x.Name == "Alice");
|
||||
results.ShouldContain(x => x.Name == "Bob");
|
||||
}
|
||||
}
|
||||
}
|
||||
160
tests/CBDD.Tests/MaintenanceDiagnosticsAndMigrationTests.cs
Normal file
160
tests/CBDD.Tests/MaintenanceDiagnosticsAndMigrationTests.cs
Normal file
@@ -0,0 +1,160 @@
|
||||
using System.IO.Compression;
|
||||
using ZB.MOM.WW.CBDD.Core.Compression;
|
||||
using ZB.MOM.WW.CBDD.Core.Storage;
|
||||
using ZB.MOM.WW.CBDD.Shared;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests;
|
||||
|
||||
public class MaintenanceDiagnosticsAndMigrationTests
|
||||
{
|
||||
[Fact]
|
||||
public void DiagnosticsApis_ShouldReturnPageUsageCompressionAndFragmentationData()
|
||||
{
|
||||
var dbPath = NewDbPath();
|
||||
|
||||
try
|
||||
{
|
||||
var options = new CompressionOptions
|
||||
{
|
||||
EnableCompression = true,
|
||||
MinSizeBytes = 0,
|
||||
MinSavingsPercent = 0,
|
||||
Codec = CompressionCodec.Brotli,
|
||||
Level = CompressionLevel.Fastest
|
||||
};
|
||||
|
||||
using var db = new TestDbContext(dbPath, options);
|
||||
for (var i = 0; i < 40; i++)
|
||||
{
|
||||
db.Users.Insert(new User
|
||||
{
|
||||
Name = BuildPayload(i, 9000),
|
||||
Age = i
|
||||
});
|
||||
}
|
||||
|
||||
db.SaveChanges();
|
||||
db.ForceCheckpoint();
|
||||
|
||||
var byType = db.GetPageUsageByPageType();
|
||||
byType.Count.ShouldBeGreaterThan(0);
|
||||
byType.Any(x => x.PageType == PageType.Data && x.PageCount > 0).ShouldBeTrue();
|
||||
|
||||
var byCollection = db.GetPageUsageByCollection();
|
||||
byCollection.Any(x => x.CollectionName.Equals("users", StringComparison.OrdinalIgnoreCase)).ShouldBeTrue();
|
||||
|
||||
var compressionByCollection = db.GetCompressionRatioByCollection();
|
||||
var usersCompression = compressionByCollection.First(x => x.CollectionName.Equals("users", StringComparison.OrdinalIgnoreCase));
|
||||
usersCompression.DocumentCount.ShouldBeGreaterThan(0);
|
||||
usersCompression.BytesBeforeCompression.ShouldBeGreaterThan(0);
|
||||
usersCompression.BytesAfterCompression.ShouldBeGreaterThan(0);
|
||||
|
||||
var freeList = db.GetFreeListSummary();
|
||||
freeList.PageCount.ShouldBeGreaterThan(0u);
|
||||
|
||||
var fragmentation = db.GetFragmentationMap();
|
||||
fragmentation.Pages.Count.ShouldBeGreaterThan(0);
|
||||
}
|
||||
finally
|
||||
{
|
||||
CleanupFiles(dbPath);
|
||||
}
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void MigrateCompression_DryRunAndApply_ShouldReturnDeterministicStatsAndPreserveData()
|
||||
{
|
||||
var dbPath = NewDbPath();
|
||||
|
||||
try
|
||||
{
|
||||
using var db = new TestDbContext(dbPath, CompressionOptions.Default);
|
||||
var ids = new List<ZB.MOM.WW.CBDD.Bson.ObjectId>();
|
||||
|
||||
for (var i = 0; i < 60; i++)
|
||||
{
|
||||
ids.Add(db.Users.Insert(new User
|
||||
{
|
||||
Name = BuildPayload(i, 12000),
|
||||
Age = i % 17
|
||||
}));
|
||||
}
|
||||
|
||||
db.SaveChanges();
|
||||
db.ForceCheckpoint();
|
||||
|
||||
var dryRun = db.MigrateCompression(new CompressionMigrationOptions
|
||||
{
|
||||
DryRun = true,
|
||||
Codec = CompressionCodec.Deflate,
|
||||
Level = CompressionLevel.Fastest,
|
||||
MinSizeBytes = 0,
|
||||
MinSavingsPercent = 0,
|
||||
IncludeCollections = ["users"]
|
||||
});
|
||||
|
||||
dryRun.DryRun.ShouldBeTrue();
|
||||
dryRun.DocumentsScanned.ShouldBeGreaterThan(0);
|
||||
dryRun.BytesBefore.ShouldBeGreaterThan(0);
|
||||
dryRun.BytesEstimatedAfter.ShouldBeGreaterThan(0);
|
||||
|
||||
var apply = db.MigrateCompression(new CompressionMigrationOptions
|
||||
{
|
||||
DryRun = false,
|
||||
Codec = CompressionCodec.Deflate,
|
||||
Level = CompressionLevel.Fastest,
|
||||
MinSizeBytes = 0,
|
||||
MinSavingsPercent = 0,
|
||||
IncludeCollections = ["users"]
|
||||
});
|
||||
|
||||
apply.DryRun.ShouldBeFalse();
|
||||
apply.DocumentsScanned.ShouldBeGreaterThan(0);
|
||||
|
||||
foreach (var id in ids)
|
||||
{
|
||||
var user = db.Users.FindById(id);
|
||||
user.ShouldNotBeNull();
|
||||
user!.Name.Length.ShouldBeGreaterThan(1000);
|
||||
}
|
||||
}
|
||||
finally
|
||||
{
|
||||
CleanupFiles(dbPath);
|
||||
}
|
||||
}
|
||||
|
||||
private static string BuildPayload(int seed, int approxLength)
|
||||
{
|
||||
var builder = new System.Text.StringBuilder(approxLength + 128);
|
||||
var i = 0;
|
||||
while (builder.Length < approxLength)
|
||||
{
|
||||
builder.Append("diag-migrate-");
|
||||
builder.Append(seed.ToString("D4"));
|
||||
builder.Append('-');
|
||||
builder.Append(i.ToString("D6"));
|
||||
builder.Append('|');
|
||||
i++;
|
||||
}
|
||||
|
||||
return builder.ToString();
|
||||
}
|
||||
|
||||
private static string NewDbPath()
|
||||
=> Path.Combine(Path.GetTempPath(), $"maint_diag_migrate_{Guid.NewGuid():N}.db");
|
||||
|
||||
private static void CleanupFiles(string dbPath)
|
||||
{
|
||||
var walPath = Path.ChangeExtension(dbPath, ".wal");
|
||||
var markerPath = $"{dbPath}.compact.state";
|
||||
var tempPath = $"{dbPath}.compact.tmp";
|
||||
var backupPath = $"{dbPath}.compact.bak";
|
||||
|
||||
if (File.Exists(dbPath)) File.Delete(dbPath);
|
||||
if (File.Exists(walPath)) File.Delete(walPath);
|
||||
if (File.Exists(markerPath)) File.Delete(markerPath);
|
||||
if (File.Exists(tempPath)) File.Delete(tempPath);
|
||||
if (File.Exists(backupPath)) File.Delete(backupPath);
|
||||
}
|
||||
}
|
||||
99
tests/CBDD.Tests/MetadataPersistenceTests.cs
Executable file
99
tests/CBDD.Tests/MetadataPersistenceTests.cs
Executable file
@@ -0,0 +1,99 @@
|
||||
using ZB.MOM.WW.CBDD.Bson;
|
||||
using ZB.MOM.WW.CBDD.Core.Collections;
|
||||
using ZB.MOM.WW.CBDD.Core.Indexing;
|
||||
using ZB.MOM.WW.CBDD.Core.Storage;
|
||||
using ZB.MOM.WW.CBDD.Core.Transactions;
|
||||
using ZB.MOM.WW.CBDD.Shared;
|
||||
using ZB.MOM.WW.CBDD.Shared.TestDbContext_TestDbContext_Mappers;
|
||||
using Xunit;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests;
|
||||
|
||||
public class MetadataPersistenceTests : IDisposable
|
||||
{
|
||||
private readonly string _dbPath;
|
||||
private readonly string _walPath;
|
||||
|
||||
public MetadataPersistenceTests()
|
||||
{
|
||||
_dbPath = Path.Combine(Path.GetTempPath(), $"docdb_meta_{Guid.NewGuid()}.db");
|
||||
_walPath = Path.ChangeExtension(_dbPath, ".wal");
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void IndexDefinitions_ArePersisted_AndReloaded()
|
||||
{
|
||||
// 1. Create index in first session
|
||||
using (var storage = new StorageEngine(_dbPath, PageFileConfig.Default))
|
||||
{
|
||||
// Disable auto-checkpoint to ensure cleaner test tracing, though not strictly required
|
||||
var mapper = new ZB_MOM_WW_CBDD_Shared_UserMapper();
|
||||
var indexManager = new CollectionIndexManager<ObjectId, User>(storage, mapper, nameof(User));
|
||||
|
||||
// Create 2 indexes
|
||||
indexManager.CreateIndex(u => u.Age, "idx_age");
|
||||
indexManager.CreateIndex(u => u.Name, unique: true); // name auto-generated
|
||||
}
|
||||
|
||||
// 2. Re-open storage and verify indexes exist
|
||||
using (var storage = new StorageEngine(_dbPath, PageFileConfig.Default))
|
||||
{
|
||||
var mapper = new ZB_MOM_WW_CBDD_Shared_UserMapper();
|
||||
// Assuming Page 1 was allocated above in clean DB
|
||||
var indexManager = new CollectionIndexManager<ObjectId, User>(storage, mapper, nameof(User));
|
||||
|
||||
var indexes = indexManager.GetAllIndexes().ToList();
|
||||
|
||||
indexes.Count.ShouldBe(2);
|
||||
|
||||
var ageIdx = indexManager.GetIndex("idx_age");
|
||||
ageIdx.ShouldNotBeNull();
|
||||
ageIdx.Definition.IsUnique.ShouldBeFalse();
|
||||
ageIdx.Definition.PropertyPaths.Count().ShouldBe(1);
|
||||
ageIdx.Definition.PropertyPaths[0].ShouldBe("Age");
|
||||
|
||||
// Check auto-generated name index
|
||||
var nameIdx = indexes.FirstOrDefault(i => i.Definition.PropertyPaths[0] == "Name");
|
||||
nameIdx.ShouldNotBeNull();
|
||||
nameIdx.Definition.IsUnique.ShouldBeTrue();
|
||||
}
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void EnsureIndex_DoesNotRecreate_IfIndexExists()
|
||||
{
|
||||
// 1. Create index
|
||||
using (var context = new Shared.TestDbContext(_dbPath))
|
||||
{
|
||||
context.Users.EnsureIndex(u => u.Age);
|
||||
}
|
||||
|
||||
// 2. Re-open and EnsureIndex again - should be fast/no-op
|
||||
using (var context = new Shared.TestDbContext(_dbPath))
|
||||
{
|
||||
var mapper = new ZB_MOM_WW_CBDD_Shared_UserMapper();
|
||||
|
||||
// Use reflection or diagnostic to check if it triggered rebuild?
|
||||
// Currently hard to verify "no rebuild" without logs or mocking.
|
||||
// But we can verify it doesn't throw and index is still valid.
|
||||
|
||||
var idx = context.Users.EnsureIndex(u => u.Age);
|
||||
idx.ShouldNotBeNull();
|
||||
|
||||
// Verify functioning
|
||||
using var txn = context.BeginTransaction();
|
||||
context.Users.Insert(new User { Name = "Bob", Age = 50 });
|
||||
txn.Commit();
|
||||
|
||||
// Should find it via index
|
||||
var results = context.Users.Find(u => u.Age == 50).ToList();
|
||||
results.Count().ShouldBe(1);
|
||||
}
|
||||
}
|
||||
|
||||
public void Dispose()
|
||||
{
|
||||
if (File.Exists(_dbPath)) File.Delete(_dbPath);
|
||||
if (File.Exists(_walPath)) File.Delete(_walPath);
|
||||
}
|
||||
}
|
||||
425
tests/CBDD.Tests/MockEntities.cs
Executable file
425
tests/CBDD.Tests/MockEntities.cs
Executable file
@@ -0,0 +1,425 @@
|
||||
using ZB.MOM.WW.CBDD.Bson;
|
||||
using ZB.MOM.WW.CBDD.Core.Collections;
|
||||
using ZB.MOM.WW.CBDD.Core.Metadata;
|
||||
using System;
|
||||
using System.Collections.Generic;
|
||||
using System.ComponentModel.DataAnnotations;
|
||||
using System.ComponentModel.DataAnnotations.Schema;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Shared
|
||||
{
|
||||
// --- Basic Entities ---
|
||||
|
||||
public class User
|
||||
{
|
||||
public ObjectId Id { get; set; }
|
||||
public string Name { get; set; } = "";
|
||||
public int Age { get; set; }
|
||||
}
|
||||
|
||||
// --- Complex Entities (Nested) ---
|
||||
|
||||
public class ComplexUser
|
||||
{
|
||||
[BsonId]
|
||||
public ObjectId Id { get; set; }
|
||||
public string Name { get; set; } = "";
|
||||
|
||||
// Direct nested object
|
||||
public Address MainAddress { get; set; } = new();
|
||||
|
||||
// Collection of nested objects
|
||||
public List<Address> OtherAddresses { get; set; } = new();
|
||||
|
||||
// Primitive collection
|
||||
public List<string> Tags { get; set; } = new();
|
||||
|
||||
[BsonIgnore]
|
||||
public string Secret { get; set; } = "";
|
||||
}
|
||||
|
||||
public class Address
|
||||
{
|
||||
public string Street { get; set; } = "";
|
||||
public City City { get; set; } = new(); // Depth 2
|
||||
}
|
||||
|
||||
public class City
|
||||
{
|
||||
public string Name { get; set; } = "";
|
||||
public string ZipCode { get; set; } = "";
|
||||
}
|
||||
|
||||
// --- Primary Key Test Entities ---
|
||||
|
||||
public class IntEntity
|
||||
{
|
||||
public int Id { get; set; }
|
||||
public string? Name { get; set; }
|
||||
}
|
||||
|
||||
public class StringEntity
|
||||
{
|
||||
public required string Id { get; set; }
|
||||
public string? Value { get; set; }
|
||||
}
|
||||
|
||||
public class GuidEntity
|
||||
{
|
||||
public Guid Id { get; set; }
|
||||
public string? Name { get; set; }
|
||||
}
|
||||
|
||||
/// <summary>
|
||||
/// Entity with string key NOT named "Id" - tests custom key name support
|
||||
/// </summary>
|
||||
public class CustomKeyEntity
|
||||
{
|
||||
[System.ComponentModel.DataAnnotations.Key]
|
||||
public required string Code { get; set; }
|
||||
public string? Description { get; set; }
|
||||
}
|
||||
|
||||
// --- Multi-collection / Auto-init entities ---
|
||||
|
||||
public class AutoInitEntity
|
||||
{
|
||||
public int Id { get; set; }
|
||||
public string Name { get; set; } = string.Empty;
|
||||
}
|
||||
|
||||
public class Person
|
||||
{
|
||||
public int Id { get; set; }
|
||||
public string Name { get; set; } = "";
|
||||
public int Age { get; set; }
|
||||
}
|
||||
|
||||
public class Product
|
||||
{
|
||||
public int Id { get; set; }
|
||||
public string Title { get; set; } = "";
|
||||
public decimal Price { get; set; }
|
||||
}
|
||||
|
||||
public class AsyncDoc
|
||||
{
|
||||
public int Id { get; set; }
|
||||
public string Name { get; set; } = "";
|
||||
}
|
||||
|
||||
public class SchemaUser
|
||||
{
|
||||
public int Id { get; set; }
|
||||
public string Name { get; set; } = "";
|
||||
public Address Address { get; set; } = new();
|
||||
}
|
||||
|
||||
public class VectorEntity
|
||||
{
|
||||
public ObjectId Id { get; set; }
|
||||
public string Title { get; set; } = "";
|
||||
public float[] Embedding { get; set; } = Array.Empty<float>();
|
||||
}
|
||||
|
||||
public class GeoEntity
|
||||
{
|
||||
public ObjectId Id { get; set; }
|
||||
public string Name { get; set; } = "";
|
||||
public (double Latitude, double Longitude) Location { get; set; }
|
||||
}
|
||||
|
||||
public record OrderId(string Value)
|
||||
{
|
||||
public OrderId() : this(string.Empty) { }
|
||||
}
|
||||
|
||||
public class OrderIdConverter : ValueConverter<OrderId, string>
|
||||
{
|
||||
public override string ConvertToProvider(OrderId model) => model?.Value ?? string.Empty;
|
||||
public override OrderId ConvertFromProvider(string provider) => new OrderId(provider);
|
||||
}
|
||||
|
||||
public class Order
|
||||
{
|
||||
public OrderId Id { get; set; } = null!;
|
||||
public string CustomerName { get; set; } = "";
|
||||
}
|
||||
|
||||
public class TestDocument
|
||||
{
|
||||
public ObjectId Id { get; set; }
|
||||
public string Category { get; set; } = string.Empty;
|
||||
public int Amount { get; set; }
|
||||
public string Name { get; set; } = string.Empty;
|
||||
}
|
||||
|
||||
public class OrderDocument
|
||||
{
|
||||
public ObjectId Id { get; set; }
|
||||
public string ItemName { get; set; } = string.Empty;
|
||||
public int Quantity { get; set; }
|
||||
}
|
||||
|
||||
public class OrderItem
|
||||
{
|
||||
public string Name { get; set; } = string.Empty;
|
||||
public int Price { get; set; }
|
||||
}
|
||||
|
||||
public class ComplexDocument
|
||||
{
|
||||
public ObjectId Id { get; set; }
|
||||
public string Title { get; set; } = string.Empty;
|
||||
public Address ShippingAddress { get; set; } = new();
|
||||
public List<OrderItem> Items { get; set; } = new();
|
||||
}
|
||||
|
||||
[Table("custom_users", Schema = "test")]
|
||||
public class AnnotatedUser
|
||||
{
|
||||
[Key]
|
||||
public ObjectId Id { get; set; }
|
||||
|
||||
[Required]
|
||||
[Column("display_name")]
|
||||
[StringLength(50, MinimumLength = 3)]
|
||||
public string Name { get; set; } = "";
|
||||
|
||||
[Range(0, 150)]
|
||||
public int Age { get; set; }
|
||||
|
||||
[NotMapped]
|
||||
public string ComputedInfo => $"{Name} ({Age})";
|
||||
|
||||
[Column(TypeName = "geopoint")]
|
||||
public (double Lat, double Lon) Location { get; set; }
|
||||
}
|
||||
public class PersonV2
|
||||
{
|
||||
public ObjectId Id { get; set; }
|
||||
public string Name { get; set; } = string.Empty;
|
||||
public int Age { get; set; }
|
||||
}
|
||||
|
||||
/// <summary>
|
||||
/// Entity used to test DbContext inheritance
|
||||
/// </summary>
|
||||
public class ExtendedEntity
|
||||
{
|
||||
public int Id { get; set; }
|
||||
public string Description { get; set; } = string.Empty;
|
||||
public DateTime CreatedAt { get; set; }
|
||||
}
|
||||
|
||||
// ===== SOURCE GENERATOR FEATURE TESTS =====
|
||||
|
||||
/// <summary>
|
||||
/// Base entity with Id property - test inheritance
|
||||
/// </summary>
|
||||
public class BaseEntityWithId
|
||||
{
|
||||
public ObjectId Id { get; set; }
|
||||
public DateTime CreatedAt { get; set; }
|
||||
}
|
||||
|
||||
/// <summary>
|
||||
/// Derived entity that inherits Id from base class
|
||||
/// </summary>
|
||||
public class DerivedEntity : BaseEntityWithId
|
||||
{
|
||||
public string Name { get; set; } = string.Empty;
|
||||
public string Description { get; set; } = string.Empty;
|
||||
}
|
||||
|
||||
/// <summary>
|
||||
/// Entity with computed getter-only properties (should be excluded from serialization)
|
||||
/// </summary>
|
||||
public class EntityWithComputedProperties
|
||||
{
|
||||
public ObjectId Id { get; set; }
|
||||
public string FirstName { get; set; } = string.Empty;
|
||||
public string LastName { get; set; } = string.Empty;
|
||||
public int BirthYear { get; set; }
|
||||
|
||||
// Computed properties - should NOT be serialized
|
||||
public string FullName => $"{FirstName} {LastName}";
|
||||
public int Age => DateTime.Now.Year - BirthYear;
|
||||
public string DisplayInfo => $"{FullName} (Age: {Age})";
|
||||
}
|
||||
|
||||
/// <summary>
|
||||
/// Entity with advanced collection types (HashSet, ISet, LinkedList, etc.)
|
||||
/// </summary>
|
||||
public class EntityWithAdvancedCollections
|
||||
{
|
||||
public ObjectId Id { get; set; }
|
||||
public string Name { get; set; } = string.Empty;
|
||||
|
||||
// Various collection types that should all be recognized
|
||||
public HashSet<string> Tags { get; set; } = new();
|
||||
public ISet<int> Numbers { get; set; } = new HashSet<int>();
|
||||
public LinkedList<string> History { get; set; } = new();
|
||||
public Queue<string> PendingItems { get; set; } = new();
|
||||
public Stack<string> UndoStack { get; set; } = new();
|
||||
|
||||
// Nested objects in collections
|
||||
public HashSet<Address> Addresses { get; set; } = new();
|
||||
public ISet<City> FavoriteCities { get; set; } = new HashSet<City>();
|
||||
}
|
||||
|
||||
/// <summary>
|
||||
/// Entity with private setters (requires reflection-based deserialization)
|
||||
/// </summary>
|
||||
public class EntityWithPrivateSetters
|
||||
{
|
||||
public ObjectId Id { get; private set; }
|
||||
public string Name { get; private set; } = string.Empty;
|
||||
public int Age { get; private set; }
|
||||
public DateTime CreatedAt { get; private set; }
|
||||
|
||||
// Factory method for creation
|
||||
public static EntityWithPrivateSetters Create(string name, int age)
|
||||
{
|
||||
return new EntityWithPrivateSetters
|
||||
{
|
||||
Id = ObjectId.NewObjectId(),
|
||||
Name = name,
|
||||
Age = age,
|
||||
CreatedAt = DateTime.UtcNow
|
||||
};
|
||||
}
|
||||
}
|
||||
|
||||
/// <summary>
|
||||
/// Entity with init-only setters (can use object initializer)
|
||||
/// </summary>
|
||||
public class EntityWithInitSetters
|
||||
{
|
||||
public ObjectId Id { get; init; }
|
||||
public required string Name { get; init; }
|
||||
public int Age { get; init; }
|
||||
public DateTime CreatedAt { get; init; }
|
||||
}
|
||||
|
||||
// ========================================
|
||||
// Circular Reference Test Entities
|
||||
// ========================================
|
||||
|
||||
/// <summary>
|
||||
/// Employee with self-referencing via ObjectIds (organizational hierarchy)
|
||||
/// Tests: self-reference using referencing (BEST PRACTICE)
|
||||
/// Recommended: Avoids embedding which can lead to large/circular documents
|
||||
/// </summary>
|
||||
public class Employee
|
||||
{
|
||||
public ObjectId Id { get; set; }
|
||||
public string Name { get; set; } = string.Empty;
|
||||
public string Department { get; set; } = string.Empty;
|
||||
public ObjectId? ManagerId { get; set; } // Reference to manager
|
||||
public List<ObjectId>? DirectReportIds { get; set; } // References to direct reports (best practice)
|
||||
}
|
||||
|
||||
/// <summary>
|
||||
/// Category with referenced products (N-N using ObjectId references)
|
||||
/// Tests: N-N relationships using referencing (BEST PRACTICE for document databases)
|
||||
/// Recommended: Avoids large documents, better for queries and updates
|
||||
/// </summary>
|
||||
public class CategoryRef
|
||||
{
|
||||
public ObjectId Id { get; set; }
|
||||
public string Name { get; set; } = string.Empty;
|
||||
public string Description { get; set; } = string.Empty;
|
||||
public List<ObjectId>? ProductIds { get; set; } // Only IDs - no embedding
|
||||
}
|
||||
|
||||
/// <summary>
|
||||
/// Product with referenced categories (N-N using ObjectId references)
|
||||
/// Tests: N-N relationships using referencing (BEST PRACTICE for document databases)
|
||||
/// Recommended: Avoids large documents, better for queries and updates
|
||||
/// </summary>
|
||||
public class ProductRef
|
||||
{
|
||||
public ObjectId Id { get; set; }
|
||||
public string Name { get; set; } = string.Empty;
|
||||
public decimal Price { get; set; }
|
||||
public List<ObjectId>? CategoryIds { get; set; } // Only IDs - no embedding
|
||||
}
|
||||
|
||||
// ========================================
|
||||
// Nullable String Key Test (UuidEntity scenario)
|
||||
// ========================================
|
||||
|
||||
/// <summary>
|
||||
/// Base entity class that simulates CleanCore's BaseEntity{TId, TEntity}
|
||||
/// This is the root of the hierarchy that causes the generator bug
|
||||
/// </summary>
|
||||
public abstract class MockBaseEntity<TId, TEntity>
|
||||
where TId : IEquatable<TId>
|
||||
where TEntity : class
|
||||
{
|
||||
[System.ComponentModel.DataAnnotations.Key]
|
||||
public virtual TId? Id { get; set; }
|
||||
|
||||
protected MockBaseEntity() { }
|
||||
|
||||
protected MockBaseEntity(TId? id)
|
||||
{
|
||||
Id = id;
|
||||
}
|
||||
}
|
||||
|
||||
/// <summary>
|
||||
/// Simulates CleanCore's UuidEntity{TEntity} which inherits from BaseEntity{string, TEntity}
|
||||
/// Tests the bug where generator incorrectly chooses ObjectIdMapperBase instead of StringMapperBase
|
||||
/// when the Id property is inherited and nullable
|
||||
/// </summary>
|
||||
public abstract class MockUuidEntity<TEntity> : MockBaseEntity<string, TEntity>
|
||||
where TEntity : class
|
||||
{
|
||||
protected MockUuidEntity() : base() { }
|
||||
|
||||
protected MockUuidEntity(string? id) : base(id) { }
|
||||
}
|
||||
|
||||
/// <summary>
|
||||
/// Concrete entity that inherits from MockUuidEntity, simulating Counter from CleanCore
|
||||
/// This is the actual entity that will be stored in the collection
|
||||
/// </summary>
|
||||
public class MockCounter : MockUuidEntity<MockCounter>
|
||||
{
|
||||
public MockCounter() : base() { }
|
||||
|
||||
public MockCounter(string? id) : base(id) { }
|
||||
|
||||
public string Name { get; set; } = string.Empty;
|
||||
public int Value { get; set; }
|
||||
}
|
||||
|
||||
/// <summary>
|
||||
/// Entity for testing temporal types: DateTimeOffset, TimeSpan, DateOnly, TimeOnly
|
||||
/// </summary>
|
||||
public class TemporalEntity
|
||||
{
|
||||
[Key]
|
||||
public ObjectId Id { get; set; }
|
||||
|
||||
public string Name { get; set; } = string.Empty;
|
||||
|
||||
// DateTime types
|
||||
public DateTime CreatedAt { get; set; }
|
||||
public DateTimeOffset UpdatedAt { get; set; }
|
||||
public DateTimeOffset? LastAccessedAt { get; set; }
|
||||
|
||||
// TimeSpan
|
||||
public TimeSpan Duration { get; set; }
|
||||
public TimeSpan? OptionalDuration { get; set; }
|
||||
|
||||
// DateOnly and TimeOnly (.NET 6+)
|
||||
public DateOnly BirthDate { get; set; }
|
||||
public DateOnly? Anniversary { get; set; }
|
||||
|
||||
public TimeOnly OpeningTime { get; set; }
|
||||
public TimeOnly? ClosingTime { get; set; }
|
||||
}
|
||||
}
|
||||
146
tests/CBDD.Tests/NullableStringIdTests.cs
Executable file
146
tests/CBDD.Tests/NullableStringIdTests.cs
Executable file
@@ -0,0 +1,146 @@
|
||||
using ZB.MOM.WW.CBDD.Shared;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests
|
||||
{
|
||||
/// <summary>
|
||||
/// Tests for entities with nullable string Id (like UuidEntity scenario from CleanCore)
|
||||
/// This reproduces the bug where the generator incorrectly chose ObjectIdMapperBase
|
||||
/// instead of StringMapperBase for inherited nullable string Id properties
|
||||
/// </summary>
|
||||
public class NullableStringIdTests : System.IDisposable
|
||||
{
|
||||
private const string DbPath = "nullable_string_id.db";
|
||||
|
||||
public NullableStringIdTests()
|
||||
{
|
||||
if (File.Exists(DbPath)) File.Delete(DbPath);
|
||||
}
|
||||
|
||||
public void Dispose()
|
||||
{
|
||||
if (File.Exists(DbPath)) File.Delete(DbPath);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void MockCounter_Collection_IsInitialized()
|
||||
{
|
||||
using var db = new Shared.TestDbContext(DbPath);
|
||||
|
||||
// Verify Collection is not null (initialized by generated method)
|
||||
db.MockCounters.ShouldNotBeNull();
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void MockCounter_Insert_And_FindById_Works()
|
||||
{
|
||||
using var db = new Shared.TestDbContext(DbPath);
|
||||
|
||||
var counter = new MockCounter("test-id-123")
|
||||
{
|
||||
Name = "TestCounter",
|
||||
Value = 42
|
||||
};
|
||||
|
||||
// Insert should work with string Id
|
||||
db.MockCounters.Insert(counter);
|
||||
|
||||
// FindById should retrieve the entity
|
||||
var stored = db.MockCounters.FindById("test-id-123");
|
||||
stored.ShouldNotBeNull();
|
||||
stored.Id.ShouldBe("test-id-123");
|
||||
stored.Name.ShouldBe("TestCounter");
|
||||
stored.Value.ShouldBe(42);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void MockCounter_Update_Works()
|
||||
{
|
||||
using var db = new Shared.TestDbContext(DbPath);
|
||||
|
||||
var counter = new MockCounter("update-test")
|
||||
{
|
||||
Name = "Original",
|
||||
Value = 10
|
||||
};
|
||||
|
||||
db.MockCounters.Insert(counter);
|
||||
|
||||
// Update the entity
|
||||
counter.Name = "Updated";
|
||||
counter.Value = 20;
|
||||
db.MockCounters.Update(counter);
|
||||
|
||||
// Verify update
|
||||
var updated = db.MockCounters.FindById("update-test");
|
||||
updated.ShouldNotBeNull();
|
||||
updated.Name.ShouldBe("Updated");
|
||||
updated.Value.ShouldBe(20);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void MockCounter_Delete_Works()
|
||||
{
|
||||
using var db = new Shared.TestDbContext(DbPath);
|
||||
|
||||
var counter = new MockCounter("delete-test")
|
||||
{
|
||||
Name = "ToDelete",
|
||||
Value = 99
|
||||
};
|
||||
|
||||
db.MockCounters.Insert(counter);
|
||||
db.MockCounters.FindById("delete-test").ShouldNotBeNull();
|
||||
|
||||
// Delete the entity
|
||||
db.MockCounters.Delete("delete-test");
|
||||
|
||||
// Verify deletion
|
||||
var deleted = db.MockCounters.FindById("delete-test");
|
||||
deleted.ShouldBeNull();
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void MockCounter_Query_Works()
|
||||
{
|
||||
using var db = new Shared.TestDbContext(DbPath);
|
||||
|
||||
db.MockCounters.Insert(new MockCounter("q1") { Name = "First", Value = 100 });
|
||||
db.MockCounters.Insert(new MockCounter("q2") { Name = "Second", Value = 200 });
|
||||
db.MockCounters.Insert(new MockCounter("q3") { Name = "Third", Value = 150 });
|
||||
|
||||
// Query all
|
||||
var all = db.MockCounters.AsQueryable().ToList();
|
||||
all.Count.ShouldBe(3);
|
||||
|
||||
// Query with condition
|
||||
var highValues = db.MockCounters.AsQueryable()
|
||||
.Where(c => c.Value > 150)
|
||||
.ToList();
|
||||
|
||||
highValues.Count().ShouldBe(1);
|
||||
highValues[0].Name.ShouldBe("Second");
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void MockCounter_InheritedId_IsStoredCorrectly()
|
||||
{
|
||||
using var db = new Shared.TestDbContext(DbPath);
|
||||
|
||||
// Test that the inherited nullable string Id from MockBaseEntity works correctly
|
||||
var counter = new MockCounter("inherited-id-test")
|
||||
{
|
||||
Name = "Inherited",
|
||||
Value = 777
|
||||
};
|
||||
|
||||
db.MockCounters.Insert(counter);
|
||||
|
||||
var stored = db.MockCounters.FindById("inherited-id-test");
|
||||
stored.ShouldNotBeNull();
|
||||
|
||||
// Verify the Id is correctly stored and retrieved through inheritance
|
||||
stored.Id.ShouldBe("inherited-id-test");
|
||||
stored.Id.ShouldBeOfType<string>();
|
||||
}
|
||||
}
|
||||
}
|
||||
52
tests/CBDD.Tests/ObjectIdTests.cs
Executable file
52
tests/CBDD.Tests/ObjectIdTests.cs
Executable file
@@ -0,0 +1,52 @@
|
||||
using ZB.MOM.WW.CBDD.Bson;
|
||||
using Xunit;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests;
|
||||
|
||||
public class ObjectIdTests
|
||||
{
|
||||
[Fact]
|
||||
public void NewObjectId_ShouldCreate12ByteId()
|
||||
{
|
||||
var oid = ObjectId.NewObjectId();
|
||||
|
||||
Span<byte> bytes = stackalloc byte[12];
|
||||
oid.WriteTo(bytes);
|
||||
|
||||
bytes.Length.ShouldBe(12);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void ObjectId_ShouldRoundTrip()
|
||||
{
|
||||
var original = ObjectId.NewObjectId();
|
||||
|
||||
Span<byte> bytes = stackalloc byte[12];
|
||||
original.WriteTo(bytes);
|
||||
|
||||
var restored = new ObjectId(bytes);
|
||||
|
||||
restored.ShouldBe(original);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void ObjectId_Equals_ShouldWork()
|
||||
{
|
||||
var oid1 = ObjectId.NewObjectId();
|
||||
var oid2 = oid1;
|
||||
var oid3 = ObjectId.NewObjectId();
|
||||
|
||||
oid2.ShouldBe(oid1);
|
||||
oid3.ShouldNotBe(oid1);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void ObjectId_Timestamp_ShouldBeRecentUtc()
|
||||
{
|
||||
var oid = ObjectId.NewObjectId();
|
||||
var timestamp = oid.Timestamp;
|
||||
|
||||
(timestamp <= DateTime.UtcNow).ShouldBeTrue();
|
||||
(timestamp >= DateTime.UtcNow.AddSeconds(-5)).ShouldBeTrue();
|
||||
}
|
||||
}
|
||||
116
tests/CBDD.Tests/PrimaryKeyTests.cs
Executable file
116
tests/CBDD.Tests/PrimaryKeyTests.cs
Executable file
@@ -0,0 +1,116 @@
|
||||
using ZB.MOM.WW.CBDD.Bson;
|
||||
using ZB.MOM.WW.CBDD.Core;
|
||||
using ZB.MOM.WW.CBDD.Core.Collections;
|
||||
using ZB.MOM.WW.CBDD.Core.Metadata;
|
||||
using ZB.MOM.WW.CBDD.Shared;
|
||||
using System;
|
||||
using System.Buffers;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests;
|
||||
|
||||
public class PrimaryKeyTests : IDisposable
|
||||
{
|
||||
private readonly string _dbPath = "primary_key_tests.db";
|
||||
|
||||
public PrimaryKeyTests()
|
||||
{
|
||||
if (File.Exists(_dbPath)) File.Delete(_dbPath);
|
||||
}
|
||||
|
||||
public void Dispose()
|
||||
{
|
||||
if (File.Exists(_dbPath)) File.Delete(_dbPath);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Test_Int_PrimaryKey()
|
||||
{
|
||||
using var db = new Shared.TestDbContext(_dbPath);
|
||||
|
||||
var entity = new IntEntity { Id = 1, Name = "Test 1" };
|
||||
db.IntEntities.Insert(entity);
|
||||
db.SaveChanges();
|
||||
|
||||
var retrieved = db.IntEntities.FindById(1);
|
||||
retrieved.ShouldNotBeNull();
|
||||
retrieved.Id.ShouldBe(1);
|
||||
retrieved.Name.ShouldBe("Test 1");
|
||||
|
||||
entity.Name = "Updated";
|
||||
db.IntEntities.Update(entity);
|
||||
|
||||
retrieved = db.IntEntities.FindById(1);
|
||||
retrieved?.Name.ShouldBe("Updated");
|
||||
|
||||
db.IntEntities.Delete(1);
|
||||
db.IntEntities.FindById(1).ShouldBeNull();
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Test_String_PrimaryKey()
|
||||
{
|
||||
using var db = new Shared.TestDbContext(_dbPath);
|
||||
|
||||
var entity = new StringEntity { Id = "key1", Value = "Value 1" };
|
||||
db.StringEntities.Insert(entity);
|
||||
db.SaveChanges();
|
||||
|
||||
var retrieved = db.StringEntities.FindById("key1");
|
||||
retrieved.ShouldNotBeNull();
|
||||
retrieved.Id.ShouldBe("key1");
|
||||
retrieved.Value.ShouldBe("Value 1");
|
||||
|
||||
db.StringEntities.Delete("key1");
|
||||
db.SaveChanges();
|
||||
db.StringEntities.FindById("key1").ShouldBeNull();
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Test_Guid_PrimaryKey()
|
||||
{
|
||||
using var db = new Shared.TestDbContext(_dbPath);
|
||||
|
||||
var id = Guid.NewGuid();
|
||||
var entity = new GuidEntity { Id = id, Name = "Guid Test" };
|
||||
db.GuidEntities.Insert(entity);
|
||||
db.SaveChanges();
|
||||
|
||||
var retrieved = db.GuidEntities.FindById(id);
|
||||
retrieved.ShouldNotBeNull();
|
||||
retrieved.Id.ShouldBe(id);
|
||||
|
||||
db.GuidEntities.Delete(id);
|
||||
db.SaveChanges();
|
||||
db.GuidEntities.FindById(id).ShouldBeNull();
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Test_String_PrimaryKey_With_Custom_Name()
|
||||
{
|
||||
// Test entity with string key NOT named "Id" (named "Code" instead)
|
||||
using var db = new Shared.TestDbContext(_dbPath);
|
||||
|
||||
var entity = new CustomKeyEntity { Code = "ABC123", Description = "Test Description" };
|
||||
db.CustomKeyEntities.Insert(entity);
|
||||
db.SaveChanges();
|
||||
|
||||
// Verify retrieval works correctly
|
||||
var retrieved = db.CustomKeyEntities.FindById("ABC123");
|
||||
retrieved.ShouldNotBeNull();
|
||||
retrieved.Code.ShouldBe("ABC123");
|
||||
retrieved.Description.ShouldBe("Test Description");
|
||||
|
||||
// Verify update works
|
||||
entity.Description = "Updated Description";
|
||||
db.CustomKeyEntities.Update(entity);
|
||||
db.SaveChanges();
|
||||
|
||||
retrieved = db.CustomKeyEntities.FindById("ABC123");
|
||||
retrieved?.Description.ShouldBe("Updated Description");
|
||||
|
||||
// Verify delete works
|
||||
db.CustomKeyEntities.Delete("ABC123");
|
||||
db.SaveChanges();
|
||||
db.CustomKeyEntities.FindById("ABC123").ShouldBeNull();
|
||||
}
|
||||
}
|
||||
177
tests/CBDD.Tests/QueryPrimitivesTests.cs
Executable file
177
tests/CBDD.Tests/QueryPrimitivesTests.cs
Executable file
@@ -0,0 +1,177 @@
|
||||
using ZB.MOM.WW.CBDD.Core.Storage;
|
||||
using ZB.MOM.WW.CBDD.Core.Indexing;
|
||||
using ZB.MOM.WW.CBDD.Bson;
|
||||
using Xunit;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests;
|
||||
|
||||
public class QueryPrimitivesTests : IDisposable
|
||||
{
|
||||
private readonly string _testFile;
|
||||
private readonly StorageEngine _storage;
|
||||
private readonly BTreeIndex _index;
|
||||
|
||||
public QueryPrimitivesTests()
|
||||
{
|
||||
_testFile = Path.Combine(Path.GetTempPath(), $"docdb_test_{Guid.NewGuid()}.db");
|
||||
_storage = new StorageEngine(_testFile, PageFileConfig.Default);
|
||||
|
||||
// Initialize simple index
|
||||
var options = IndexOptions.CreateBTree("test");
|
||||
_index = new BTreeIndex(_storage, options);
|
||||
|
||||
SeedData();
|
||||
}
|
||||
|
||||
private void SeedData()
|
||||
{
|
||||
// Insert keys: 10, 20, 30, 40, 50
|
||||
// And strings: "A", "AB", "ABC", "B", "C"
|
||||
|
||||
var txnId = _storage.BeginTransaction().TransactionId;
|
||||
|
||||
Insert(10, txnId);
|
||||
Insert(20, txnId);
|
||||
Insert(30, txnId);
|
||||
Insert(40, txnId);
|
||||
Insert(50, txnId);
|
||||
|
||||
Insert("A", txnId);
|
||||
Insert("AB", txnId);
|
||||
Insert("ABC", txnId);
|
||||
Insert("B", txnId);
|
||||
Insert("C", txnId);
|
||||
|
||||
_storage.CommitTransaction(txnId);
|
||||
}
|
||||
|
||||
private void Insert(dynamic value, ulong txnId)
|
||||
{
|
||||
IndexKey key;
|
||||
if (value is int i) key = IndexKey.Create(i);
|
||||
else if (value is string s) key = IndexKey.Create(s);
|
||||
else throw new ArgumentException();
|
||||
|
||||
_index.Insert(key, new DocumentLocation(1, 1), txnId);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Equal_ShouldFindExactMatch()
|
||||
{
|
||||
var key = IndexKey.Create(30);
|
||||
var result = _index.Equal(key, 0).ToList();
|
||||
|
||||
result.Count().ShouldBe(1);
|
||||
result[0].Key.ShouldBe(key);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Equal_ShouldReturnEmpty_WhenNotFound()
|
||||
{
|
||||
var key = IndexKey.Create(25);
|
||||
var result = _index.Equal(key, 0).ToList();
|
||||
|
||||
result.ShouldBeEmpty();
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void GreaterThan_ShouldReturnMatches()
|
||||
{
|
||||
var key = IndexKey.Create(30);
|
||||
var result = _index.GreaterThan(key, orEqual: false, 0).ToList();
|
||||
|
||||
(result.Count >= 2).ShouldBeTrue();
|
||||
result[0].Key.ShouldBe(IndexKey.Create(40));
|
||||
result[1].Key.ShouldBe(IndexKey.Create(50));
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void GreaterThanOrEqual_ShouldReturnMatches()
|
||||
{
|
||||
var key = IndexKey.Create(30);
|
||||
var result = _index.GreaterThan(key, orEqual: true, 0).ToList();
|
||||
|
||||
(result.Count >= 3).ShouldBeTrue();
|
||||
result[0].Key.ShouldBe(IndexKey.Create(30));
|
||||
result[1].Key.ShouldBe(IndexKey.Create(40));
|
||||
result[2].Key.ShouldBe(IndexKey.Create(50));
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void LessThan_ShouldReturnMatches()
|
||||
{
|
||||
var key = IndexKey.Create(30);
|
||||
var result = _index.LessThan(key, orEqual: false, 0).ToList();
|
||||
|
||||
result.Count.ShouldBe(2); // 20, 10 (Order is backward?)
|
||||
// LessThan yields backward?
|
||||
// Implementation: MovePrev(). So yes, 20 then 10.
|
||||
result[0].Key.ShouldBe(IndexKey.Create(20));
|
||||
result[1].Key.ShouldBe(IndexKey.Create(10));
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Between_ShouldReturnRange()
|
||||
{
|
||||
var start = IndexKey.Create(20);
|
||||
var end = IndexKey.Create(40);
|
||||
var result = _index.Between(start, end, startInclusive: true, endInclusive: true, 0).ToList();
|
||||
|
||||
result.Count.ShouldBe(3); // 20, 30, 40
|
||||
result[0].Key.ShouldBe(IndexKey.Create(20));
|
||||
result[1].Key.ShouldBe(IndexKey.Create(30));
|
||||
result[2].Key.ShouldBe(IndexKey.Create(40));
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void StartsWith_ShouldReturnPrefixMatches()
|
||||
{
|
||||
var result = _index.StartsWith("AB", 0).ToList();
|
||||
|
||||
result.Count.ShouldBe(2); // AB, ABC
|
||||
result[0].Key.ShouldBe(IndexKey.Create("AB"));
|
||||
result[1].Key.ShouldBe(IndexKey.Create("ABC"));
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Like_ShouldSupportWildcards()
|
||||
{
|
||||
// "A%" -> A, AB, ABC
|
||||
var result = _index.Like("A%", 0).ToList();
|
||||
result.Count.ShouldBe(3);
|
||||
|
||||
// "%B%" -> AB, ABC, B
|
||||
var result2 = _index.Like("%B%", 0).ToList();
|
||||
// A (no), AB (yes), ABC (yes), B (yes), C (no)
|
||||
result2.Count.ShouldBe(3); // AB, ABC, B. Wait, order?
|
||||
// Index order: A, AB, ABC, B, C.
|
||||
// AB ok. ABC ok. B ok.
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Like_Underscore_ShouldMatchSingleChar()
|
||||
{
|
||||
// "_B" -> AB (yes), B (no: len 1), ABC (no)
|
||||
var result = _index.Like("_B", 0).ToList();
|
||||
result.Count().ShouldBe(1);
|
||||
result[0].Key.ShouldBe(IndexKey.Create("AB"));
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void In_ShouldReturnSpecificKeys()
|
||||
{
|
||||
var keys = new[] { IndexKey.Create(10), IndexKey.Create(30), IndexKey.Create(50), IndexKey.Create(99) };
|
||||
var result = _index.In(keys, 0).ToList();
|
||||
|
||||
result.Count.ShouldBe(3); // 10, 30, 50. (99 missing)
|
||||
result[0].Key.ShouldBe(IndexKey.Create(10));
|
||||
result[1].Key.ShouldBe(IndexKey.Create(30));
|
||||
result[2].Key.ShouldBe(IndexKey.Create(50));
|
||||
}
|
||||
|
||||
public void Dispose()
|
||||
{
|
||||
_storage.Dispose();
|
||||
File.Delete(_testFile);
|
||||
}
|
||||
}
|
||||
62
tests/CBDD.Tests/RobustnessTests.cs
Executable file
62
tests/CBDD.Tests/RobustnessTests.cs
Executable file
@@ -0,0 +1,62 @@
|
||||
using ZB.MOM.WW.CBDD.Bson;
|
||||
using ZB.MOM.WW.CBDD.Core.Collections;
|
||||
using Xunit;
|
||||
using System.Collections.Generic;
|
||||
using System;
|
||||
using System.Linq;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests;
|
||||
|
||||
public class RobustnessTests
|
||||
{
|
||||
public struct Point
|
||||
{
|
||||
public int X { get; set; }
|
||||
public int Y { get; set; }
|
||||
}
|
||||
|
||||
public class RobustEntity
|
||||
{
|
||||
public List<int?> NullableInts { get; set; } = new();
|
||||
public Dictionary<string, int> Map { get; set; } = new();
|
||||
public IEnumerable<string> EnumerableStrings { get; set; } = Array.Empty<string>();
|
||||
public Point Location { get; set; }
|
||||
public Point? NullableLocation { get; set; }
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void GenerateSchema_RobustnessChecks()
|
||||
{
|
||||
var schema = BsonSchemaGenerator.FromType<RobustEntity>();
|
||||
|
||||
// 1. Nullable Value Types in List
|
||||
var nullableInts = schema.Fields.First(f => f.Name == "nullableints");
|
||||
nullableInts.Type.ShouldBe(BsonType.Array);
|
||||
nullableInts.ArrayItemType.ShouldBe(BsonType.Int32);
|
||||
// Note: Current Schema doesn't capture "ItemIsNullable", but verifying it doesn't crash/return Undefined
|
||||
|
||||
// 2. Dictionary (likely treated as Array of KVPs currently, or Undefined if structs fail)
|
||||
// With current logic: Dictionary implements IEnumerable<KVP>. KVP is struct.
|
||||
// If generator doesn't handle structs as Documents, item type might be Undefined.
|
||||
var map = schema.Fields.First(f => f.Name == "map");
|
||||
map.Type.ShouldBe(BsonType.Array);
|
||||
|
||||
// 3. IEnumerable property
|
||||
var enumerable = schema.Fields.First(f => f.Name == "enumerablestrings");
|
||||
enumerable.Type.ShouldBe(BsonType.Array);
|
||||
enumerable.ArrayItemType.ShouldBe(BsonType.String);
|
||||
|
||||
// 4. Struct
|
||||
var location = schema.Fields.First(f => f.Name == "location");
|
||||
// Structs should be treated as Documents in BSON if not primitive
|
||||
location.Type.ShouldBe(BsonType.Document);
|
||||
location.NestedSchema.ShouldNotBeNull();
|
||||
location.NestedSchema.Fields.ShouldContain(f => f.Name == "x");
|
||||
|
||||
// 5. Nullable Struct
|
||||
var nullableLocation = schema.Fields.First(f => f.Name == "nullablelocation");
|
||||
nullableLocation.Type.ShouldBe(BsonType.Document);
|
||||
nullableLocation.IsNullable.ShouldBeTrue();
|
||||
nullableLocation.NestedSchema.ShouldNotBeNull();
|
||||
}
|
||||
}
|
||||
113
tests/CBDD.Tests/ScanTests.cs
Executable file
113
tests/CBDD.Tests/ScanTests.cs
Executable file
@@ -0,0 +1,113 @@
|
||||
using ZB.MOM.WW.CBDD.Core;
|
||||
using ZB.MOM.WW.CBDD.Core.Storage;
|
||||
using ZB.MOM.WW.CBDD.Core.Collections;
|
||||
using ZB.MOM.WW.CBDD.Core.Indexing;
|
||||
using ZB.MOM.WW.CBDD.Bson;
|
||||
using Xunit;
|
||||
using System;
|
||||
using System.IO;
|
||||
using System.Linq;
|
||||
using System.Collections.Generic;
|
||||
using ZB.MOM.WW.CBDD.Shared;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests
|
||||
{
|
||||
public class ScanTests : IDisposable
|
||||
{
|
||||
private readonly string _testFile;
|
||||
private readonly Shared.TestDbContext _db;
|
||||
|
||||
public ScanTests()
|
||||
{
|
||||
_testFile = Path.Combine(Path.GetTempPath(), $"scan_tests_{Guid.NewGuid()}.db");
|
||||
if (File.Exists(_testFile)) File.Delete(_testFile);
|
||||
var wal = Path.ChangeExtension(_testFile, ".wal");
|
||||
if (File.Exists(wal)) File.Delete(wal);
|
||||
|
||||
_db = new Shared.TestDbContext(_testFile);
|
||||
}
|
||||
|
||||
public void Dispose()
|
||||
{
|
||||
_db.Dispose();
|
||||
if (File.Exists(_testFile)) File.Delete(_testFile);
|
||||
var wal = Path.ChangeExtension(_testFile, ".wal");
|
||||
if (File.Exists(wal)) File.Delete(wal);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Scan_FindsMatchingDocuments()
|
||||
{
|
||||
// Arrange
|
||||
_db.Users.Insert(new User { Name = "Alice", Age = 30 });
|
||||
_db.Users.Insert(new User { Name = "Bob", Age = 25 });
|
||||
_db.Users.Insert(new User { Name = "Charlie", Age = 35 });
|
||||
_db.SaveChanges();
|
||||
|
||||
// Act: Find users older than 28
|
||||
var results = _db.Users.Scan(reader => ParseAge(reader) > 28).ToList();
|
||||
|
||||
// Assert
|
||||
results.Count.ShouldBe(2);
|
||||
results.ShouldContain(d => d.Name == "Alice");
|
||||
results.ShouldContain(d => d.Name == "Charlie");
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Repro_Insert_Loop_Hang()
|
||||
{
|
||||
// Reproduce hang reported by user at 501 documents
|
||||
int count = 600;
|
||||
for (int i = 0; i < count; i++)
|
||||
{
|
||||
_db.Users.Insert(new User { Name = $"User_{i}", Age = i });
|
||||
}
|
||||
_db.SaveChanges();
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void ParallelScan_FindsMatchingDocuments()
|
||||
{
|
||||
// Arrange
|
||||
int count = 1000;
|
||||
for (int i = 0; i < count; i++)
|
||||
{
|
||||
_db.Users.Insert(new User { Name = $"User_{i}", Age = i });
|
||||
}
|
||||
_db.SaveChanges();
|
||||
|
||||
// Act: Find users with Age >= 500
|
||||
// Parallelism 2 to force partitioning
|
||||
var results = _db.Users.ParallelScan(reader => ParseAge(reader) >= 500, degreeOfParallelism: 2).ToList();
|
||||
|
||||
// Assert
|
||||
results.Count.ShouldBe(500);
|
||||
}
|
||||
|
||||
private int ParseAge(BsonSpanReader reader)
|
||||
{
|
||||
try
|
||||
{
|
||||
reader.ReadDocumentSize();
|
||||
while (reader.Remaining > 0)
|
||||
{
|
||||
var type = reader.ReadBsonType();
|
||||
if (type == 0) break; // End of doc
|
||||
|
||||
var name = reader.ReadElementHeader();
|
||||
|
||||
if (name == "age")
|
||||
{
|
||||
return reader.ReadInt32();
|
||||
}
|
||||
else
|
||||
{
|
||||
reader.SkipValue(type);
|
||||
}
|
||||
}
|
||||
}
|
||||
catch { return -1; }
|
||||
return -1;
|
||||
}
|
||||
}
|
||||
}
|
||||
237
tests/CBDD.Tests/SchemaPersistenceTests.cs
Executable file
237
tests/CBDD.Tests/SchemaPersistenceTests.cs
Executable file
@@ -0,0 +1,237 @@
|
||||
using System;
|
||||
using System.IO;
|
||||
using System.Linq;
|
||||
using Xunit;
|
||||
using ZB.MOM.WW.CBDD.Bson;
|
||||
using ZB.MOM.WW.CBDD.Bson.Schema;
|
||||
using ZB.MOM.WW.CBDD.Core.Storage;
|
||||
using ZB.MOM.WW.CBDD.Core.Collections;
|
||||
using ZB.MOM.WW.CBDD.Core.Indexing;
|
||||
using ZB.MOM.WW.CBDD.Shared.TestDbContext_TestDbContext_Mappers;
|
||||
using ZB.MOM.WW.CBDD.Shared;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests;
|
||||
|
||||
public class SchemaPersistenceTests : IDisposable
|
||||
{
|
||||
private readonly string _dbPath;
|
||||
private readonly Shared.TestDbContext _db;
|
||||
|
||||
public SchemaPersistenceTests()
|
||||
{
|
||||
_dbPath = Path.Combine(Path.GetTempPath(), $"schema_test_{Guid.NewGuid()}.db");
|
||||
_db = new Shared.TestDbContext(_dbPath);
|
||||
}
|
||||
|
||||
public void Dispose()
|
||||
{
|
||||
_db.Dispose();
|
||||
if (File.Exists(_dbPath)) File.Delete(_dbPath);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void BsonSchema_Serialization_RoundTrip()
|
||||
{
|
||||
var schema = new BsonSchema
|
||||
{
|
||||
Title = "Person",
|
||||
Fields =
|
||||
{
|
||||
new BsonField { Name = "id", Type = BsonType.ObjectId },
|
||||
new BsonField { Name = "name", Type = BsonType.String, IsNullable = true },
|
||||
new BsonField { Name = "age", Type = BsonType.Int32 },
|
||||
new BsonField
|
||||
{
|
||||
Name = "address",
|
||||
Type = BsonType.Document,
|
||||
NestedSchema = new BsonSchema
|
||||
{
|
||||
Fields =
|
||||
{
|
||||
new BsonField { Name = "city", Type = BsonType.String }
|
||||
}
|
||||
}
|
||||
}
|
||||
}
|
||||
};
|
||||
|
||||
var buffer = new byte[1024];
|
||||
var keyMap = new System.Collections.Concurrent.ConcurrentDictionary<string, ushort>(StringComparer.OrdinalIgnoreCase);
|
||||
var keys = new System.Collections.Concurrent.ConcurrentDictionary<ushort, string>();
|
||||
|
||||
// Manual registration for schema keys
|
||||
ushort id = 1;
|
||||
foreach (var k in new[] { "person", "id", "name", "age", "address", "city", "fields", "title", "type", "isnullable", "nestedschema", "t", "v", "f", "n", "b", "s", "a", "_v", "0", "1", "2", "3", "4", "5" })
|
||||
{
|
||||
keyMap[k] = id;
|
||||
keys[id] = k;
|
||||
id++;
|
||||
}
|
||||
|
||||
var writer = new BsonSpanWriter(buffer, keyMap);
|
||||
schema.ToBson(ref writer);
|
||||
|
||||
var reader = new BsonSpanReader(buffer.AsSpan(0, writer.Position), keys);
|
||||
var roundTrip = BsonSchema.FromBson(ref reader);
|
||||
|
||||
roundTrip.Title.ShouldBe(schema.Title);
|
||||
roundTrip.Fields.Count.ShouldBe(schema.Fields.Count);
|
||||
roundTrip.Fields[0].Name.ShouldBe(schema.Fields[0].Name);
|
||||
roundTrip.Fields[3].NestedSchema!.Fields[0].Name.ShouldBe(schema.Fields[3].NestedSchema!.Fields[0].Name);
|
||||
schema.Equals(roundTrip).ShouldBeTrue();
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void StorageEngine_Collections_Metadata_Persistence()
|
||||
{
|
||||
var meta = new CollectionMetadata
|
||||
{
|
||||
Name = "users",
|
||||
PrimaryRootPageId = 10,
|
||||
SchemaRootPageId = 20
|
||||
};
|
||||
meta.Indexes.Add(new IndexMetadata { Name = "age", IsUnique = false, Type = IndexType.BTree, PropertyPaths = ["Age"] });
|
||||
|
||||
_db.Storage.SaveCollectionMetadata(meta);
|
||||
|
||||
var loaded = _db.Storage.GetCollectionMetadata("users");
|
||||
loaded.ShouldNotBeNull();
|
||||
loaded.Name.ShouldBe(meta.Name);
|
||||
loaded.PrimaryRootPageId.ShouldBe(meta.PrimaryRootPageId);
|
||||
loaded.SchemaRootPageId.ShouldBe(meta.SchemaRootPageId);
|
||||
loaded.Indexes.Count().ShouldBe(1);
|
||||
loaded.Indexes[0].Name.ShouldBe("age");
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void StorageEngine_Schema_Versioning()
|
||||
{
|
||||
var schema1 = new BsonSchema { Title = "V1", Fields = { new BsonField { Name = "f1", Type = BsonType.String } } };
|
||||
var schema2 = new BsonSchema { Title = "V2", Fields = { new BsonField { Name = "f1", Type = BsonType.String }, new BsonField { Name = "f2", Type = BsonType.Int32 } } };
|
||||
|
||||
var rootId = _db.Storage.AppendSchema(0, schema1);
|
||||
rootId.ShouldNotBe(0u);
|
||||
|
||||
var schemas = _db.Storage.GetSchemas(rootId);
|
||||
schemas.Count().ShouldBe(1);
|
||||
schemas[0].Title.ShouldBe("V1");
|
||||
|
||||
var updatedRoot = _db.Storage.AppendSchema(rootId, schema2);
|
||||
updatedRoot.ShouldBe(rootId);
|
||||
|
||||
schemas = _db.Storage.GetSchemas(rootId);
|
||||
schemas.Count.ShouldBe(2, $"Expected 2 schemas but found {schemas.Count}. Titles: {(schemas.Count > 0 ? string.Join(", ", schemas.Select(s => s.Title)) : "None")}");
|
||||
schemas[0].Title.ShouldBe("V1");
|
||||
schemas[1].Title.ShouldBe("V2");
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void DocumentCollection_Integrates_Schema_Versioning_On_Startup()
|
||||
{
|
||||
// Use a dedicated database for this test to avoid schema pollution from _db
|
||||
var testDbPath = Path.Combine(Path.GetTempPath(), $"schema_versioning_test_{Guid.NewGuid()}.db");
|
||||
|
||||
try
|
||||
{
|
||||
var mapper1 = new ZB_MOM_WW_CBDD_Shared_PersonMapper();
|
||||
var schema1 = mapper1.GetSchema();
|
||||
|
||||
// 1. First startup - create DB and initialize Person collection
|
||||
using (var db1 = new Shared.TestDbContext(testDbPath))
|
||||
{
|
||||
// Access only People collection to avoid initializing others
|
||||
var coll = db1.People;
|
||||
var meta = db1.Storage.GetCollectionMetadata("people_collection");
|
||||
meta.ShouldNotBeNull();
|
||||
var schemas = db1.Storage.GetSchemas(meta.SchemaRootPageId);
|
||||
schemas.Count().ShouldBe(1);
|
||||
schema1.Equals(schemas[0]).ShouldBeTrue("Persisted schema 1 should equal current schema 1");
|
||||
|
||||
coll.CurrentSchemaVersion.ShouldNotBeNull();
|
||||
coll.CurrentSchemaVersion!.Value.Version.ShouldBe(1);
|
||||
coll.CurrentSchemaVersion!.Value.Hash.ShouldBe(schema1.GetHash());
|
||||
}
|
||||
|
||||
// 2. Restart with SAME schema (should NOT append)
|
||||
using (var db2 = new Shared.TestDbContext(testDbPath))
|
||||
{
|
||||
var coll = db2.People;
|
||||
var meta = db2.Storage.GetCollectionMetadata("people_collection");
|
||||
var schemas = db2.Storage.GetSchemas(meta!.SchemaRootPageId);
|
||||
schemas.Count().ShouldBe(1); // Still 1
|
||||
|
||||
coll.CurrentSchemaVersion!.Value.Version.ShouldBe(1);
|
||||
coll.CurrentSchemaVersion!.Value.Hash.ShouldBe(schema1.GetHash());
|
||||
}
|
||||
|
||||
// 3. Simulate schema evolution: Person with an additional field
|
||||
// Since we can't change the actual Person class at runtime, this test verifies
|
||||
// that the same schema doesn't get re-appended.
|
||||
// A real-world scenario would involve deploying a new mapper version.
|
||||
using (var db3 = new Shared.TestDbContext(testDbPath))
|
||||
{
|
||||
var coll = db3.People;
|
||||
var meta = db3.Storage.GetCollectionMetadata("people_collection");
|
||||
var schemas = db3.Storage.GetSchemas(meta!.SchemaRootPageId);
|
||||
|
||||
// Schema should still be 1 since we're using the same Person type
|
||||
schemas.Count().ShouldBe(1);
|
||||
schemas[0].Title.ShouldBe("Person");
|
||||
coll.CurrentSchemaVersion!.Value.Version.ShouldBe(1);
|
||||
}
|
||||
}
|
||||
finally
|
||||
{
|
||||
if (File.Exists(testDbPath)) File.Delete(testDbPath);
|
||||
}
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Document_Contains_Schema_Version_Field()
|
||||
{
|
||||
var mapper = new ZB_MOM_WW_CBDD_Shared_PersonMapper();
|
||||
using (var coll = _db.People)
|
||||
{
|
||||
var person = new Person { Name = "John" };
|
||||
var id = coll.Insert(person);
|
||||
_db.SaveChanges();
|
||||
|
||||
coll.Count().ShouldBe(1);
|
||||
coll.CurrentSchemaVersion.ShouldNotBeNull();
|
||||
coll.CurrentSchemaVersion!.Value.Version.ShouldBe(1);
|
||||
|
||||
// Verify that the document in storage contains _v
|
||||
var meta = _db.Storage.GetCollectionMetadata("persons"); // person lowercase
|
||||
// meta.ShouldNotBeNull();
|
||||
|
||||
// Get location from primary index (internal access enabled by InternalsVisibleTo)
|
||||
var key = mapper.ToIndexKey(id);
|
||||
coll._primaryIndex.TryFind(key, out var location, 0).ShouldBeTrue();
|
||||
|
||||
// Read raw bytes from page
|
||||
var pageBuffer = new byte[_db.Storage.PageSize];
|
||||
_db.Storage.ReadPage(location.PageId, 0, pageBuffer);
|
||||
var slotOffset = SlottedPageHeader.Size + (location.SlotIndex * SlotEntry.Size);
|
||||
var slot = SlotEntry.ReadFrom(pageBuffer.AsSpan(slotOffset));
|
||||
var docData = pageBuffer.AsSpan(slot.Offset, slot.Length);
|
||||
|
||||
// Print some info if it fails (using Assert messages)
|
||||
string hex = BitConverter.ToString(docData.ToArray()).Replace("-", "");
|
||||
|
||||
// Look for _v (BsonType.Int32 + 2-byte ID)
|
||||
ushort vId = _db.Storage.GetKeyMap()["_v"];
|
||||
string vIdHex = vId.ToString("X4");
|
||||
// Reverse endian for hex string check (ushort is LE)
|
||||
string vIdHexLE = vIdHex.Substring(2, 2) + vIdHex.Substring(0, 2);
|
||||
string pattern = "10" + vIdHexLE;
|
||||
|
||||
bool found = hex.Contains(pattern);
|
||||
found.ShouldBeTrue($"Document should contain _v field ({pattern}). Raw BSON: {hex}");
|
||||
|
||||
// Verify the value (1) follows the key
|
||||
int index = hex.IndexOf(pattern);
|
||||
string valueHex = hex.Substring(index + 6, 8); // 4 bytes = 8 hex chars (pattern is 6 hex chars: 10 + ID_LE)
|
||||
valueHex.ShouldBe("01000000");
|
||||
}
|
||||
}
|
||||
}
|
||||
57
tests/CBDD.Tests/SchemaTests.cs
Executable file
57
tests/CBDD.Tests/SchemaTests.cs
Executable file
@@ -0,0 +1,57 @@
|
||||
using ZB.MOM.WW.CBDD.Bson;
|
||||
using ZB.MOM.WW.CBDD.Core.Collections;
|
||||
using ZB.MOM.WW.CBDD.Core.Indexing;
|
||||
using ZB.MOM.WW.CBDD.Shared.TestDbContext_TestDbContext_Mappers;
|
||||
using System.Text;
|
||||
using Xunit;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests;
|
||||
|
||||
public class SchemaTests
|
||||
{
|
||||
private static readonly System.Collections.Concurrent.ConcurrentDictionary<string, ushort> _testKeyMap = new(StringComparer.OrdinalIgnoreCase);
|
||||
static SchemaTests()
|
||||
{
|
||||
ushort id = 1;
|
||||
foreach (var k in new[] { "_id", "name", "mainaddress", "otheraddresses", "tags", "secret", "street", "city" }) _testKeyMap[k] = id++;
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void UsedKeys_ShouldReturnAllKeys()
|
||||
{
|
||||
var mapper = new ZB_MOM_WW_CBDD_Shared_ComplexUserMapper();
|
||||
var keys = mapper.UsedKeys.ToList();
|
||||
|
||||
keys.ShouldContain("_id");
|
||||
keys.ShouldContain("name");
|
||||
keys.ShouldContain("mainaddress");
|
||||
keys.ShouldContain("otheraddresses");
|
||||
keys.ShouldContain("tags");
|
||||
keys.ShouldContain("secret");
|
||||
keys.ShouldContain("street");
|
||||
keys.ShouldContain("city");
|
||||
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void GetSchema_ShouldReturnBsonSchema()
|
||||
{
|
||||
var mapper = new ZB_MOM_WW_CBDD_Shared_ComplexUserMapper();
|
||||
var schema = mapper.GetSchema();
|
||||
|
||||
var idField = schema.Fields.FirstOrDefault(f => f.Name == "_id");
|
||||
idField.ShouldNotBeNull();
|
||||
idField.Type.ShouldBe(BsonType.ObjectId);
|
||||
|
||||
var nameField = schema.Fields.FirstOrDefault(f => f.Name == "name");
|
||||
nameField.ShouldNotBeNull();
|
||||
nameField.Type.ShouldBe(BsonType.String);
|
||||
|
||||
var addressField = schema.Fields.FirstOrDefault(f => f.Name == "mainaddress");
|
||||
addressField.ShouldNotBeNull();
|
||||
addressField.Type.ShouldBe(BsonType.Document);
|
||||
addressField.NestedSchema.ShouldNotBeNull();
|
||||
// Address in MockEntities has City (Nested)
|
||||
addressField.NestedSchema.Fields.ShouldContain(f => f.Name == "city");
|
||||
}
|
||||
}
|
||||
229
tests/CBDD.Tests/SetMethodTests.cs
Executable file
229
tests/CBDD.Tests/SetMethodTests.cs
Executable file
@@ -0,0 +1,229 @@
|
||||
using ZB.MOM.WW.CBDD.Bson;
|
||||
using ZB.MOM.WW.CBDD.Core.Collections;
|
||||
using ZB.MOM.WW.CBDD.Shared;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests;
|
||||
|
||||
public class SetMethodTests : IDisposable
|
||||
{
|
||||
private readonly string _dbPath;
|
||||
private readonly Shared.TestDbContext _db;
|
||||
|
||||
public SetMethodTests()
|
||||
{
|
||||
_dbPath = Path.Combine(Path.GetTempPath(), $"cbdd_set_{Guid.NewGuid()}.db");
|
||||
_db = new Shared.TestDbContext(_dbPath);
|
||||
}
|
||||
|
||||
public void Dispose()
|
||||
{
|
||||
_db.Dispose();
|
||||
if (File.Exists(_dbPath)) File.Delete(_dbPath);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Set_ObjectId_ReturnsCorrectCollection()
|
||||
{
|
||||
var collection = _db.Set<ObjectId, User>();
|
||||
collection.ShouldNotBeNull();
|
||||
collection.ShouldBeSameAs(_db.Users);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Set_Shorthand_ReturnsCorrectCollection()
|
||||
{
|
||||
var collection = _db.Set<User>();
|
||||
collection.ShouldNotBeNull();
|
||||
collection.ShouldBeSameAs(_db.Users);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Set_Int_ReturnsCorrectCollection()
|
||||
{
|
||||
var collection = _db.Set<int, Person>();
|
||||
collection.ShouldNotBeNull();
|
||||
collection.ShouldBeSameAs(_db.People);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Set_String_ReturnsCorrectCollection()
|
||||
{
|
||||
var collection = _db.Set<string, StringEntity>();
|
||||
collection.ShouldNotBeNull();
|
||||
collection.ShouldBeSameAs(_db.StringEntities);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Set_Guid_ReturnsCorrectCollection()
|
||||
{
|
||||
var collection = _db.Set<Guid, GuidEntity>();
|
||||
collection.ShouldNotBeNull();
|
||||
collection.ShouldBeSameAs(_db.GuidEntities);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Set_CustomKey_ReturnsCorrectCollection()
|
||||
{
|
||||
var collection = _db.Set<OrderId, Order>();
|
||||
collection.ShouldNotBeNull();
|
||||
collection.ShouldBeSameAs(_db.Orders);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Set_AllObjectIdCollections_ReturnCorrectInstances()
|
||||
{
|
||||
_db.Set<ObjectId, AnnotatedUser>().ShouldBeSameAs(_db.AnnotatedUsers);
|
||||
_db.Set<ObjectId, ComplexUser>().ShouldBeSameAs(_db.ComplexUsers);
|
||||
_db.Set<ObjectId, TestDocument>().ShouldBeSameAs(_db.TestDocuments);
|
||||
_db.Set<ObjectId, OrderDocument>().ShouldBeSameAs(_db.OrderDocuments);
|
||||
_db.Set<ObjectId, ComplexDocument>().ShouldBeSameAs(_db.ComplexDocuments);
|
||||
_db.Set<ObjectId, PersonV2>().ShouldBeSameAs(_db.PeopleV2);
|
||||
_db.Set<ObjectId, VectorEntity>().ShouldBeSameAs(_db.VectorItems);
|
||||
_db.Set<ObjectId, GeoEntity>().ShouldBeSameAs(_db.GeoItems);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Set_AllIntCollections_ReturnCorrectInstances()
|
||||
{
|
||||
_db.Set<int, AutoInitEntity>().ShouldBeSameAs(_db.AutoInitEntities);
|
||||
_db.Set<int, Product>().ShouldBeSameAs(_db.Products);
|
||||
_db.Set<int, IntEntity>().ShouldBeSameAs(_db.IntEntities);
|
||||
_db.Set<int, AsyncDoc>().ShouldBeSameAs(_db.AsyncDocs);
|
||||
_db.Set<int, SchemaUser>().ShouldBeSameAs(_db.SchemaUsers);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Set_StringKeyCollections_ReturnCorrectInstances()
|
||||
{
|
||||
_db.Set<string, CustomKeyEntity>().ShouldBeSameAs(_db.CustomKeyEntities);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Set_UnregisteredEntity_ThrowsInvalidOperationException()
|
||||
{
|
||||
Should.Throw<InvalidOperationException>(() => _db.Set<ObjectId, Address>());
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Set_WrongKeyType_ThrowsInvalidOperationException()
|
||||
{
|
||||
Should.Throw<InvalidOperationException>(() => _db.Set<string, User>());
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Set_CanPerformOperations()
|
||||
{
|
||||
var users = _db.Set<User>();
|
||||
|
||||
var user = new User { Name = "Alice", Age = 30 };
|
||||
var id = users.Insert(user);
|
||||
|
||||
var found = users.FindById(id);
|
||||
found.ShouldNotBeNull();
|
||||
found.Name.ShouldBe("Alice");
|
||||
found.Age.ShouldBe(30);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Set_WithIntKey_CanPerformOperations()
|
||||
{
|
||||
var products = _db.Set<int, Product>();
|
||||
|
||||
var product = new Product { Id = 1, Title = "Widget", Price = 9.99m };
|
||||
products.Insert(product);
|
||||
|
||||
var found = products.FindById(1);
|
||||
found.ShouldNotBeNull();
|
||||
found.Title.ShouldBe("Widget");
|
||||
found.Price.ShouldBe(9.99m);
|
||||
}
|
||||
}
|
||||
|
||||
public class SetMethodInheritanceTests : IDisposable
|
||||
{
|
||||
private readonly string _dbPath;
|
||||
private readonly Shared.TestExtendedDbContext _db;
|
||||
|
||||
public SetMethodInheritanceTests()
|
||||
{
|
||||
_dbPath = Path.Combine(Path.GetTempPath(), $"cbdd_set_inherit_{Guid.NewGuid()}.db");
|
||||
_db = new Shared.TestExtendedDbContext(_dbPath);
|
||||
}
|
||||
|
||||
public void Dispose()
|
||||
{
|
||||
_db.Dispose();
|
||||
if (File.Exists(_dbPath)) File.Delete(_dbPath);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Set_OwnCollection_ReturnsCorrectInstance()
|
||||
{
|
||||
var collection = _db.Set<int, ExtendedEntity>();
|
||||
collection.ShouldNotBeNull();
|
||||
collection.ShouldBeSameAs(_db.ExtendedEntities);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Set_ParentCollection_ReturnsCorrectInstance()
|
||||
{
|
||||
var collection = _db.Set<ObjectId, User>();
|
||||
collection.ShouldNotBeNull();
|
||||
collection.ShouldBeSameAs(_db.Users);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Set_ParentShorthand_ReturnsCorrectInstance()
|
||||
{
|
||||
var collection = _db.Set<User>();
|
||||
collection.ShouldNotBeNull();
|
||||
collection.ShouldBeSameAs(_db.Users);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Set_ParentIntCollection_ReturnsCorrectInstance()
|
||||
{
|
||||
_db.Set<int, Person>().ShouldBeSameAs(_db.People);
|
||||
_db.Set<int, Product>().ShouldBeSameAs(_db.Products);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Set_ParentCustomKey_ReturnsCorrectInstance()
|
||||
{
|
||||
var collection = _db.Set<OrderId, Order>();
|
||||
collection.ShouldNotBeNull();
|
||||
collection.ShouldBeSameAs(_db.Orders);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Set_UnregisteredEntity_ThrowsInvalidOperationException()
|
||||
{
|
||||
Should.Throw<InvalidOperationException>(() => _db.Set<ObjectId, Address>());
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Set_OwnCollection_CanPerformOperations()
|
||||
{
|
||||
var entities = _db.Set<int, ExtendedEntity>();
|
||||
|
||||
var entity = new ExtendedEntity { Id = 1, Description = "Test", CreatedAt = DateTime.UtcNow };
|
||||
entities.Insert(entity);
|
||||
|
||||
var found = entities.FindById(1);
|
||||
found.ShouldNotBeNull();
|
||||
found.Description.ShouldBe("Test");
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Set_ParentCollection_CanPerformOperations()
|
||||
{
|
||||
var users = _db.Set<User>();
|
||||
|
||||
var user = new User { Name = "Bob", Age = 25 };
|
||||
var id = users.Insert(user);
|
||||
|
||||
var found = users.FindById(id);
|
||||
found.ShouldNotBeNull();
|
||||
found.Name.ShouldBe("Bob");
|
||||
}
|
||||
}
|
||||
538
tests/CBDD.Tests/SourceGeneratorFeaturesTests.cs
Executable file
538
tests/CBDD.Tests/SourceGeneratorFeaturesTests.cs
Executable file
@@ -0,0 +1,538 @@
|
||||
using ZB.MOM.WW.CBDD.Bson;
|
||||
using ZB.MOM.WW.CBDD.Shared;
|
||||
using System.Linq;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests;
|
||||
|
||||
/// <summary>
|
||||
/// Tests for Source Generator enhancements:
|
||||
/// 1. Property inheritance from base classes (including Id)
|
||||
/// 2. Exclusion of computed getter-only properties
|
||||
/// 3. Recognition of advanced collection types (HashSet, ISet, LinkedList, etc.)
|
||||
/// </summary>
|
||||
public class SourceGeneratorFeaturesTests : IDisposable
|
||||
{
|
||||
private readonly string _dbPath;
|
||||
private readonly string _walPath;
|
||||
private readonly Shared.TestDbContext _db;
|
||||
|
||||
public SourceGeneratorFeaturesTests()
|
||||
{
|
||||
_dbPath = Path.Combine(Path.GetTempPath(), $"test_sg_features_{Guid.NewGuid()}.db");
|
||||
_walPath = Path.Combine(Path.GetTempPath(), $"test_sg_features_{Guid.NewGuid()}.wal");
|
||||
|
||||
_db = new Shared.TestDbContext(_dbPath);
|
||||
}
|
||||
|
||||
#region Inheritance Tests
|
||||
|
||||
[Fact]
|
||||
public void DerivedEntity_InheritsId_FromBaseClass()
|
||||
{
|
||||
// Arrange
|
||||
var entity = new DerivedEntity
|
||||
{
|
||||
Name = "Test Entity",
|
||||
Description = "Testing inheritance",
|
||||
CreatedAt = DateTime.UtcNow
|
||||
};
|
||||
|
||||
// Act
|
||||
var id = _db.DerivedEntities.Insert(entity);
|
||||
_db.SaveChanges();
|
||||
var retrieved = _db.DerivedEntities.FindById(id);
|
||||
|
||||
// Assert
|
||||
retrieved.ShouldNotBeNull();
|
||||
retrieved.Id.ShouldBe(id); // Id from base class should work
|
||||
retrieved.Name.ShouldBe("Test Entity");
|
||||
retrieved.Description.ShouldBe("Testing inheritance");
|
||||
retrieved.CreatedAt.Date.ShouldBe(entity.CreatedAt.Date); // Compare just date part
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void DerivedEntity_Update_WorksWithInheritedId()
|
||||
{
|
||||
// Arrange
|
||||
var entity = new DerivedEntity
|
||||
{
|
||||
Name = "Original",
|
||||
Description = "Original Description",
|
||||
CreatedAt = DateTime.UtcNow
|
||||
};
|
||||
var id = _db.DerivedEntities.Insert(entity);
|
||||
_db.SaveChanges();
|
||||
|
||||
// Act
|
||||
var retrieved = _db.DerivedEntities.FindById(id);
|
||||
retrieved.ShouldNotBeNull();
|
||||
retrieved.Name = "Updated";
|
||||
retrieved.Description = "Updated Description";
|
||||
_db.DerivedEntities.Update(retrieved);
|
||||
_db.SaveChanges();
|
||||
|
||||
var updated = _db.DerivedEntities.FindById(id);
|
||||
|
||||
// Assert
|
||||
updated.ShouldNotBeNull();
|
||||
updated.Id.ShouldBe(id);
|
||||
updated.Name.ShouldBe("Updated");
|
||||
updated.Description.ShouldBe("Updated Description");
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void DerivedEntity_Query_WorksWithInheritedProperties()
|
||||
{
|
||||
// Arrange
|
||||
var now = DateTime.UtcNow;
|
||||
var entity1 = new DerivedEntity { Name = "Entity1", CreatedAt = now.AddDays(-2) };
|
||||
var entity2 = new DerivedEntity { Name = "Entity2", CreatedAt = now.AddDays(-1) };
|
||||
var entity3 = new DerivedEntity { Name = "Entity3", CreatedAt = now };
|
||||
|
||||
_db.DerivedEntities.Insert(entity1);
|
||||
_db.DerivedEntities.Insert(entity2);
|
||||
_db.DerivedEntities.Insert(entity3);
|
||||
_db.SaveChanges();
|
||||
|
||||
// Act - Query using inherited property
|
||||
var recent = _db.DerivedEntities.Find(e => e.CreatedAt >= now.AddDays(-1.5)).ToList();
|
||||
|
||||
// Assert
|
||||
recent.Count.ShouldBe(2);
|
||||
recent.ShouldContain(e => e.Name == "Entity2");
|
||||
recent.ShouldContain(e => e.Name == "Entity3");
|
||||
}
|
||||
|
||||
#endregion
|
||||
|
||||
#region Computed Properties Tests
|
||||
|
||||
[Fact]
|
||||
public void ComputedProperties_AreNotSerialized()
|
||||
{
|
||||
// Arrange
|
||||
var entity = new EntityWithComputedProperties
|
||||
{
|
||||
FirstName = "John",
|
||||
LastName = "Doe",
|
||||
BirthYear = 1990
|
||||
};
|
||||
|
||||
// Act
|
||||
var id = _db.ComputedPropertyEntities.Insert(entity);
|
||||
_db.SaveChanges();
|
||||
var retrieved = _db.ComputedPropertyEntities.FindById(id);
|
||||
|
||||
// Assert
|
||||
retrieved.ShouldNotBeNull();
|
||||
retrieved.FirstName.ShouldBe("John");
|
||||
retrieved.LastName.ShouldBe("Doe");
|
||||
retrieved.BirthYear.ShouldBe(1990);
|
||||
|
||||
// Computed properties should still work after deserialization
|
||||
retrieved.FullName.ShouldBe("John Doe");
|
||||
(retrieved.Age >= 34).ShouldBeTrue(); // Born in 1990, so at least 34 in 2024+
|
||||
retrieved.DisplayInfo.ShouldContain("John Doe");
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void ComputedProperties_UpdateDoesNotBreak()
|
||||
{
|
||||
// Arrange
|
||||
var entity = new EntityWithComputedProperties
|
||||
{
|
||||
FirstName = "Jane",
|
||||
LastName = "Smith",
|
||||
BirthYear = 1985
|
||||
};
|
||||
var id = _db.ComputedPropertyEntities.Insert(entity);
|
||||
_db.SaveChanges();
|
||||
|
||||
// Act - Update stored properties
|
||||
var retrieved = _db.ComputedPropertyEntities.FindById(id);
|
||||
retrieved.ShouldNotBeNull();
|
||||
retrieved.FirstName = "Janet";
|
||||
retrieved.BirthYear = 1986;
|
||||
_db.ComputedPropertyEntities.Update(retrieved);
|
||||
_db.SaveChanges();
|
||||
|
||||
var updated = _db.ComputedPropertyEntities.FindById(id);
|
||||
|
||||
// Assert
|
||||
updated.ShouldNotBeNull();
|
||||
updated.FirstName.ShouldBe("Janet");
|
||||
updated.LastName.ShouldBe("Smith");
|
||||
updated.BirthYear.ShouldBe(1986);
|
||||
updated.FullName.ShouldBe("Janet Smith"); // Computed property reflects new data
|
||||
}
|
||||
|
||||
#endregion
|
||||
|
||||
#region Advanced Collections Tests
|
||||
|
||||
[Fact]
|
||||
public void HashSet_SerializesAndDeserializes()
|
||||
{
|
||||
// Arrange
|
||||
var entity = new EntityWithAdvancedCollections
|
||||
{
|
||||
Name = "Test HashSet"
|
||||
};
|
||||
entity.Tags.Add("tag1");
|
||||
entity.Tags.Add("tag2");
|
||||
entity.Tags.Add("tag3");
|
||||
|
||||
// Act
|
||||
var id = _db.AdvancedCollectionEntities.Insert(entity);
|
||||
_db.SaveChanges();
|
||||
var retrieved = _db.AdvancedCollectionEntities.FindById(id);
|
||||
|
||||
// Assert
|
||||
retrieved.ShouldNotBeNull();
|
||||
retrieved.Tags.ShouldNotBeNull();
|
||||
retrieved.Tags.ShouldBeOfType<HashSet<string>>();
|
||||
retrieved.Tags.Count.ShouldBe(3);
|
||||
retrieved.Tags.ShouldContain("tag1");
|
||||
retrieved.Tags.ShouldContain("tag2");
|
||||
retrieved.Tags.ShouldContain("tag3");
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void ISet_SerializesAndDeserializes()
|
||||
{
|
||||
// Arrange
|
||||
var entity = new EntityWithAdvancedCollections
|
||||
{
|
||||
Name = "Test ISet"
|
||||
};
|
||||
entity.Numbers.Add(10);
|
||||
entity.Numbers.Add(20);
|
||||
entity.Numbers.Add(30);
|
||||
entity.Numbers.Add(10); // Duplicate - should be ignored by set
|
||||
|
||||
// Act
|
||||
var id = _db.AdvancedCollectionEntities.Insert(entity);
|
||||
_db.SaveChanges();
|
||||
var retrieved = _db.AdvancedCollectionEntities.FindById(id);
|
||||
|
||||
// Assert
|
||||
retrieved.ShouldNotBeNull();
|
||||
retrieved.Numbers.ShouldNotBeNull();
|
||||
retrieved.Numbers.ShouldBeAssignableTo<ISet<int>>();
|
||||
retrieved.Numbers.Count.ShouldBe(3); // Only unique values
|
||||
retrieved.Numbers.ShouldContain(10);
|
||||
retrieved.Numbers.ShouldContain(20);
|
||||
retrieved.Numbers.ShouldContain(30);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void LinkedList_SerializesAndDeserializes()
|
||||
{
|
||||
// Arrange
|
||||
var entity = new EntityWithAdvancedCollections
|
||||
{
|
||||
Name = "Test LinkedList"
|
||||
};
|
||||
entity.History.AddLast("first");
|
||||
entity.History.AddLast("second");
|
||||
entity.History.AddLast("third");
|
||||
|
||||
// Act
|
||||
var id = _db.AdvancedCollectionEntities.Insert(entity);
|
||||
_db.SaveChanges();
|
||||
var retrieved = _db.AdvancedCollectionEntities.FindById(id);
|
||||
|
||||
// Assert
|
||||
retrieved.ShouldNotBeNull();
|
||||
retrieved.History.ShouldNotBeNull();
|
||||
// LinkedList may be deserialized as List, then need conversion
|
||||
var historyList = retrieved.History.ToList();
|
||||
historyList.Count.ShouldBe(3);
|
||||
historyList[0].ShouldBe("first");
|
||||
historyList[1].ShouldBe("second");
|
||||
historyList[2].ShouldBe("third");
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Queue_SerializesAndDeserializes()
|
||||
{
|
||||
// Arrange
|
||||
var entity = new EntityWithAdvancedCollections
|
||||
{
|
||||
Name = "Test Queue"
|
||||
};
|
||||
entity.PendingItems.Enqueue("item1");
|
||||
entity.PendingItems.Enqueue("item2");
|
||||
entity.PendingItems.Enqueue("item3");
|
||||
|
||||
// Act
|
||||
var id = _db.AdvancedCollectionEntities.Insert(entity);
|
||||
_db.SaveChanges();
|
||||
var retrieved = _db.AdvancedCollectionEntities.FindById(id);
|
||||
|
||||
// Assert
|
||||
retrieved.ShouldNotBeNull();
|
||||
retrieved.PendingItems.ShouldNotBeNull();
|
||||
retrieved.PendingItems.Count.ShouldBe(3);
|
||||
var items = retrieved.PendingItems.ToList();
|
||||
items.ShouldContain("item1");
|
||||
items.ShouldContain("item2");
|
||||
items.ShouldContain("item3");
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Stack_SerializesAndDeserializes()
|
||||
{
|
||||
// Arrange
|
||||
var entity = new EntityWithAdvancedCollections
|
||||
{
|
||||
Name = "Test Stack"
|
||||
};
|
||||
entity.UndoStack.Push("action1");
|
||||
entity.UndoStack.Push("action2");
|
||||
entity.UndoStack.Push("action3");
|
||||
|
||||
// Act
|
||||
var id = _db.AdvancedCollectionEntities.Insert(entity);
|
||||
_db.SaveChanges();
|
||||
var retrieved = _db.AdvancedCollectionEntities.FindById(id);
|
||||
|
||||
// Assert
|
||||
retrieved.ShouldNotBeNull();
|
||||
retrieved.UndoStack.ShouldNotBeNull();
|
||||
retrieved.UndoStack.Count.ShouldBe(3);
|
||||
var items = retrieved.UndoStack.ToList();
|
||||
items.ShouldContain("action1");
|
||||
items.ShouldContain("action2");
|
||||
items.ShouldContain("action3");
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void HashSet_WithNestedObjects_SerializesAndDeserializes()
|
||||
{
|
||||
// Arrange
|
||||
var entity = new EntityWithAdvancedCollections
|
||||
{
|
||||
Name = "Test Nested HashSet"
|
||||
};
|
||||
entity.Addresses.Add(new Address { Street = "123 Main St", City = new City { Name = "NYC", ZipCode = "10001" } });
|
||||
entity.Addresses.Add(new Address { Street = "456 Oak Ave", City = new City { Name = "LA", ZipCode = "90001" } });
|
||||
|
||||
// Act
|
||||
var id = _db.AdvancedCollectionEntities.Insert(entity);
|
||||
_db.SaveChanges();
|
||||
var retrieved = _db.AdvancedCollectionEntities.FindById(id);
|
||||
|
||||
// Assert
|
||||
retrieved.ShouldNotBeNull();
|
||||
retrieved.Addresses.ShouldNotBeNull();
|
||||
retrieved.Addresses.ShouldBeOfType<HashSet<Address>>();
|
||||
retrieved.Addresses.Count.ShouldBe(2);
|
||||
|
||||
var addressList = retrieved.Addresses.ToList();
|
||||
addressList.ShouldContain(a => a.Street == "123 Main St" && a.City.Name == "NYC");
|
||||
addressList.ShouldContain(a => a.Street == "456 Oak Ave" && a.City.Name == "LA");
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void ISet_WithNestedObjects_SerializesAndDeserializes()
|
||||
{
|
||||
// Arrange
|
||||
var entity = new EntityWithAdvancedCollections
|
||||
{
|
||||
Name = "Test Nested ISet"
|
||||
};
|
||||
entity.FavoriteCities.Add(new City { Name = "Paris", ZipCode = "75001" });
|
||||
entity.FavoriteCities.Add(new City { Name = "Tokyo", ZipCode = "100-0001" });
|
||||
entity.FavoriteCities.Add(new City { Name = "London", ZipCode = "SW1A" });
|
||||
|
||||
// Act
|
||||
var id = _db.AdvancedCollectionEntities.Insert(entity);
|
||||
_db.SaveChanges();
|
||||
var retrieved = _db.AdvancedCollectionEntities.FindById(id);
|
||||
|
||||
// Assert
|
||||
retrieved.ShouldNotBeNull();
|
||||
retrieved.FavoriteCities.ShouldNotBeNull();
|
||||
retrieved.FavoriteCities.ShouldBeAssignableTo<ISet<City>>();
|
||||
retrieved.FavoriteCities.Count.ShouldBe(3);
|
||||
|
||||
var cityNames = retrieved.FavoriteCities.Select(c => c.Name).ToList();
|
||||
cityNames.ShouldContain("Paris");
|
||||
cityNames.ShouldContain("Tokyo");
|
||||
cityNames.ShouldContain("London");
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void AdvancedCollections_AllTypesInSingleEntity()
|
||||
{
|
||||
// Arrange - Test all collection types at once
|
||||
var entity = new EntityWithAdvancedCollections
|
||||
{
|
||||
Name = "Complete Test"
|
||||
};
|
||||
|
||||
entity.Tags.Add("tag1");
|
||||
entity.Tags.Add("tag2");
|
||||
|
||||
entity.Numbers.Add(1);
|
||||
entity.Numbers.Add(2);
|
||||
|
||||
entity.History.AddLast("h1");
|
||||
entity.History.AddLast("h2");
|
||||
|
||||
entity.PendingItems.Enqueue("p1");
|
||||
entity.PendingItems.Enqueue("p2");
|
||||
|
||||
entity.UndoStack.Push("u1");
|
||||
entity.UndoStack.Push("u2");
|
||||
|
||||
entity.Addresses.Add(new Address { Street = "Street1" });
|
||||
entity.FavoriteCities.Add(new City { Name = "City1" });
|
||||
|
||||
// Act
|
||||
var id = _db.AdvancedCollectionEntities.Insert(entity);
|
||||
_db.SaveChanges();
|
||||
var retrieved = _db.AdvancedCollectionEntities.FindById(id);
|
||||
|
||||
// Assert
|
||||
retrieved.ShouldNotBeNull();
|
||||
retrieved.Name.ShouldBe("Complete Test");
|
||||
retrieved.Tags.Count.ShouldBe(2);
|
||||
retrieved.Numbers.Count.ShouldBe(2);
|
||||
retrieved.History.Count.ShouldBe(2);
|
||||
retrieved.PendingItems.Count.ShouldBe(2);
|
||||
retrieved.UndoStack.Count.ShouldBe(2);
|
||||
retrieved.Addresses.Count().ShouldBe(1);
|
||||
retrieved.FavoriteCities.Count().ShouldBe(1);
|
||||
}
|
||||
|
||||
#endregion
|
||||
|
||||
#region Private Setters Tests
|
||||
|
||||
[Fact]
|
||||
public void EntityWithPrivateSetters_CanBeDeserialized()
|
||||
{
|
||||
// Arrange
|
||||
var entity = EntityWithPrivateSetters.Create("John Doe", 30);
|
||||
|
||||
// Act
|
||||
var id = _db.PrivateSetterEntities.Insert(entity);
|
||||
_db.SaveChanges();
|
||||
var retrieved = _db.PrivateSetterEntities.FindById(id);
|
||||
|
||||
// Assert
|
||||
retrieved.ShouldNotBeNull();
|
||||
retrieved.Id.ShouldBe(id);
|
||||
retrieved.Name.ShouldBe("John Doe");
|
||||
retrieved.Age.ShouldBe(30);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void EntityWithPrivateSetters_Update_Works()
|
||||
{
|
||||
// Arrange
|
||||
var entity1 = EntityWithPrivateSetters.Create("Alice", 25);
|
||||
var id1 = _db.PrivateSetterEntities.Insert(entity1);
|
||||
|
||||
var entity2 = EntityWithPrivateSetters.Create("Bob", 35);
|
||||
entity2.GetType().GetProperty("Id")!.SetValue(entity2, id1); // Force same Id
|
||||
|
||||
_db.PrivateSetterEntities.Update(entity2);
|
||||
_db.SaveChanges();
|
||||
|
||||
// Act
|
||||
var retrieved = _db.PrivateSetterEntities.FindById(id1);
|
||||
|
||||
// Assert
|
||||
retrieved.ShouldNotBeNull();
|
||||
retrieved.Id.ShouldBe(id1);
|
||||
retrieved.Name.ShouldBe("Bob");
|
||||
retrieved.Age.ShouldBe(35);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void EntityWithPrivateSetters_Query_Works()
|
||||
{
|
||||
// Arrange
|
||||
var entity1 = EntityWithPrivateSetters.Create("Charlie", 20);
|
||||
var entity2 = EntityWithPrivateSetters.Create("Diana", 30);
|
||||
var entity3 = EntityWithPrivateSetters.Create("Eve", 40);
|
||||
|
||||
_db.PrivateSetterEntities.Insert(entity1);
|
||||
_db.PrivateSetterEntities.Insert(entity2);
|
||||
_db.PrivateSetterEntities.Insert(entity3);
|
||||
_db.SaveChanges();
|
||||
|
||||
// Act
|
||||
var adults = _db.PrivateSetterEntities.Find(e => e.Age >= 30).ToList();
|
||||
|
||||
// Assert
|
||||
adults.Count.ShouldBe(2);
|
||||
adults.ShouldContain(e => e.Name == "Diana");
|
||||
adults.ShouldContain(e => e.Name == "Eve");
|
||||
}
|
||||
|
||||
#endregion
|
||||
|
||||
#region Init-Only Setters Tests
|
||||
|
||||
[Fact]
|
||||
public void EntityWithInitSetters_CanBeDeserialized()
|
||||
{
|
||||
// Arrange
|
||||
var entity = new EntityWithInitSetters
|
||||
{
|
||||
Id = ObjectId.NewObjectId(),
|
||||
Name = "Jane Doe",
|
||||
Age = 28,
|
||||
CreatedAt = DateTime.UtcNow
|
||||
};
|
||||
|
||||
// Act
|
||||
var id = _db.InitSetterEntities.Insert(entity);
|
||||
_db.SaveChanges();
|
||||
var retrieved = _db.InitSetterEntities.FindById(id);
|
||||
|
||||
// Assert
|
||||
retrieved.ShouldNotBeNull();
|
||||
retrieved.Id.ShouldBe(id);
|
||||
retrieved.Name.ShouldBe("Jane Doe");
|
||||
retrieved.Age.ShouldBe(28);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void EntityWithInitSetters_Query_Works()
|
||||
{
|
||||
// Arrange
|
||||
var entity1 = new EntityWithInitSetters { Id = ObjectId.NewObjectId(), Name = "Alpha", Age = 20, CreatedAt = DateTime.UtcNow };
|
||||
var entity2 = new EntityWithInitSetters { Id = ObjectId.NewObjectId(), Name = "Beta", Age = 30, CreatedAt = DateTime.UtcNow };
|
||||
var entity3 = new EntityWithInitSetters { Id = ObjectId.NewObjectId(), Name = "Gamma", Age = 40, CreatedAt = DateTime.UtcNow };
|
||||
|
||||
_db.InitSetterEntities.Insert(entity1);
|
||||
_db.InitSetterEntities.Insert(entity2);
|
||||
_db.InitSetterEntities.Insert(entity3);
|
||||
_db.SaveChanges();
|
||||
|
||||
// Act
|
||||
var results = _db.InitSetterEntities.Find(e => e.Age > 25).ToList();
|
||||
|
||||
// Assert
|
||||
results.Count.ShouldBe(2);
|
||||
results.ShouldContain(e => e.Name == "Beta");
|
||||
results.ShouldContain(e => e.Name == "Gamma");
|
||||
}
|
||||
|
||||
#endregion
|
||||
|
||||
public void Dispose()
|
||||
{
|
||||
_db?.Dispose();
|
||||
|
||||
if (File.Exists(_dbPath))
|
||||
File.Delete(_dbPath);
|
||||
if (File.Exists(_walPath))
|
||||
File.Delete(_walPath);
|
||||
}
|
||||
}
|
||||
104
tests/CBDD.Tests/StorageEngineDictionaryTests.cs
Executable file
104
tests/CBDD.Tests/StorageEngineDictionaryTests.cs
Executable file
@@ -0,0 +1,104 @@
|
||||
using ZB.MOM.WW.CBDD.Core.Storage;
|
||||
using Xunit;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests;
|
||||
|
||||
public class StorageEngineDictionaryTests
|
||||
{
|
||||
private string GetTempDbPath() => Path.Combine(Path.GetTempPath(), $"test_storage_dict_{Guid.NewGuid()}.db");
|
||||
|
||||
private void Cleanup(string path)
|
||||
{
|
||||
if (File.Exists(path)) File.Delete(path);
|
||||
if (File.Exists(Path.ChangeExtension(path, ".wal"))) File.Delete(Path.ChangeExtension(path, ".wal"));
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void StorageEngine_ShouldInitializeDictionary()
|
||||
{
|
||||
var path = GetTempDbPath();
|
||||
try
|
||||
{
|
||||
using (var storage = new StorageEngine(path, PageFileConfig.Default))
|
||||
{
|
||||
// Should generate ID > 100
|
||||
var id = storage.GetOrAddDictionaryEntry("TestKey");
|
||||
(id > DictionaryPage.ReservedValuesEnd).ShouldBeTrue();
|
||||
|
||||
var key = storage.GetDictionaryKey(id);
|
||||
key.ShouldBe("testkey");
|
||||
}
|
||||
}
|
||||
finally { Cleanup(path); }
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void StorageEngine_ShouldPersistDictionary()
|
||||
{
|
||||
var path = GetTempDbPath();
|
||||
try
|
||||
{
|
||||
ushort id1, id2;
|
||||
using (var storage = new StorageEngine(path, PageFileConfig.Default))
|
||||
{
|
||||
id1 = storage.GetOrAddDictionaryEntry("Key1");
|
||||
id2 = storage.GetOrAddDictionaryEntry("Key2");
|
||||
}
|
||||
|
||||
// Reopen
|
||||
using (var storage = new StorageEngine(path, PageFileConfig.Default))
|
||||
{
|
||||
var val1 = storage.GetOrAddDictionaryEntry("Key1");
|
||||
var val2 = storage.GetOrAddDictionaryEntry("Key2");
|
||||
|
||||
val1.ShouldBe(id1);
|
||||
val2.ShouldBe(id2);
|
||||
|
||||
storage.GetDictionaryKey(val1).ShouldBe("key1");
|
||||
storage.GetDictionaryKey(val2).ShouldBe("key2");
|
||||
}
|
||||
}
|
||||
finally { Cleanup(path); }
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void StorageEngine_ShouldHandleManyKeys()
|
||||
{
|
||||
var path = GetTempDbPath();
|
||||
try
|
||||
{
|
||||
const int keyCount = 3000;
|
||||
var expectedIds = new Dictionary<string, ushort>();
|
||||
|
||||
using (var storage = new StorageEngine(path, PageFileConfig.Default))
|
||||
{
|
||||
for (int i = 0; i < keyCount; i++)
|
||||
{
|
||||
var key = $"Key_{i}";
|
||||
var id = storage.GetOrAddDictionaryEntry(key);
|
||||
expectedIds[key] = id;
|
||||
}
|
||||
}
|
||||
|
||||
// Reopen and Verify
|
||||
using (var storage = new StorageEngine(path, PageFileConfig.Default))
|
||||
{
|
||||
for (int i = 0; i < keyCount; i++)
|
||||
{
|
||||
var key = $"Key_{i}";
|
||||
var id = storage.GetOrAddDictionaryEntry(key); // Should get existing
|
||||
id.ShouldBe(expectedIds[key]);
|
||||
|
||||
var loadedKey = storage.GetDictionaryKey(id);
|
||||
loadedKey.ShouldBe(key.ToLowerInvariant());
|
||||
}
|
||||
|
||||
// Add new one
|
||||
var newId = storage.GetOrAddDictionaryEntry("NewKeyAfterReopen");
|
||||
(newId > 0).ShouldBeTrue();
|
||||
expectedIds.ContainsValue(newId).ShouldBeFalse();
|
||||
}
|
||||
}
|
||||
finally { Cleanup(path); }
|
||||
}
|
||||
}
|
||||
211
tests/CBDD.Tests/StorageEngineTransactionProtocolTests.cs
Normal file
211
tests/CBDD.Tests/StorageEngineTransactionProtocolTests.cs
Normal file
@@ -0,0 +1,211 @@
|
||||
using ZB.MOM.WW.CBDD.Core.Storage;
|
||||
using ZB.MOM.WW.CBDD.Core.Transactions;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests;
|
||||
|
||||
public class StorageEngineTransactionProtocolTests
|
||||
{
|
||||
[Fact]
|
||||
public void PrepareTransaction_Should_ReturnFalse_For_Unknown_Transaction()
|
||||
{
|
||||
var dbPath = NewDbPath();
|
||||
try
|
||||
{
|
||||
using var storage = new StorageEngine(dbPath, PageFileConfig.Default);
|
||||
storage.PrepareTransaction(999_999).ShouldBeFalse();
|
||||
}
|
||||
finally
|
||||
{
|
||||
CleanupFiles(dbPath);
|
||||
}
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void CommitTransaction_With_TransactionObject_Should_Throw_When_Not_Active()
|
||||
{
|
||||
var dbPath = NewDbPath();
|
||||
try
|
||||
{
|
||||
using var storage = new StorageEngine(dbPath, PageFileConfig.Default);
|
||||
var detached = new Transaction(123, storage);
|
||||
|
||||
Should.Throw<InvalidOperationException>(() => storage.CommitTransaction(detached));
|
||||
}
|
||||
finally
|
||||
{
|
||||
CleanupFiles(dbPath);
|
||||
}
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void CommitTransaction_With_TransactionObject_Should_Commit_Writes()
|
||||
{
|
||||
var dbPath = NewDbPath();
|
||||
try
|
||||
{
|
||||
using var storage = new StorageEngine(dbPath, PageFileConfig.Default);
|
||||
using var txn = storage.BeginTransaction();
|
||||
|
||||
var pageId = storage.AllocatePage();
|
||||
var data = new byte[storage.PageSize];
|
||||
data[0] = 0xAB;
|
||||
|
||||
storage.WritePage(pageId, txn.TransactionId, data);
|
||||
storage.CommitTransaction(txn);
|
||||
|
||||
storage.ActiveTransactionCount.ShouldBe(0);
|
||||
|
||||
var readBuffer = new byte[storage.PageSize];
|
||||
storage.ReadPage(pageId, 0, readBuffer);
|
||||
readBuffer[0].ShouldBe((byte)0xAB);
|
||||
}
|
||||
finally
|
||||
{
|
||||
CleanupFiles(dbPath);
|
||||
}
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void CommitTransaction_ById_With_NoWrites_Should_Not_Throw()
|
||||
{
|
||||
var dbPath = NewDbPath();
|
||||
try
|
||||
{
|
||||
using var storage = new StorageEngine(dbPath, PageFileConfig.Default);
|
||||
storage.CommitTransaction(424242);
|
||||
}
|
||||
finally
|
||||
{
|
||||
CleanupFiles(dbPath);
|
||||
}
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void MarkTransactionCommitted_Should_Move_Cache_And_Clear_ActiveCount()
|
||||
{
|
||||
var dbPath = NewDbPath();
|
||||
try
|
||||
{
|
||||
using var storage = new StorageEngine(dbPath, PageFileConfig.Default);
|
||||
using var txn = storage.BeginTransaction();
|
||||
|
||||
var pageId = storage.AllocatePage();
|
||||
var data = new byte[storage.PageSize];
|
||||
data[5] = 0x5A;
|
||||
storage.WritePage(pageId, txn.TransactionId, data);
|
||||
|
||||
storage.ActiveTransactionCount.ShouldBe(1);
|
||||
storage.MarkTransactionCommitted(txn.TransactionId);
|
||||
storage.ActiveTransactionCount.ShouldBe(0);
|
||||
|
||||
var readBuffer = new byte[storage.PageSize];
|
||||
storage.ReadPage(pageId, 0, readBuffer);
|
||||
readBuffer[5].ShouldBe((byte)0x5A);
|
||||
}
|
||||
finally
|
||||
{
|
||||
CleanupFiles(dbPath);
|
||||
}
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void RollbackTransaction_Should_Discard_Uncommitted_Write()
|
||||
{
|
||||
var dbPath = NewDbPath();
|
||||
try
|
||||
{
|
||||
using var storage = new StorageEngine(dbPath, PageFileConfig.Default);
|
||||
|
||||
var pageId = storage.AllocatePage();
|
||||
var baseline = new byte[storage.PageSize];
|
||||
baseline[0] = 0x11;
|
||||
storage.WritePageImmediate(pageId, baseline);
|
||||
|
||||
using var txn = storage.BeginTransaction();
|
||||
var changed = new byte[storage.PageSize];
|
||||
changed[0] = 0x99;
|
||||
storage.WritePage(pageId, txn.TransactionId, changed);
|
||||
|
||||
storage.ActiveTransactionCount.ShouldBe(1);
|
||||
storage.RollbackTransaction(txn.TransactionId);
|
||||
storage.ActiveTransactionCount.ShouldBe(0);
|
||||
|
||||
var readBuffer = new byte[storage.PageSize];
|
||||
storage.ReadPage(pageId, 0, readBuffer);
|
||||
readBuffer[0].ShouldBe((byte)0x11);
|
||||
}
|
||||
finally
|
||||
{
|
||||
CleanupFiles(dbPath);
|
||||
}
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Transaction_MarkCommitted_Should_Transition_State()
|
||||
{
|
||||
var dbPath = NewDbPath();
|
||||
try
|
||||
{
|
||||
using var storage = new StorageEngine(dbPath, PageFileConfig.Default);
|
||||
using var txn = storage.BeginTransaction();
|
||||
|
||||
var pageId = storage.AllocatePage();
|
||||
var data = new byte[storage.PageSize];
|
||||
data[3] = 0x33;
|
||||
storage.WritePage(pageId, txn.TransactionId, data);
|
||||
|
||||
txn.MarkCommitted();
|
||||
|
||||
txn.State.ShouldBe(TransactionState.Committed);
|
||||
storage.ActiveTransactionCount.ShouldBe(0);
|
||||
|
||||
var readBuffer = new byte[storage.PageSize];
|
||||
storage.ReadPage(pageId, 0, readBuffer);
|
||||
readBuffer[3].ShouldBe((byte)0x33);
|
||||
}
|
||||
finally
|
||||
{
|
||||
CleanupFiles(dbPath);
|
||||
}
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Transaction_Prepare_Should_Write_Wal_And_Transition_State()
|
||||
{
|
||||
var dbPath = NewDbPath();
|
||||
try
|
||||
{
|
||||
using var storage = new StorageEngine(dbPath, PageFileConfig.Default);
|
||||
using var txn = storage.BeginTransaction();
|
||||
|
||||
var pageId = storage.AllocatePage();
|
||||
var data = new byte[storage.PageSize];
|
||||
data[11] = 0x7B;
|
||||
storage.WritePage(pageId, txn.TransactionId, data);
|
||||
|
||||
txn.Prepare().ShouldBeTrue();
|
||||
txn.State.ShouldBe(TransactionState.Preparing);
|
||||
|
||||
txn.Commit();
|
||||
txn.State.ShouldBe(TransactionState.Committed);
|
||||
}
|
||||
finally
|
||||
{
|
||||
CleanupFiles(dbPath);
|
||||
}
|
||||
}
|
||||
|
||||
private static string NewDbPath()
|
||||
=> Path.Combine(Path.GetTempPath(), $"storage_txn_{Guid.NewGuid():N}.db");
|
||||
|
||||
private static void CleanupFiles(string dbPath)
|
||||
{
|
||||
if (File.Exists(dbPath)) File.Delete(dbPath);
|
||||
|
||||
var walPath = Path.ChangeExtension(dbPath, ".wal");
|
||||
if (File.Exists(walPath)) File.Delete(walPath);
|
||||
|
||||
var altWalPath = dbPath + "-wal";
|
||||
if (File.Exists(altWalPath)) File.Delete(altWalPath);
|
||||
}
|
||||
}
|
||||
229
tests/CBDD.Tests/TemporalTypesTests.cs
Executable file
229
tests/CBDD.Tests/TemporalTypesTests.cs
Executable file
@@ -0,0 +1,229 @@
|
||||
using System;
|
||||
using ZB.MOM.WW.CBDD.Bson;
|
||||
using ZB.MOM.WW.CBDD.Shared;
|
||||
using ZB.MOM.WW.CBDD.Tests;
|
||||
using Xunit;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests
|
||||
{
|
||||
public class TemporalTypesTests : IDisposable
|
||||
{
|
||||
private readonly Shared.TestDbContext _db;
|
||||
private readonly string _dbPath;
|
||||
|
||||
public TemporalTypesTests()
|
||||
{
|
||||
_dbPath = $"temporal_test_{Guid.NewGuid()}.db";
|
||||
_db = new Shared.TestDbContext(_dbPath);
|
||||
}
|
||||
|
||||
public void Dispose()
|
||||
{
|
||||
_db?.Dispose();
|
||||
if (File.Exists(_dbPath))
|
||||
File.Delete(_dbPath);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void TemporalEntity_Collection_IsInitialized()
|
||||
{
|
||||
_db.TemporalEntities.ShouldNotBeNull();
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void TemporalEntity_Insert_And_FindById_Works()
|
||||
{
|
||||
// Arrange
|
||||
var now = DateTime.UtcNow;
|
||||
var offset = DateTimeOffset.UtcNow;
|
||||
var duration = TimeSpan.FromHours(5.5);
|
||||
var birthDate = new DateOnly(1990, 5, 15);
|
||||
var openingTime = new TimeOnly(9, 30, 0);
|
||||
|
||||
var entity = new TemporalEntity
|
||||
{
|
||||
Id = ObjectId.NewObjectId(),
|
||||
Name = "Test Entity",
|
||||
CreatedAt = now,
|
||||
UpdatedAt = offset,
|
||||
LastAccessedAt = offset.AddDays(1),
|
||||
Duration = duration,
|
||||
OptionalDuration = TimeSpan.FromMinutes(30),
|
||||
BirthDate = birthDate,
|
||||
Anniversary = new DateOnly(2020, 6, 10),
|
||||
OpeningTime = openingTime,
|
||||
ClosingTime = new TimeOnly(18, 0, 0)
|
||||
};
|
||||
|
||||
// Act
|
||||
_db.TemporalEntities.Insert(entity);
|
||||
var retrieved = _db.TemporalEntities.FindById(entity.Id);
|
||||
|
||||
// Assert
|
||||
retrieved.ShouldNotBeNull();
|
||||
retrieved.Name.ShouldBe(entity.Name);
|
||||
|
||||
// DateTime comparison (allowing some millisecond precision loss)
|
||||
(retrieved.CreatedAt.Ticks / 10000).ShouldBe(entity.CreatedAt.Ticks / 10000); // millisecond precision
|
||||
|
||||
// DateTimeOffset comparison
|
||||
(retrieved.UpdatedAt.UtcDateTime.Ticks / 10000).ShouldBe(entity.UpdatedAt.UtcDateTime.Ticks / 10000);
|
||||
retrieved.LastAccessedAt.ShouldNotBeNull();
|
||||
(retrieved.LastAccessedAt!.Value.UtcDateTime.Ticks / 10000).ShouldBe(entity.LastAccessedAt!.Value.UtcDateTime.Ticks / 10000);
|
||||
|
||||
// TimeSpan comparison
|
||||
retrieved.Duration.ShouldBe(entity.Duration);
|
||||
retrieved.OptionalDuration.ShouldNotBeNull();
|
||||
retrieved.OptionalDuration!.Value.ShouldBe(entity.OptionalDuration!.Value);
|
||||
|
||||
// DateOnly comparison
|
||||
retrieved.BirthDate.ShouldBe(entity.BirthDate);
|
||||
retrieved.Anniversary.ShouldNotBeNull();
|
||||
retrieved.Anniversary!.Value.ShouldBe(entity.Anniversary!.Value);
|
||||
|
||||
// TimeOnly comparison
|
||||
retrieved.OpeningTime.ShouldBe(entity.OpeningTime);
|
||||
retrieved.ClosingTime.ShouldNotBeNull();
|
||||
retrieved.ClosingTime!.Value.ShouldBe(entity.ClosingTime!.Value);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void TemporalEntity_Insert_WithNullOptionalFields_Works()
|
||||
{
|
||||
// Arrange
|
||||
var entity = new TemporalEntity
|
||||
{
|
||||
Id = ObjectId.NewObjectId(),
|
||||
Name = "Minimal Entity",
|
||||
CreatedAt = DateTime.UtcNow,
|
||||
UpdatedAt = DateTimeOffset.UtcNow,
|
||||
Duration = TimeSpan.FromHours(1),
|
||||
BirthDate = new DateOnly(1985, 3, 20),
|
||||
OpeningTime = new TimeOnly(8, 0, 0),
|
||||
// Optional fields left null
|
||||
LastAccessedAt = null,
|
||||
OptionalDuration = null,
|
||||
Anniversary = null,
|
||||
ClosingTime = null
|
||||
};
|
||||
|
||||
// Act
|
||||
_db.TemporalEntities.Insert(entity);
|
||||
var retrieved = _db.TemporalEntities.FindById(entity.Id);
|
||||
|
||||
// Assert
|
||||
retrieved.ShouldNotBeNull();
|
||||
retrieved.Name.ShouldBe(entity.Name);
|
||||
retrieved.LastAccessedAt.ShouldBeNull();
|
||||
retrieved.OptionalDuration.ShouldBeNull();
|
||||
retrieved.Anniversary.ShouldBeNull();
|
||||
retrieved.ClosingTime.ShouldBeNull();
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void TemporalEntity_Update_Works()
|
||||
{
|
||||
// Arrange
|
||||
var entity = new TemporalEntity
|
||||
{
|
||||
Id = ObjectId.NewObjectId(),
|
||||
Name = "Original",
|
||||
CreatedAt = DateTime.UtcNow,
|
||||
UpdatedAt = DateTimeOffset.UtcNow,
|
||||
Duration = TimeSpan.FromHours(1),
|
||||
BirthDate = new DateOnly(1990, 1, 1),
|
||||
OpeningTime = new TimeOnly(9, 0, 0)
|
||||
};
|
||||
|
||||
_db.TemporalEntities.Insert(entity);
|
||||
|
||||
// Act - Update temporal fields
|
||||
entity.Name = "Updated";
|
||||
entity.UpdatedAt = DateTimeOffset.UtcNow.AddDays(1);
|
||||
entity.Duration = TimeSpan.FromHours(2);
|
||||
entity.BirthDate = new DateOnly(1991, 2, 2);
|
||||
entity.OpeningTime = new TimeOnly(10, 0, 0);
|
||||
|
||||
_db.TemporalEntities.Update(entity);
|
||||
var retrieved = _db.TemporalEntities.FindById(entity.Id);
|
||||
|
||||
// Assert
|
||||
retrieved.ShouldNotBeNull();
|
||||
retrieved.Name.ShouldBe("Updated");
|
||||
retrieved.Duration.ShouldBe(entity.Duration);
|
||||
retrieved.BirthDate.ShouldBe(entity.BirthDate);
|
||||
retrieved.OpeningTime.ShouldBe(entity.OpeningTime);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void TemporalEntity_Query_Works()
|
||||
{
|
||||
// Arrange
|
||||
var birthDate1 = new DateOnly(1990, 1, 1);
|
||||
var birthDate2 = new DateOnly(1995, 6, 15);
|
||||
|
||||
var entity1 = new TemporalEntity
|
||||
{
|
||||
Id = ObjectId.NewObjectId(),
|
||||
Name = "Person 1",
|
||||
CreatedAt = DateTime.UtcNow,
|
||||
UpdatedAt = DateTimeOffset.UtcNow,
|
||||
Duration = TimeSpan.FromHours(1),
|
||||
BirthDate = birthDate1,
|
||||
OpeningTime = new TimeOnly(9, 0, 0)
|
||||
};
|
||||
|
||||
var entity2 = new TemporalEntity
|
||||
{
|
||||
Id = ObjectId.NewObjectId(),
|
||||
Name = "Person 2",
|
||||
CreatedAt = DateTime.UtcNow,
|
||||
UpdatedAt = DateTimeOffset.UtcNow,
|
||||
Duration = TimeSpan.FromHours(2),
|
||||
BirthDate = birthDate2,
|
||||
OpeningTime = new TimeOnly(10, 0, 0)
|
||||
};
|
||||
|
||||
_db.TemporalEntities.Insert(entity1);
|
||||
_db.TemporalEntities.Insert(entity2);
|
||||
|
||||
// Act
|
||||
var results = _db.TemporalEntities.AsQueryable()
|
||||
.Where(e => e.BirthDate == birthDate1)
|
||||
.ToList();
|
||||
|
||||
// Assert
|
||||
results.Count().ShouldBe(1);
|
||||
results[0].Name.ShouldBe("Person 1");
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void TimeSpan_EdgeCases_Work()
|
||||
{
|
||||
// Arrange - Test various TimeSpan values
|
||||
var entity = new TemporalEntity
|
||||
{
|
||||
Id = ObjectId.NewObjectId(),
|
||||
Name = "TimeSpan Test",
|
||||
CreatedAt = DateTime.UtcNow,
|
||||
UpdatedAt = DateTimeOffset.UtcNow,
|
||||
Duration = TimeSpan.Zero,
|
||||
OptionalDuration = TimeSpan.MaxValue,
|
||||
BirthDate = DateOnly.MinValue,
|
||||
OpeningTime = TimeOnly.MinValue
|
||||
};
|
||||
|
||||
// Act
|
||||
_db.TemporalEntities.Insert(entity);
|
||||
var retrieved = _db.TemporalEntities.FindById(entity.Id);
|
||||
|
||||
// Assert
|
||||
retrieved.ShouldNotBeNull();
|
||||
retrieved.Duration.ShouldBe(TimeSpan.Zero);
|
||||
retrieved.OptionalDuration.ShouldNotBeNull();
|
||||
retrieved.OptionalDuration!.Value.ShouldBe(TimeSpan.MaxValue);
|
||||
retrieved.BirthDate.ShouldBe(DateOnly.MinValue);
|
||||
retrieved.OpeningTime.ShouldBe(TimeOnly.MinValue);
|
||||
}
|
||||
}
|
||||
}
|
||||
135
tests/CBDD.Tests/TestDbContext.cs
Executable file
135
tests/CBDD.Tests/TestDbContext.cs
Executable file
@@ -0,0 +1,135 @@
|
||||
using ZB.MOM.WW.CBDD.Bson;
|
||||
using ZB.MOM.WW.CBDD.Core;
|
||||
using ZB.MOM.WW.CBDD.Core.Collections;
|
||||
using ZB.MOM.WW.CBDD.Core.Compression;
|
||||
using ZB.MOM.WW.CBDD.Core.Indexing;
|
||||
using ZB.MOM.WW.CBDD.Core.Metadata;
|
||||
using ZB.MOM.WW.CBDD.Core.Storage;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Shared;
|
||||
|
||||
/// <summary>
|
||||
/// Test context with manual collection initialization
|
||||
/// (Source Generator will automate this in the future)
|
||||
/// </summary>
|
||||
public partial class TestDbContext : DocumentDbContext
|
||||
{
|
||||
public DocumentCollection<ObjectId, AnnotatedUser> AnnotatedUsers { get; set; } = null!;
|
||||
public DocumentCollection<OrderId, Order> Orders { get; set; } = null!;
|
||||
public DocumentCollection<ObjectId, TestDocument> TestDocuments { get; set; } = null!;
|
||||
public DocumentCollection<ObjectId, OrderDocument> OrderDocuments { get; set; } = null!;
|
||||
public DocumentCollection<ObjectId, ComplexDocument> ComplexDocuments { get; set; } = null!;
|
||||
public DocumentCollection<ObjectId, User> Users { get; set; } = null!;
|
||||
public DocumentCollection<ObjectId, ComplexUser> ComplexUsers { get; set; } = null!;
|
||||
public DocumentCollection<int, AutoInitEntity> AutoInitEntities { get; set; } = null!;
|
||||
public DocumentCollection<int, Person> People { get; set; } = null!;
|
||||
public DocumentCollection<ObjectId, PersonV2> PeopleV2 { get; set; } = null!;
|
||||
public DocumentCollection<int, Product> Products { get; set; } = null!;
|
||||
public DocumentCollection<int, IntEntity> IntEntities { get; set; } = null!;
|
||||
public DocumentCollection<string, StringEntity> StringEntities { get; set; } = null!;
|
||||
public DocumentCollection<Guid, GuidEntity> GuidEntities { get; set; } = null!;
|
||||
public DocumentCollection<string, CustomKeyEntity> CustomKeyEntities { get; set; } = null!;
|
||||
public DocumentCollection<int, AsyncDoc> AsyncDocs { get; set; } = null!;
|
||||
public DocumentCollection<int, SchemaUser> SchemaUsers { get; set; } = null!;
|
||||
public DocumentCollection<ObjectId, VectorEntity> VectorItems { get; set; } = null!;
|
||||
public DocumentCollection<ObjectId, GeoEntity> GeoItems { get; set; } = null!;
|
||||
|
||||
// Source Generator Feature Tests
|
||||
public DocumentCollection<ObjectId, DerivedEntity> DerivedEntities { get; set; } = null!;
|
||||
public DocumentCollection<ObjectId, EntityWithComputedProperties> ComputedPropertyEntities { get; set; } = null!;
|
||||
public DocumentCollection<ObjectId, EntityWithAdvancedCollections> AdvancedCollectionEntities { get; set; } = null!;
|
||||
public DocumentCollection<ObjectId, EntityWithPrivateSetters> PrivateSetterEntities { get; set; } = null!;
|
||||
public DocumentCollection<ObjectId, EntityWithInitSetters> InitSetterEntities { get; set; } = null!;
|
||||
|
||||
// Circular Reference Tests
|
||||
public DocumentCollection<ObjectId, Employee> Employees { get; set; } = null!;
|
||||
public DocumentCollection<ObjectId, CategoryRef> CategoryRefs { get; set; } = null!;
|
||||
public DocumentCollection<ObjectId, ProductRef> ProductRefs { get; set; } = null!;
|
||||
|
||||
// Nullable String Id Test (UuidEntity scenario with inheritance)
|
||||
public DocumentCollection<string, MockCounter> MockCounters { get; set; } = null!;
|
||||
|
||||
// Temporal Types Test (DateTimeOffset, TimeSpan, DateOnly, TimeOnly)
|
||||
public DocumentCollection<ObjectId, TemporalEntity> TemporalEntities { get; set; } = null!;
|
||||
|
||||
public TestDbContext(string databasePath)
|
||||
: this(databasePath, PageFileConfig.Default, CompressionOptions.Default)
|
||||
{
|
||||
}
|
||||
|
||||
public TestDbContext(string databasePath, CompressionOptions compressionOptions)
|
||||
: this(databasePath, PageFileConfig.Default, compressionOptions)
|
||||
{
|
||||
}
|
||||
|
||||
public TestDbContext(string databasePath, PageFileConfig pageFileConfig)
|
||||
: this(databasePath, pageFileConfig, CompressionOptions.Default)
|
||||
{
|
||||
}
|
||||
|
||||
public TestDbContext(
|
||||
string databasePath,
|
||||
PageFileConfig pageFileConfig,
|
||||
CompressionOptions? compressionOptions,
|
||||
MaintenanceOptions? maintenanceOptions = null)
|
||||
: base(databasePath, pageFileConfig, compressionOptions, maintenanceOptions)
|
||||
{
|
||||
}
|
||||
|
||||
protected override void OnModelCreating(ModelBuilder modelBuilder)
|
||||
{
|
||||
modelBuilder.Entity<AnnotatedUser>();
|
||||
modelBuilder.Entity<User>().ToCollection("users");
|
||||
modelBuilder.Entity<ComplexUser>().ToCollection("complex_users");
|
||||
modelBuilder.Entity<AutoInitEntity>().ToCollection("auto_init_entities");
|
||||
modelBuilder.Entity<Person>().ToCollection("people_collection").HasIndex(p => p.Age);
|
||||
modelBuilder.Entity<PersonV2>().ToCollection("peoplev2_collection").HasIndex(p => p.Age);
|
||||
modelBuilder.Entity<Product>().ToCollection("products_collection").HasIndex(p => p.Price);
|
||||
modelBuilder.Entity<IntEntity>().HasKey(e => e.Id);
|
||||
modelBuilder.Entity<StringEntity>().HasKey(e => e.Id);
|
||||
modelBuilder.Entity<GuidEntity>().HasKey(e => e.Id);
|
||||
modelBuilder.Entity<CustomKeyEntity>().HasKey(e => e.Code);
|
||||
modelBuilder.Entity<AsyncDoc>().ToCollection("async_docs");
|
||||
modelBuilder.Entity<SchemaUser>().ToCollection("schema_users").HasKey(e => e.Id);
|
||||
modelBuilder.Entity<TestDocument>();
|
||||
modelBuilder.Entity<OrderDocument>();
|
||||
modelBuilder.Entity<ComplexDocument>();
|
||||
|
||||
modelBuilder.Entity<VectorEntity>()
|
||||
.ToCollection("vector_items")
|
||||
.HasVectorIndex(x => x.Embedding, dimensions: 3, metric: VectorMetric.L2, name: "idx_vector");
|
||||
|
||||
modelBuilder.Entity<GeoEntity>()
|
||||
.ToCollection("geo_items")
|
||||
.HasSpatialIndex(x => x.Location, name: "idx_spatial");
|
||||
|
||||
modelBuilder.Entity<Order>()
|
||||
.HasKey(x => x.Id)
|
||||
.HasConversion<OrderIdConverter>();
|
||||
|
||||
// Source Generator Feature Tests
|
||||
modelBuilder.Entity<DerivedEntity>().ToCollection("derived_entities");
|
||||
modelBuilder.Entity<EntityWithComputedProperties>().ToCollection("computed_property_entities");
|
||||
modelBuilder.Entity<EntityWithAdvancedCollections>().ToCollection("advanced_collection_entities");
|
||||
modelBuilder.Entity<EntityWithPrivateSetters>().ToCollection("private_setter_entities");
|
||||
modelBuilder.Entity<EntityWithInitSetters>().ToCollection("init_setter_entities");
|
||||
|
||||
// Circular Reference Tests
|
||||
modelBuilder.Entity<Employee>().ToCollection("employees");
|
||||
modelBuilder.Entity<CategoryRef>().ToCollection("category_refs");
|
||||
modelBuilder.Entity<ProductRef>().ToCollection("product_refs");
|
||||
|
||||
// Nullable String Id Test (UuidEntity scenario)
|
||||
modelBuilder.Entity<MockCounter>().ToCollection("mock_counters").HasKey(e => e.Id);
|
||||
|
||||
// Temporal Types Test
|
||||
modelBuilder.Entity<TemporalEntity>().ToCollection("temporal_entities").HasKey(e => e.Id);
|
||||
}
|
||||
|
||||
public void ForceCheckpoint()
|
||||
{
|
||||
Engine.Checkpoint();
|
||||
}
|
||||
|
||||
public StorageEngine Storage => Engine;
|
||||
}
|
||||
27
tests/CBDD.Tests/TestExtendedDbContext.cs
Executable file
27
tests/CBDD.Tests/TestExtendedDbContext.cs
Executable file
@@ -0,0 +1,27 @@
|
||||
using ZB.MOM.WW.CBDD.Core.Collections;
|
||||
using ZB.MOM.WW.CBDD.Core.Metadata;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Shared;
|
||||
|
||||
/// <summary>
|
||||
/// Extended test context that inherits from TestDbContext.
|
||||
/// Used to verify that collection initialization works correctly with inheritance.
|
||||
/// </summary>
|
||||
public partial class TestExtendedDbContext : TestDbContext
|
||||
{
|
||||
public DocumentCollection<int, ExtendedEntity> ExtendedEntities { get; set; } = null!;
|
||||
|
||||
public TestExtendedDbContext(string databasePath) : base(databasePath)
|
||||
{
|
||||
InitializeCollections();
|
||||
}
|
||||
|
||||
protected override void OnModelCreating(ModelBuilder modelBuilder)
|
||||
{
|
||||
base.OnModelCreating(modelBuilder);
|
||||
|
||||
modelBuilder.Entity<ExtendedEntity>()
|
||||
.ToCollection("extended_entities")
|
||||
.HasKey(e => e.Id);
|
||||
}
|
||||
}
|
||||
44
tests/CBDD.Tests/ValueObjectIdTests.cs
Executable file
44
tests/CBDD.Tests/ValueObjectIdTests.cs
Executable file
@@ -0,0 +1,44 @@
|
||||
using ZB.MOM.WW.CBDD.Bson;
|
||||
using ZB.MOM.WW.CBDD.Core;
|
||||
using ZB.MOM.WW.CBDD.Core.Collections;
|
||||
using ZB.MOM.WW.CBDD.Core.Metadata;
|
||||
using ZB.MOM.WW.CBDD.Shared;
|
||||
using Xunit;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests;
|
||||
|
||||
public class ValueObjectIdTests : IDisposable
|
||||
{
|
||||
private readonly string _dbPath = "value_object_ids.db";
|
||||
private readonly Shared.TestDbContext _db;
|
||||
|
||||
public ValueObjectIdTests()
|
||||
{
|
||||
if (File.Exists(_dbPath)) File.Delete(_dbPath);
|
||||
_db = new Shared.TestDbContext(_dbPath);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Should_Support_ValueObject_Id_Conversion()
|
||||
{
|
||||
var order = new Order
|
||||
{
|
||||
Id = new OrderId("ORD-123"),
|
||||
CustomerName = "John Doe"
|
||||
};
|
||||
|
||||
_db.Orders.Insert(order);
|
||||
|
||||
var retrieved = _db.Orders.FindById(new OrderId("ORD-123"));
|
||||
|
||||
retrieved.ShouldNotBeNull();
|
||||
retrieved.Id.Value.ShouldBe("ORD-123");
|
||||
retrieved.CustomerName.ShouldBe("John Doe");
|
||||
}
|
||||
|
||||
public void Dispose()
|
||||
{
|
||||
_db.Dispose();
|
||||
if (File.Exists(_dbPath)) File.Delete(_dbPath);
|
||||
}
|
||||
}
|
||||
50
tests/CBDD.Tests/VectorMathTests.cs
Normal file
50
tests/CBDD.Tests/VectorMathTests.cs
Normal file
@@ -0,0 +1,50 @@
|
||||
using ZB.MOM.WW.CBDD.Core.Indexing;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests;
|
||||
|
||||
public class VectorMathTests
|
||||
{
|
||||
[Fact]
|
||||
public void Distance_Should_Cover_All_Metrics()
|
||||
{
|
||||
float[] v1 = [1f, 2f];
|
||||
float[] v2 = [3f, 4f];
|
||||
|
||||
var cosineDistance = VectorMath.Distance(v1, v2, VectorMetric.Cosine);
|
||||
var l2Distance = VectorMath.Distance(v1, v2, VectorMetric.L2);
|
||||
var dotDistance = VectorMath.Distance(v1, v2, VectorMetric.DotProduct);
|
||||
|
||||
l2Distance.ShouldBe(8f);
|
||||
dotDistance.ShouldBe(-11f);
|
||||
|
||||
var expectedCosine = 1f - (11f / (MathF.Sqrt(5f) * 5f));
|
||||
MathF.Abs(cosineDistance - expectedCosine).ShouldBeLessThan(0.0001f);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void CosineSimilarity_Should_Return_Zero_For_ZeroMagnitude_Vector()
|
||||
{
|
||||
float[] v1 = [0f, 0f, 0f];
|
||||
float[] v2 = [1f, 2f, 3f];
|
||||
|
||||
VectorMath.CosineSimilarity(v1, v2).ShouldBe(0f);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void DotProduct_Should_Throw_For_Length_Mismatch()
|
||||
{
|
||||
float[] v1 = [1f, 2f];
|
||||
float[] v2 = [1f];
|
||||
|
||||
Should.Throw<ArgumentException>(() => VectorMath.DotProduct(v1, v2));
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void EuclideanDistanceSquared_Should_Throw_For_Length_Mismatch()
|
||||
{
|
||||
float[] v1 = [1f, 2f, 3f];
|
||||
float[] v2 = [1f, 2f];
|
||||
|
||||
Should.Throw<ArgumentException>(() => VectorMath.EuclideanDistanceSquared(v1, v2));
|
||||
}
|
||||
}
|
||||
31
tests/CBDD.Tests/VectorSearchTests.cs
Executable file
31
tests/CBDD.Tests/VectorSearchTests.cs
Executable file
@@ -0,0 +1,31 @@
|
||||
using ZB.MOM.WW.CBDD.Bson;
|
||||
using ZB.MOM.WW.CBDD.Core;
|
||||
using ZB.MOM.WW.CBDD.Core.Indexing;
|
||||
using ZB.MOM.WW.CBDD.Shared;
|
||||
using Xunit;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests;
|
||||
|
||||
public class VectorSearchTests
|
||||
{
|
||||
[Fact]
|
||||
public void Test_VectorSearch_Basic()
|
||||
{
|
||||
string dbPath = "vector_test.db";
|
||||
if (File.Exists(dbPath)) File.Delete(dbPath);
|
||||
|
||||
using (var db = new Shared.TestDbContext(dbPath))
|
||||
{
|
||||
db.VectorItems.Insert(new VectorEntity { Title = "Near", Embedding = [1.0f, 1.0f, 1.0f] });
|
||||
db.VectorItems.Insert(new VectorEntity { Title = "Far", Embedding = [10.0f, 10.0f, 10.0f] });
|
||||
|
||||
var query = new[] { 0.9f, 0.9f, 0.9f };
|
||||
var results = db.VectorItems.AsQueryable().Where(x => x.Embedding.VectorSearch(query, 1)).ToList();
|
||||
|
||||
results.Count().ShouldBe(1);
|
||||
results[0].Title.ShouldBe("Near");
|
||||
}
|
||||
|
||||
File.Delete(dbPath);
|
||||
}
|
||||
}
|
||||
47
tests/CBDD.Tests/VisibilityTests.cs
Executable file
47
tests/CBDD.Tests/VisibilityTests.cs
Executable file
@@ -0,0 +1,47 @@
|
||||
using ZB.MOM.WW.CBDD.Bson;
|
||||
using ZB.MOM.WW.CBDD.Core.Collections;
|
||||
using Xunit;
|
||||
using System.Linq;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests;
|
||||
|
||||
public class VisibilityTests
|
||||
{
|
||||
public class VisibilityEntity
|
||||
{
|
||||
// Should be included
|
||||
public int NormalProp { get; set; }
|
||||
|
||||
// Should be included (serialization usually writes it)
|
||||
public int PrivateSetProp { get; private set; }
|
||||
|
||||
// Should be included
|
||||
public int InitProp { get; init; }
|
||||
|
||||
// Fields - typically included in BSON if public, but reflection need GetFields
|
||||
public string PublicField = string.Empty;
|
||||
|
||||
// Should NOT be included
|
||||
private int _privateField;
|
||||
|
||||
// Helper to set private
|
||||
public void SetPrivate(int val) => _privateField = val;
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void GenerateSchema_VisibilityChecks()
|
||||
{
|
||||
var schema = BsonSchemaGenerator.FromType<VisibilityEntity>();
|
||||
|
||||
schema.Fields.ShouldContain(f => f.Name == "normalprop");
|
||||
schema.Fields.ShouldContain(f => f.Name == "privatesetprop");
|
||||
schema.Fields.ShouldContain(f => f.Name == "initprop");
|
||||
|
||||
// Verify assumption about fields
|
||||
// Current implementation uses GetProperties, so publicfield might be missing.
|
||||
// We will assert current status and then fix if requested/failed.
|
||||
schema.Fields.ShouldContain(f => f.Name == "publicfield"); // This will likely fail currently
|
||||
|
||||
schema.Fields.ShouldNotContain(f => f.Name == "_privatefield");
|
||||
}
|
||||
}
|
||||
167
tests/CBDD.Tests/WalIndexTests.cs
Executable file
167
tests/CBDD.Tests/WalIndexTests.cs
Executable file
@@ -0,0 +1,167 @@
|
||||
using ZB.MOM.WW.CBDD.Bson;
|
||||
using ZB.MOM.WW.CBDD.Core;
|
||||
using ZB.MOM.WW.CBDD.Core.Collections;
|
||||
using ZB.MOM.WW.CBDD.Core.Indexing;
|
||||
using ZB.MOM.WW.CBDD.Core.Storage;
|
||||
using ZB.MOM.WW.CBDD.Core.Transactions;
|
||||
using ZB.MOM.WW.CBDD.Shared;
|
||||
using ZB.MOM.WW.CBDD.Shared.TestDbContext_TestDbContext_Mappers;
|
||||
using System.Buffers;
|
||||
using Xunit;
|
||||
|
||||
namespace ZB.MOM.WW.CBDD.Tests;
|
||||
|
||||
public class WalIndexTests : IDisposable
|
||||
{
|
||||
private readonly string _dbPath;
|
||||
private readonly string _walPath;
|
||||
private readonly Shared.TestDbContext _db;
|
||||
private readonly ITestOutputHelper _output;
|
||||
|
||||
public WalIndexTests(ITestOutputHelper output)
|
||||
{
|
||||
_output = output;
|
||||
_dbPath = Path.Combine(Path.GetTempPath(), $"test_wal_index_{Guid.NewGuid()}.db");
|
||||
// WAL defaults to .wal next to db
|
||||
_walPath = Path.ChangeExtension(_dbPath, ".wal");
|
||||
|
||||
_db = new Shared.TestDbContext(_dbPath);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void IndexWritesAreLoggedToWal()
|
||||
{
|
||||
// 2. Start a transaction
|
||||
using var txn = _db.BeginTransaction();
|
||||
_output.WriteLine($"Started Transaction: {txn.TransactionId}");
|
||||
|
||||
// 3. Insert a user
|
||||
var user = new User { Name = "Alice", Age = 30 };
|
||||
_db.Users.Insert(user);
|
||||
|
||||
// 4. Commit
|
||||
txn.Commit();
|
||||
_output.WriteLine("Committed Transaction");
|
||||
|
||||
// 5. Verify WAL
|
||||
// Dispose current storage to release file locks, BUT skip checkpoint/truncate
|
||||
_db.Dispose();
|
||||
|
||||
File.Exists(_walPath).ShouldBeTrue("WAL file should exist");
|
||||
|
||||
using var walReader = new WriteAheadLog(_walPath);
|
||||
var records = walReader.ReadAll();
|
||||
|
||||
_output.WriteLine($"Found {records.Count} WAL records");
|
||||
|
||||
// Filter for this transaction
|
||||
var txnRecords = records.Where(r => r.TransactionId == txn.TransactionId).ToList();
|
||||
|
||||
txnRecords.ShouldContain(r => r.Type == WalRecordType.Begin);
|
||||
txnRecords.ShouldContain(r => r.Type == WalRecordType.Commit);
|
||||
|
||||
var writeRecords = txnRecords.Where(r => r.Type == WalRecordType.Write).ToList();
|
||||
_output.WriteLine($"Found {writeRecords.Count} Write records for Txn {txn.TransactionId}");
|
||||
|
||||
// Analyze pages
|
||||
int indexPageCount = 0;
|
||||
int dataPageCount = 0;
|
||||
|
||||
foreach (var record in writeRecords)
|
||||
{
|
||||
var pageType = ParsePageType(record.AfterImage);
|
||||
_output.WriteLine($"Page {record.PageId}: Type={pageType}, Size={record.AfterImage?.Length}");
|
||||
|
||||
if (pageType == PageType.Index) indexPageCount++;
|
||||
else if (pageType == PageType.Data) dataPageCount++;
|
||||
}
|
||||
|
||||
(indexPageCount > 0).ShouldBeTrue($"Expected at least 1 Index page write, found {indexPageCount}");
|
||||
(dataPageCount > 0).ShouldBeTrue($"Expected at least 1 Data page write, found {dataPageCount}");
|
||||
}
|
||||
|
||||
private PageType ParsePageType(byte[]? pageData)
|
||||
{
|
||||
if (pageData == null || pageData.Length < 32) return (PageType)0;
|
||||
// PageType is at offset 4 (1 byte)
|
||||
return (PageType)pageData[4]; // Casting byte to PageType
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Compact_ShouldLeaveWalEmpty_AfterOfflineRun()
|
||||
{
|
||||
for (var i = 0; i < 100; i++)
|
||||
{
|
||||
_db.Users.Insert(new User { Name = $"wal-compact-{i:D3}", Age = i % 30 });
|
||||
}
|
||||
|
||||
_db.SaveChanges();
|
||||
_db.Storage.GetWalSize().ShouldBeGreaterThan(0);
|
||||
|
||||
var stats = _db.Compact(new CompactionOptions
|
||||
{
|
||||
EnableTailTruncation = true,
|
||||
NormalizeFreeList = true,
|
||||
DefragmentSlottedPages = true
|
||||
});
|
||||
|
||||
stats.OnlineMode.ShouldBeFalse();
|
||||
_db.Storage.GetWalSize().ShouldBe(0);
|
||||
new FileInfo(_walPath).Length.ShouldBe(0);
|
||||
}
|
||||
|
||||
[Fact]
|
||||
public void Recover_WithCommittedWal_ThenCompact_ShouldPreserveData()
|
||||
{
|
||||
var dbPath = Path.Combine(Path.GetTempPath(), $"test_wal_recover_compact_{Guid.NewGuid():N}.db");
|
||||
var walPath = Path.ChangeExtension(dbPath, ".wal");
|
||||
var markerPath = $"{dbPath}.compact.state";
|
||||
var expectedIds = new List<ObjectId>();
|
||||
|
||||
try
|
||||
{
|
||||
using (var writer = new Shared.TestDbContext(dbPath))
|
||||
{
|
||||
for (var i = 0; i < 48; i++)
|
||||
{
|
||||
expectedIds.Add(writer.Users.Insert(new User { Name = $"recover-{i:D3}", Age = i % 10 }));
|
||||
}
|
||||
|
||||
writer.SaveChanges();
|
||||
writer.Storage.GetWalSize().ShouldBeGreaterThan(0);
|
||||
}
|
||||
|
||||
new FileInfo(walPath).Length.ShouldBeGreaterThan(0);
|
||||
|
||||
using (var recovered = new Shared.TestDbContext(dbPath))
|
||||
{
|
||||
recovered.Users.Count().ShouldBe(expectedIds.Count);
|
||||
recovered.Compact();
|
||||
recovered.Storage.GetWalSize().ShouldBe(0);
|
||||
|
||||
foreach (var id in expectedIds)
|
||||
{
|
||||
recovered.Users.FindById(id).ShouldNotBeNull();
|
||||
}
|
||||
}
|
||||
}
|
||||
finally
|
||||
{
|
||||
if (File.Exists(dbPath)) File.Delete(dbPath);
|
||||
if (File.Exists(walPath)) File.Delete(walPath);
|
||||
if (File.Exists(markerPath)) File.Delete(markerPath);
|
||||
}
|
||||
}
|
||||
|
||||
public void Dispose()
|
||||
{
|
||||
try
|
||||
{
|
||||
_db?.Dispose(); // Safe to call multiple times
|
||||
}
|
||||
catch { }
|
||||
|
||||
try { if (File.Exists(_dbPath)) File.Delete(_dbPath); } catch { }
|
||||
try { if (File.Exists(_walPath)) File.Delete(_walPath); } catch { }
|
||||
}
|
||||
}
|
||||
39
tests/CBDD.Tests/ZB.MOM.WW.CBDD.Tests.csproj
Executable file
39
tests/CBDD.Tests/ZB.MOM.WW.CBDD.Tests.csproj
Executable file
@@ -0,0 +1,39 @@
|
||||
<Project Sdk="Microsoft.NET.Sdk">
|
||||
|
||||
<PropertyGroup>
|
||||
<TargetFramework>net10.0</TargetFramework>
|
||||
<AssemblyName>ZB.MOM.WW.CBDD.Tests</AssemblyName>
|
||||
<RootNamespace>ZB.MOM.WW.CBDD.Tests</RootNamespace>
|
||||
<ImplicitUsings>enable</ImplicitUsings>
|
||||
<Nullable>enable</Nullable>
|
||||
<TreatWarningsAsErrors>true</TreatWarningsAsErrors>
|
||||
<IsPackable>false</IsPackable>
|
||||
</PropertyGroup>
|
||||
|
||||
<ItemGroup>
|
||||
<PackageReference Include="coverlet.collector" Version="6.0.4" />
|
||||
<PackageReference Include="coverlet.msbuild" Version="6.0.4">
|
||||
<PrivateAssets>all</PrivateAssets>
|
||||
<IncludeAssets>runtime; build; native; contentfiles; analyzers; buildtransitive</IncludeAssets>
|
||||
</PackageReference>
|
||||
<PackageReference Include="Microsoft.NET.Test.Sdk" Version="17.14.1" />
|
||||
<PackageReference Include="NSubstitute" Version="5.3.0" />
|
||||
<PackageReference Include="Shouldly" Version="4.3.0" />
|
||||
<PackageReference Include="xunit.runner.visualstudio" Version="3.1.4">
|
||||
<IncludeAssets>runtime; build; native; contentfiles; analyzers; buildtransitive</IncludeAssets>
|
||||
<PrivateAssets>all</PrivateAssets>
|
||||
</PackageReference>
|
||||
<PackageReference Include="xunit.v3" Version="3.2.2" />
|
||||
</ItemGroup>
|
||||
|
||||
<ItemGroup>
|
||||
<Using Include="Xunit" />
|
||||
</ItemGroup>
|
||||
|
||||
<ItemGroup>
|
||||
<ProjectReference Include="..\..\src\CBDD.SourceGenerators\ZB.MOM.WW.CBDD.SourceGenerators.csproj" OutputItemType="Analyzer" ReferenceOutputAssembly="false" />
|
||||
<ProjectReference Include="..\..\src\CBDD.Bson\ZB.MOM.WW.CBDD.Bson.csproj" />
|
||||
<ProjectReference Include="..\..\src\CBDD.Core\ZB.MOM.WW.CBDD.Core.csproj" />
|
||||
</ItemGroup>
|
||||
|
||||
</Project>
|
||||
23828
tests/CBDD.Tests/coverage.cobertura.xml
Normal file
23828
tests/CBDD.Tests/coverage.cobertura.xml
Normal file
File diff suppressed because it is too large
Load Diff
16
tests/CBDD.Tests/coverage.runsettings
Normal file
16
tests/CBDD.Tests/coverage.runsettings
Normal file
@@ -0,0 +1,16 @@
|
||||
<?xml version="1.0" encoding="utf-8"?>
|
||||
<RunSettings>
|
||||
<DataCollectionRunSettings>
|
||||
<DataCollectors>
|
||||
<DataCollector friendlyName="XPlat code coverage">
|
||||
<Configuration>
|
||||
<Format>cobertura</Format>
|
||||
<Include>[ZB.MOM.WW.CBDD.Core*]*,[ZB.MOM.WW.CBDD.Bson*]*</Include>
|
||||
<Exclude>[*.Tests]*</Exclude>
|
||||
<ExcludeByAttribute>GeneratedCodeAttribute,CompilerGeneratedAttribute,ExcludeFromCodeCoverageAttribute</ExcludeByAttribute>
|
||||
<SkipAutoProps>true</SkipAutoProps>
|
||||
</Configuration>
|
||||
</DataCollector>
|
||||
</DataCollectors>
|
||||
</DataCollectionRunSettings>
|
||||
</RunSettings>
|
||||
Reference in New Issue
Block a user