Implement in-process multi-dataset sync isolation across core, network, persistence, and tests
All checks were successful
NuGet Package Publish / nuget (push) Successful in 1m14s

This commit is contained in:
Joseph Doherty
2026-02-22 11:58:34 -05:00
parent c06b56172a
commit 8e97061ab8
60 changed files with 4519 additions and 559 deletions

View File

@@ -0,0 +1,440 @@
using System.Diagnostics;
using System.Text;
using BenchmarkDotNet.Attributes;
using SurrealDb.Net.Models;
using SurrealDb.Net.Models.Response;
using ZB.MOM.WW.CBDDC.Persistence.Surreal;
namespace ZB.MOM.WW.CDBBC.E2E.Benchmark.Tests;
[MemoryDiagnoser]
[SimpleJob(launchCount: 1, warmupCount: 1, iterationCount: 3)]
public class SurrealLogStorageBenchmarks
{
private const int LogRecordCount = 100_000;
private const int InsertBatchSize = 500;
private const string LogTable = "benchmark_log_entry";
private const string LogKvTable = "benchmark_log_kv";
private static readonly string[] LoggerNames =
[
"Api.RequestHandler",
"Api.AuthController",
"Api.OrderController",
"Api.InventoryController",
"Api.CustomerController",
"Workers.OutboxPublisher",
"Workers.NotificationDispatcher",
"Infrastructure.SqlRepository",
"Infrastructure.CacheService",
"Infrastructure.HttpClient",
"Domain.OrderService",
"Domain.PricingService"
];
private static readonly string[] TenantIds =
[
"tenant-01",
"tenant-02",
"tenant-03",
"tenant-04",
"tenant-05",
"tenant-06",
"tenant-07",
"tenant-08"
];
private CBDDCSurrealEmbeddedClient _surrealClient = null!;
private string _databasePath = string.Empty;
private string _workDir = string.Empty;
private DateTime _seedBaseUtc;
private DateTime _queryRangeStartUtc;
private DateTime _queryRangeEndUtc;
private string _contextIdQueryValue = string.Empty;
private string _loggerQueryValue = string.Empty;
private string _contextStringKeyQueryValue = string.Empty;
private string _contextStringValueQueryValue = string.Empty;
private string _contextNumericKeyQueryValue = string.Empty;
private int _contextNumericValueQueryValue;
[GlobalSetup]
public async Task GlobalSetupAsync()
{
_workDir = Path.Combine(Path.GetTempPath(), $"cbddc-serilog-benchmark-{Guid.NewGuid():N}");
Directory.CreateDirectory(_workDir);
_databasePath = Path.Combine(_workDir, "serilog.rocksdb");
_surrealClient = new CBDDCSurrealEmbeddedClient(
new CBDDCSurrealEmbeddedOptions
{
Endpoint = "rocksdb://local",
DatabasePath = _databasePath,
Namespace = "cbddc_benchmark",
Database = $"serilog_{Guid.NewGuid():N}",
Cdc = new CBDDCSurrealCdcOptions { Enabled = false }
});
await _surrealClient.InitializeAsync();
await DefineSchemaAndIndexesAsync();
_seedBaseUtc = DateTime.UtcNow.AddDays(-1);
_contextIdQueryValue = BuildContextId(LogRecordCount / 2);
_loggerQueryValue = LoggerNames[3];
_contextStringKeyQueryValue = "tenantId";
_contextStringValueQueryValue = TenantIds[5];
_contextNumericKeyQueryValue = "statusCode";
_contextNumericValueQueryValue = 500;
_queryRangeStartUtc = _seedBaseUtc.AddMinutes(6);
_queryRangeEndUtc = _queryRangeStartUtc.AddSeconds(30);
var seedTimer = Stopwatch.StartNew();
await InsertLogRecordsAsync();
seedTimer.Stop();
long sizeBytes = CalculatePathSizeBytes(_databasePath);
Console.WriteLine(
$"Seeded {LogRecordCount:N0} records in {seedTimer.Elapsed.TotalSeconds:F2}s. " +
$"RocksDB size: {sizeBytes / (1024d * 1024d):F2} MiB ({sizeBytes:N0} bytes). Path: {_databasePath}");
}
[GlobalCleanup]
public Task GlobalCleanupAsync()
{
// Avoid explicit Surreal embedded disposal in benchmark child processes due known native callback race.
return Task.CompletedTask;
}
[Benchmark(Description = "Query by contextId (latest 200 rows)")]
public async Task QueryByContextIdAsync()
{
await ExecuteQueryAsync(
$"""
SELECT * FROM {LogTable}
WHERE contextId = $contextId
ORDER BY timestamp DESC
LIMIT 200;
""",
new Dictionary<string, object?> { ["contextId"] = _contextIdQueryValue });
}
[Benchmark(Description = "Query by loggerName + timestamp range (latest 200 rows)")]
public async Task QueryByLoggerAndTimestampAsync()
{
await ExecuteQueryAsync(
$"""
SELECT * FROM {LogTable}
WHERE loggerName = $loggerName
AND timestamp >= $fromTs
AND timestamp <= $toTs
ORDER BY timestamp DESC
LIMIT 200;
""",
new Dictionary<string, object?>
{
["loggerName"] = _loggerQueryValue,
["fromTs"] = _queryRangeStartUtc,
["toTs"] = _queryRangeEndUtc
});
}
[Benchmark(Description = "Query by loggerName + timestamp + arbitrary context string key/value")]
public async Task QueryByLoggerTimestampAndContextKeyAsync()
{
await ExecuteQueryAsync(
$"""
LET $logIds = (
SELECT VALUE logId FROM {LogKvTable}
WHERE loggerName = $loggerName
AND key = $contextKey
AND valueStr = $contextValueStr
AND timestamp >= $fromTs
AND timestamp <= $toTs
ORDER BY timestamp DESC
LIMIT 200
);
SELECT * FROM {LogTable}
WHERE id INSIDE $logIds
ORDER BY timestamp DESC
LIMIT 200;
""",
new Dictionary<string, object?>
{
["loggerName"] = _loggerQueryValue,
["fromTs"] = _queryRangeStartUtc,
["toTs"] = _queryRangeEndUtc,
["contextKey"] = _contextStringKeyQueryValue,
["contextValueStr"] = _contextStringValueQueryValue
});
}
[Benchmark(Description = "Query by loggerName + timestamp + arbitrary context number key/value")]
public async Task QueryByLoggerTimestampAndNumericContextKeyAsync()
{
await ExecuteQueryAsync(
$"""
LET $logIds = (
SELECT VALUE logId FROM {LogKvTable}
WHERE loggerName = $loggerName
AND key = $contextKey
AND valueNum = $contextValueNum
AND timestamp >= $fromTs
AND timestamp <= $toTs
ORDER BY timestamp DESC
LIMIT 200
);
SELECT * FROM {LogTable}
WHERE id INSIDE $logIds
ORDER BY timestamp DESC
LIMIT 200;
""",
new Dictionary<string, object?>
{
["loggerName"] = _loggerQueryValue,
["fromTs"] = _queryRangeStartUtc,
["toTs"] = _queryRangeEndUtc,
["contextKey"] = _contextNumericKeyQueryValue,
["contextValueNum"] = _contextNumericValueQueryValue
});
}
[Benchmark(Description = "RocksDB size (bytes)")]
public long GetDatabaseFileSizeBytes()
{
return CalculatePathSizeBytes(_databasePath);
}
private async Task DefineSchemaAndIndexesAsync()
{
string schemaSql =
$"""
DEFINE TABLE OVERWRITE {LogTable} SCHEMAFULL;
DEFINE FIELD OVERWRITE timestamp ON TABLE {LogTable} TYPE datetime;
DEFINE FIELD OVERWRITE level ON TABLE {LogTable} TYPE string;
DEFINE FIELD OVERWRITE loggerName ON TABLE {LogTable} TYPE option<string>;
DEFINE FIELD OVERWRITE contextId ON TABLE {LogTable} TYPE option<string>;
DEFINE FIELD OVERWRITE messageTemplate ON TABLE {LogTable} TYPE string;
DEFINE FIELD OVERWRITE renderedMessage ON TABLE {LogTable} TYPE option<string>;
DEFINE FIELD OVERWRITE exception ON TABLE {LogTable} TYPE option<string>;
DEFINE FIELD OVERWRITE contextValues ON TABLE {LogTable} TYPE object FLEXIBLE;
DEFINE INDEX OVERWRITE idx_log_contextid_ts
ON TABLE {LogTable} COLUMNS contextId, timestamp;
DEFINE INDEX OVERWRITE idx_log_logger_ts
ON TABLE {LogTable} COLUMNS loggerName, timestamp;
DEFINE TABLE OVERWRITE {LogKvTable} SCHEMAFULL;
DEFINE FIELD OVERWRITE logId ON TABLE {LogKvTable} TYPE record<{LogTable}>;
DEFINE FIELD OVERWRITE loggerName ON TABLE {LogKvTable} TYPE string;
DEFINE FIELD OVERWRITE timestamp ON TABLE {LogKvTable} TYPE datetime;
DEFINE FIELD OVERWRITE key ON TABLE {LogKvTable} TYPE string;
DEFINE FIELD OVERWRITE valueStr ON TABLE {LogKvTable} TYPE option<string>;
DEFINE FIELD OVERWRITE valueNum ON TABLE {LogKvTable} TYPE option<number>;
DEFINE FIELD OVERWRITE valueBool ON TABLE {LogKvTable} TYPE option<bool>;
DEFINE INDEX OVERWRITE idx_logkv_logger_key_str_ts
ON TABLE {LogKvTable} COLUMNS loggerName, key, valueStr, timestamp;
DEFINE INDEX OVERWRITE idx_logkv_logger_key_num_ts
ON TABLE {LogKvTable} COLUMNS loggerName, key, valueNum, timestamp;
DEFINE INDEX OVERWRITE idx_logkv_logger_key_bool_ts
ON TABLE {LogKvTable} COLUMNS loggerName, key, valueBool, timestamp;
DEFINE INDEX OVERWRITE idx_logkv_logid
ON TABLE {LogKvTable} COLUMNS logId;
""";
var response = await _surrealClient.RawQueryAsync(schemaSql);
EnsureSuccessfulResponse(response, "Schema definition");
}
private async Task InsertLogRecordsAsync()
{
for (var batchStart = 0; batchStart < LogRecordCount; batchStart += InsertBatchSize)
{
int batchCount = Math.Min(InsertBatchSize, LogRecordCount - batchStart);
var sqlBuilder = new StringBuilder();
sqlBuilder.AppendLine("BEGIN TRANSACTION;");
var parameters = new Dictionary<string, object?>(batchCount * 2);
for (var offset = 0; offset < batchCount; offset++)
{
int sequence = batchStart + offset;
string idParameterName = $"id{offset}";
string recordParameterName = $"record{offset}";
string logId = $"log-{sequence:D8}";
RecordId logRecordId = RecordId.From(LogTable, logId);
IReadOnlyDictionary<string, object?> logRecord = CreateLogRecord(sequence);
parameters[idParameterName] = logRecordId;
parameters[recordParameterName] = logRecord;
sqlBuilder.Append("UPSERT $")
.Append(idParameterName)
.Append(" CONTENT $")
.Append(recordParameterName)
.AppendLine(";");
int kvOrdinal = 0;
foreach (IReadOnlyDictionary<string, object?> kvRow in CreateKvRows(logId, logRecordId, logRecord))
{
string kvIdParameterName = $"kvid{offset}_{kvOrdinal}";
string kvRecordParameterName = $"kvrecord{offset}_{kvOrdinal}";
parameters[kvIdParameterName] = RecordId.From(LogKvTable, $"{logId}-{kvOrdinal:D2}");
parameters[kvRecordParameterName] = kvRow;
sqlBuilder.Append("UPSERT $")
.Append(kvIdParameterName)
.Append(" CONTENT $")
.Append(kvRecordParameterName)
.AppendLine(";");
kvOrdinal++;
}
}
sqlBuilder.AppendLine("COMMIT TRANSACTION;");
var response = await _surrealClient.RawQueryAsync(sqlBuilder.ToString(), parameters);
EnsureSuccessfulResponse(response, $"Insert batch starting at row {batchStart}");
}
}
private IReadOnlyDictionary<string, object?> CreateLogRecord(int sequence)
{
DateTime timestamp = _seedBaseUtc.AddMilliseconds(sequence * 10L);
string loggerName = LoggerNames[sequence % LoggerNames.Length];
string tenantId = TenantIds[sequence % TenantIds.Length];
bool isBackground = sequence % 7 == 0;
string? exception = sequence % 2_500 == 0
? "System.InvalidOperationException: simulated benchmark exception."
: null;
var record = new Dictionary<string, object?>
{
["timestamp"] = timestamp,
["level"] = ResolveLogLevel(sequence),
["loggerName"] = loggerName,
["contextId"] = BuildContextId(sequence),
["messageTemplate"] = "Processed request {RequestId} for {Route}",
["renderedMessage"] = $"Processed request req-{sequence:D8} for /api/items/{sequence % 250}",
["contextValues"] = new Dictionary<string, object?>
{
["tenantId"] = tenantId,
["requestId"] = $"req-{sequence:D8}",
["route"] = $"/api/items/{sequence % 250}",
["statusCode"] = sequence % 20 == 0 ? 500 : 200,
["elapsedMs"] = 5 + (sequence % 200),
["nodeId"] = $"node-{sequence % 8:D2}",
["isBackground"] = isBackground
}
};
if (!string.IsNullOrEmpty(exception))
record["exception"] = exception;
return record;
}
private static IEnumerable<IReadOnlyDictionary<string, object?>> CreateKvRows(
string logId,
RecordId logRecordId,
IReadOnlyDictionary<string, object?> logRecord)
{
if (!logRecord.TryGetValue("loggerName", out object? loggerNameValue) || loggerNameValue is not string loggerName)
yield break;
if (!logRecord.TryGetValue("timestamp", out object? timestampValue) || timestampValue is not DateTime timestamp)
yield break;
if (!logRecord.TryGetValue("contextValues", out object? contextValuesObject) ||
contextValuesObject is not IReadOnlyDictionary<string, object?> contextValues)
yield break;
foreach ((string key, object? value) in contextValues)
{
if (value == null) continue;
var row = new Dictionary<string, object?>
{
["logId"] = logRecordId,
["loggerName"] = loggerName,
["timestamp"] = timestamp,
["key"] = key
};
switch (value)
{
case string stringValue:
row["valueStr"] = stringValue;
break;
case bool boolValue:
row["valueBool"] = boolValue;
break;
case sbyte or byte or short or ushort or int or uint or long or ulong or float or double or decimal:
row["valueNum"] = Convert.ToDouble(value);
break;
default:
row["valueStr"] = value.ToString();
break;
}
yield return row;
}
}
private async Task ExecuteQueryAsync(string query, IReadOnlyDictionary<string, object?> parameters)
{
var response = await _surrealClient.RawQueryAsync(query, parameters);
EnsureSuccessfulResponse(response, "Query execution");
}
private static string BuildContextId(int sequence)
{
return $"ctx-{sequence / 10:D6}";
}
private static string ResolveLogLevel(int sequence)
{
if (sequence % 2_500 == 0) return "Error";
if (sequence % 500 == 0) return "Warning";
return "Information";
}
private static long CalculatePathSizeBytes(string path)
{
if (File.Exists(path)) return new FileInfo(path).Length;
if (!Directory.Exists(path)) return 0;
long size = 0;
foreach (string file in Directory.EnumerateFiles(path, "*", SearchOption.AllDirectories))
size += new FileInfo(file).Length;
return size;
}
private static void EnsureSuccessfulResponse(SurrealDbResponse response, string operation)
{
if (!response.HasErrors) return;
string errorSummary = string.Join(
" | ",
response.Errors.Take(3).Select((error, index) => DescribeError(error, index)));
throw new InvalidOperationException(
$"{operation} failed with SurrealDB errors. Details: {errorSummary}");
}
private static string DescribeError(object error, int index)
{
Type errorType = error.GetType();
string[] fieldsToExtract = ["Status", "Details", "Description", "Information", "Code"];
var extracted = new List<string>();
foreach (string field in fieldsToExtract)
{
object? value = errorType.GetProperty(field)?.GetValue(error);
if (value != null) extracted.Add($"{field}={value}");
}
if (extracted.Count == 0) return $"error[{index}] type={errorType.Name}";
return $"error[{index}] type={errorType.Name} {string.Join(", ", extracted)}";
}
}