mirror of
https://github.com/seaweedfs/seaweedfs.git
synced 2026-05-14 05:41:29 +00:00
* fix(s3/shell): include EC volumes in bucket size metrics and collection.list S3 bucket size metrics exported to Prometheus (and fed through stats.UpdateBucketSizeMetrics) are computed by collectCollectionInfoFromTopology, which only walked diskInfo.VolumeInfos. As soon as a volume was encoded to EC it dropped out of every aggregate, so Grafana showed bucket sizes shrinking while physical disk usage kept climbing. The shell helper collectCollectionInfo — used by collection.list and s3.bucket.quota.enforce — had the same gap, with the EC branch left as a commented-out TODO. Fold EC shards into both paths using the same approach the admin dashboard already uses (PR #9093): - PhysicalSize / Size sum across shard holders: EC shards are node-local (not replicas), so per-node TotalSize() and MinusParityShards().TotalSize() sum to the whole-volume physical and logical sizes respectively. - FileCount is deduped via max across reporters (every shard holder reports the same .ecx count; a slow node with a not-yet-loaded .ecx reports 0 and must not pin the aggregate). - DeleteCount is summed (each delete tombstones exactly one node's .ecj). - VolumeCount increments once per unique EC volume id. Adds regression tests covering pure-EC, mixed regular+EC, and the slow-reporter FileCount dedupe case. Refs #9086 * Address PR review feedback: EC size helpers, composite key, VolumeCount dedupe - Add EcShardsTotalSize / EcShardsDataSize helpers in the erasure_coding package that walk the shard bitmap directly instead of materializing a ShardsInfo and copying it via MinusParityShards(). Keeps the DataShardsCount dependency encapsulated in one place and avoids the per-shard allocation/copy overhead in the metrics hot path. - Switch shell collectCollectionInfo ecVolumes map to a composite {collection, volumeId} key, matching the bucket_size_metrics collector and defending against any cross-collection volume id aliasing. - Dedupe VolumeCount in shell addToCollection by volume id so regular volumes aren't counted once per replica presence. Aligns the shell's collection.list output with the S3 metrics collector and the EC branch, all of which now report logical volume counts. - Add unit tests for the new helpers and for the regular-volume VolumeCount dedupe. * Parameterize EcShardsDataSize with dataShards for custom EC ratios Add a dataShards parameter to EcShardsDataSize so forks with per-volume ratio metadata (e.g. the enterprise data_shards field carried on an extended VolumeEcShardInformationMessage) can pass the configured value and get accurate logical sizes under custom EC policies like 6+3 or 16+6. Passing 0 or a negative value falls back to the upstream DataShardsCount default, which is correct for the fixed 10+4 layout — so OSS callers in s3api and shell pass 0 and keep their current behavior. Added table cases covering the custom 6+3 and 16+6 paths so the parameterization is pinned by tests.
207 lines
5.8 KiB
Go
207 lines
5.8 KiB
Go
package shell
|
|
|
|
import (
|
|
"testing"
|
|
|
|
"github.com/seaweedfs/seaweedfs/weed/pb/master_pb"
|
|
)
|
|
|
|
// TestCollectCollectionInfoEC verifies that EC-encoded volumes contribute to
|
|
// collection.list totals and s3.bucket.quota.enforce sums. Before this fix
|
|
// the EC branch was a no-op, so encoded volumes silently dropped out of
|
|
// collection size accounting.
|
|
func TestCollectCollectionInfoEC(t *testing.T) {
|
|
// One 10+4 EC volume split across two nodes: 14 shards * 1000 bytes.
|
|
// Data shards 0..9 live on (nodeA: 0..6, nodeB: 7..9), parity 10..13
|
|
// on nodeB. Every shard holder reports file_count=100; node-local
|
|
// delete counts are 2 + 3 = 5.
|
|
nodeA := &master_pb.DataNodeInfo{
|
|
DiskInfos: map[string]*master_pb.DiskInfo{
|
|
"disk1": {
|
|
EcShardInfos: []*master_pb.VolumeEcShardInformationMessage{
|
|
{
|
|
Id: 42,
|
|
Collection: "bucket-a",
|
|
EcIndexBits: (1 << 0) | (1 << 1) | (1 << 2) | (1 << 3) | (1 << 4) | (1 << 5) | (1 << 6),
|
|
ShardSizes: []int64{1000, 1000, 1000, 1000, 1000, 1000, 1000},
|
|
FileCount: 100,
|
|
DeleteCount: 2,
|
|
},
|
|
},
|
|
},
|
|
},
|
|
}
|
|
nodeB := &master_pb.DataNodeInfo{
|
|
DiskInfos: map[string]*master_pb.DiskInfo{
|
|
"disk1": {
|
|
EcShardInfos: []*master_pb.VolumeEcShardInformationMessage{
|
|
{
|
|
Id: 42,
|
|
Collection: "bucket-a",
|
|
EcIndexBits: (1 << 7) | (1 << 8) | (1 << 9) | (1 << 10) | (1 << 11) | (1 << 12) | (1 << 13),
|
|
ShardSizes: []int64{1000, 1000, 1000, 1000, 1000, 1000, 1000},
|
|
FileCount: 100,
|
|
DeleteCount: 3,
|
|
},
|
|
},
|
|
},
|
|
},
|
|
}
|
|
|
|
topo := &master_pb.TopologyInfo{
|
|
DataCenterInfos: []*master_pb.DataCenterInfo{
|
|
{
|
|
RackInfos: []*master_pb.RackInfo{
|
|
{
|
|
DataNodeInfos: []*master_pb.DataNodeInfo{nodeA, nodeB},
|
|
},
|
|
},
|
|
},
|
|
},
|
|
}
|
|
|
|
infos := make(map[string]*CollectionInfo)
|
|
collectCollectionInfo(topo, infos)
|
|
|
|
cif, ok := infos["bucket-a"]
|
|
if !ok {
|
|
t.Fatalf("expected collection bucket-a in infos, got: %v", infos)
|
|
}
|
|
// 10 data shards * 1000 bytes.
|
|
if cif.Size != 10000 {
|
|
t.Errorf("Size: got %.0f, want 10000", cif.Size)
|
|
}
|
|
if cif.FileCount != 100 {
|
|
t.Errorf("FileCount: got %.0f, want 100 (max across reporters)", cif.FileCount)
|
|
}
|
|
if cif.DeleteCount != 5 {
|
|
t.Errorf("DeleteCount: got %.0f, want 5 (sum across reporters)", cif.DeleteCount)
|
|
}
|
|
if cif.VolumeCount != 1 {
|
|
t.Errorf("VolumeCount: got %d, want 1", cif.VolumeCount)
|
|
}
|
|
}
|
|
|
|
// TestCollectCollectionInfoMixed verifies that a collection with both a
|
|
// regular volume and an EC volume sums both branches without either clobbering
|
|
// the other, matching the state mid-EC-conversion.
|
|
func TestCollectCollectionInfoMixed(t *testing.T) {
|
|
node := &master_pb.DataNodeInfo{
|
|
DiskInfos: map[string]*master_pb.DiskInfo{
|
|
"disk1": {
|
|
VolumeInfos: []*master_pb.VolumeInformationMessage{
|
|
{
|
|
Id: 1,
|
|
Collection: "bucket-mix",
|
|
Size: 5000,
|
|
FileCount: 50,
|
|
DeleteCount: 1,
|
|
},
|
|
},
|
|
EcShardInfos: []*master_pb.VolumeEcShardInformationMessage{
|
|
{
|
|
Id: 2,
|
|
Collection: "bucket-mix",
|
|
EcIndexBits: (1 << 0) | (1 << 1) | (1 << 10), // 2 data + 1 parity
|
|
ShardSizes: []int64{3000, 3000, 3000},
|
|
FileCount: 80,
|
|
DeleteCount: 4,
|
|
},
|
|
},
|
|
},
|
|
},
|
|
}
|
|
|
|
topo := &master_pb.TopologyInfo{
|
|
DataCenterInfos: []*master_pb.DataCenterInfo{
|
|
{
|
|
RackInfos: []*master_pb.RackInfo{
|
|
{
|
|
DataNodeInfos: []*master_pb.DataNodeInfo{node},
|
|
},
|
|
},
|
|
},
|
|
},
|
|
}
|
|
|
|
infos := make(map[string]*CollectionInfo)
|
|
collectCollectionInfo(topo, infos)
|
|
|
|
cif, ok := infos["bucket-mix"]
|
|
if !ok {
|
|
t.Fatalf("expected collection bucket-mix in infos, got: %v", infos)
|
|
}
|
|
// Regular: 5000 (replicaCount=1). EC data shards: 6000.
|
|
if cif.Size != 5000+6000 {
|
|
t.Errorf("Size: got %.0f, want 11000", cif.Size)
|
|
}
|
|
if cif.FileCount != 50+80 {
|
|
t.Errorf("FileCount: got %.0f, want 130", cif.FileCount)
|
|
}
|
|
if cif.DeleteCount != 1+4 {
|
|
t.Errorf("DeleteCount: got %.0f, want 5", cif.DeleteCount)
|
|
}
|
|
if cif.VolumeCount != 2 {
|
|
t.Errorf("VolumeCount: got %d, want 2", cif.VolumeCount)
|
|
}
|
|
}
|
|
|
|
// TestCollectCollectionInfoRegularVolumeDedupesReplicas verifies that a
|
|
// regular volume replicated across three nodes is counted once in
|
|
// VolumeCount (logical, like the S3 metrics path and the EC branch) rather
|
|
// than three times (one per replica presence).
|
|
func TestCollectCollectionInfoRegularVolumeDedupesReplicas(t *testing.T) {
|
|
// 001 = one-copy replication placement byte; three copies of volume id=7
|
|
// on three distinct nodes. Per-replica Size/FileCount are divided by
|
|
// copyCount so summing yields the whole-volume totals.
|
|
makeNode := func() *master_pb.DataNodeInfo {
|
|
return &master_pb.DataNodeInfo{
|
|
DiskInfos: map[string]*master_pb.DiskInfo{
|
|
"disk1": {
|
|
VolumeInfos: []*master_pb.VolumeInformationMessage{
|
|
{
|
|
Id: 7,
|
|
Collection: "bucket-rep",
|
|
Size: 3000,
|
|
FileCount: 30,
|
|
DeleteCount: 3,
|
|
DeletedByteCount: 300,
|
|
ReplicaPlacement: 002, // 0x02 = 3 total copies
|
|
},
|
|
},
|
|
},
|
|
},
|
|
}
|
|
}
|
|
|
|
topo := &master_pb.TopologyInfo{
|
|
DataCenterInfos: []*master_pb.DataCenterInfo{
|
|
{
|
|
RackInfos: []*master_pb.RackInfo{
|
|
{
|
|
DataNodeInfos: []*master_pb.DataNodeInfo{makeNode(), makeNode(), makeNode()},
|
|
},
|
|
},
|
|
},
|
|
},
|
|
}
|
|
|
|
infos := make(map[string]*CollectionInfo)
|
|
collectCollectionInfo(topo, infos)
|
|
|
|
cif, ok := infos["bucket-rep"]
|
|
if !ok {
|
|
t.Fatalf("expected collection bucket-rep in infos, got: %v", infos)
|
|
}
|
|
if cif.VolumeCount != 1 {
|
|
t.Errorf("VolumeCount: got %d, want 1 (deduped by volume id)", cif.VolumeCount)
|
|
}
|
|
// Per-replica values are Size/3, summed across 3 replicas gives 3000.
|
|
if cif.Size != 3000 {
|
|
t.Errorf("Size: got %.0f, want 3000", cif.Size)
|
|
}
|
|
if cif.FileCount != 30 {
|
|
t.Errorf("FileCount: got %.0f, want 30", cif.FileCount)
|
|
}
|
|
}
|