W2.5.B — first authoring of the three Catalyst Blueprint wrapper charts
that fill bootstrap-kit slots 18 (seaweedfs), 19 (harbor) and 29 (vpa).
Each wraps an upstream chart as a Helm subchart and ships Catalyst-
curated overlay templates (NetworkPolicy + ServiceMonitor) gated behind
opt-in toggles, per docs/BLUEPRINT-AUTHORING.md §11 and
docs/INVIOLABLE-PRINCIPLES.md.
bp-seaweedfs (slot 18 — storage foundation)
- Wraps seaweedfs/seaweedfs 4.22.0; Chart name `bp-seaweedfs`.
- Catalyst defaults: 1 master + 3 volume + 1 filer + 2 s3 replicas.
- S3 API on 8333 — single S3 surface every consumer talks to per
docs/PLATFORM-TECH-STACK.md §3.5 (no per-app MinIO).
- Overlay templates: NetworkPolicy (cross-namespace S3 reachability,
cold-tier egress allowlist), ServiceMonitor (Capabilities-gated,
DEFAULT FALSE per §11.2).
- Default helm template kinds: ClusterRole, ClusterRoleBinding,
ConfigMap, Deployment, Secret, Service, ServiceAccount, StatefulSet.
bp-harbor (slot 19 — per-Sovereign OCI registry)
- Wraps goharbor/harbor 1.18.3 (appVersion 2.14.3); Chart name
`bp-harbor`.
- Catalyst defaults: blob backend = SeaweedFS S3 (regionendpoint
seaweedfs-s3.seaweedfs.svc:8333), metadata DB = bp-cnpg external
Postgres, ingress class `cilium`, expose.tls.enabled true (cert-
manager-issued Secret).
- Overlay templates: NetworkPolicy (CNPG/SeaweedFS/Keycloak egress),
ServiceMonitor (Capabilities-gated, DEFAULT FALSE).
- Trivy + SSO + pull-mirror are operator-flag opt-ins per per-
Sovereign overlay (default false; trivy/keycloak/cnpg deps land on
later slots).
- Default helm template kinds: ConfigMap, Deployment, Ingress,
PersistentVolumeClaim, Secret, Service, StatefulSet.
bp-vpa (slot 29 — vertical autoscaling)
- Wraps cowboysysop/vertical-pod-autoscaler 11.1.1 (appVersion
1.5.0); Chart name `bp-vpa`.
- Catalyst defaults: 1 replica each of recommender + updater +
admission-controller. Default mode `Off` (recommend only).
- Admission webhook self-signs via init Job (cluster-internal); per-
Sovereign overlay MAY swap to cert-manager.
- Overlay templates: NetworkPolicy (apiserver + metrics-server
egress, admission webhook ingress).
- Upstream metrics.serviceMonitor / metrics.prometheusRule defaulted
false per §11.2.
- Default helm template kinds: ClusterRole, ClusterRoleBinding,
ConfigMap, Deployment, Job, Pod, Secret, Service, ServiceAccount.
Lint + observability-toggle results
helm lint: 1 chart(s) linted, 0 chart(s) failed (each)
tests/observability-toggle.sh: PASS on all three (default render has
zero monitoring.coreos.com/v1 references; opt-in render produces a
ServiceMonitor; explicit-off render is clean).
Path isolation: only platform/seaweedfs/, platform/harbor/, and
platform/vpa/ — no HR slot files or other charts touched.
Refs: bootstrap-kit slots 18, 19, 29 reconcile against
ghcr.io/openova-io/bp-seaweedfs:1.0.0, bp-harbor:1.0.0, bp-vpa:1.0.0
which this commit produces on next blueprint-release CI run.
Co-authored-by: hatiyildiz <hatice.yildiz@openova.io>
|
||
|---|---|---|
| .. | ||
| chart | ||
| blueprint.yaml | ||
| README.md | ||
SeaweedFS
S3-compatible object storage with native archive tiering. Per-host-cluster infrastructure (see docs/PLATFORM-TECH-STACK.md §3.5) — runs on every host cluster Catalyst manages. Acts as the unified S3 encapsulation layer in front of cloud archival object storage (Cloudflare R2 / AWS S3 Glacier / Hetzner Object Storage / etc.), so every Catalyst component sees a single S3 API while SeaweedFS transparently routes hot/warm/cold tiers.
Status: Accepted | Updated: 2026-04-28
Overview
SeaweedFS provides a single S3-compatible API that Catalyst components and Application Blueprints consume uniformly, while internally:
- Stores hot data on local NVMe across the host cluster's volume servers
- Tiers warm data to in-cluster bulk storage
- Tiers cold/archival data to a cloud object-storage backend (configured at Sovereign provisioning time)
The encapsulation property is the architectural point: no Catalyst component talks to cloud S3 directly. Every consumer talks to seaweedfs.storage.svc:8333. SeaweedFS is the policy boundary for tiering, lifecycle, replication, and audit.
Why SeaweedFS
| Factor | SeaweedFS |
|---|---|
| License | Apache 2.0 (truly open source) |
| API | S3-compatible (drop-in for any S3 client; no app changes when migrating off MinIO) |
| Architecture | Master + volume server split — scales to millions of files per node |
| Native tiering | Built-in tiered storage with cloud backends (S3, R2, Glacier, GCS, Azure Blob) — no MinIO ILM-style external policy engine needed |
| Encryption | AES-256 at rest, TLS in transit |
| Replication | Cross-region async, per-bucket replication policy |
| Erasure coding | Reed-Solomon (4+2, 6+3, 8+4) for cost-efficient durability |
| FUSE / WebDAV / HDFS / Filer APIs | Optional surfaces for non-S3 consumers |
Architecture
flowchart TB
subgraph SeaweedFS["SeaweedFS — unified S3 layer"]
S3API[S3 API :8333]
Filer[Filer]
Master[Master]
Volumes[Volume Servers]
TierMgr[Tier Manager]
end
subgraph LocalTiers["Local tiers (in-cluster)"]
Hot[Hot: NVMe volume servers]
Warm[Warm: bulk volume servers]
end
subgraph CloudArchive["Cloud archive backend (chosen at provisioning)"]
R2[Cloudflare R2]
S3G[AWS S3 Glacier]
Hetzner[Hetzner Object Storage]
end
subgraph Consumers["All S3 consumers — only see seaweedfs.storage.svc:8333"]
Loki[Loki]
Mimir[Mimir]
Tempo[Tempo]
Velero[Velero]
CNPG[CNPG WAL archive]
Harbor[Harbor blob store]
OpenSearch[OpenSearch snapshots]
Iceberg[Iceberg tables]
Apps[Application Blueprints]
end
Consumers --> S3API
S3API --> Filer
Filer --> Master
Master --> Volumes
Volumes --> Hot
TierMgr --> Warm
TierMgr --> R2
TierMgr --> S3G
TierMgr --> Hetzner
Tiered Storage Strategy
| Tier | Duration | Storage backend | Cost |
|---|---|---|---|
| Hot | 0–7 days | Local NVMe (in-cluster volume servers) | $ |
| Warm | 7–30 days | Local bulk (in-cluster volume servers) | |
| Cold | 30d+ | Cloud archive backend (R2 / Glacier / Hetzner) | $ |
Tiering is automatic and transparent — the consumer's S3 GET/PUT call goes to the same endpoint regardless of which tier the object lives in. SeaweedFS resolves the tier transparently, fetching from cloud archive on demand if the object has aged past warm.
Cloudflare R2 is the default cold backend (zero egress cost). AWS S3 Glacier and Hetzner Object Storage are first-class alternatives chosen at Sovereign provisioning time.
Configuration
Deployment (StatefulSet via SeaweedFS Operator)
apiVersion: seaweed.seaweedfs.com/v1
kind: Seaweed
metadata:
name: seaweedfs
namespace: storage
spec:
image: chrislusf/seaweedfs:3.71
master:
replicas: 3
volumeSizeLimitMB: 30000
persistentVolumeClaim:
storageClassName: fast-nvme
resources: { requests: { storage: 10Gi } }
volume:
replicas: 6
requests: { cpu: 1, memory: 4Gi }
persistentVolumeClaim:
storageClassName: bulk
resources: { requests: { storage: 1Ti } }
filer:
replicas: 2
config: |
[postgres]
enabled = true
hostname = "seaweedfs-meta-rw.storage.svc"
port = 5432
username = "seaweedfs"
password = "${POSTGRES_PASSWORD}"
database = "seaweedfs"
s3:
replicas: 2
enabled: true # S3 API on :8333 — what every consumer talks to
Cloud archive tier (cold backend)
# Configured by sovereign-admin at provisioning time
apiVersion: v1
kind: ConfigMap
metadata:
name: seaweedfs-tier-config
namespace: storage
data:
remote.s3.cold: |
[s3.cold]
type = "s3"
aws_access_key_id = "${R2_ACCESS_KEY}"
aws_secret_access_key = "${R2_SECRET_KEY}"
region = "auto"
bucket = "openova-archive"
endpoint = "https://<account>.r2.cloudflarestorage.com"
storage_class = "STANDARD"
Lifecycle policies (per-bucket)
# Applied via SeaweedFS S3 API or CRD wrapper
apiVersion: seaweed.seaweedfs.com/v1
kind: BucketLifecycle
metadata:
name: loki-data
spec:
bucket: loki-data
rules:
- id: TierToCold
status: Enabled
filter: { prefix: "logs/" }
transitions:
- days: 7
storageClass: WARM
- days: 30
storageClass: COLD # routes to cold tier (R2/Glacier/etc.) automatically
- id: ExpireAfterRetention
status: Enabled
filter: { prefix: "logs/" }
expiration: { days: 365 }
Buckets (canonical layout — same regardless of cloud backend)
| Bucket | Consumer | Default lifecycle |
|---|---|---|
loki-data |
Loki | Tier to cold after 7d, expire after 365d |
mimir-data |
Mimir (long-term metrics) | Tier to cold after 7d, expire after 365d |
tempo-data |
Tempo | Tier to cold after 7d, expire after 30d |
velero-backups |
Velero | Tier to cold after 30d, expire after 90d |
cnpg-wal |
CNPG WAL archive | Tier to warm after 1d, expire after 7d |
cnpg-base-backups |
CNPG base backups | Tier to cold after 7d, expire after 90d |
harbor-data |
Harbor blob store | Tier to cold after 90d, no expiry |
opensearch-snapshots |
OpenSearch snapshots | Tier to cold after 30d, expire after 365d |
iceberg-warehouse |
Iceberg analytic tables | No automatic transition (Iceberg manages compaction); cold after 90d |
livekit-recordings |
LiveKit egress | Tier to cold after 7d, expire after 90d |
ai-hub-models |
KServe / vLLM model weights | No automatic transition; pinned to warm |
<org>-<app>-* |
Per-Application buckets (auto-created on App install) | Inherits Org default; overridable |
Multi-Region Replication
flowchart LR
subgraph Region1["Region 1 (primary)"]
SW1[SeaweedFS]
end
subgraph Region2["Region 2 (replica)"]
SW2[SeaweedFS]
end
subgraph CloudArchive["Cloud archive (single, shared)"]
R2[Cloudflare R2]
end
SW1 -->|"Async bucket replication"| SW2
SW1 -->|"Cold tier"| R2
SW2 -.->|"Cold tier read-through"| R2
The cold backend is shared across regions — each region writes warm-aged objects to the same archive bucket, deduplicated by content hash. On region failover, the surviving region reads cold objects directly from cloud archive without needing to re-replicate.
Monitoring
| Metric | Description |
|---|---|
seaweedfs_volume_server_disk_usage_bytes |
Per-volume-server disk usage |
seaweedfs_s3_request_total |
S3 request count by op |
seaweedfs_s3_request_seconds |
S3 request latency histogram |
seaweedfs_tier_transition_total |
Objects transitioned per tier |
seaweedfs_tier_remote_storage_bytes |
Bytes stored in each cold backend |
seaweedfs_master_topology_volume_count |
Volumes per data center / rack |
Migration from MinIO
SeaweedFS is a drop-in S3 replacement:
- No application changes — same S3 API surface (PUT, GET, MULTIPART, presigned URLs, ListObjectsV2, etc.).
- Same client libraries — boto3, AWS SDK for Go/Java/JS, mc client all work unchanged against
seaweedfs.storage.svc:8333. - Bucket migration —
mc mirrorfrom MinIO → SeaweedFS, or use SeaweedFS'sweed filer.remote.syncto ingest existing data. - Tiering simplification — MinIO's external ILM + tier configuration is replaced by SeaweedFS's native tiered storage. One configuration boundary instead of two.
- Encapsulation upgrade — direct S3 calls to cloud providers (which were scattered across components) are replaced by uniform calls to SeaweedFS, which routes to the right tier. Audit log, encryption, and lifecycle policy are now applied at one boundary.
Part of OpenOva