Channel Summary
Managed bucket slices extend archive drops by appending fresh partitions every 5-60 minutes. BlockDB owns the orchestration, writes directly into your bucket, and attaches manifests so you can reuse existing ETL jobs.How It Works
- Provide a destination bucket + prefix (S3, Azure Blob, or GCS) and grant write permissions.
- BlockDB appends new partitions per dataset and chain, tagged with ingestion timestamps.
- Every batch includes a
manifest.jsonlfile containing_tracing_idranges, row counts, and sequence numbers. - Duplicate protection: when BlockDB retries a batch the same
seqis reused so loaders can dedupe safely.
Partition Layout
Consumption Pattern
- Configure Glue, dbt, Data Factory, or Databricks jobs to watch for new
seqfolders. - Upsert into warehouse tables using dataset primary keys +
_tracing_id. - Persist the latest processed
seqto resume after outages.
Monitoring & Verification
- Compare manifest timestamps with Data Freshness SLAs.
- Sample rows and run Verification endpoints for periodic audits.
- Hook bucket notifications (EventBridge, Event Grid, Pub/Sub) into your alerting stack to detect stalled feeds.
Pairing With Other Channels
- Let bucket slices feed analytics stores while WebSocket streams power alerting.
- Maintain a rolling buffer (e.g., 7 days) in hot storage and archive older partitions to Glacier/Coldline to manage costs.
The folder layout matches archive deliveries, so no new ingestion code is required—just run the same loaders more frequently.