Compute Unit (CU) Pricing Model
BlockDB uses a Compute Unit (CU) pricing model that charges based on the computational complexity of each API endpoint. This ensures you only pay for what you use, without being forced into expensive plans for simple queries.How Compute Units Work
Each API request consumes a certain number of Compute Units based on the dataset level and endpoint complexity:Need the per-endpoint CU matrix? Open pricing/compute-units for the detailed cost breakdown. This page focuses on the technical mechanics of CU enforcement.
CU Metering Details
- The CU charge is evaluated before the API gateway streams data, so rejected requests still consume CUs.
- Pagination depth, response size, or filter complexity do not change the CU cost—only the endpoint family matters.
- Retries count as new requests; use idempotent mutation endpoints and exponential backoff to avoid duplicate CU consumption.
- Background jobs calling multiple endpoints must account for the cumulative CU cost per workflow; there is no bulk discount at the API layer.
Distributed CU Budgets
CU allowances mirror the Distributed Rate Limits model: you can carve up the monthly allowance per key and per region.- Assign each API key a CU slice (static or auto-scaling) in the Admin Panel.
- Bind the key to a region (
US,EU,AP) to keep CU accounting local to that gateway. - Rotate keys for separate applications or environments so noisy workloads cannot drain the shared CU pool.
Tracking CU Consumption
Usage Endpoint
CallGET /usage to retrieve per-key CU data.
Usage Snapshot
The
/usage endpoint is free to call and does not consume Compute Units. Poll it to drive dashboards or automated alerts.Best Practices
- Split workloads: Give ingestion, research, and production separate keys to prevent noisy neighbors.
- Pin to region: Keep workloads local to reduce latency and avoid double-counting during multi-region retries.
- Use cheaper datasets first: Filter with 01xx datasets before querying higher-CU analytics endpoints.
- Automate alerts: Pair
/usagewith your observability stack to catch spikes early.
Think of CUs as the compute equivalent of your distributed rate limits—allocate, monitor, and adjust per key and per region to keep mission-critical systems online.