Skip to content

Commit eb1d32d

Browse files
authored
Update scalability-overview.md
1 parent 6ca150f commit eb1d32d

File tree

1 file changed

+8
-7
lines changed

1 file changed

+8
-7
lines changed

articles/cosmos-db/mongodb/vcore/scalability-overview.md

Lines changed: 8 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -30,7 +30,7 @@ Eventually, the application grows to a point where scaling vertically is not suf
3030
- Geo-distributed clusters have a homogeneous multi-node configuration. Thus logical to physical shard mappings are consistent across the primary and replica regions of a cluster.
3131

3232

33-
# Compute and Storage scaling
33+
## Compute and Storage scaling
3434
Compute and memory resources influence read operations in the vCore based service for Azure Cosmos DB for MongoDB more than disk IOPS.
3535
- Read operations first consult the cache in the compute layer and fall back to the disk when data could not be retrieved from the cache. For workloads with a higher rate of read operations per second, scaling up the cluster tier to get more CPU and memory resources leads to higher throughput.
3636
- In addition to read throughput, workloads with a high volume of data per read operation also benefit from scaling the compute resources of the cluster. For instance, cluster tiers with more memory facilitate larger payload sizes per document and a larger number of smaller documents per response.
@@ -47,12 +47,13 @@ As mentioned earlier, storage and compute resources are decoupled for billing an
4747
### Lower TCO with large disks (32 TB and beyond)
4848
Typically, NoSQL databases with a vCore based model limit the storage per physical shard to 4 TB. The vCore based service for Azure Cosmos DB for MongoDB provides upto 8x that capacity with 32 TB disks and plans to expand to 64 TB and 128 TB disks per shard soon. For storage heavy workloads, a 4 TB storage capacity per physical shard requires a massive fleet of compute resources just to sustain the storage requirements of the workload. Compute is more expensive than storage and over provisioning compute due to capacity limits in a service can inflate costs rapidly.
4949

50-
Let's consider a storage heavy workload with 200 TB of data.
51-
| Storage size per shard | Min shards needed to sustain 200 TB |
52-
|-----------------------------|-------------------------------------|
53-
| 4 TB | 50 |
54-
| 32 TiB | 7 |
55-
| 64 TiB (Coming soon) | 4 |
50+
Let's consider a storage heavy workload with 200 TB of data.
51+
52+
| Storage size per shard | Min shards needed to sustain 200 TB |
53+
|------------------------|-------------------------------------|
54+
| 4 TB | 50 |
55+
| 32 TiB | 7 |
56+
| 64 TiB (Coming soon) | 4 |
5657

5758
The reduction in Compute requirements reduces sharply with larger disks. While more than the minimum number of physical shards may be needed sustain the throughput requirements of the workload, even doubling or tripling the number of shards are more cost effective than a 50 shard cluster with smaller disks.
5859

0 commit comments

Comments
 (0)