Limit identifier | Limit | Comments |
---|---|---|
Maximum number of storage account credentials | 64 | |
Maximum number of volume containers | 64 | |
Maximum number of volumes | 255 | |
Maximum number of schedules per bandwidth template | 168 | A schedule for every hour, every day of the week (24*7). |
Maximum size of a tiered volume on physical devices | 64 TB for 8100 and 8600 | 8100 and 8600 are physical devices. |
Maximum size of a tiered volume on virtual devices in Azure | 30 TB for 8010 64 TB for 8020 |
8010 and 8020 are virtual devices in Azure that use Standard Storage and Premium Storage respectively. |
Maximum size of a locally pinned volume on physical devices | 9 TB for 8100 24 TB for 8600 |
8100 and 8600 are physical devices. |
Maximum number of iSCSI connections | 512 | |
Maximum number of iSCSI connections from initiators | 512 | |
Maximum number of access control records per device | 64 | |
Maximum number of volumes per backup policy | 24 | |
Maximum number of backups retained per backup policy | 64 | |
Maximum number of schedules per backup policy | 10 | |
Maximum number of snapshots of any type that can be retained per volume | 256 | This includes local snapshots and cloud snapshots. |
Maximum number of snapshots that can be present in any device | 10,000 | |
Maximum number of volumes that can be processed in parallel for backup, restore, or clone | 16 |
|
Restore and clone recover time for tiered volumes | < 2 minutes |
|
Restore recover time for locally pinned volumes | < 2 minutes |
|
Thin-restore availability | Last failover | |
Maximum client read/write throughput (when served from the SSD tier)* | 920/720 MB/s with a single 10GbE network interface | Up to 2x with MPIO and two network interfaces. |
Maximum client read/write throughput (when served from the HDD tier)* | 120/250 MB/s | |
Maximum client read/write throughput (when served from the cloud tier)* | 11/41 MB/s | Read throughput depends on clients generating and maintaining sufficient I/O queue depth. |
* Maximum throughput per I/O type was measured with 100 percent read and 100 percent write scenarios. Actual throughput may be lower and depends on I/O mix and network conditions.