Use QNAP QES series to build a large-capacity storage system


最后修订日期: 2018-06-20

Technical overview and application scenarios

High-availability file servers, enterprise cloud storage, online video streaming, digital monitoring storage and similar applications all require large-capacity storage devices. General-use NAS has limitations in hardware and software architecture, and it is difficult to adapt to enterprise-class high-pressure, high availability environments. Traditional file systems are also not optimized for large capacities with size/capacity limits, they feature no integrated data integrity checks, and also lack other features.

The QNAP ES NAS has features such as nearly non-stop high-availability, self-healing, data compression and deduplication, unlimited snapshots, instantaneous construction of disk arrays and more. In addition to all the necessary functions, it also provides the most cost-effective large capacity storage solution.

Example scenario

Construction of a Large Capacity Storage Pool

Prepare the disk expansion enclosure

The QNAP QES system provides cross-expansion closure storage pool construction. The ES1640dc has 16 built-in drive bays, and can be connected to multiple expansion enclosures to scale up the storage capacity. The dual-controller ES NAS and its expansion enclosure both have 16 drive bays. Based on the dual-controller ES1640dc series design, up to 7 expansion enclosures can be connected to provide a single system with up to 128 drive bays.

The ES1640dc v2 supports the EJ1600 v2 expansion enclosure. This expansion enclosure connects to the ES NAS via SAS 12Gb/s, providing lower latency than USB. The EJ series expansion enclosures also use a dual-controller design and can support each other. Each group of controllers has 2-port SAS expansion ports, so that ES NAS and EJ expansion enclosures can be connected in a dual-loop, 4-SAS-port architecture. The total interconnected bandwidth can reach up to 48 Gbps to enable a high-performance and reliable high-availability environment.

Recommended connection architecture between QNAP ES NAS and EJ expansion enclosures

ES NAS and EJ series expansion enclosures both have a dual-controller architecture. Each controller group has built-in 2 external SAS ports. To effectively build a dual-loop topology, it is recommended that each group chassis be connected according to the above diagram to ensure the storage system is still able to maintain normal operations in the event of a single point of failure. After completing the dual-loop connection, the NAS controller on either side of the system will be fault tolerant with the expansion enclosure controller on either side. Even if any of the connections are interrupted, the system will still be able to access data from another group.

Note: The bottom JBOD will require two longer SAS lines to connect back to the ES NAS host (Loop Back). For detailed specifications of hardware and cables, please refer to the materials list table below.

Installable disks and corresponding materials list

QNAP ES NAS QNAP
EJ expansion enclosure
0.5M SAS short cable Loop back SAS long cable Total number of SAS cables Installable number of disks Max. raw capacity*
1 0 0 0 0 16 160 TB
1 1 2 2 (0.5M) 4 32 320 TB
1 2 4 2 (1M) 6 48 480 TB
1 3 6 2 (1M) 8 64 640 TB
1 4 8 2 (1M) 10 80 880 TB
1 5 10 2 (2M) 12 96 960 TB
1 6 12 2 (2M) 14 112 1120 TB (1.12PB)
1 7 14 2 (2M) 16 128 1280 TB (1.28PB)

* Based on using 10 TB hard drives.
To estimate the available capacity with specific RAID types, use the ES NAS ZFS Capacity Calculation Tool:
(https://enterprise-nas.qnap.com/en/calculator/)

Need SSD?

Dual-control series ES NAS system supports using SSD as a read cache. When configuring an ES NAS, the small capacity, cost, and use of a drive bay must be taken into consideration when using SSD. SSD provides low-latency responses and IOPS that vastly outperforms hard drives, making them useful for applications such as databases and virtual desktops. If the main application of the ES NAS is a file server, high capacity and read/write performance should be prioritized bandwidth, making hard drives more suitable than SSD.

Planning the RAID architecture and hot spares

As storage systems store more data, the chance of data loss through hardware failure also increases. Reliability and backup plans thus become exceedingly important. Before building the overall system, we can plan RAID architectures according to the importance of data, capacity utilization, and tolerance level for data loss.

As QES is based on ZFS, there is greater focus on data protection, including Global Hot Spare technology, which allows for the simultaneous failure of up to three drives in RAID TP mode, 3-times data protection Triple-Mirror mode and more. In environments that require multiple backups, QES allows you to sacrifice performance and capacity utilization to achieve absolute reliability.

For general use, it is recommended to use ZFS RAID Group architecture. Using 10 to 15 disks for each NAS/expansion enclosure as a unit to establish RAID 6 mode and set aside a blank disk with global hot backup technology, you will be able to achieve both complete data protection and disaster fault tolerance. The recommended setting is as shown in the table and figures below.

128 disk RAID configuration example

Where Installed number of disks RAID type Hot spare
ES NAS 16 11 RAID 6
(4 system disks are not included in the storage pool)
1
EJ JBOD#1 16 15 RAID 6 1
EJ JBOD#2 16 15 RAID 6 1
EJ JBOD#3 16 15 RAID 6 1
EJ JBOD#4 16 15 RAID 6 1
EJ JBOD#5 16 15 RAID 6 1
EJ JBOD#6 16 15 RAID 6 1
EJ JBOD#7 16 15 RAID 6 1
Total 128 8 groups of RAID 6 groups
RAID 60 storage pool
8

When creating a RAID, it is recommended to reserve a blank hard disk without including it into any array. Depending on the mechanism of the QES system architecture, these reserved disks will be treated as hot spares, without additional special set up. If the system encounters a hard disk failure, the global hot backup mechanism will be triggered, it will automatically capture any blank disk on ES NAS or JBOD and actively rebuild the RAID.

NAS Drive configuration:

JBOD # 1 drive configuration:

JBOD # 7 drive configuration:

From top to bottom, the figure respectively shows the recommended configuration for 1 ES NAS and 7 EJ JBOD. If the first ES NAS is configured with four SSD as a system cache, the RAID 6 array can be made up of 11 disks and 1 reserved hot spare hard disk. If no SSD cache is deployed in the system, then 15 disks can be used to build a RAID 6 array; the rest of expansion closures all use a 15 disk RAID 6 array, with 1 hot spare hard disk in its configuration.

How to build RAID 60

Based on the above architecture, after setting up the first RAID group, more RAID groups can be added through storage pool expansion (Expand Pool). As each RAID group is Striped in relation to each other, when data is written, it will be written to several RAID groups at the same time, and this is also true for read, so RAID 60 storage pool can be created in QES to achieve both efficiency and reliability.


Basic application and performance data

In this section, we create a super storage pool with capacity up to hundreds of terabytes based on the above architecture, and mount it for corresponding application scenarios through the shared folder as an example. The test scenario simulates basic application scenario and measures its performance data.

Test platform

Storage system
NAS 1 x QNAP ES1640dc v2
Expansion enclosures 7 x QNAP EJ1600 v2
Interconnection method 4 x Mini-SAS 12Gb/s, Dual Loop
SSD model 4 x Samsung PM863 240GB
HDD model 1 12 x HGST HUH721010AL5200 10TB
HDD model 2 112 x Seagate ST8000NM0075 8TB
Total number of drives 4 x SSD + 124 x HDD
NAS data port network card Intel X520-SR2 SFP+
SFP+ connection method Optical fiber
SFP+ GBIC Intel FTLX8571D3BCV
System version QES 1.1.3
Storage pool structure 8 RAID 6 groups, a total of 124 HDDs

Ten Windows Server 2012 R2 hosts mount the same shared folder and concurrently run IO tests, and test the throughput provided by ES NAS simulating concurrent write/access to/from multiple hosts.

Low capacity level read/write performance

It is clear from the following measured data that SSD cache acceleration is significantly helpful for block random read performance.

Single client test

Test item Load mode No SSD Caching With SSD Caching
AJA System Test sequential read HD-1080i, 16GB, 10bit YUV 603 MB/s 410 MB/s
Iometer 256K sequential read Thread=1, Queue depth=1 304.36 MB/s 236.71 MB/s
Iometer 4K random read Thread=1, Queue depth=1 48.8 IOPS 108.29 IOPS
Iometer 4K random read Thread=1, Queue depth=16 974.5 IOPS 4434 IOPS
AJA System Test sequential write HD-1080i, 16GB, 10bit YUV 520 MB/s 500 MB/s
Iometer 256K sequential write Thread=1, Queue depth=1 233.18 MB/s 295.78 MB/s
Iometer 4K random write Thread=1, Queue depth=1 1474 IOPS 2959.64 IOPS
Iometer 4K random write Thread=1, Queue depth=16 2281 IOPS 2801.89 IOPS

Multi-client concurrent access test

Test item Load mode No SSD Caching With SSD Caching
AJA System Test sequential read HD-1080i, 16GB, 10bit YUV N/A(note) N/A(note)
Iometer 256K sequential read Thread=1, Queue depth=1 1017.34 MB/s 1017.17 MB/s
Iometer 4K random read Thread=1, Queue depth=1 476.1 IOPS 7027.98 IOPS
Iometer 4K random read Thread=1, Queue depth=16 974.76 IOPS 18821.64 IOPS
AJA System Test sequential write HD-1080i, 16GB, 10bit YUV N/A(note) N/A(note)
Iometer 256K sequential write Thread=1, Queue depth=1 796.38 MB/s 659.79 MB/s
Iometer 4K random write Thread=1, Queue depth=1 2628.61 IOPS 3200.6 IOPS
Iometer 4K random write Thread=1, Queue depth=16 2602.41 IOPS 2863.7 IOPS

Note: AJA System Test does not support multi-client concurrent access testing, so no data is available.

Test writing large number of files

To meet the requirements for large number of file loads, the test in this section uses Disk tools (File Generator) to adjust file size, quantity and content (duplicate or random text), to simulate the performance in writing a large number of files.

Network connection method

In terms of network topology, two shared folders on the ES NAS are mounted on the server host via two 10Gb network ports.

Test load source

Path 1
(\\33.31.1.102)
(Z:\)
Path 2
(\\33.31.1.102)
(Z:\1)
Path 3 (\\33.32.1.102)
(Y:\2)
Write load 1
Single file size 30MB
Write load 3
Single file size 40MB
Write load 5
Single file size 20MB
Write load 7
Single file size 60MB
Write load 9
Single file size 50MB
Write load 2
Single file size 30MB
Write load 4
Single file size 50MB
Write load 6
Single file size 50MB
Write load 8
Single file size 80MB
Write load 10
Single file size 100MB

To verify high-intensity write load, the generated files will range between 20-100MB, with completely random file content. The files are then written to the storage system for a long time, and the results are as follows.

Write test result

  • Total amount of data written: 722TB
  • Total write time: 713 hours
  • Average write speed: 294.75MB/s (about 1TB per hour)
  • NAS CPU usage: 50 ~ 55%
  • Host CPU usage: 80 ~ 90%
  • The actual low-level writing bandwidth of ZFS file system is 850 ~ 1000MB/s, IOPS is about 9K ~ 11K, the average IO size is 94KB
  • ZFS low-level write magnification is around 3 times

High capacity level performance

After a long period of intensive writing, the total capacity and the number of files will reach a massive number. Monitoring the storage system's read performance in such a condition, the simulated stability under intensive use is as follows.

Read Performance measurement

Test item Load mode Result
Iometer 256K sequential read Thread=1, Queue depth=1 421.40 MB/s
Thread=1, Queue depth=2 719.17 MB/s
Thread=1, Queue depth=4 751.45 MB/s
Thread=4, Queue depth=4 1084.06 MB/s
Iometer 4K random read Thread=1, Queue depth=1 2240.47 IOPS
Thread=1, Queue depth=4 8225.18 IOPS
Thread=4, Queue depth=4 14141.09 IOPS

In the actual measurement, under a near saturation level of storage capacity, the NAS read performance can still be maintained at a certain level, and the highest sequential read performance is still close to the upper limit of 10Gb network bandwidth, effectively providing long-term stable output performance.

Disaster recovery and array reconstruction

While building a large capacity storage system where storage pool require a lot of hard disk support and dozens or even hundreds of hard disks might be running concurrently, responding to hard disk failure becomes an  important issue. As mentioned in the above section, QES has multi-level data protection through hardware and software technology, even if a hard disk fails, it still has 2 security mechanisms to ensure data reliability: global hot backup and fast data recovery technology.

Disk failure handling process

Using the above test architecture as an example, the storage pool consists of 8 RAID groups composed of a total of 116 hard disks, with 8 hard disks set aside as hot spares. If disk usage reaches up to 722 TB and the capacity level has reached 99.97% saturation, if a hard disk fails occurs, QES will immediately grab the standby hard disk through the global hot backup mechanism, and use fast data recovery technology to reconstruct data to minimize the impact of the disk failure.

Array reconstruction measurement

In a near-saturated storage pool, if an 8TB hard disk fails, the system will be reconstructed from another 8 TB hot spare. The following test data shows the approximate time required for data recovery.

Storage pool capacity level Used space 8TB HDD failure recovery time
99.97% 722.0 TB/722.2 TB 276H 23M (about 11.5 days)


Expansion enclosure failure handling

In the event of an expansion enclosure failure, in order to ensure integrity of existing data, QES will actively protect the storage pool through offline data protection mechanisms. When the expansion enclosure returns online and the storage pool is restored, it will continue to provide service as normal.


Conclusion

QNAP QES has many features, such as multi-level data protection technology, large single storage space, real-time data compression, data deduplication and more. These are all useful features for creating a large capacity application environment. In combination with a variety of RAID and storage pool configurations, it can effectively help users flexibly plan for their storage. Even in the event of accidents and disasters, it can use rapid reconstruction technology to reduce the impact of hardware failure as far as possible through the recovery process. It has multi-port 10Gb network interface, supports iSCSI, NFS, CIFS and other communication protocols, and is well suited for a variety of application scenarios.

这篇文章有帮助吗?

100% 的人认为有帮助。
谢谢您,我们已经收到您的意见。

请告诉我们如何改进这篇文章:

如果您想提供其他意见,请于下方输入。

选择规格

      显示更多 隐藏更多
      open menu
      back to top