官术网_书友最值得收藏!

Failure domains

If your cluster will have less than 10 nodes, this is probably the most important point.

With legacy scale-up storage, the hardware is expected to be 100% reliable. All components are redundant, and the failure of a complete component such as a system board or disk JBOD would likely cause an outage. Therefore, there is no real knowledge of how such a failure might impact the operation of the system, just the hope that it doesn't happen! With Ceph, there is an underlying assumption that complete failure of a section of your infrastructure, be that a disk, node, or even rack should be considered as normal and should not make your cluster unavailable.

Let's take two Ceph clusters both comprising 240 disks. Cluster A comprises 20x12 disk nodes; Cluster B comprises 4x60 disk nodes. Now, let's take a scenario where for whatever reason a Ceph OSD node goes offline. It could be due to planned maintenance or unexpected failure, but that node is now down and any data on it is unavailable. Ceph is designed to mask this situation and will even recover from it whilst maintaining full data access.

In the case of cluster A, we have now lost 5% of our disks and in the event of a permanent loss would have to reconstruct 72 TB of data. Cluster B has lost 25% of its disks and would have to reconstruct 360 TB. The latter would severely impact the performance of the cluster, and in the case of data reconstruction, this period of degraded performance could last for many days.

It's clear that on smaller sized clusters, these very large dense nodes are not a good idea. A 10 Ceph node cluster is probably the minimum size if you want to reduce the impact of node failure, and so in the case of 60 drive JBODs, you would need a cluster that at minimum is measured in petabytes.

主站蜘蛛池模板: 交口县| 长治县| 莱芜市| 东海县| 日土县| 朝阳区| 聂拉木县| 休宁县| 中卫市| 伊宁县| 徐闻县| 滦平县| 临沧市| 台安县| 韶关市| 石柱| 沧州市| 佛山市| 辽源市| 双桥区| 宿松县| 曲松县| 龙江县| 安阳县| 格尔木市| 通辽市| 吉隆县| 自贡市| 东至县| 增城市| 扬州市| 嘉义县| 清新县| 德阳市| 丰城市| 永和县| 曲麻莱县| 竹北市| 兴宁市| 芦山县| 德昌县|