分布式存储系统中基于纠错码的容错机制研究
详细信息    本馆镜像全文|  推荐本文 |  |   获取CNKI官网全文
摘要
随着分布式系统与网络技术的不断发展,分布式存储系统在存储容量、性能、可扩展性等方面已取得了飞速进步,但同时在容错性方面也面临着巨大的挑战:现代的分布式存储系统往往由成百上千个存储节点组成,多个节点出现故障的可能性大大增加,因节点故障导致数据丢失会造成灾难性的后果。因此,如何设计高效、可靠的容错机制,已经是分布式存储系统中迫切需要解决的问题。
     纠错码是通信领域中广泛采用的检测和纠正通信误差、提高信息传输可靠性的技术。而组通信服务是分布式系统中构建自治区域的重要方法,可保证系统在局部故障的情况下仍能稳定正常地运行。目前,组通信服务是分布式系统中实现容错的一项关键技术。
     基于纠错码的容错机制将纠错码技术与组通信服务相结合,提出了基于纠错码的容错组,解决分布式存储系统中的容错问题。容错组是实现纠错码的存储节点集合,数据经过纠错码编码后保存在容错组相应的存储节点中。容错组中的存储节点利用组通信服务组织在一起,并通过协调合作提供高可用与高可靠的数据服务。出现节点故障时,可以利用容错组中其它节点保存的数据恢复出原始数据。容错组利用组视图记录当前组中成员列表,组中存储节点加入、退出、故障都将引起组视图的更新。为了方便数据的读/写与管理,容错组中引入了基于对象存储技术。系统运行时,容错组中的存储节点与元数据服务器、管理服务器进行消息通信,更新元数据服务器与管理服务器保存的容错组信息。
     采取纠错码作为冗余策略的容错组需要减少计算校验数据带来的计算开销,并保证容错组中各存储节点数据一致。设计了容错组中基于组视图的I/O操作,包括读操作、写操作与更新操作。容错组中每一个存储节点都可以完成存储控制器的功能,客户端将I/O请求发送给任意存储节点,存储节点作为协调者再访问组中节点,完成请求。容错组的纠错码都采用系统码形式,编码后的数据块分为信息数据与校验数据。读操作中协调者可以直接访问保存信息数据的节点,只是节点故障时才需要译码恢复数据。数据第一次写入容错组需要编码,然后将编码后的数据存储到相应的节点中。根据纠错码理论,更新操作并不需要再次对数据进行编码,而是先读取已保存的信息数据,与更新的数据计算异或,再将信息数据与校验数据分别更新。为了保证更新数据的一致性,更新操作还需应用时间戳排序、两阶段提交、日志等技术。
     为了保证容错组的稳定性与容错组中I/O操作的正确性,容错组需要及时检测出故障节点。设计了容错组中基于服务质量的QB-AFD故障检测器。为了探测出组中的故障节点,容错组的每个成员都要向其它成员发送心跳消息,同时监听来自其它成员的心跳消息。如果存储节点在超时时刻还没有接收到某节点的心跳消息,即开始怀疑该节点。故障检测需要满足速度与精度的要求,因此,QB-AFD故障检测器动态调整预期心跳消息到达时间与安全时间余量,以满足服务质量的评价指标。
     利用组成员关系服务完成容错组的重构任务,设计了基于两阶段提交的视图更新算法2pc-VUA保证组中存储节点维护一致的全局组视图,并设计了节点加入算法负责在新加入容错组的存储节点中生成与其它成员一致的组视图。
With the development of distributed systems and network technology, the storage capacity, performance, and scalability of distributed storage systems have increased rapidly. However, distributed storage systems are confronted with great challenges in fault tolerance. As storage systems scale up, they consist of thousands of storage nodes. Multiple failures will occur frequently, which would induce a disaster because of data loss. How to design a efficient and reliable fault tolerance mechanism has become a issue that needs to be resolved urgently in distributed storage systems.
     Fault tolerance mechanism based on erasure code integrates erasure code and group communication services, which aims to resolve the fault tolerance in distributed storage systems. By using group communication services, storage nodes are organized in a fault tolerance group. Through negotiating and cooperating, storage nodes in a group can provide high available and reliable data services. After encoding, data are stored in corresponding storage nodes. When a storage node fails, original data can be recovered by decoding the data stored in other nodes. To keep fault tolerance group stabilizing, fault nodes should be detected in time and the fault tolerance group need to reconfigure itself.
     Information of fault tolerance groups are kept in metadata server, and user can fetch relative fault tolerance group when sending request to metadata server. While distributed storage system is running, storage nodes in a fault tolerance group communicates with metadata server and manage server, and update the information of fault tolerance group in them. By updating these information, metadata server can ensure that I/O requests are delivered to correct store nodes, and manage server can ensure the validity of history record in a fault tolerating group.
     The major contributions of fault tolerance mechanism based on erasure code are as follows:
     Present fault tolerance groups in distributed storage systems that merge erasure code and group communication services. A fault tolerance group is a set of storage nodes that implement erasure code, and the Identifier of a node correspond to a serial number of the data after encoding. Data are chosen storage nodes to store according to the serial number. A fault tolerance group uses a group view to record members of the group, and the events that a storage node joins, leaves and fails will cause update of the group view. To make read/write and management more convenience, object-based storage and storage virtualization are introduced in fault tolerance groups. Because of the effect of cache, the performance of a storage node is better than that of a disk.
     Propose I/O operation based on erasure code in a fault tolerance group, include read, write, and update. Unlike the traditional method where a client accesses a storage controller, a storage node picked randomly acts as a coordinator. A client sends its read or write request to the coordinator, and the coordinator forwards the request to members in the group. Responses are returned through the same coordinator to the client. The I/O operation based on erasure code needs to avoid computation of encoding and decoding, and keep data in every storage node consistent. So the erasure codes in fault tolerance groups are systematic code, original data will be contained in data after encoding. In a read operation, coordinator directly accesses data, and decoding is only needed when node fails and data loss. If data are stored into fault tolerance group in the first time, encoding process are absolutely necessary. However, when updating data, delta that is exclusive-or of the last data and newest data can be calculated and then sends to the storage nodes storing related checksum data. To keep data consistent, timestamp, two-phase protocol and log are applied in update operation.
     Propose a QB-AFD failure detector in fault tolerance based on QoS. To detect failure nodes in a group, each node sends heartbeat messages to each other member and listens others heartbeat messages at the same time. A failure detector may suspect a storage node when it can’t receive heartbeat messages from that node at a timeout point. A failure detector should also satisfy the requirement of speed and accuracy, so the QB-AFD failure detector dynamically adapts the heartbeat period and timeout delay to fulfill the metrics of QoS.
     Present reconfiguration of fault tolerance group through group membership services. When nodes join, leave and fail, a view update algorithm based on two-phase commit maintains a consistent global view in every storage node. And an adding algorithm generates a view when a node joins a group.
引文
[1] Tanenbaum A S, Steen M v.分布式系统原理与范型.北京:清华大学出版社, 2004.
    [2] Wu J.分布式系统设计.北京:机械工业出版社, 2001.
    [3]杨向东.数据存储——金融电子化的基石.中国金融电脑, 2002(3): 30-32
    [4] Schmuck F, Haskin R. GPFS: A Shared-Disk File System for Large Computing Clusters. In: Proceedings of the 1st USENIX Conference on File and Storage Technologies. 2002. 19-32
    [5] Satyanarayanan M, Kistler J J, Kumar P, et al. Coda: A Highly Available File System for Distributed Workstation Environment. IEEE Transactions on Computers, 1990, 39(4): 447-459
    [6] Ghemawat S, Gobioff H, Leung S-T. The Google File System. In: Proceedings of the 9th ACM Symposium on Operating System Principles. 2003. 29-43
    [7] Yu W, Noronha R, Liang S, et al. Benefits of High Speed Interconnects to Cluster File Systems: A Case Study with Lustre. In: Proceedings of the 20th International Symposium on Parallel and Distributed Processing. 2006. 385-392
    [8] Saito Y, Frolund S, Veitch A, et al. FAB: Building Distributed Enterprise Disk Arrays from Commodity Components. ACM SIGARCH Computer Architecture News, 2004, 32(5): 48-58
    [9] Frolund S, Merchant A, Saito Y, et al. A Decentralized Algorithm for Erasure-Code Virtual Disks. In: Proceedings of the 2004 International Conference on Dependable Systems and Networks (DSN'04). 2004. 125-134
    [10] Rhea S, Wells C, Geels D, et al. Maintaince-Free Global Data Storage. IEEE Internet Computing, 2001, 5(5): 40-50
    [11] Kubiatowicz J, Bindel D, Chen Y, et al. OceanStore: An Architecture for Global-Scale Persistent Storage. In: Proceedings of the Ninth InternationalConference on Architecutual Support for Programming Language and Operating Systems (ASPLOS 2000). 2000. 190-201
    [12] Rhea S, Eaton P, Geels D, et al. Pond: the OceanStore Prototype. In: Proceedings of the 2nd USENIX Conference on File and Storage Technologies (FAST'03). 2003. 1-14
    [13] Lamport L, Shostak R, Pease M. The Byzantine Generals Problem. ACM Transactions on Programming Languages and Systems, 1982, 4(3): 382-401
    [14] Adya A, Bolosky W J, Castro M, et al. FARSITE: Federated, Available, and Reliable Storage for an Incompletely Trusted Environment. In: Proceedings of the 5th Symposium on Operating Systems Design and Implemention. 2002. 1-12
    [15] Malkhi D, Reiter M K. Byzantine quorum systems. Distributed Computing, 1998, 11(4): 203-213
    [16] Tang H, Gulbeden A, Zhou J, et al. A Self-Organizing Storage Cluster for Parallel Data-Intensive Applications. In: Proceedings of the ACM/IEEE SC2004 Conference on Supercomputing. 2004. 635-647
    [17] Wylie J J, Bigrigg M W, Strunk J D, et al. Survivable information storage systems. Computer, 2000, 33(8): 61-68
    [18] Ganger G R, Khosla P K, Bakkaloglu M, et al. Survivable Storage Systems. In: Proceedings of the 2001 DPRPA Information Survivability Conference & Exposition. 2001. 184-195
    [19] Goodson G R, Wylie J J, Ganger G R, et al. Efficient Byzantine-tolerance Erasure-coded Storage. In: Proceedings of the 2004 International Conference on Dependable Systems and Networks (DSN'04). 2004. 135-144
    [20] Bohossian V, Fan C C, LeMahieu P S, et al. Computing in the RAIN: A Reliable Array of Independent Nodes. IEEE Transactions on Parallel and Distributed Systems, 2001, 12(2): 99-114
    [21] Li J. The Efficient Implementation of Reed-Solomon High Rate Erasure ResilientCodes. In: Proceedings of the 2005 IEEE International Conference on Acoustics, Speech and Signal Processing. 2005. 1097-1100
    [22] Hsieh P-H, Chen I-Y, Lin Y-T, et al. An XOR Based Reed-Solomon Algorithm for Advanced RAID Systems. In: Proceedings of the 19th IEEE International Symposium on Defect and Fault Tolerance in VLSI Systems. 2004. 165-172
    [23] Shokrollahi A. An Introduction to Low-Density Parity-Check Codes. In: Theoretical Aspects of Computer Science: Advanced Lectures. 2000. 175-197
    [24] Sobe P. Reconfiguration of RAID-like Data Layouts in Distributed Storage Systems. In: Proceedings of the 18th International Parallel and Distributed Processing Symposium (IPDPS'04). 2004. 205-212
    [25] Rhee I, Cheung S Y, Hutto P W, et al. Group Communication Support for Distributed Collaboration Systems. In: Proceedings of the 17th International Conference on Distributed Computing Systems. 1997. 43-50
    [26] Mesnier M, Ganger G R, Riedel E. Object-Based Storage. IEEE Communications Magazine, 2003, 41(8): 84-90
    [27] Azagury A, Dreizin V, Factor M, et al. Towards an Object Store. In: Proceedings of the 20 th IEEE/11 th NASA Goddard Conference on Mass Storage Systems and Technologies. 2003. 165-176
    [28] Rodeh O, Teperman A. zFS - A Scalable Distributed File System Using Object Disks. In: 20th IEEE/11 th NASN Goddard Conference on Mass Storage Systems and Technologies. 2003. 207-218
    [29] Wilkes J, Golding R, Staelin C, et al. The HP AutoRAID Hierarchical Storage System. ACM Transactions on Computer Systems, 1996, 14: 108-136
    [30] Alvarez G A, Borowsky E, Go S, et al. MINERVA: An Automated Resource Provisioning Tool for Large-Scale Storage Systems. ACM Transactions on Computer Systems, 2001, 19(4): 483-518
    [31] Anderson E, Hobbs M, Keeton K, et al. Hippodrome: running circles aroundstorage administration. In: Proceedings of the 1st USENIX Conference on File and Storage Technologies. 2002. 175-188
    [32] Anderson E. Simple table-based modeling of storage devices. 2001, http://www.hpl.hp.com/SSP/papers.
    [33] Hafner J L. HoVer Erasure Codes For Disk Arrays. In: Proceedings of the 2006 International Conference on Dependable Systems and Networks (DSN'06). 2006. 217-226
    [34]裴先登,李伟,谢长生.用于体全息存储系统的纠错码编码及数据交错技术.计算机研究与发展, 2002, 39(10): 1393-1398
    [35] Malluhi Q M, Johnson W E. Coding for High Availability of a Distributed-Parallel Storage System. IEEE Transactions on Parallel and Distributed Systems, 1998, 9(12): 1237-1252
    [36] Marco A D, Chiola G, Ciaccio G. Using a Gigabit Ethernet Cluster as a Distributed Disk Array with Multiple Fault Tolerance. In: Proceedings of the 28th Annual IEEE International Conference on Local Computer Networks (LCN'03). 2003. 605-613
    [37] Li X, Liu C. Towards a Reliable and Efficient Distributed Storage System. In: Proceeding of the 38th Hawaii International Conference on System Sciences - 2005. 2005. 301-311
    [38] Zhang Z, Lian Q. Reperasure: Replication Protocol using Erasure-code in Peer-to-Peer Storage Network. In: Proceedings of the 21st IEEE Symposium on Reliable Distributed Systems (SRDS'02). 2002. 330-335
    [39] Zhang Z, Lin S, Lian Q, et al. RepStore: A Self-Managing and Self-Tuning Storage Backend with Smart Bricks. In: Proceedings of the International Conference on Autonomic Computing. 2004. 122-129
    [40] Pitkanen M, Moussa R. Erasure Codes for Increasing the Availability of Grid Data Storage. In: Proceedings of the Advanced International Conference onTelecommunications and International Conference on Internet and Web Applications and Services (AICT/ICIW 2006). 2006. 185-195
    [41]刘小军.容错存储编码算法的分析研究.陕西科技大学学报, 2006, 24(1): 100-104
    [42] Chockler G V, Keidar I, Vitenberg R. Group Communication Specifications: A Comprehensive Study. ACM Computing Surveys, 2001, 33(4): 427-469
    [43] Babaoglu O, Davoli R, Montresor A. Group Communication in Partitionable Systems: Specification and Algorithms. IEEE Transactions on Software Engineering, 2001, 27(4): 308-336
    [44] Mishra S, Fetzer C, Cristian F. The Timewheel Group Communication System. IEEE Transactions on Computers, 2002, 51(8): 883-899
    [45] Pereira J, Rodrigues L, Oliveira R. Reducing the Cost of Group Communication with Semantic View Synchrony. In: Proceedings of the International Conference on Dependable Systems and Networks (DSN'02). 2002. 293-302
    [46] Jain N K. Group Formation Mechanisms for Transactions in Isis. In: Proceedings of the third International Conference on Information and Knowledge management. 1995. 203-210
    [47] Dolev D, Malki D. The Transis Approach to High Availability Cluster Communication. Communications of the ACM, 1996, 39(4): 64-70
    [48] Amir Y, Moser L E, Melliar-Smith P M, et al. The Totem Single-Ring Ordering and Membership Protocol. ACM Transactions on Computer Systems, 1995, 13(4): 311-342
    [49] Pease D A, Rees R M, Hineman W C, et al. IBM Storage Tank: A Distributed Storage System: http://www.usenix.org/events/fast02/wips, 2002.
    [50] Schiper A, Birman K, Stephenson P. Lightweight Causal and Atomic Group Multicast. ACM transactions on Computer Systems, 1991, 9(3): 272-314
    [51] Dolev S, Schiller E. Communication Adaptive Self-Stabilizing Group MembershipService. IEEE Transactions on Parallel and Distributed Systems, 2003, 14(7): 709-720
    [52]宋平,孙建伶,何志均.基于Quorum系统容错技术综述.计算机研究与发展, 2004, 41(5): 513-523
    [53] Martin J-P, Alvisi L. A Framework for Dynamic Byzantine Storage. In: International Conference on Dependable Systems and Networks. Florence, Italy, 2004. 325-334
    [54] Alvisi L, Malkhi D, Pierce E, et al. Dynamic Byzantine Quorum Systems. In: Proceedings of Internationa conference on Dependable Systems and Networks. 2000. 283-292
    [55] Kong L, J.Manohar D, Subbian A, et al. Agile Store: Experience with Quorum-based Data Replication Techniques for Adaptive Byzantine Fault Tolerance. In: Proceedings of the 2005 24th IEEE Symposium on Reliable Distributed Systems. 2005. 143-154
    [56] Kong L, Subbian A, Ahamad M, et al. A Reconfigurable Byzantine Quorum Approach for the Agile Store. In: the 22nd International Symposium for Reliable Distributed Systems. 2003. 219-228
    [57] Chandra T D, Hadzilacos V, Toueg S, et al. On the Impossibility of Group Membership. In: Proceedings of the fifteenth Annual ACM Symposium on Principlesof Distributed Computing. 1996. 322-330
    [58] Lohmeyer J B, Penokie G O, Aloisi P D. Information technology - SCSI Object-Based Storage Device Commands (OSD). 2005, http://www.t10.org/drafts.htm.
    [59] Wang F, Brandt S A, Miller E L, et al. OBFS: A File System for Object-based Storage Devices. In: Proceedings of the 21st IEEE/1 NASA Goddard Conference on Mass Storage and Technologies. 2004. 8-14
    [60] Uysal M, Alvarez G A, Merchant A. A Modular, Analytical Throughput Model forModern Disk Arrays. In: Proceedings of the Ninth International Symposium on Modeling, Analysis and Simulation of Computer and Telecommunication Systems. 2001. 183-192
    [61] Varki E, Merchant A, Xu J, et al. Issues and Chanllenges in the Performance Analysis of Real Disk Arrays. IEEE Transactions on Parallel and Distributed Systems, 2004, 15(6): 559-574
    [62]林闯.计算机网络和计算机系统的性能评价.北京:清华大学出版社, 2001.
    [63] Iozone Filesystem Benchmark. http://www.iozone.org.
    [64]张宗橙.纠错码原理和应用.北京:电子工业出版社, 2003.
    [65] Kurtas E M, Kuznetsov A V, Djurdjevic I. System Perspectives for the Application of Structured LDPC Codes to Data Storage Devices. IEEE Transactions on Magnetics, 2006, 42(2): 200-207
    [66] Plank J S, Thomason M G. A Practical Analysis of Low-Density Parity-Check Erasure Codes for Wide-Area Storage Applications. In: Proceedings of the 2004 International Conference on Dependable Systems and Networks (DSN'04). 2004. 115-124
    [67] Hu X-Y, Eleftheriou E, Arnold D-M. Progressive Edge-Growth Tanner Graphs. In: Proceedings of the 2001 IEEE Conference on Global Telecommunications. 2001. 995-1001
    [68] Hu X-Y, Eleftheriou E, Arnold D M. Regular and Irregular Progressive Edge-Growth Tanner Graphs. IEEE Transactions on Information Theory, 2004, 51(1): 386-398
    [69] Yang M, Ryan W E, Li Y. Design of Efficiently Encoded Moderated-Length High-Rate Irregular LDPC Codes. IEEE Transactions on Communications, 2004, 52(4): 564-571
    [70] Sankaranarayanan S, Vasic B, Kurtes E M. A Systematic Construction of Irregular Low-Density Parity-Check Codes from Combinatorial Designs. In: Proceedings ofthe 2003 IEEE International Symposium on Information Theory. 2003. 1-1
    [71] Zarrinkhat P, Banihashemi A H. Hybrid Decoding of Irregular LDPC Codes. In: Proceedings of the 2005 International Symposium on Information Theory. 2005. 312-316
    [72] Luby M G, Mitzenmacher M, Shokrollahi M A, et al. Efficient Erasure Correcting Codes. IEEE Transactions on Information Theory, 2001, 47(2): 569-584
    [73] Woitaszek M, Tufo H M. Fault Tolerance of Tornado Codes for Archival Storage. In: Proceedings of the 15th IEEE International Symposium on High Performance Distributed Computing. 2006. 83-92
    [74] Nybom K, Bjorkqvist J. Designing Tornado Codes as Hyper Codes for Improving Error Correcting Performance. In: Proceedings of the Advanced International Conference on Telecommunications and International Conference on Internet and Web Applications and Services (AICT/CIW 2006). 2006. 41-46
    [75] MacKay D J C. Fountain codes. IEE Proc-Commun, 2005, 156(6): 1062-1068
    [76] Dimarkis A G, Prabhakaran V, Ramchandran K. Distributed Fountain Codes For Networked Storage. In: Proceedings of the 2006 IEEE International Conference on Acoustics, Speech and Signal Processing. 2006. 1149-1152
    [77] Plank J S, Buchsbaum A L, Collins R L, et al. Small Parity-Check Erasure Codes-Exploration and Obserations. In: Proceedings of the 2005 International Conference on Dependable Systems and Networks (DSN'05). 2005. 326-335
    [78] Collins R L, Plank J S. Accessing the Performance of Erasure Codes in the Wide-Area. In: Proceedings of the 2005 International Conference on Dependable Systems and Network (DSN'05). 2005. 182-187
    [79] Litwin W, Schwarz T. LH*RS: A High-Availability Scalable Distributed Data Structure using Reed Solomon Codes. In: Proceedings of the 2000 ACM SIGMOD International Conference on Management of Data. 2000. 237-248
    [80] Chang F, Ji M, Leung S-T A, et al. Myriad: Cost-effective Disaster Tolerance. In:Proceedings of the 1st USENIX Conference on File and Storage Technologies. 2002. 103-116
    [81] Aguilera M K, Janakiraman R, Xu L. Using Erasure Codes Efficiently for Storage in a Distributed System. In: Proceedings of the 2005 International Conference on Dependable Systems and Networks (DSN'05). 2005. 336-345
    [82] Roesler M, Burkhard W A. Semantic Lock Models in Object-oriented Distributed Systems and Deadlock Resolution. In: Proceedings of the 1988 ACM SIGMOD International Conference on Management of Data. Chicago, illinois, United States, 1988. 361-370
    [83] Sinha M K, Nandikar P D, Mehndiratta S L. Timestamp based Certification Schemes for Transactions in Distributed Database Systems. ACM SIGMOD Record, 1985, 14(4): 402-411
    [84] Chandra T D, Toueg S. Unreliable Failure Detectors for Reliable Distributed Systems. Journal of the ACM, 1996, 43(2): 225-267
    [85] Chen W, Toueg S, Aguilera M K. On the Quality of Service of Failure Detectors. IEEE Transactions on Computers, 2002, 51(5): 561-580
    [86] Nunes R C, Jansch-Porto I. QoS of Timeout-based Self-Tuned Failure Detectors: the Effects of the Communication Delay Predictor and the Safety Margin. In: Proceedings of the 2004 International Conference on Dependable Systems and Networks (DSN'04). 2004. 753-761
    [87] Bertier M, Marin O, Sens P. Implementation and Performance Evaluation of an Adaptable Failure Detector. In: Proceedings of the International Conference on Dependable Systems and Networks (DSN'02). 2002. 354-363
    [88] Jain A, Shyamasundar R K. Failure Detection and Membership Management in Grid Environment. In: Proceedings of the Fifth IEEE/ACM International Workshop on Grid Computing. 2004. 44-52
    [89] Hayashibara N, Cherif A. Failure Detectors for Large-Scale Distributed Systems.In: Proceedings of the 21st IEEE Symposium on Reliable Distributed Systems. 2002. 404-409
    [90] Horita Y, Taura K, Chikayarna T. A Scalable and Efficient Self-Organizing Failure Detector for Grid Applications. In: Proceedings of the 6th IEEE/ACM International workshop on Grid Computing. 2005. 202-210
    [91] Gupta I, Chandra T D, Goldszmidt G S. On Scalable and Efficient Distributed Failure Detectors. In: Proceedings of the Twentieth Annual ACM Symposium on Principles of Distributed Computing. 2001. 170-179
    [92] Mostefaoui A, Powell D, Raynal M. A Hybrid Approach for Building Eventually Accurate Failure Detectors. In: Proceedings of the 10th IEEE Pacific Rim International Symposium on Dependable Computing (PRDC'04). 2004. 57-65
    [93] Mostefaoui A, Rajsbaum S, Raynal M, et al. From Failure Detectors with Limited Scope Accuracy to System-wide Leadership. In: Proceedings of the 20th International Conference on Advanced Information Networking and Applications (AINA'06). 2006. 81-86
    [94] Raynal M. A Short Introduction to Failure Detectors for Asynchronous Distributed Systems. ACM SIGACT News, 2005, 36(1): 53-70
    [95]董剑,左德承,刘宏伟, et al.一种基于QoS自适应网格失效检测器.软件学报, 2006, 17(11): 2362-2372
    [96] Carson M, Santay D. NIST Net - A Linux-based Network Emulation Tool. http://www-x.antd.nist.gov/nistnet/nistnet.pdf.
    [97] Birman K P. The Process Group Approach to Reliable Distributed Computing. Communications of the ACM, 1993, 36(12): 37-53
    [98] Cristian F. Group, Majority, and Strict Agreement in Timed Asynchronous Distributed Systems. In: Proceedings of the 26th International Symposium on Fault Tolerance Computing. 1996. 178-187
    [99] Neiger G. A New Look at Membership Services. In: Proceedings of the FifteenthAnnual ACM Symposium on Principles of Distributed Computing. 1996. 331-340
    [100] Franceschetti M, Bruck J. A Group Membership Algorithm with a Practical Specification. IEEE Transactions on Parallel and Distributed Systems, 2001, 12(11): 1190-1200
    [101]汪芸.分布环境下容错组成员主动退出组行为的研究.中国科学(E辑), 2003, 33(12): 1077-1086
    [102]邓延军,徐学洲.支持扩展虚拟同步的组通讯算法形式描述与验证.计算机研究与发展, 2005, 42(4): 676-683
    [103] Khazan R I. Group Membership: A Novel Approach and the First Single-Round Algorithm. In: Proceedings of the 23 Annual ACM Symposium on Principles of Distributed Computing. 2004. 347-356
    [104] Schiper A, Toueg S. From Set Membership to Group Membership A Separation of Concerns. IEEE Transactions on Dependable and Secure Computing, 2006, 3(1): 2-12

© 2004-2018 中国地质图书馆版权所有 京ICP备05064691号 京公网安备11010802017129号

地址:北京市海淀区学院路29号 邮编:100083

电话:办公室:(+86 10)66554848;文献借阅、咨询服务、科技查新:66554700