scholarly journals REVERT: A Network Failure Recovery Method for Data Center Networks

Electronics ◽  
2020 ◽  
Vol 9 (8) ◽  
pp. 1187
Author(s):  
Yunhe Cui ◽  
Qing Qian ◽  
Guowei Shen ◽  
Chun Guo ◽  
Saifei Li

As a repository that holds computing facilities, storage facilities, network facilities and other facilities, the Software Defined Data Center (SDDC) can provide computing and storage resources for users. For a SDDC, it is important to provide continuous services for users. Hence, in order to achieve high reliability in Software Defined Data Center Networks (SDDCNs), a network failure recovery method for software defined data center networks (REVERT) is proposed to recover failures in SDDCNs. In REVERT, the network failures that occurred in SDDCNs are classified into three types, which are switch failure, failure of links among switches and failure of links between switches and servers. Specially, except recovering the switch failure and failure of links between switches, REVERT can also recover the failures of links between the switches and servers. To achieve that, a failure preprocessing method used to classify the network failures, a data structure for storing and finding the affected flows, a server cluster agent for communicating with the server clustering algorithm and a routing path calculation method are designed in REVERT. Meanwhile, REVERT has been implemented and evaluated on RYU controller and Mininet using three routing algorithms. Compared with the link usage before recovering the network failures, when there are more than 200 flows in the network, the mean link usages only slightly increase at about 1.83 percent. More importantly, the evaluation results also demonstrate that except recovering switch failures, intra-topo link failures, REVERT has the ability of recovering failures of links between servers and edge switches successfully.

Author(s):  
Tariq Emad Ali ◽  
Ameer Hussein Morad ◽  
Mohammed A. Abdala

<span>In the last two decades, networks had been changed according to the rapid changing in its requirements.  The current Data Center Networks have large number of hosts (tens or thousands) with special needs of bandwidth as the cloud network and the multimedia content computing is increased. The conventional Data Center Networks (DCNs) are highlighted by the increased number of users and bandwidth requirements which in turn have many implementation limitations.  The current networking devices with its control and forwarding planes coupling result in network architectures are not suitable for dynamic computing and storage needs.  Software Defined networking (SDN) is introduced to change this notion of traditional networks by decoupling control and forwarding planes. So, due to the rapid increase in the number of applications, websites, storage space, and some of the network resources are being underutilized due to static routing mechanisms. To overcome these limitations, a Software Defined Network based Openflow Data Center network architecture is used to obtain better performance parameters and implementing traffic load balancing function. The load balancing distributes the traffic requests over the connected servers, to diminish network congestions, and reduce underutilization problem of servers. As a result, SDN is developed to afford more effective configuration, enhanced performance, and more flexibility to deal with huge network designs</span>


Author(s):  
Aymen Hasan Alawadi ◽  
Sándor Molnár

AbstractData center networks (DCNs) act as critical infrastructures for emerging technologies. In general, a DCN involves a multi-rooted tree with various shortest paths of equal length from end to end. The DCN fabric must be maintained and monitored to guarantee high availability and better QoS. Traditional traffic engineering (TE) methods frequently reroute large flows based on the shortest and least-congested paths to maintain high service availability. This procedure results in a weak link utilization with frequent packet reordering. Moreover, DCN link failures are typical problems. State-of-the-art approaches address such challenges by modifying the network components (switches or hosts) to discover and avoid broken connections. This study proposes Oddlab (Odds labels), a novel deployable TE method to guarantee the QoS of multi-rooted data center (DC) traffic in symmetric and asymmetric modes. Oddlab creatively builds a heuristic model for efficient flow scheduling and faulty link detection by exclusively using the gathered statistics from the DCN data plane, such as residual bandwidth and the number of installed elephant flows. Besides, the proposed method is implemented in an SDN-based DCN without altering the network components. Our findings indicate that Oddlab can minimize the flow completion time, maximize bisection bandwidth, improve network utilization, and recognize faulty links with sufficient accuracy to improve DC productivity.


2021 ◽  
pp. 1-10
Author(s):  
Linlin Zhang ◽  
Sujuan Zhang

In order to overcome the problems of long time and low accuracy of traditional methods, a cloud computing data center information classification and storage method based on group collaborative intelligent clustering was proposed. The cloud computing data center information is collected in real time through the information acquisition terminal, and the collected information is transmitted. The optimization function of information classification storage location was constructed by using the group collaborative intelligent clustering algorithm, and the optimal solutions of all storage locations were evolved to obtain the elite set. According to the information attribute characteristics, different information was allocated to different elite sets to realize the classified storage of information in the cloud computing data center. The experimental results show that the longest time of information classification storage is only 0.6 s, the highest information loss rate is 10.0%, and the highest accuracy rate is more than 80%.


2017 ◽  
Vol 25 (4) ◽  
pp. 1940-1953 ◽  
Author(s):  
Guo Chen ◽  
Youjian Zhao ◽  
Hailiang Xu ◽  
Dan Pei ◽  
Dan Li

2016 ◽  
Vol E99.B (11) ◽  
pp. 2361-2372 ◽  
Author(s):  
Chang RUAN ◽  
Jianxin WANG ◽  
Jiawei HUANG ◽  
Wanchun JIANG

Author(s):  
Jiawei Huang ◽  
Shiqi Wang ◽  
Shuping Li ◽  
Shaojun Zou ◽  
Jinbin Hu ◽  
...  

AbstractModern data center networks typically adopt multi-rooted tree topologies such leaf-spine and fat-tree to provide high bisection bandwidth. Load balancing is critical to achieve low latency and high throughput. Although the per-packet schemes such as Random Packet Spraying (RPS) can achieve high network utilization and near-optimal tail latency in symmetric topologies, they are prone to cause significant packet reordering and degrade the network performance. Moreover, some coding-based schemes are proposed to alleviate the problem of packet reordering and loss. Unfortunately, these schemes ignore the traffic characteristics of data center network and cannot achieve good network performance. In this paper, we propose a Heterogeneous Traffic-aware Partition Coding named HTPC to eliminate the impact of packet reordering and improve the performance of short and long flows. HTPC smoothly adjusts the number of redundant packets based on the multi-path congestion information and the traffic characteristics so that the tailing probability of short flows and the timeout probability of long flows can be reduced. Through a series of large-scale NS2 simulations, we demonstrate that HTPC reduces average flow completion time by up to 60% compared with the state-of-the-art mechanisms.


Sign in / Sign up

Export Citation Format

Share Document