Comparison of high-performance packet processing frameworks on NUMA

Author(s):  
Haipeng Wang ◽  
Dazhong He ◽  
Huan Wang
2016 ◽  
Vol 51 (4) ◽  
pp. 67-81 ◽  
Author(s):  
Antoine Kaufmann ◽  
SImon Peter ◽  
Naveen Kr. Sharma ◽  
Thomas Anderson ◽  
Arvind Krishnamurthy

2016 ◽  
Vol 44 (2) ◽  
pp. 67-81 ◽  
Author(s):  
Antoine Kaufmann ◽  
SImon Peter ◽  
Naveen Kr. Sharma ◽  
Thomas Anderson ◽  
Arvind Krishnamurthy

2021 ◽  
Vol 13 (3) ◽  
pp. 78
Author(s):  
Chuanhong Li ◽  
Lei Song ◽  
Xuewen Zeng

The continuous increase in network traffic has sharply increased the demand for high-performance packet processing systems. For a high-performance packet processing system based on multi-core processors, the packet scheduling algorithm is critical because of the significant role it plays in load distribution, which is related to system throughput, attracting intensive research attention. However, it is not an easy task since the canonical flow-level packet scheduling algorithm is vulnerable to traffic locality, while the packet-level packet scheduling algorithm fails to maintain cache affinity. In this paper, we propose an adaptive throughput-first packet scheduling algorithm for DPDK-based packet processing systems. Combined with the feature of DPDK burst-oriented packet receiving and transmitting, we propose using Subflow as the scheduling unit and the adjustment unit making the proposed algorithm not only maintain the advantages of flow-level packet scheduling algorithms when the adjustment does not happen but also avoid packet loss as much as possible when the target core may be overloaded Experimental results show that the proposed method outperforms Round-Robin, HRW (High Random Weight), and CRC32 on system throughput and packet loss rate.


Author(s):  
Salvatore Di Girolamo ◽  
Andreas Kurth ◽  
Alexandru Calotoiu ◽  
Thomas Benz ◽  
Timo Schneider ◽  
...  

Electronics ◽  
2019 ◽  
Vol 9 (1) ◽  
pp. 59 ◽  
Author(s):  
Junnan Li ◽  
Zhigang Sun ◽  
Jinli Yan ◽  
Xiangrui Yang ◽  
Yue Jiang ◽  
...  

In the public cloud, FPGA-based SmartNICs are widely deployed to accelerate network functions (NFs) for datacenter operators. We argue that with the trend of network as a service (NaaS) in the cloud is also meaningful to accelerate tenant NFs to meet performance requirements. However, in pursuit of high performance, existing work such as AccelNet is carefully designed to accelerate specific NFs for datacenter providers, which sacrifices the flexibility of rapidly deploying new NFs. For most tenants with limited hardware design ability, it is time-consuming to develop NFs from scratch due to the lack of a rapidly reconfigurable framework. In this paper, we present a reconfigurable network processing pipeline, i.e., DrawerPipe, which abstracts packet processing into multiple “drawers” connected by the same interface. NF developers can easily share existing modules with other NFs and simply load core application logic in the appropriate “drawer” to implement new NFs. Furthermore, we propose a programmable module indexing mechanism, namely PMI, which can connect “drawers” in any logical order, to perform distinct NFs for different tenants or flows. Finally, we implemented several highly reusable modules for low-level packet processing, and extended four example NFs (firewall, stateful firewall, load balancer, IDS) based on DrawerPipe. Our evaluation shows that DrawerPipe can easily offload customized packet processing to FPGA with high performance up to 100 Mpps and ultra-low latency (<10 µs). Moreover, DrawerPipe enables modular development of NFs, which is suitable for rapid deployment of NFs. Compared with individual NF development, DrawerPipe reduces the line of code (LoC) of the four NFs above by 68%.


Sign in / Sign up

Export Citation Format

Share Document