Concept of parallel graph processing system for large-scale network science

Author(s):  
Mikhail Chernoskutov
2022 ◽  
Vol 27 (1) ◽  
pp. 1-30
Author(s):  
Mengke Ge ◽  
Xiaobing Ni ◽  
Xu Qi ◽  
Song Chen ◽  
Jinglei Huang ◽  
...  

Brain network is a large-scale complex network with scale-free, small-world, and modularity properties, which largely supports this high-efficiency massive system. In this article, we propose to synthesize brain-network-inspired interconnections for large-scale network-on-chips. First, we propose a method to generate brain-network-inspired topologies with limited scale-free and power-law small-world properties, which have a low total link length and extremely low average hop count approximately proportional to the logarithm of the network size. In addition, given the large-scale applications, considering the modularity of the brain-network-inspired topologies, we present an application mapping method, including task mapping and deterministic deadlock-free routing, to minimize the power consumption and hop count. Finally, a cycle-accurate simulator BookSim2 is used to validate the architecture performance with different synthetic traffic patterns and large-scale test cases, including real-world communication networks for the graph processing application. Experiments show that, compared with other topologies and methods, the brain-network-inspired network-on-chips (NoCs) generated by the proposed method present significantly lower average hop count and lower average latency. Especially in graph processing applications with a power-law and tightly coupled inter-core communication, the brain-network-inspired NoC has up to 70% lower average hop count and 75% lower average latency than mesh-based NoCs.


MIS Quarterly ◽  
2016 ◽  
Vol 40 (4) ◽  
pp. 849-868 ◽  
Author(s):  
Kunpeng Zhang ◽  
◽  
Siddhartha Bhattacharyya ◽  
Sudha Ram ◽  
◽  
...  

2014 ◽  
Vol 26 (7) ◽  
pp. 1377-1389 ◽  
Author(s):  
Bo-Cheng Kuo ◽  
Mark G. Stokes ◽  
Alexandra M. Murray ◽  
Anna Christina Nobre

In the current study, we tested whether representations in visual STM (VSTM) can be biased via top–down attentional modulation of visual activity in retinotopically specific locations. We manipulated attention using retrospective cues presented during the retention interval of a VSTM task. Retrospective cues triggered activity in a large-scale network implicated in attentional control and led to retinotopically specific modulation of activity in early visual areas V1–V4. Importantly, shifts of attention during VSTM maintenance were associated with changes in functional connectivity between pFC and retinotopic regions within V4. Our findings provide new insights into top–down control mechanisms that modulate VSTM representations for flexible and goal-directed maintenance of the most relevant memoranda.


Sign in / Sign up

Export Citation Format

Share Document