fast computation
Recently Published Documents


TOTAL DOCUMENTS

1239
(FIVE YEARS 148)

H-INDEX

52
(FIVE YEARS 5)

Sensors ◽  
2022 ◽  
Vol 22 (2) ◽  
pp. 428
Author(s):  
Vincent Sircoulomb ◽  
Houcine Chafouk

This paper presents a constrained Kalman filter for Wi-Fi-based indoor localization. The contribution of this work is to introduce constraints on the object speed and to provide a numerically optimized form for fast computation. The proposed approach is suitable to flexible space organization, as in warehouses, and when objects can be spun around, for example barcode readers in a hand. We experimented with the proposed technique using a robot and three devices, on five different journeys, in a 6000 m2 warehouse equipped with six Wi-Fi access points. The results highlight that the proposed approach provides a 19% improvement in localization accuracy.


2021 ◽  
Vol 2021 ◽  
pp. 1-12
Author(s):  
Yongho Kim ◽  
Gilnam Ryu ◽  
Yongho Choi

Simulation speed depends on code structures. Hence, it is crucial how to build a fast algorithm. We solve the Allen–Cahn equation by an explicit finite difference method, so it requires grid calculations implemented by many for-loops in the simulation code. In terms of programming, many for-loops make the simulation speed slow. We propose a model architecture containing a pad and a convolution operation on the Allen–Cahn equation for fast computation while maintaining accuracy. Also, the GPU operation is used to boost up the speed more. In this way, the simulation of other differential equations can be improved. In this paper, various numerical simulations are conducted to confirm that the Allen–Cahn equation follows motion by mean curvature and phase separation in two-dimensional and three-dimensional spaces. Finally, we demonstrate that our algorithm is much faster than an unoptimized code and the CPU operation.


Electronics ◽  
2021 ◽  
Vol 10 (23) ◽  
pp. 2960
Author(s):  
Youngbin Son ◽  
Seokwon Kang ◽  
Hongjun Um ◽  
Seokho Lee ◽  
Jonghyun Ham ◽  
...  

Most modern processors contain a vector accelerator or internal vector units for the fast computation of large target workloads. However, accelerating applications using vector units is difficult because the underlying data parallelism should be uncovered explicitly using vector-specific instructions. Therefore, vector units are often underutilized or remain idle because of the challenges faced in vector code generation. To solve this underutilization problem of existing vector units, we propose the Vector Offloader for executing scalar programs, which considers the vector unit as a scalar operation unit. By using vector masking, an appropriate partition of the vector unit can be utilized to support scalar instructions. To efficiently utilize all execution units, including the vector unit, the Vector Offloader suggests running the target applications concurrently in both the central processing unit (CPU) and the decoupled vector units, by offloading some parts of the program to the vector unit. Furthermore, a profile-guided optimization technique is employed to determine the optimal offloading ratio for balancing the load between the CPU and the vector unit. We implemented the Vector Offloader on a RISC-V infrastructure with a Hwacha vector unit, and evaluated its performance using a Polybench benchmark set. Experimental results showed that the proposed technique achieved performance improvements up to 1.31× better than the simple, CPU-only execution on a field programmable gate array (FPGA)-level evaluation.


Sign in / Sign up

Export Citation Format

Share Document