Online midcourse guidance method for boost phase interception via adaptive convex programming

2021 ◽  
pp. 107037
Author(s):  
Biao Yang ◽  
Wuxing Jing ◽  
Changsheng Gao
Author(s):  
Mario A. Rotea ◽  
Pramod P. Khargonekar
Keyword(s):  

2021 ◽  
Vol 0 (0) ◽  
Author(s):  
Darina Dvinskikh ◽  
Alexander Gasnikov

Abstract We introduce primal and dual stochastic gradient oracle methods for decentralized convex optimization problems. Both for primal and dual oracles, the proposed methods are optimal in terms of the number of communication steps. However, for all classes of the objective, the optimality in terms of the number of oracle calls per node takes place only up to a logarithmic factor and the notion of smoothness. By using mini-batching technique, we show that the proposed methods with stochastic oracle can be additionally parallelized at each node. The considered algorithms can be applied to many data science problems and inverse problems.


2015 ◽  
Vol 20 (1) ◽  
pp. 457-468 ◽  
Author(s):  
Xiaosong Hu ◽  
Nikolce Murgovski ◽  
Lars Mardh Johannesson ◽  
Bo Egardt

Sign in / Sign up

Export Citation Format

Share Document