Reduction of the Matrix Polynomial Decomposition of the Transfer Function to a Coprime form Using the Sylvester Matrix

Author(s):  
Alexander A. Voevoda ◽  
Kurbonmurod M. Bobobekov
2021 ◽  
Vol 71 (2) ◽  
pp. 301-316
Author(s):  
Reshma Sanjhira

Abstract We propose a matrix analogue of a general inverse series relation with an objective to introduce the generalized Humbert matrix polynomial, Wilson matrix polynomial, and the Rach matrix polynomial together with their inverse series representations. The matrix polynomials of Kiney, Pincherle, Gegenbauer, Hahn, Meixner-Pollaczek etc. occur as the special cases. It is also shown that the general inverse matrix pair provides the extension to several inverse pairs due to John Riordan [An Introduction to Combinatorial Identities, Wiley, 1968].


2021 ◽  
Vol 2021 (1) ◽  
Author(s):  
Adisorn Kittisopaporn ◽  
Pattrawut Chansangiam ◽  
Wicharn Lewkeeratiyutkul

AbstractWe derive an iterative procedure for solving a generalized Sylvester matrix equation $AXB+CXD = E$ A X B + C X D = E , where $A,B,C,D,E$ A , B , C , D , E are conforming rectangular matrices. Our algorithm is based on gradients and hierarchical identification principle. We convert the matrix iteration process to a first-order linear difference vector equation with matrix coefficient. The Banach contraction principle reveals that the sequence of approximated solutions converges to the exact solution for any initial matrix if and only if the convergence factor belongs to an open interval. The contraction principle also gives the convergence rate and the error analysis, governed by the spectral radius of the associated iteration matrix. We obtain the fastest convergence factor so that the spectral radius of the iteration matrix is minimized. In particular, we obtain iterative algorithms for the matrix equation $AXB=C$ A X B = C , the Sylvester equation, and the Kalman–Yakubovich equation. We give numerical experiments of the proposed algorithm to illustrate its applicability, effectiveness, and efficiency.


Author(s):  
Nikta Shayanfar ◽  
Heike Fassbender

The polynomial eigenvalue problem is to find the eigenpair of $(\lambda,x) \in \mathbb{C}\bigcup \{\infty\} \times \mathbb{C}^n \backslash \{0\}$ that satisfies $P(\lambda)x=0$, where $P(\lambda)=\sum_{i=0}^s P_i \lambda ^i$ is an $n\times n$ so-called matrix polynomial of degree $s$, where the coefficients $P_i, i=0,\cdots,s$, are $n\times n$ constant matrices, and $P_s$ is supposed to be nonzero. These eigenvalue problems arise from a variety of physical applications including acoustic structural coupled systems, fluid mechanics, multiple input multiple output systems in control theory, signal processing, and constrained least square problems. Most numerical approaches to solving such eigenvalue problems proceed by linearizing the matrix polynomial into a matrix pencil of larger size. Such methods convert the eigenvalue problem into a well-studied linear eigenvalue problem, and meanwhile, exploit and preserve the structure and properties of the original eigenvalue problem. The linearizations have been extensively studied with respect to the basis that the matrix polynomial is expressed in. If the matrix polynomial is expressed in a special basis, then it is desirable that its linearization be also expressed in the same basis. The reason is due to the fact that changing the given basis ought to be avoided \cite{H1}. The authors in \cite{ACL} have constructed linearization for different bases such as degree-graded ones (including monomial, Newton and Pochhammer basis), Bernstein and Lagrange basis. This contribution is concerned with polynomial eigenvalue problems in which the matrix polynomial is expressed in Hermite basis. In fact, Hermite basis is used for presenting matrix polynomials designed for matching a series of points and function derivatives at the prescribed nodes. In the literature, the linearizations of matrix polynomials of degree $s$, expressed in Hermite basis, consist of matrix pencils with $s+2$ blocks of size $n \times n$. In other words, additional eigenvalues at infinity had to be introduced, see e.g. \cite{CSAG}. In this research, we try to overcome this difficulty by reducing the size of linearization. The reduction scheme presented will gradually reduce the linearization to its minimal size making use of ideas from \cite{VMM1}. More precisely, for $n \times n$ matrix polynomials of degree $s$, we present linearizations of smaller size, consisting of $s+1$ and $s$ blocks of $n \times n$ matrices. The structure of the eigenvectors is also discussed.


2012 ◽  
Vol 2012 ◽  
pp. 1-28 ◽  
Author(s):  
Feng Yin ◽  
Guang-Xin Huang

An iterative algorithm is constructed to solve the generalized coupled Sylvester matrix equations(AXB-CYD,EXF-GYH)=(M,N), which includes Sylvester and Lyapunov matrix equations as special cases, over generalized reflexive matricesXandY. When the matrix equations are consistent, for any initial generalized reflexive matrix pair[X1,Y1], the generalized reflexive solutions can be obtained by the iterative algorithm within finite iterative steps in the absence of round-off errors, and the least Frobenius norm generalized reflexive solutions can be obtained by choosing a special kind of initial matrix pair. The unique optimal approximation generalized reflexive solution pair[X̂,Ŷ]to a given matrix pair[X0,Y0]in Frobenius norm can be derived by finding the least-norm generalized reflexive solution pair[X̃*,Ỹ*]of a new corresponding generalized coupled Sylvester matrix equation pair(AX̃B-CỸD,EX̃F-GỸH)=(M̃,Ñ), whereM̃=M-AX0B+CY0D,Ñ=N-EX0F+GY0H. Several numerical examples are given to show the effectiveness of the presented iterative algorithm.


2007 ◽  
Vol 2007 ◽  
pp. 1-16 ◽  
Author(s):  
Wang Guo-Sheng ◽  
Lv Qiang ◽  
Duan Guang-Ren

This paper considers the solution to a class of the second-order Sylvester matrix equationEVF2−AVF−CV=BW. Under the controllability of the matrix triple(E,A,B), a complete, general, and explicit parametric solution to the second-order Sylvester matrix equation, with the matrixFin a diagonal form, is proposed. The results provide great convenience to the analysis of the solution to the second-order Sylvester matrix equation, and can perform important functions in many analysis and design problems in control systems theory. As a demonstration, an illustrative example is given to show the effectiveness of the proposed solution.


1987 ◽  
Vol 109 (2) ◽  
pp. 120-123 ◽  
Author(s):  
S. G. Braun ◽  
Y. M. Ram

This paper deals with the fitting of a rational function to an experimentally determined frequency domain transfer function. Examples show that the use of overdetermined fitted systems improve the identification in noisy situations. A method is presented, enabling the determination of the number of zero/poles of the system, and the relation to the rank of the matrix used in a least square identification method.


Sign in / Sign up

Export Citation Format

Share Document