Inverse Scattering on the Half Line for the Matrix Schrödinger Equation

The matrix Schrödinger equation is considered on the half line with the general selfadjoint boundary condition at the origin described by two boundary matrices satisfying certain appropriate conditions. It is assumed that the matrix potential is integrable, is selfadjoint, and has a finite first mom...

Full description

Saved in:
Bibliographic Details
Date:2018
Main Authors: Aktosun, Tuncay, Weder, Ricardo
Format: Article
Language:English
Published: Фізико-технічний інститут низьких температур ім. Б.І. Вєркіна НАН України 2018
Series:Журнал математической физики, анализа, геометрии
Online Access:http://dspace.nbuv.gov.ua/handle/123456789/145874
Tags: Add Tag
No Tags, Be the first to tag this record!
Journal Title:Digital Library of Periodicals of National Academy of Sciences of Ukraine
Cite this:Inverse Scattering on the Half Line for the Matrix Schrödinger Equation / Tuncay Aktosun, Ricardo Weder // Журнал математической физики, анализа, геометрии. — 2018. — Т. 14, № 3. — С. 237-269. — Бібліогр.: 24 назв. — англ.

Institution

Digital Library of Periodicals of National Academy of Sciences of Ukraine
id irk-123456789-145874
record_format dspace
spelling irk-123456789-1458742019-02-03T01:22:59Z Inverse Scattering on the Half Line for the Matrix Schrödinger Equation Aktosun, Tuncay Weder, Ricardo The matrix Schrödinger equation is considered on the half line with the general selfadjoint boundary condition at the origin described by two boundary matrices satisfying certain appropriate conditions. It is assumed that the matrix potential is integrable, is selfadjoint, and has a finite first moment. The corresponding scattering data set is constructed, and such scattering data sets are characterized by providing a set of necessary and sufficient conditions assuring the existence and uniqueness of the one-toone correspondence between the scattering data set and the input data set containing the potential and boundary matrices. The work presented here provides a generalization of the classic result by Agranovich and Marchenko from the Dirichlet boundary condition to the general selfadjoint boundary condition. На пiвпрямiй розглянуто матричне рiвняння Шредiнгера iз загальною самоспряженою крайовою умовою в нулi, яка задається двома матрицями, що задовольняють певнi умови. Вважається, що матричний потенцiал є самоспряженим, iнтегровним та має скiнченний перший момент. Побудовано вiдповiдну множину даних розсiяння. Цю множину даних розсiювання характеризовано набором необхiдних i достатнiх умов, якi гарантують єдинiсть та взаємно однозначну вiдповiднiсть мiж множиною даних розсiяння та множиною вхiдних даних, яка мiстить потенцiал та крайовi матрицi. Ця робота надає узагальнення з крайової умови Дiрiхле на загальну самоспряжену крайову умову для класичного результату Аграновича та Марченка. 2018 Article Inverse Scattering on the Half Line for the Matrix Schrödinger Equation / Tuncay Aktosun, Ricardo Weder // Журнал математической физики, анализа, геометрии. — 2018. — Т. 14, № 3. — С. 237-269. — Бібліогр.: 24 назв. — англ. 1812-9471 DOI: https://doi.org/10.15407/mag14.03.237 Mathematics Subject Classification 2000: 34L25, 34L40, 81U05 http://dspace.nbuv.gov.ua/handle/123456789/145874 en Журнал математической физики, анализа, геометрии Фізико-технічний інститут низьких температур ім. Б.І. Вєркіна НАН України
institution Digital Library of Periodicals of National Academy of Sciences of Ukraine
collection DSpace DC
language English
description The matrix Schrödinger equation is considered on the half line with the general selfadjoint boundary condition at the origin described by two boundary matrices satisfying certain appropriate conditions. It is assumed that the matrix potential is integrable, is selfadjoint, and has a finite first moment. The corresponding scattering data set is constructed, and such scattering data sets are characterized by providing a set of necessary and sufficient conditions assuring the existence and uniqueness of the one-toone correspondence between the scattering data set and the input data set containing the potential and boundary matrices. The work presented here provides a generalization of the classic result by Agranovich and Marchenko from the Dirichlet boundary condition to the general selfadjoint boundary condition.
format Article
author Aktosun, Tuncay
Weder, Ricardo
spellingShingle Aktosun, Tuncay
Weder, Ricardo
Inverse Scattering on the Half Line for the Matrix Schrödinger Equation
Журнал математической физики, анализа, геометрии
author_facet Aktosun, Tuncay
Weder, Ricardo
author_sort Aktosun, Tuncay
title Inverse Scattering on the Half Line for the Matrix Schrödinger Equation
title_short Inverse Scattering on the Half Line for the Matrix Schrödinger Equation
title_full Inverse Scattering on the Half Line for the Matrix Schrödinger Equation
title_fullStr Inverse Scattering on the Half Line for the Matrix Schrödinger Equation
title_full_unstemmed Inverse Scattering on the Half Line for the Matrix Schrödinger Equation
title_sort inverse scattering on the half line for the matrix schrödinger equation
publisher Фізико-технічний інститут низьких температур ім. Б.І. Вєркіна НАН України
publishDate 2018
url http://dspace.nbuv.gov.ua/handle/123456789/145874
citation_txt Inverse Scattering on the Half Line for the Matrix Schrödinger Equation / Tuncay Aktosun, Ricardo Weder // Журнал математической физики, анализа, геометрии. — 2018. — Т. 14, № 3. — С. 237-269. — Бібліогр.: 24 назв. — англ.
series Журнал математической физики, анализа, геометрии
work_keys_str_mv AT aktosuntuncay inversescatteringonthehalflineforthematrixschrodingerequation
AT wederricardo inversescatteringonthehalflineforthematrixschrodingerequation
first_indexed 2025-07-10T22:46:30Z
last_indexed 2025-07-10T22:46:30Z
_version_ 1837301918639063040
fulltext Journal of Mathematical Physics, Analysis, Geometry 2018, Vol. 14, No. 3, pp. 237–269 doi: https://doi.org/10.15407/mag14.03.237 Inverse Scattering on the Half Line for the Matrix Schrödinger Equation Tuncay Aktosun and Ricardo Weder Dedicated to Professor V.A. Marchenko for his 95th birthday The matrix Schrödinger equation is considered on the half line with the general selfadjoint boundary condition at the origin described by two boundary matrices satisfying certain appropriate conditions. It is assumed that the matrix potential is integrable, is selfadjoint, and has a finite first moment. The corresponding scattering data set is constructed, and such scattering data sets are characterized by providing a set of necessary and sufficient conditions assuring the existence and uniqueness of the one-to- one correspondence between the scattering data set and the input data set containing the potential and boundary matrices. The work presented here provides a generalization of the classic result by Agranovich and Marchenko from the Dirichlet boundary condition to the general selfadjoint boundary condition. Key words: matrix Schrödinger equation, selfadjoint boundary condi- tion, Marchenko method, matrix Marchenko method, Jost matrix, scattering matrix, inverse scattering, characterization. Mathematical Subject Classification 2010: 34L25, 34L40, 81U05. 1. Introduction Our aim in this paper is to describe the direct and inverse scattering prob- lems for the half-line matrix Schrödinger operator with a selfadjoint boundary condition. In the direct problem we are given an input data set D consisting of an n× n matrix-valued potential V (x) and a selfadjoint boundary condition at x = 0, and our goal is to determine the corresponding scattering data set S consisting of the scattering matrix S(k) and the bound-state data. In the inverse problem, we are given a scattering data set S, and our goal is to determine the corresponding input data set D. We would like to have a one-to-one correspon- dence between an input data set D and a scattering data set S so that both the direct and inverse problems are well posed. Thus, some restrictions are needed on D and S for a one-to-one correspondence. Since the scattering and inverse scattering problems in the scalar case, i.e. when n = 1, are well understood, it is desirable that the analysis in the matrix c© Tuncay Aktosun and Ricardo Weder, 2018 https://doi.org/10.15407/mag14.03.237 238 Tuncay Aktosun and Ricardo Weder case reduces to the scalar case when n = 1. However, as elaborated in Section 8, the current formulation of the scattering and inverse scattering problems in the scalar case presents a problem. As a consequence, it becomes impossible to have a one-to-one correspondence between an input data set D and a scattering data set S, unless the Dirichlet and non-Dirichlet boundary conditions are analyzed separately and they are not mixed with each other. Although not ideal, this could perhaps be done in the scalar case because a given boundary condition in the scalar case is either Dirichlet or non-Dirichlet. On the other hand, in the matrix case with n ≥ 2, a given boundary condition may partly be Dirichlet and partly non-Dirichlet, and this may be as a result of constraints in a physical problem. It turns out that the proper way to deal with the issue is to modify the definition of the scattering matrix in such a way that it is defined the same way regardless of the boundary condition, i.e. one should avoid defining the scattering matrix in one way in the Dirichlet case and in another way in the non-Dirichlet case. There are four aspects related to the direct and inverse problems. These are the existence, uniqueness, construction, and characterization. In the existence aspect in the direct problem, given D in a specified class we determine whether a corresponding S exists in some specific class. The uniqueness aspect is concerned with whether there exists a unique S corresponding to a given D, or two or more distinct sets S may correspond to the same D. The construction deals with the recovery of S from D. In the inverse problem the existence problem deals with the existence of some D corresponding to a given S belonging to a particular class. The uniqueness deals with the question whether D corresponding to a given S is unique, and the construction consists of the recovery of D from S. After the existence and uniqueness aspects in the direct and inverse problems are settled, one then turns the attention to the characterization problem, which consists of the identification of the class to which D belongs and the identification of the class to which S belongs so that there is a one-to-one correspondence between D and S in the respective classes. One also needs to ensure that the scattering data set S uniquely constructed from a given D in the direct problem in turn uniquely constructs the same D in the inverse problem. A viable characterization in the literature for the matrix Schrödinger operator on the half line can be found in the seminal work by Agranovich and Marchenko [1]. However, the analysis in [1] is restricted to the Dirichlet boundary condition, and hence our study can be viewed as a generalization of the characterization in [1]. A characterization for the case of the general selfadjoint boundary condition is recently provided [9,10] by the authors of this paper, and in the current paper we present a summary of some of the results in [9, 10]. For brevity, we do not include any proofs because such proofs are already available in [9, 10]. We present the existence, uniqueness, reconstruction, and characterization issues related to the relevant direct and inverse problems under the assumption that D belongs to the Faddeev class and that S belongs to the Marchenko class. The Faddeev class consists of input data sets D as in (2.1), where the potential V (x) and the boundary matrices A and B are as specified in Definition 2.1. Inverse Scattering on the Half Line for the Matrix Schrödinger Equation 239 The Marchenko class consists of scattering data sets S as in (3.12), where the scattering matrix S(k) and the bound-state data {κj ,Mj}Nj=1 are as specified in Definition 4.1. Let us mention the relevant references [14–16], where the direct and inverse problems for (2.2) are formally studied with the general selfadjoint boundary condition, not as in (2.5)–(2.7) but in a form equivalent to (2.5)–(2.7). However, the study in [14–16] lacks the large-k analysis beyond the leading term and also lacks the small-k analysis of the scattering data, which are both essential for the analysis of the relevant inverse problem. Thus, our study can also be considered as a complement to the work by Harmer [14–16]. In our paper, which is essentially a brief summary of [9,10], we rely on results from previous work [1,4,5,8,22–24], in particular [1, 4, 8, 22]. Our paper is organized as follows. In Section 2 we introduce the matrix Schrödinger equation on the half line, describe the general selfadjoint boundary condition in terms of two constant matrices A and B. We then describe the Faddeev class of input data sets D consisting of the matrix potential V (x) and the boundary matrices A and B. In Section 3 we describe the solution to the direct problem, which uses an input data set D in the Faddeev class. We outline the construction of various quantities such as the Jost solution, the physical solution, the regular solution, the Jost matrix, the scattering matrix, and the bound-state data. In Section 4 we introduce the Marchenko class of scattering data sets. We present the solution to the inverse problem by starting with a scattering data set S in the Marchenko class, and we describe the construction of the potential and the boundary matrices. In Section 5 we provide a characterization of the scattering data by showing that there is a one-to-one correspondence between the Faddeev class of input data sets D and the Marchenko class of scattering data sets S. In Section 6 we provide an equivalent description of the Marchenko class, and we provide an alternate characterization of the scattering data with the help of Levinson’s theorem. In Section 7 we provide yet another description of the Marchenko class based on an approach utilizing the so-called generalized Fourier map. Finally, in Section 8 we contrast our definition of the Jost matrix and the scattering matrix with those definitions in the previous literature. We indicate the similarities and differences occurring when the boundary condition used is Dirichlet or non-Dirichlet. We elaborate on the resulting nonuniqueness issue if the scattering matrix is defined differently when the Dirichlet boundary condition is used, as commonly done in the previous literature. 2. The matrix Schrödinger equation In this section we introduce the matrix Schrödinger equation (2.2), the matrix potential V (x), and the boundary matrices A and B used to describe the general selfadjoint boundary condition. We also indicate that the boundary matrices A and B can be uniquely specified modulo a postmultiplication by an invertible matrix. Our input data set D is defined as D := {V,A,B}. (2.1) 240 Tuncay Aktosun and Ricardo Weder Consider the matrix Schrödinger equation on the half line − ψ′′ + V (x)ψ = k2 ψ, x ∈ R+, (2.2) where R+ := (0,+∞), the prime denotes the derivative with respect to the spatial coordinate x, k2 is the complex-valued spectral parameter, the potential V (x) is an n×n selfadjoint matrix-valued function of x and belongs to class L1 1(R +), and n is any positive integer. We assume that the value of n is fixed and is known. The selfadjointness of V (x) is expressed as V (x) = V (x)†, x ∈ R+, (2.3) where the dagger denotes the matrix adjoint (complex conjugate and matrix transpose). We equivalently say Hermitian to describe a selfadjoint matrix. We remark that, unless we are in the scalar case, i.e. unless n = 1, the potential is not necessarily real valued. The condition V ∈ L1 1(R +) means that each entry of the matrix V (x) is Lebesgue measurable on R+ and∫ ∞ 0 dx (1 + x) |V (x)| < +∞, (2.4) where |V (x)| denotes the matrix operator norm. Clearly, a matrix-valued function belongs to L1 1(R +) if and only if each entry of that matrix belongs to L1 1(R +). The wavefunction ψ(k, x) appearing in (2.2) may be either an n× n matrix- valued function or it may be a column vector with n components. We use C for the complex plane, R for the real line (−∞,+∞), R− for the left-half line (−∞, 0), C+ for the open upper-half complex plane, C+ for C+ ∪R, C− for the open lower-half complex plane, and C− for C− ∪R. We are interested in studying (2.2) with an n×n selfadjoint matrix potential V (x) in L1 1(R +) under the general selfadjoint boundary condition at x = 0. There are various equivalent formulations [4,8,14–18] of the general selfadjoint boundary condition at x = 0, and we find it convenient to state it [4, 8] in terms of two constant n× n matrices A and B as −B† ψ(0) +A† ψ′(0) = 0, (2.5) where A and B satisfy −B†A+A†B = 0, (2.6) A†A+B†B > 0. (2.7) The condition in (2.7) means that the n×n matrix (A†A+B†B) is positive, which is also called positive definite. One can easily verify that (2.5)–(2.7) remain invariant if the boundary matrices A and B are replaced with AT and BT, respectively, where T is an arbitrary n×n invertible matrix. We express this fact by saying that the selfadjoint boundary condition (2.5) is uniquely determined by the matrix pair (A,B) modulo an invertible matrix T, and we equivalently state that (2.5) is equivalent to the knowledge of (A,B) modulo T. We remark Inverse Scattering on the Half Line for the Matrix Schrödinger Equation 241 that the positivity condition (2.7) is equivalent to having the rank of the 2n× n matrix [ A B ] equal to n. In our analysis of the direct problem related to (2.2) and (2.5), we assume that our input data set D belongs to the Faddeev class defined below. Definition 2.1. The input data set D given in (2.1) is said to belong to the Faddeev class if the potential V (x) satisfies (2.3) and (2.4) and the boundary matrices A and B appearing in (2.5) satisfy (2.6) and (2.7). In other words, D belongs to the Faddeev class if the n×n matrix-valued potential V (x) appearing in (2.2) is Hermitian and belongs to class L1 1(R +) and the constant n×n matrices A and B appearing in (2.5) satisfy (2.6) and (2.7). It is possible to formulate the general selfadjoint boundary condition by using a unique n × n constant matrix instead of using the pair of matrices A and B appearing in (2.5)–(2.7). For example, in [15] a unitary n × n matrix U is used to describe the selfadjoint boundary condition as i 2 ( U † − I ) ψ(0) + 1 2 ( U † + I ) ψ′(0) = 0, where I is the n × n identity matrix. Without loss of any generality, one could also use [4] a diagonal representation of the selfadjoint boundary condition by choosing the matrices A and B as A = diag{− sin θ1, . . . ,− sin θn}, B = diag{cos θ1, . . . , cos θn}, (2.8) where the θj are some real constants in the interval (0, π]. In fact, through the representation (2.8), one can directly identify [5] the three integers nD, nN , and nM , where nD is the number of θj-values equal to π, nN is the number of θj- values equal to π/2, and nM is the number of θj-values in the union (0, π/2) ∪ (π/2, π). One can informally call nD the number of Dirichlet boundary conditions, nN the number of Neumann boundary conditions, and nM the number of mixed boundary conditions. We find it more convenient to write the general selfadjoint boundary condition in terms of the two constant n × n matrices A and B, with the understanding that A and B are unique up to a postmultiplication by an invertible matrix T . For example, the so-called Kirchhoff boundary condition is easier to recognize if expressed in terms of A and B, rather than written in terms of a single unique n× n constant matrix. 3. The solution to the direct problem In this section we summarize the solution to the direct scattering problem associated with (2.2) and (2.5) when the related input data set D given in (2.1) belongs to the Faddeev class. In other words, we start with an n× n Hermitian potential V (x) belonging to L1 1(R +) and a pair of constant boundary matrices A and B satisfying (2.6) and (2.7), and we construct the relevant quantities 242 Tuncay Aktosun and Ricardo Weder leading to the scattering data set S. The unique construction of the scattering data set S also enables us to determine the basic properties of S. The steps of the construction are given below: (a) When our input data set D belongs to the Faddeev class, regardless of the boundary matrices A and B, the matrix Schrödinger equation (2.2) has an n×n matrix-valued solution, usually called the Jost solution and denoted by f(k, x), satisfying the asymptotic condition f(k, x) = eikx [I + o(1)], x→ +∞. (3.1) The solution f(k, x) is uniquely determined by the potential V (x). For each fixed x ∈ [0,+∞), the Jost solution f(k, x) has an extension from k ∈ R to k ∈ C+, and such an extension is continuous in k ∈ C+ and analytic in k ∈ C+ and has the asymptotic behavior e−ikxf(k, x) = I + o(1), k →∞ in C+. (b) In terms of the boundary matrices A and B in D and the Jost solution f(k, x) obtained as in (a), we construct the Jost matrix J(k) as J(k) := f(−k∗, 0)†B − f ′(−k∗, 0)†A, k ∈ R, (3.2) where the asterisk denotes complex conjugation. We remark that J(k) is an n × n matrix-valued function of k. The redundant appearance of k∗ instead of k in (3.2) when k ∈ R is useful in extending the Jost matrix analytically from k ∈ R to k ∈ C+. We recall that the boundary matrices A and B can be postmultiplied by any invertible matrix T without affecting (2.5)–(2.7) and hence the definition given in (3.2) yields the Jost matrix J(k), which is unique up to a postmultiplication by T. (c) In terms of the Jost matrix J(k), obtained from D as indicated in (3.2), we construct the scattering matrix S(k) as S(k) := −J(−k) J(k)−1, k ∈ R. (3.3) We remark that S(k) is an n × n matrix-valued function of k. Even though the Jost matrix in (3.2) is uniquely determined up to a postmultiplication by an invertible matrix T, from (3.3) we see that the scattering matrix S(k) is uniquely determined irrespective of T. (d) In terms of the Jost solution f(k, x) obtained in (a) and the scattering matrix S(k) obtained in (c), we construct the so-called physical solution to (2.2). The physical solution, denoted by Ψ(k, x), is constructed as Ψ(k, x) := f(−k, x) + f(k, x)S(k), k ∈ R. (3.4) We remark that Ψ(k, x) is an n × n matrix-valued function of k and x. The physical solution, as the name implies, has the physical interpretation of a Inverse Scattering on the Half Line for the Matrix Schrödinger Equation 243 scattering solution; namely, the initial n×n matrix-valued plane wave e−ikxI sent from x = +∞ onto the potential yields the n×n matrix-valued scattered wave S(k) eikx at x = +∞ with the amplitude S(k). This interpretation is seen by using (3.1) in (3.4), i.e. for each fixed k ∈ R, we get Ψ(k, x) = e−ikx + S(k) eikx + o(1), x→ +∞. We also remark that each column of the physical solution satisfies the bound- ary condition (2.5), and hence the physical solution itself satisfies (2.5) and we have −B†Ψ(k, 0) +A†Ψ′(k, 0) = 0. (3.5) Even though the boundary matrices A and B appearing in (2.5)–(2.7) can be modified by a postmultiplication by an invertible matrix T, the definition given in (3.4) uniquely determines the physical solution irrespective of T. In the definition of the physical solution given in (3.4), one could multiply the right-hand side of (3.4) by a scalar function of k without affecting the physical interpretation of a physical solution. Nevertheless, we prefer to use (3.4) to define the physical solution in a unique manner. (e) Instead of constructing the physical solution via (3.4), one can alternatively construct it in an equivalent way as follows. When our input data set D belongs to the Faddeev class, there exists [4] an n×n matrix-valued solution to (2.1), called the regular solution and denoted by ϕ(k, x), satisfying the initial conditions ϕ(k, 0) = A, ϕ′(k, 0) = B. The solution ϕ(k, x) is uniquely determined by the input data set D given in (2.1). We remark that ϕ(k, x) depends on the choice of A and B. The solution ϕ(k, x) is known as the regular solution because it is entire in k for each fixed x ∈ [0,+∞). In terms of the regular solution ϕ(k, x) and the Jost matrix J(k) appearing in (3.2) we can introduce the physical solution as Ψ(k, x) = −2ik ϕ(k, x) J(k)−1. (3.6) One can show that the expressions given in (3.4) and (3.6) are equivalent, and this can be shown by using the relationship given in (3.5) of [4], i.e. ϕ(k, x) = 1 2ik f(k, x) J(−k)− 1 2ik f(−k, x) J(k), (3.7) where we recall that f(k, x) is the Jost solution appearing in (3.1). (f) When the input data set D belongs to the Faddeev class, the Jost matrix J(k) constructed as in (3.2) has an analytic extension from k ∈ R to k ∈ C+ and its determinant det[J(k)] is nonzero in C+ except perhaps at a finite number of k-values on the positive imaginary axis. Let us use N to denote the number of distinct zeros of det[J(k)] in C+ without counting multiplicities of those zeros, by realizing that the integer N may be zero for some input data sets D. 244 Tuncay Aktosun and Ricardo Weder Let us use N distinct positive numbers κj so that the zeros of det[J(k)] occur at k = iκj and use mj to denote the multiplicity of the zero of det[J(k)] at k = iκj . Thus, the nonnegative integer N, the set of distinct positive values {κj}Nj=1, and the set of positive integers {mj}Nj=1 are all uniquely determined by the input data set D. Each mj satisfies 1 ≤ mj ≤ n. It is appropriate to call N the number of bound states without counting the multiplicities. The nonnegative integer N defined as N := N∑ j=1 mj , (3.8) can be referred to as the number of bound states including the multiplicities. (g) Having determined the sets {κj}Nj=1 and {mj}Nj=1, let us use Ker[J(iκj) †] to denote the kernel of the n × n constant matrix J(iκj) †. Next, we construct the orthogonal projection matrices Pj onto Ker[J(iκj) †] for j = 1, . . . , N. The n× n matrices Pj are Hermitian and idempotent, i.e. P †j = Pj , P 2 j = Pj , j = 1, . . . , N. Furthermore, the rank of Pj is equal to mj . We then construct the constant n× n matrices Aj , Bj , and Mj defined as Aj := ∫ ∞ 0 dx f(iκj , x)† f(iκj , x), j = 1, . . . , N, Bj := (I − Pj) + Pj Aj Pj , j = 1, . . . , N, (3.9) Mj := B −1/2 j Pj , j = 1, . . . , N, (3.10) where f(k, x) is the Jost solution constructed in (a). We remark that when D belongs to the Faddeev class, the matrices Bj given in (3.9) are Hermitian and positive definite and hence the matrices B −1/2 j are well defined as positive definite matrices. Since each projection matrix Pj has rank mj , it follows from (3.10) that each matrix Mj is Hermitian, nonnegative, and has rank mj . The matrices Mj are usually called the bound-state normalization matrices. (h) When the input data set D belongs to the Faddeev class, at each k = iκj with j = 1, . . . , N, the Schrödinger equation (2.2) has mj linearly indepen- dent column-vector solutions, where each of those column-vector solutions is square integrable in x ∈ R+. It is possible to rearrange those mj linearly independent column-vector solutions to form an n× n matrix Ψj(x), in such a way that Ψj(x) can be uniquely constructed as Ψj(x) := f(iκj , x)Mj , j = 1, . . . , N, (3.11) where Mj is the n × n normalization matrix defined in (3.10). We can refer to Ψj(x) as the normalized bound-state matrix solution to (2.1) at k = iκj . Inverse Scattering on the Half Line for the Matrix Schrödinger Equation 245 We remark that each Ψj(x) satisfies the boundary condition (2.5) and has rank equal to mj . Having constructed all the relevant quantities starting with the input data set D, we now define the scattering data set S as S := {S, {κj ,Mj}Nj=1}, (3.12) where S denotes the scattering matrix S(k) for k ∈ R constructed as in (3.3), the N distinct positive constants κj are as described in (f), and the N Hermitian, nonnegative, rank-mj matrices Mj are as in (3.10). 4. The solution to the inverse problem In this section, given the scattering data set S in (3.12), our goal is to con- struct the input data set D given in (2.1), with the understanding that the potential V (x) is uniquely constructed and that the boundary matrices A and B are uniquely constructed up to a postmultiplication by an invertible matrix. The construction is given when S belongs to the so-called Marchenko class. We first present the construction and provide the definition of the Marchenko class at the end of the construction procedure. Later in the section we show that the Marchenko class can also be described in various equivalent ways. We summarize the steps in the construction of D from S as follows: (a) From the large-k asymptotics of the scattering matrix S(k), we determine the constant n× n matrix S∞ via S∞ := lim k→±∞ S(k), (4.1) and the constant n× n matrix G1 via S(k) = S∞ + G1 ik + o ( 1 k ) , k → ±∞. (4.2) (b) Using S(k) and S∞, we uniquely construct the n× n matrix Fs(y) via Fs(y) := 1 2π ∫ ∞ −∞ dk [S(k)− S∞] eiky, y ∈ R. (4.3) (c) Using Fs(y) constructed as in (4.3) and the bound-state data {κj ,Mj}Nj=1 appearing in S, we construct the n× n matrix F (y) via F (y) := Fs(y) + N∑ j=1 M2 j e −κjy, y ∈ R+. (4.4) Note that we have Fs(y) for y ∈ R, but we need F (y) only for y ∈ R+. 246 Tuncay Aktosun and Ricardo Weder (d) We use the matrix F (y) given in (4.4) as input to the Marchenko integral equation K(x, y) + F (x+ y) + ∫ ∞ x dz K(x, z)F (z + y) = 0, 0 ≤ x < y, (4.5) and uniquely solve (4.5) and obtain K(x, y) for 0 ≤ x < y < +∞. We remark that K(x, y) is continuous in the region 0 ≤ x < y < +∞. We note that K(0, 0), which is used to denote K(0, 0+), is well defined as a constant n×n matrix. (e) Having obtained K(x, y) for 0 ≤ x < y < +∞ uniquely from S as described in (d), we construct the potential V (x) via V (x) = −2 dK(x, x) dx , x ∈ R+. (4.6) By K(x, x) we mean K(x, x+). We remark that, in general, V (x) constructed as in (4.6) may exists only a.e. and it may not be continuous in x. (f) Having constructed the potential V (x) from the scattering data set S, we turn our attention to the construction of the boundary matrices A and B appearing in (2.1). We recall that we need to construct A and B uniquely, where the uniqueness is understood in the sense of being up to a postmultiplication by an arbitrary invertible n× n matrix T. Such a construction is carried out as follows. We use the already-constructed n×n constant matrices S∞, G1, and K(0, 0) as input in the linear, homogeneous algebraic system{ (I − S∞)A = 0, (I + S∞)B = [G1 − S∞K(0, 0)−K(0, 0)S∞]A, (4.7) and determine A and B as the general solution to (4.7). Such a general solution is equivalent to finding A and B satisfying (4.7) in such a way that the rank of the 2n× n matrix [ A B ] is equal to n. Even though the steps outlined above complete the construction of the input data set D from the scattering data set S, we can construct various auxiliary quantities relevant to the corresponding direct and inverse scattering problems as described below. (g) Having constructed the solution K(x, y) to the Marchenko integral equation (4.5), we obtain the Jost solution f(k, x) via f(k, x) = eikxI + ∫ ∞ x dyK(x, y) eiky. (4.8) (h) Having the Jost solution f(k, x) and the scattering matrix S(k) at hand, we construct the physical solution Ψ(k, x) as in (3.4). Inverse Scattering on the Half Line for the Matrix Schrödinger Equation 247 (i) Having the Jost solution f(k, x) and the boundary matrices A and B at hand, we construct Jost matrix J(k) as in (3.2). Note that the constructed A and B are unique up to a postmultiplication by an arbitrary invertible matrix T, and hence the constructed Jost matrix J(k) is also unique up to a postmultiplication by T. (j) Having the Jost solution f(k, x) and the Jost matrix J(k) at hand, we con- struct the regular solution ϕ(k, x) as in (3.7). Since the constructed A and B as well as the constructed J(k) are each unique up to a postmultiplication by an arbitrary invertible matrix T, the constructed regular solution ϕ(k, x) is also unique up to a postmultiplication by T. For each particular choice of the pair (A,B), we have a particular choice of the regular solution. (k) Having the Jost solution f(k, x) and the bound-state data {κj ,Mj}Nj=1 ap- pearing in S, we construct the normalized bound-state matrix solutions Ψj(x) as in (3.11). Next we define the Marchenko class of scattering data sets S. The importance of the Marchenko class is that there exists [9, 10] a one-to-one correspondence between the Faddeev class of input data sets D and the Marchenko class of scattering data sets S. Definition 4.1. Consider a scattering data set S as in (3.12), which consists of an n×n scattering matrix S(k) for k ∈ R, a set of N distinct positive constants κj , and a set of N constant n× n Hermitian and nonnegative matrices Mj with respective positive ranks mj , where N is a nonnegative integer. In case N = 0, it is understood that S consists only of S(k) for k ∈ R. We say that S belongs to the Marchenko class if S satisfies the following four conditions, listed below as (1), (2), (3a), (4a): (1) The scattering matrix S(k) satisfies S(−k) = S(k)† = S(k)−1, k ∈ R, (4.9) and there exist constant n×n matrices S∞ and G1 in such a way that (4.2) holds. Furthermore, the n × n matrix quantity Fs(y) defined in (4.3) is bounded in y ∈ R and integrable in y ∈ R+. (2) For the matrix Fs(y) defined in (4.3), the derivative F ′s(y) exists a.e. for y ∈ R+ and it satisfies ∫ ∞ 0 dy (1 + y) |F ′s(y)| < +∞, (4.10) where we recall that the norm in the integrand of (4.10) is the operator norm of a matrix. (3a) The physical solution Ψ(k, x) satisfies the boundary condition (2.5), i.e. it satisfies (3.5). We clarify this property as follows: The scattering matrix 248 Tuncay Aktosun and Ricardo Weder appearing in S yields a particular n × n matrix-valued solution Ψ(k, x) to (2.2) known as the physical solution given in (3.4) and also yields a pair of matrices A and B (modulo an invertible matrix) satisfying (2.6) and (2.7). Our statement (3a) is equivalent to saying that (2.5) is satisfied if we use in (2.5) the quantities Ψ(k, x), A, and B constructed from S. (4a) The Marchenko equation (4.5) at x = 0 given by K(0, y) + F (y) + ∫ ∞ 0 dz K(0, z)F (z + y) = 0, y ∈ R+, has a unique solution K(0, y) in L1(R+). Here, F (y) is the n × n matrix related to Fs(y) as in (4.4). Let us mention a slight drawback in the description of the Marchenko class given in Definition 4.1. The property (3a) cannot be checked from the scattering data set S directly because it requires the construction of the corresponding boundary matrices A and B as well as the physical solution Ψ(k, x). It is already known [9,10] that one can replace (3a) by an equivalent pair of conditions, listed as (IIIa) and (Vc) as indicated in the next theorem. Theorem 4.2. Consider a scattering data set S as in (3.12), which consists of an n×n scattering matrix S(k) for k ∈ R, a set of N distinct positive constants κj , and a set of N constant n× n Hermitian and nonnegative matrices Mj with respective positive ranks mj , where N is a nonnegative integer. In case N = 0, it is understood that S consists only of S(k) for k ∈ R and that N appearing in (3.8) is zero. The scattering data set S belongs to the Marchenko class if and only if S satisfies the five conditions, three of which are listed as (1), (2), and (4a) in Definition 4.1, and the two additional conditions (IIIa) and (Vc) are given as: (IIIa) For the matrix-valued function Fs(y) given in (4.3), the derivative F ′s(y) for y ∈ R− can be written as a sum of two matrix-valued functions, one of which is integrable and the other is square integrable in y ∈ R−. Further- more, the only solution X(y), which is a row vector with n square-integrable components in y ∈ R−, to the linear homogeneous integral equation −X(y) + ∫ 0 −∞ dz X(z)Fs(z + y) = 0, y ∈ R−, is the trivial solution X(y) ≡ 0. (Vc) The linear homogeneous integral equation X(y) + ∫ ∞ 0 dz X(z)Fs(z + y) = 0, y ∈ R+, (4.11) has precisely N linearly independent row-vector solutions for some non- negative integer N , with n components which are integrable in y ∈ R+. Here Fs(y) is the matrix defined in (4.3) and N is the nonnegative integer readily constructed from S as in (3.8). If N = 0, it is understood that the only solution in L1(R+) to (4.11) is the trivial solution X(y) ≡ 0. Inverse Scattering on the Half Line for the Matrix Schrödinger Equation 249 We remark that Theorem 4.2 is a special case of Theorem 4.5, but we still prefer to state it as a separate result. This is because Theorem 4.2 is closely related to the characterization result stated by Agranovich and Marchenko in the Dirichlet case on pp. 4–5 of their manuscript [1]. The next theorem shows that in the description of the Marchenko class spec- ified in Definition 4.1, we can replace the condition (3a) with another equivalent condition. Theorem 4.3. Consider a scattering data set S as in (3.12), which consists of an n×n scattering matrix S(k) for k ∈ R, a set of N distinct positive constants κj , and a set of N constant n× n Hermitian and nonnegative matrices Mj with respective positive ranks mj , where N is a nonnegative integer. In case N = 0, it is understood that S consists only of S(k) for k ∈ R. The scattering data set S belongs to the Marchenko class if and only if S satisfies the four conditions, three of which are listed as (1), (2), and (4a) in Definition 4.1 and one additional condition (3b) replacing (3a), which is given by (3b) The Jost matrix J(k) satisfies J(−k) + S(k) J(k) = 0, k ∈ R. (4.12) We clarify this property as follows: The scattering matrix S(k) given in S yields a Jost matrix J(k) constructed as in (3.2), unique up to a postmulti- plication by an invertible matrix. Using the scattering matrix S(k) given in S and the Jost matrix constructed from S(k), we find that (4.12) is satisfied. Let us use L̂1(C+) to denote the Banach space of all complex-valued functions ξ(k) that are analytic in k ∈ C+ in such a way that there exists a corresponding function η(x) belonging to L1(R+) satisfying ξ(k) = ∫ ∞ 0 dx η(x) eikx. We note that if ξ(k) belongs to L̂1(C+), then ξ(k) is continuous in k ∈ R and it satisfies ξ(k) = o(1) as k → ∞ in C+. If ξ(k) is vector valued or matrix valued instead of being scalar valued, then it belongs to L̂1(C+) if and only if each entry of ξ(k) belongs to L̂1(C+). We remark that the result of Theorem 4.3 is included in the next theorem presented. However, we have stated Theorem 4.3 separately in order to emphasize the importance of (3b) of Theorem 4.3 and its connection to (3a) of Definition 4.1. Theorem 4.4. Consider a scattering data set S as in (3.12), which consists of an n×n scattering matrix S(k) for k ∈ R, a set of N distinct positive constants κj , and a set of N constant n× n Hermitian and nonnegative matrices Mj with respective positive ranks mj , where N is a nonnegative integer. In case N = 0, it is understood that S consists only of S(k) for k ∈ R. The scattering data set S belongs to the Marchenko class if and only if S satisfies the four conditions (1), (2), (3), and (4), where (3) can be either one of (3a) and (3b); and (4) can be 250 Tuncay Aktosun and Ricardo Weder any one of (4a), (4b), (4c), (4d), (4e). Note that (1), (2), (3a), (4a) are listed in Definition 4.1; (3b) is listed in Theorem 4.3; and the remaining conditions (4b), (4c), (4d), (4e) are listed below: (4b) The only solution in L1(R+) to the homogeneous Marchenko integral equa- tion at x = 0 given by K(0, y) + ∫ ∞ 0 dz K(0, z)F (z + y) = 0, y ∈ R+, (4.13) is the trivial solution K(0, y) ≡ 0. Note that (4.13) is the homogeneous version at x = 0 of the Marchenko equation given by (4.5). We remark that F (y) appearing in (4.13) is the quantity defined in (4.4). (4c) The only integrable solution X(y), which is a row vector with n integrable components in y ∈ R+, to the linear homogeneous integral equation X(y) + ∫ ∞ 0 dz X(z)F (z + y) = 0, y ∈ R+, (4.14) is the trivial solution X(y) ≡ 0. Again, we recall that F (y) is the quantity defined in (4.4). (4d) The only solution X̂(k) to the system{ X̂(iκj)Mj = 0 , j = 1, . . . , N, X̂(−k) + X̂(k)S(k) = 0 , k ∈ R, (4.15) where X̂(k) is a row vector with n components belonging to the class L̂1(C+), is the trivial solution X̂(k) ≡ 0. (4e) The only solution h(k) to the system{ Mj h(iκj) = 0 , j = 1, . . . , N, h(−k) + S(k)h(k) = 0 , k ∈ R, (4.16) where h(k) is a column vector with n components belonging to the class L̂1(C+), is the trivial solution h(k) ≡ 0. We use H2(C±) to denote the Hardy space of all complex-valued functions ξ(k) that are analytic in k ∈ C± with a finite norm defined as ‖ξ‖H2(C±) := sup ρ>0 [∫ ∞ −∞ dα |ξ(α± iρ)|2 ]1/2 . Thus, ξ(k) is square integrable along all lines in C± that are parallel to the real axis. The value of ξ(k) for k ∈ R is defined to be the non-tangential limit of ξ(k± iρ) as ρ→ 0+. Such a non-tangential limit exists a.e. in k ∈ R and belongs Inverse Scattering on the Half Line for the Matrix Schrödinger Equation 251 to L2(R). It is known that ξ(k) belongs to H2(C+) if and only if there exists a corresponding function η(x) belonging to L2(R+) in such a way that ξ(k) = ∫ ∞ 0 dx η(x) eikx. Similarly, ξ(k) belongs to H2(C−) if and only if there exists a corresponding function η(x) belonging to L2(R−) in such a way that ξ(k) = ∫ 0 −∞ dx η(x) eikx. If ξ(k) is vector valued or matrix valued instead of being scalar valued, then it belongs to H2(C±) if and only if each entry of ξ(k) belongs to H2(C±). The next theorem shows that in the equivalent description of the Marchenko class specified in Theorem 4.2, we can replace the condition (IIIa) with one of two other equivalent conditions and we can also replace the condition (Vc) with any one of various other equivalent conditions. Theorem 4.5. Consider a scattering data set S as in (3.12), which consists of an n×n scattering matrix S(k) for k ∈ R, a set of N distinct positive constants κj , and a set of N constant n× n Hermitian and nonnegative matrices Mj with respective positive ranks mj , where N is a nonnegative integer. In case N = 0, it is understood that S consists only of S(k) for k ∈ R and that N appearing in (3.8) is zero. The scattering data set S belongs to the Marchenko class if and only if S satisfies the five conditions (1), (2), (III), (4), and (V), where (III) represents any one of the three conditions (IIIa), (IIIb), (IIIc); (4) represents any one of the five conditions (4a), (4b), (4c), (4d), (4e); and (V) represents any one of the eight conditions (Va), (Vb), (Vc), (Vd), (Ve), (Vf ), (Vg), (Vh). We remark that (1), (2), and (4a) are listed in Definition 4.1; (IIIa) and (Vc) are listed in Theorem 4.2; (4b), (4c), (4d), (4e) are listed in Theorem 4.4; and the remaining conditions are listed below: (IIIb) For the matrix-valued function Fs(y) given in (4.3), the derivative F ′s(y) for y ∈ R− can be written as a sum of two matrix-valued functions, one of which is integrable and the other is square integrable in y ∈ R−. Fur- thermore, the only solution X̂(k) to the homogeneous Riemann–Hilbert problem −X̂(−k) + X̂(k)S(k) = 0, k ∈ R, where X̂(k) is a row vector with n components belonging to the class H2(C−), is the trivial solution X̂(k) ≡ 0. (IIIc) For the matrix-valued function Fs(y) given in (4.3), the derivative F ′s(y) for y ∈ R− can be written as a sum of two matrix-valued functions, one of which is integrable and the other is square integrable in y ∈ R−. Further- more, the only solution h(k) to the homogeneous Riemann–Hilbert problem − h(−k) + S(k)h(k) = 0, k ∈ R, (4.17) 252 Tuncay Aktosun and Ricardo Weder where h(k) is a column vector with n components belonging to the class H2(C−), is the trivial solution h(k) ≡ 0. (Va) Each of the N normalized bound-state matrix solutions Ψj(x) constructed as in (3.11) satisfies the boundary condition (2.5), i.e. −B†Ψj(0) +A†Ψ′j(0) = 0, j = 1, . . . , N. (4.18) We clarify this statement as follows. The scattering matrix S(k) and the bound-state data {κj ,Mj}Nj=1 given in S yield n × n matrices Ψj(x) as in (3.11), where each Ψj(x) is a solution to (2.2) at k = iκj . As stated in (3a) of Definition 4.1, the scattering matrix given in S yields a pair of matrices A and B (modulo an invertible matrix) satisfying (2.6) and (2.7). The statement (Va) is equivalent to saying that (2.5) is satisfied if we use in (2.5) the quantities Ψj(x), A, and B constructed from the quantities appearing in S. If N = 0, then the condition (4.18) is redundant. (Vb) The normalization matrices Mj appearing in S satisfy J(iκj) †Mj = 0, j = 1, . . . , N. (4.19) We clarify this condition as follows. As indicated in (3b) of Theorem 4.3, the scattering matrix S(k) given in S yields a Jost matrix J(k). Using in (4.19) the matrix Mj given in S and the Jost matrix constructed from S(k), at each κj-value listed in S the matrix equation (4.19) holds. If N = 0, then the condition (4.19) is redundant. (Vd) The homogeneous Riemann–Hilbert problem given by X̂(−k) + X̂(k)S(k) = 0, k ∈ R, (4.20) has precisely N linearly independent row-vector solutions with n compo- nents in L̂1(C+). Here, N is the nonnegative integer given in (3.8). If N = 0, it is understood that the only solution in L̂1(C+) to (4.20) is the trivial solution X̂(k) ≡ 0. (Ve) The homogeneous Riemann–Hilbert problem given by h(−k) + S(k)h(k) = 0, k ∈ R, (4.21) has precisely N linearly independent column-vector solutions with n com- ponents in L̂1(C+), where N is the nonnegative integer given in (3.8). If N = 0, it is understood that the only solution in L̂1(C+) to (4.21) is the trivial solution h(k) ≡ 0. (Vf) The integral equation (4.11) has precisely N linearly independent row- vector solutions X(y) with n components in L2(R+), where N is the non- negative integer given in (3.8). If N = 0, it is understood that the only solution in L2(R+) to (4.11) is the trivial solution X(y) ≡ 0. We remark that the matrix Fs(y) appearing in the kernel of (4.11) is defined in (4.3). Inverse Scattering on the Half Line for the Matrix Schrödinger Equation 253 (Vg) The homogeneous Riemann–Hilbert problem given in (4.20) has precisely N linearly independent row-vector solutions X̂(k) with n components in H2(C+), Here, N is the nonnegative integer given in (3.8). If N = 0, it is understood that the only solution in H2(C+) to (4.20) is the trivial solution X̂(k) ≡ 0. (Vh) The homogeneous Riemann–Hilbert problem given in (4.21) has pre- cisely N linearly independent row-vector solutions with n components in H2(C+). Here, N is the nonnegative integer given in (3.8). If N = 0, it is understood that the only solution in H2(C+) to (4.21) is the trivial solution h(k) ≡ 0. 5. The characterization of the scattering data In this section we consider the characterization of the scattering data. In the next theorem we present one of our main characterization results. It shows that the four conditions given in Definition 4.1 for the Marchenko class form a characterization of the scattering data sets S so that there exists a one-to-one correspondence between a scattering data set in the Marchenko class and an input data set D in the Faddeev class specified in Definition 2.1. From Section 4 we know that the Marchenko class can be described in various equivalent ways, and hence it is possible to present the characterization in various different ways. Theorem 5.1. Consider a scattering data set S as in (3.12), which consists of an n×n scattering matrix S(k) for k ∈ R, a set of N distinct positive constants κj , and a set of N constant n× n Hermitian and nonnegative matrices Mj with respective positive ranks mj , where N is a nonnegative integer. In case N = 0, it is understood that S consists only of S(k) for k ∈ R and that N appearing in (3.8) is zero. Consider also an input data set D as in (2.1) consisting of an n× n matrix potential V (x) satisfying (2.3) and (2.4) and a pair of constant n × n matrices A and B satisfying (2.6) and (2.7). Then, we have the following: (a) For each input data set D in the Faddeev class specified in Definition 2.1, there exists and uniquely exists a scattering data set S in the Marchenko class specified in Definition 4.1. (b) Conversely, for each S in the Marchenko class, there exists and uniquely exists an input data set D in the Faddeev class, where the boundary matrices A and B are uniquely determined up to a postmultiplication by an invertible n× n matrix T. (c) Let S̃ be the scattering data set corresponding to D given in the previous step (b), where D is constructed from the scattering data set S. Then, we have S̃ = S, i.e. the scattering data set constructed from D is equal to the scattering data set used to construct D. (d) The characterization outlined in the steps (a)–(c) given above can equivalently be stated as follows. A set S as in (3.12) is the scattering data set correspond- 254 Tuncay Aktosun and Ricardo Weder ing to an input data set D in the Faddeev class if and only if S satisfies (1), (2), (3a), and (4a) stated in Definition 4.1. (e) The characterization outlined in the steps (a)–(c) given above can equivalently be stated as follows. A set S as in (3.12) is the scattering data set correspond- ing to an input data set D in the Faddeev class if and only if S satisfies (1), (2), (4a) of Definition 4.1 and (IIIa) and (Vc) of Theorem 4.2. (f) The characterization outlined in the steps (a)–(c) given above can equivalently be stated as follows. A set S as in (3.12) is the scattering data set correspond- ing to an input data set D in the Faddeev class if and only if S satisfies (1), (2), (3), and (4), where (3) can be either one of (3a) and (3b); and (4) can be any one of (4a), (4b), (4c), (4d), (4e). We recall that (1), (2), (3a), (4a) are listed in Definition 4.1; (3b) is listed in Theorem 4.3; and (4b), (4c), (4d), (4e) are listed in Theorem 4.4. (g) The characterization outlined in the steps (a)–(c) given above can equivalently be stated as follows. A set S as in (3.12) is the scattering data set correspond- ing to an input data set D in the Faddeev class if and only if S satisfies the five conditions (1), (2), (III), (4), and (V), where (III) can be any one of (IIIa), (IIIb), (IIIc); (4) can be any one of (4a), (4b), (4c), (4d), (4e); and (V) can be any one of (Va), (Vb), (Vc), (Vd), (Ve), (Vf ), (Vg), (Vh). We recall that (1), (2), (4a) are listed in Definition 4.1; (IIIa) and (Vc) are listed in Theorem 4.2; (4b), (4c), (4d), (4e) are listed in Theorem 4.4; and (IIIb), (IIIc), (Va), (Vb), (Vd), (Ve), (Vf ), (Vg), (Vh) are listed in Theorem 4.5. We have the following remarks on the results presented in Theorem 5.1. The characterization result stated in Theorem 5.1(e) follows from Theorem 4.2. The result stated in Theorem 5.1(f) is a consequence of Theorem 4.4. The result in Theorem 5.1(e) is a particular case of the result in Theorem 5.1(g), but we prefer to state it separately because it resembles the characterization result stated by Agranovich and Marchenko [1] in the Dirichlet case. Finally we remark that Theorem 5.1(g) is a direct consequence of Theorem 4.5. 6. An alternate characterization of the scattering data It is possible to present an alternate characterization of the scattering data us- ing Levinson’s theorem. This characterization again establishes a one-to-one cor- respondence between the Faddeev class of input data sets D and the Marchenko class of scattering data sets S. Hence, such an alternate characterization can also be viewed as an alternate description of the Marchenko class of scattering data sets with the help of Levinson’s theorem. In general, the bound-state data {κj ,Mj}Nj=1 appearing in the scattering data set S of (3.12) and the scattering matrix S(k) are independent, and they need to be specified separately. On the other hand, the determinant of the scattering matrix contains the information of the number of bound states including the multiplicities, which is the nonnegative integer N appearing in (3.8). The change Inverse Scattering on the Half Line for the Matrix Schrödinger Equation 255 in the argument of the determinant of S(k) as k changes from k = 0+ to k = +∞ in the k-interval (0,+∞) is related to the number of bound states including multiplicities. This general fact is usually known as Levinson’s theorem. When the input data set D belongs to the Faddeev class, we have [8] Levin- son’s theorem stated in the following. Theorem 6.1. Consider the matrix Schrödinger equation (2.2) with the self- adjoint boundary condition (2.5). Assume that the corresponding input data set D given in (2.1) belongs to the Faddeev class. Let S appearing in (3.12) be the scattering data set corresponding to D. Then, the number N of bound states including the multiplicities appearing in (3.8) is related to the argument of the determinant of the scattering matrix S(k) as arg [ det[S(0+)] ] − arg [det[S(+∞)]] = π (2N + µ+ nD − n) , (6.1) where µ is the (algebraic and geometric) multiplicity of the eigenvalue +1 of the zero-energy scattering matrix S(0), n is the positive integer appearing in the matrix size n×n of the scattering matrix S(k), and nD is the number of Dirichlet boundary conditions in the diagonal representation (2.8) of the boundary matrices A and B. We remark that nD is the same as the nonzero integer which is equal to the multiplicity of the eigenvalue −1 of the constant n× n matrix S∞ appearing in (4.1). In some cases, by using Levinson’s theorem we may be able to quickly de- termine if a given scattering data set S does not belong to the Marchenko class. Using the scattering matrix S(k), we readily know the positive integer n appear- ing in the matrix size n×n of the matrix S(k). The zero-energy scattering matrix S(0) has eigenvalues equal to either −1 or +1. Thus, we can identify µ as the multiplicity of the eigenvalue +1 of S(0). From the large-k limit of S(k) given in (4.1) we can easily construct the constant matrix S∞ and we already know that S∞ has eigenvalues equal to either −1 or +1. Thus, we can identify nD as the multiplicity of the eigenvalue −1 of S∞. Then, from the scattering matrix S(k) we can evaluate the change in the argument of det[S(k)] given on the left-hand side of (6.1). We can then use (6.1) to determine the value of N predicted by Levinson’s theorem. If that value of N evaluated from (6.1) does not turn out to be a nonnegative integer, we know that the corresponding S does not belong to the Marchenko class. The next theorem shows that we can obtain an equivalent description of the Marchenko class specified in Definition 4.1, by replacing (3a) by a set of two conditions one of which is related to Levinson’s theorem, and at the same time by replacing (4a) by any one of three other conditions. Theorem 6.2. Consider a scattering data set S as in (3.12), which consists of an n×n scattering matrix S(k) for k ∈ R, a set of N distinct positive constants κj , and a set of N constant n× n Hermitian and nonnegative matrices Mj with respective positive ranks mj , where N is a nonnegative integer. In case N = 0, it is understood that S consists only of S(k) for k ∈ R and that N appearing 256 Tuncay Aktosun and Ricardo Weder in (3.8) is zero. The scattering data set S belongs to the Marchenko class if and only if S satisfies the five conditions, two of which are listed as (1) and (2) in Definition 4.1, the third and the fourth are the respective conditions listed as (L) and (5◦) below, and the fifth is any one of the three conditions listed as (4◦c), (4◦d), and (4◦e) below: (L) The scattering matrix S(k) appearing in S is continuous for k ∈ R, and the equality (6.1) of Levinson’s theorem is satisfied with µ, nD, and N coming from S. Here, µ is the (algebraic and geometric) multiplicity of the eigen- value +1 of the zero-energy scattering matrix S(0), nD is the (algebraic and geometric) multiplicity of the eigenvalue −1 of the Hermitian matrix S∞ appearing in (4.1), and N is the nonnegative integer in (3.8) which is equal to the sum of the ranks mj of the matrices Mj appearing in S. (4◦c) The only square-integrable solution X(y), which is a row vector with n square-integrable components in y ∈ R+, to the linear homogeneous integral equation X(y) + ∫ ∞ 0 dz X(z)F (z + y) = 0, y ∈ R+, (6.2) is the trivial solution X(y) ≡ 0. Here, F (y) is the quantity defined in (4.4). (4◦d) The only solution X̂(k) to the system{ X̂(iκj)Mj = 0 , j = 1, . . . , N, X̂(−k) + X̂(k)S(k) = 0 , k ∈ R, (6.3) where X̂(k) is a row vector with n components belonging to the Hardy space H2(C+), is the trivial solution X̂(k) ≡ 0. (4◦e) The only solution h(k) to the system{ Mj h(iκj) = 0 , j = 1, . . . , N, h(−k) + S(k)h(k) = 0 , k ∈ R, (6.4) where h(k) is a column vector with n components belonging to the Hardy space H2(C+), is the trivial solution h(k) ≡ 0. (5◦) For the matrix-valued function Fs(y) given in (4.3), the derivative F ′s(y) for y ∈ R− can be written as a sum of two matrix-valued functions, one of which is integrable and the other is square integrable in y ∈ R−. We remark that the conditions (4◦c), (4◦d), (4◦e) listed in Theorem 6.2 are somehow similar to the respective conditions (4c), (4d), (4e) of Theorem 4.4. However, there are also some differences; for example, X(y) appearing in (4.14) belongs to L1(R+) whereas X(y) appearing in (6.2) belongs to L2(R+), X̂(k) of (4.15) belongs to L̂1(C+) whereas X̂(k) of (6.3) belongs to H2(C+), and h(k) of (4.16) belongs to L̂1(C+) whereas h(k) of (6.4) belongs to H2(C+). Inverse Scattering on the Half Line for the Matrix Schrödinger Equation 257 Let us also remark that the condition (5◦) in Theorem 6.2 is the same as the first sentence given in (IIIa) of Theorem 4.2. We note that Theorem 6.2 is the generalization of a characterization result by Agranovich and Marchenko presented in [1, Theorem 2, p. 281], which utilizes Levinson’s theorem in the purely Dirichlet case. That characterization result by Agranovich and Marchenko is valid only in the case of the Dirichlet boundary condition and does not include the condition stated in (5◦) in Theorem 6.2. In the special case of the purely Dirichlet boundary condition, it turns out that (5◦) in Theorem 6.2 is not needed. This has something to do with the fact that in the purely Dirichlet case the Marchenko integral equation (4.5) alone plays a key role in the solution to the inverse problem whereas in the non-Dirichlet case not only the Marchenko integral equation but also the derivative Marchenko integral equation both play a key role in the solution to the inverse problem, in particular in the satisfaction of the selfadjoint boundary condition given in (3.5). The derivative Marchenko integral equation is obtained by taking the x-derivative of (4.5), and hence the quantity F ′s(y) appears in the nonhomogeneous term of the derivative Marchenko integral equation. That presence of F ′s(y) somehow results in the condition stated in (5◦) of Theorem 6.2. The presence of (5◦) in Theorem 6.2 also has something to do with the fact that the boundary condition stated in (3.5) must hold for all k ∈ R. By taking the Fourier transform of both sides of (3.5), we end up with the requirement that the Fourier transform of the left-hand side of (3.5) must identically vanish. For this, one needs the necessity of the satisfaction of (5◦) in Theorem 6.2, unless A = 0 in (3.5). Since the case A = 0 is the same as having the purely Dirichlet boundary condition, (5◦) in Theorem 6.2 is relevant only in the non-Dirichlet case. For the mathematical elaboration on (5◦) we refer the reader to [10]. The presence of (5◦) in Theorem 6.2 is an indication of one of several rea- sons why the characterization of scattering data sets with the general selfadjoint boundary condition is more involved than the characterization with the Dirichlet boundary condition. We conclude that the result presented in Theorem 6.2, compared to Theo- rem 5.1, constitutes an alternate characterization of the scattering data sets S. Recall that Theorem 5.1 characterizes the scattering data sets that are in a one- to-one correspondence with the input data sets D in the Faddeev class. With the help of Theorem 6.2 we have the following alternate characterization of the scattering data sets. Theorem 6.3. Consider a scattering data set S as in (3.12), which consists of an n×n scattering matrix S(k) for k ∈ R, a set of N distinct positive constants κj , and a set of N constant n× n Hermitian and nonnegative matrices Mj with respective positive ranks mj , where N is a nonnegative integer. In case N = 0, it is understood that S consists only of S(k) for k ∈ R. Consider also an input data set D as in (2.1) consisting of an n×n matrix potential V (x) satisfying (2.3) and (2.4) and a pair of constant n × n matrices A and B satisfying (2.6) and (2.7), where it is understood that the boundary matrices A and B are unique up to a postmultiplication by an invertible n × n matrix T. Then, we have the following 258 Tuncay Aktosun and Ricardo Weder characterization of the scattering data sets. A set S as in (3.12) is the scattering data set corresponding to an input data set D in the Faddeev class if and only if S satisfies (1) and (2) of Definition 4.1, both (L) and (5◦) of Theorem 6.2, and any one of the three conditions listed as (4◦c), (4◦d), (4◦e) in Theorem 6.2. 7. Another characterization of the scattering data In this section we provide yet another description of the Marchenko class of scattering data sets S so that there exists a one-to-one correspondence between an input data set D in the Faddeev class and a scattering data set S in the Marchenko class. Such a description allows us to have yet another characterization of the scattering data sets S in a one-to-one correspondence with the input data sets D in the Faddeev class. The characterization stated in this section resulting from a new description of the Marchenko class has some similarities and differences compared to the first characterization presented in Theorem 5.1 and the alternate characterization pre- sented in Theorem 6.3. Related to this new characterization, the construction of the potential in the solution to the inverse problem is the same as in the previous characterizations; namely, one constructs the potential by solving the Marchenko equation. Hence, the conditions (1), (2), (4a) in the first characterization, the conditions (1), (2), (4◦c) in the alternate characterization, and the conditions (I), (2), (4c) in this new characterization are essentially used to construct the poten- tial. This new characterization differs from the two earlier ones in regard to the satisfaction of the boundary condition by the physical solution Ψ(k, x) and by the normalized bound-state matrix solutions Ψj(x). It is based on the alternate solu- tion to the inverse problem by using the generalized Fourier map [22]. This new characterization uses six conditions, indicated as (I), (2), (A), (4c), either one of (Ve) and (Vh), and (VI). Recall that (2) is described in Definition 4.1, (4c) is described in Theorem 4.4, and (Ve) and (Vh) are described in Theorem 4.5. In the following definition we describe the conditions (I), (A), and (VI). Definition 7.1. The properties (I), (A), and (VI) for the scattering data set S in (3.12) are defined as follows: (I) The scattering matrix S(k) satisfies (4.9), the quantity S∞ appearing in (4.1) exists, the quantity S(k) − S∞ is square integrable in k ∈ R, and the quantity Fs(y) defined in (4.3) is bounded in y ∈ R and integrable in y ∈ R+. (A) Consider the nonhomogeneous Riemann–Hilbert problem given by h(k) + S(−k)h(−k) = g(k), k ∈ R, (7.1) where the nonhomogeneous term g(k) belongs to a dense subset ◦ Υ of the vector space Υ of column vectors with n square-integrable components and satisfying g(−k) = S(k) g(k) for k ∈ R. Then, for each such given g(k), the equation (7.1) has a solution h(k) as a column vector with n components belonging to the Hardy space H2(C+). Inverse Scattering on the Half Line for the Matrix Schrödinger Equation 259 (VI) The scattering matrix S(k) is continuous in k ∈ R. We remark that (I) of Definition 7.1 is weaker than (1) of Definition 4.1. The quantity G1 and hence (4.2) appearing in (1) are used to construct the boundary matrices A and B. In order to construct the potential V (x) only, it is enough to use the weaker condition (I). The condition (A) of Definition 7.1 somehow resembles (IIIc) of Theorem 4.5, but there are also some major differences. In (IIIc) a solution is sought to the homogeneous Riemann–Hilbert problem (4.17) as a column vector with n components where each of those components belongs to H2(C−), and the only solution is expected to be the trivial solution h(k) ≡ 0. On the other hand, in (A) one solves a nonhomogeneous Riemann–Hilbert problem and the solution is sought as a column vector where each of the n com- ponents belongs the Hardy space H2(C+). The solution h(k) to (7.1) is in general nontrivial because the nonhomogeneous term g(k) there is in general nontrivial, and the existence of a solution to (7.1) is more relevant than its uniqueness. The condition (VI) of Definition 7.1, which is the continuity of the scattering matrix S(k), is mainly needed to prove that the physical solution Ψ(k, x) satisfies the boundary condition (2.5). Let us first describe the solution to the inverse scattering problem related to this new characterization and then present the characterization itself. As already indicated, the part of the solution to the inverse problem involving the construction of the potential is practically the same as the solution outlined in Section 4. However, the part of the solution related to the boundary condition is different than the procedure outlined in Section 4. We summarize below the construction of D from S in this new method, where the existence and uniqueness are implicit at each step: (a) From the large-k asymptotics of the scattering matrix S(k), with the help of (4.1), we determine the n× n constant matrix S∞. Contrary to the method of Section 4, we do not deal with the determination of the constant n × n matrix G1 appearing in (4.2). It follows from (4.9) that the matrix S∞ is Hermitian when S satisfies the condition (I) described in Definition 7.1. (b) In terms of the quantities in S, we uniquely construct the n×n matrix Fs(y) by using (4.3) and the n×n matrix F (y) by using (4.4). This step is the same as steps (b) and (c) of the summary of the method outlined in Section 4. (c) If the condition (4c) of Theorem 4.4 is also satisfied, then one uses the matrix F (y) as input to the Marchenko integral equation (4.5). If F (y) is integrable in y ∈ (x,+∞) for each x ≥ 0, then for each fixed x ≥ 0 there exists a solution K(x, y) integrable in y ∈ (x,+∞) to (4.5) and such a solution is unique. The solution K(x, y) can be constructed by iterating (4.5). We remark that this step is the same as step (d) of the summary of the method outlined in Section 4. Even though K(x, y) is constructed only for 0 ≤ x < y, one can extend K(x, y) to y ∈ R+ by letting K(x, y) = 0 for 0 ≤ y < x. (d) Having obtained K(x, y) uniquely from S, one constructs the potential V (x) via (4.6) and also constructs the Jost solution f(k, x) via (4.8). Then, by using 260 Tuncay Aktosun and Ricardo Weder (I), (2), and (4c), one proves that the constructed V (x) satisfies (2.3) and (2.4) and that the constructed f(k, x) satisfies (2.2) used with the constructed potential V (x). (e) Having constructed the Jost solution f(k, x), one then constructs the physical solution Ψ(k, x) via (3.4) and the normalized bound-state matrix solutions Ψj(x) via (3.11). One then proves that the constructed matrix Ψ(k, x) satis- fies (2.2) and that the constructed Ψj(x) satisfies (2.2) at k = iκj , with the understanding that the constructed potential V (x) is used in (2.2). (f) Having constructed the potential V (x), one forms a matrix-valued differential operator denoted by Lmin, which acts as (−D2 xI+V ) with Dx := d/dx, with a domain that is a dense subset of L2(R+). More precisely, the domain of Lmin consists of the column vectors with n components each of which is a function of x belonging to a dense subset of L2(R+). The constructed operator Lmin is symmetric, i.e. it satisfies Lmin ⊂ L†min, but is not selfadjoint, i.e. it does not satisfy Lmin = L†min. The operator inclusion Lmin ⊂ L†min indicates that the domain of the operator Lmin is a subset of the domain of the operator L†min and these two operators have the same value on the domain of Lmin. (g) One then constructs a selfadjoint realization of Lmin, namely an operator L in such a way that Lmin ⊂ L and L = L†. The constructed operator L is a restriction of L†min, i.e. we have L ⊂ L†min but not L = L†min. (h) The construction of the operator L is achieved [9,10,22] by using the so-called generalized Fourier map F and its adjoint F†. The generalized Fourier map F corresponds to a generalization of the Fourier transform between the space of square-integrable functions of x and the space of square-integrable functions of k. (i) Once the selfadjoint operator L is constructed, it follows [9, 10] that the do- main of L is a maximal isotropic subspace, which is sometimes also called a Lagrange plane. Once we know that the domain of L is a maximal isotropic subspace, it follows [9,10] that the functions in the domain of L must satisfy the boundary condition (2.5) for some boundary matrices A and B satisfying (2.6) and (2.7), where A and B are uniquely determined up to a postmulti- plication by an invertible matrix T. (j) Finally, one proves that the constructed physical solution Ψ(k, x) and the con- structed normalized bound-state matrix solutions Ψj(x) satisfy the boundary condition (2.5) with the boundary matrices A and B specified in the previ- ous step; however, such a proof is different in nature than the proofs for the previous characterizations. For the constructed matrices Ψj(x), it is imme- diate that they satisfy the boundary condition because they belong to the domain of L. Thus, it remains to prove that the constructed Ψ(k, x) satisfies the boundary condition. We note that the matrix Ψ(k, x) does not belong to the domain of L because its entries do not belong to L2(R+). On the other Inverse Scattering on the Half Line for the Matrix Schrödinger Equation 261 hand, Ψ(k, x) is locally square integrable in x ∈ [0,+∞), i.e. it is square integrable in every compact subset of [0,+∞). Hence, it is possible to use a simple limiting argument to prove that Ψ(k, x) satisfies the boundary condi- tion (2.5), and the condition (VI) is utilized in the aforementioned limiting argument. (k) As in the previous characterization given in Theorem 5.1(c), we still need to prove that the input data set D of (2.1) constructed from the scattering data set S of (3.12) yields S. The proof of this step is the same as in the proof of Theorem 5.1(c). Based on the procedure outlined above, we next present another description of the Marchenko class of scattering data sets S. Recall that (I), (A), and (VI) are described in Definition 7.1, (2) is described in Definition 4.1, (4c) is described in Theorem 4.4, and (Ve) and (Vh) are described in Theorem 4.5. Theorem 7.2. Consider a scattering data set S as in (3.12), which consists of an n×n scattering matrix S(k) for k ∈ R, a set of N distinct positive constants κj , and a set of N constant n× n Hermitian and nonnegative matrices Mj with respective positive ranks mj , where N is a nonnegative integer. In case N = 0, it is understood that S consists only of S(k) for k ∈ R and that N appearing in (3.8) is zero. The set S is the scattering data set corresponding to a unique input data set D as in (4.2) in the Faddeev class specified in Definition 2.1 if and only if S satisfies the six conditions consisting of (I), (2), (A), (4c), either one of (Ve) and (Vh), and (VI). We recall that the uniqueness of the input data set D is understood in the sense that the boundary matrices A and B in (4.2) are unique up to a postmultiplication by an arbitrary invertible n× n matrix T. 8. Some elaborations In this section we make a comparison with the definitions of the Jost ma- trix and the scattering matrix in the scalar case appearing in the literature. We also elaborate on the nonuniqueness issue arising if the scattering matrix is de- fined differently when the Dirichlet boundary condition is used. The reader is referred to Section 4 of [6] and Example 6.3 of [6] for further elaborations on the nonuniqueness issue. In the scalar case, i.e. when n = 1, from (2.8) we see that we can choose A = − sin θ, B = cos θ, θ ∈ (0, π], (8.1) where θ represents the boundary parameter. We can write the boundary condition (2.5) in the equivalent form −A† ψ′(0) +B† ψ(0) = 0. (8.2) Using (8.1) in (8.2) we see that our boundary condition (2.5) in the scalar case is equivalent to (sin θ)ψ′(0) + (cos θ)ψ(0) = 0, θ ∈ (0, π]. (8.3) 262 Tuncay Aktosun and Ricardo Weder We remark that the boundary condition (8.3) agrees with the boundary condition used in the literature [7,19,20] in the scalar case. Since θ = π corresponds to the Dirichlet boundary condition, we can write (8.3) in the equivalent form{ ψ(0) = 0 , Dirichlet case, ψ′(0) + (cot θ)ψ(0) = 0 , non-Dirichlet case, (8.4) where θ ∈ (0, π) in the non-Dirichlet case. The boundary condition (8.4) is also identical [7, 19, 20] to that used in the literature in the scalar case. As stated below (2.7), the boundary matrices A and B in (2.5) can be postmultiplied by any invertible matrix T without affecting (2.5)–(2.7). Hence, the constants A and B appearing in (8.1) can be multiplied by any nonzero constant. In any case, the boundary condition (2.5) we use is the same as the boundary condition used in the literature [7, 19,20] in the scalar case. Using (8.1) we see that the Jost matrix defined in (3.2) is given by J(k) = f(−k∗, 0)† (cos θ) + f ′(−k∗, 0)† (sin θ), k ∈ R, (8.5) where we recall that θ = π in the Dirichlet case and θ ∈ (0, π) in the non-Dirichlet case. Using (2.2), (2.3), and (3.1), for each fixed x ≥ 0 one can prove that f(k, x) and f ′(k, x) in the scalar case satisfy f(−k∗, x)∗ = f(k, x), f ′(−k∗, x)∗ = f ′(k, x), k ∈ C+. (8.6) Informally speaking, f(k, x) and f ′(k, x) each contain k as ik, and hence we have (8.6). Using (8.6) in (8.5) we see that the Jost matrix in the scalar case is given by J(k) = f(k, 0) (cos θ) + f ′(k, 0) (sin θ), k ∈ R, (8.7) which is equivalent to J(k) = { −f(k, 0) , Dirichlet case, (sin θ) [f ′(k, 0) + (cot θ) f(k, 0)] , non-Dirichlet case. (8.8) In the literature in the scalar case the Jost matrix is usually called the Jost function and is defined [7, 19,20] as J(k) = { f(k, 0) , Dirichlet case, −i [f ′(k, 0) + (cot θ) f(k, 0)] , non-Dirichlet case. (8.9) The primary motivation behind the definition in (8.9) is to define the Jost function J(k) in the scalar case in such a way that as k →∞ in C+ we have J(k) = 1 + O(1/k) in the Dirichlet case and J(k) = k + O(1) in the non-Dirichlet case. We remark that (8.8) and (8.9) do not agree, and we further elaborate on this disagreement. We know from (b) in Section 3 that the right-hand side of (8.7) can be multiplied by any nonzero constant because the boundary matrices A and B appearing in (3.2) can be postmultiplied by any invertible matrix T without Inverse Scattering on the Half Line for the Matrix Schrödinger Equation 263 affecting (2.5)–(2.7). Comparing (8.8) and (8.9) we see that it is impossible to modify the right-hand side of (8.8) through a multiplication by a nonzero scalar so that the right-hand sides of (8.8) and (8.9) agree. In other words, we cannot use the same multiplicative constant both in the Dirichlet case and in the non- Dirichlet case so that (8.8) and (8.9) can agree. Using (8.8) in (3.3) we obtain the scattering matrix in the scalar case as S(k) =  −f(−k, 0) f(k, 0) , Dirichlet case, −f ′(−k, 0) + (cot θ) f(−k, 0) f ′(k, 0) + (cot θ) f(k, 0) , non-Dirichlet case. (8.10) On the other hand, the scattering matrix in the scalar case is defined in the literature [7, 19,20] as S(k) =  f(−k, 0) f(k, 0) , Dirichlet case, −f ′(−k, 0) + (cot θ) f(−k, 0) f ′(k, 0) + (cot θ) f(k, 0) , non-Dirichlet case. (8.11) Thus, the first lines of (8.10) and (8.11) differ by a minus sign and their second lines are identical. Note that (8.9) and (8.11) indicate that the scattering matrix in the literature [7, 19,20] in the scalar case is related to the Jost matrix as S(k) = { J(−k) J(k)−1 , Dirichlet case, −J(−k) J(k)−1 , non-Dirichlet case. (8.12) The definition (8.12) of the scattering matrix in the scalar case in the literature is motivated by the fact that (8.12) ensures that S∞ defined in (4.1) is equal to 1, regardless of the Dirichlet case or the non-Dirichlet case. Comparing (8.12) with (3.3) we see that (3.3) and the first line of (8.12) differ by a minus sign and that (3.3) and the second line of (8.12) agree with each other. In the previous literature [1, 21], the scattering matrix in the Dirichlet case is defined as in the first line of (8.12) even in the nonscalar case, i.e. when n ≥ 2. Again, this ensures that S∞ = I, where we recall that I is the n × n identity matrix. However, defining the scattering matrix in the Dirichlet case as in (8.12) and not as (3.3) makes it impossible to have a unique solution to the inverse scattering problem unless the boundary condition is already known as a part of the scattering data. If the physical problem arises mainly from quantum mechanics and hence the boundary condition is the purely Dirichlet condition, which corresponds to having A = 0 in (2.5), this does not present a problem. On the other hand, if the determination of the selfadjoint boundary condition is a part of the solution to the inverse problem, then the definition of the scattering matrix S(k) given in (8.12) is problematic and that is one of the reasons why we use the definition of S(k) given in (3.3) regardless of the boundary condition. Note that we define the scattering matrix as in (3.3) so that the associated Schrödinger operator for the unperturbed problem has the 264 Tuncay Aktosun and Ricardo Weder Neumann boundary condition. This definition is motivated by the theory of quantum graphs, where the Neumann boundary condition is usually used for the unperturbed problem. We refer the reader to [15,17,18,22] for further details. In the following three examples, we illustrate the drawback of using (8.12) and not (3.3) as the definition of the scattering matrix. Example 8.1. Let us use (8.12) as the definition of the scattering matrix, instead of using (3.3). Let us assume that we are in the scalar case. Let us consider the input data set D given in (2.1) and the scattering data set S given in (3.12). The input data set D1 corresponding to the trivial potential V1(x) ≡ 0 and the Dirichlet boundary condition with θ1 = π yields the Jost solution f1(k, x) = eikx, and hence the corresponding Jost matrix is evaluated by using the first line of (8.9) as J1(k) = f1(k, 0) = 1. There are no bound states because J1(k) does not vanish on the positive imaginary axis in the complex k-plane. Thus, using the first line of (8.11), we evaluate the scattering matrix as S1(k) ≡ 1, and hence the corresponding scattering data set S1 consists of S1(k) ≡ 1 without any bound states. On the other hand, the input data set D2 corresponding to the trivial potential V2(x) ≡ 0 and the Neumann boundary condition with θ2 = π/2 corresponds to the Jost solution f2(k, x) = eikx and hence, by using the second line of (8.9), the corresponding Jost matrix is evaluated as J2(k) = −if ′2(k, 0) = k. There are no bound states because J2(k) does not vanish on the positive imaginary axis in the complex k-plane. Thus, using the second line of (8.11) or equivalently using the second line of (8.12), we evaluate the scattering matrix as S2(k) ≡ 1, and hence the corresponding scattering data set S2 consists of S2(k) ≡ 1 without any bound states. Thus, we have shown that S1 = S2 even though D1 6= D2. This nonuniqueness would not occur if we used (3.3) as the definition of the scattering matrix S(k). We would then get S1(k) ≡ −1 and S2(k) ≡ 1, and hence S1 6= S2. Next, we further illustrate the nonuniqueness encountered in Example 8.1 with a nontrivial example. Example 8.2. Let us use (8.12) as the definition of the scattering matrix, instead of using (3.3). Let us assume that we are in the scalar case. Let us choose a nontrivial potential V1(x) so that it is real valued and satisfies (2.4). Let us also view V1(x) as a full-line potential with support on x ∈ R+. We refer the reader to any reference on the scattering theory for the full-line Schrödinger equation such as [2,11–13,19,20] for the description of the corresponding scattering coefficients. As a full-line potential, let us also assume that V1(x) has no bound states and corresponds to the full-line exceptional case. The assumption of the absence of bound states on the full line is the same as assuming that the transmission coefficient has no poles on the positive imaginary axis in the complex k-plane, and the exceptional case on the full line is equivalent to the assumption that the transmission coefficient does not vanish at k = 0. Corresponding to V1(x) as a full-line potential we have the full-line scattering data {T1(k), R1(k), L1(k)}, where T1(k) is the transmission coefficient, R1(k) is the reflection coefficient from Inverse Scattering on the Half Line for the Matrix Schrödinger Equation 265 the right, and L1(k) is the reflection coefficient from the left. It is known [2, 11– 13,19,20] that the full-line scattering data {T2(k), R2(k), L2(k)}, where we have T2(k) := T1(k), R2(k) := −R1(k), L2(k) := −L1(k), (8.13) corresponds to a nontrivial full-line potential V2(x) so that V2(x) is real valued, vanishes when x < 0, has no bound states on the full line, and corresponds to the full-line exceptional case. Furthermore, V2(x) satisfies (2.4). Viewing V1(x) and V2(x) as half-line potentials, let us now evaluate the corresponding half- line scattering data sets S1 and S2 associated with the full-line scattering data sets {T1(k), R1(k), L1(k)} and {T2(k), R2(k), L2(k)}, respectively. Since V1(x) and V2(x) both vanish when x < 0, the corresponding respective Jost solutions f1(k, x) and f2(k, x) yield f1(k, 0) = 1 + L1(k) T1(k) , f2(k, 0) = 1 + L2(k) T2(k) , (8.14) f ′1(k, 0) = ik 1− L1(k) T1(k) , f ′2(k, 0) = ik 1− L2(k) T2(k) . (8.15) Let us now view V1(x) as a half-line potential, associate it with the Dirichlet boundary condition θ1 = π, and use D1 to denote the resulting input data set. Similarly, let us view V2(x) as a half-line potential, associate it with the Neumann boundary condition θ2 = π/2, and use D2 to denote the resulting input data set. Clearly, we have D1 6= D2 because θ1 6= θ2. Using (8.14) in the first lines of (8.9) and (8.11) we obtain the Jost matrix J1(k) and the scattering matrix S1(k) as J1(k) = f1(k, 0) = 1 + L1(k) T1(k) , S1(k) = f1(−k, 0) f1(k, 0) = T1(k) T1(−k) 1 + L1(−k) 1 + L1(k) . (8.16) On the other hand, using (8.15) and the second lines of (8.9) and (8.11) with θ = π/2, we obtain the Jost matrix J2(k) and the scattering matrix S2(k) as J2(k) = −i f ′2(k, 0) = k 1− L2(k) T2(k) , S2(k) = −f ′ 2(−k, 0) f ′2(k, 0) = T2(k) T2(−k) 1− L2(−k) 1− L2(k) . (8.17) Using (8.13) in (8.16) and (8.17) we see that S1(k) ≡ S2(k), and hence the corresponding scattering data sets S1 and S2 coincide. Thus, we get D1 6= D2 and S1 = S2. This nonuniqueness can be fixed by using (3.3) and not (8.12) as the definition of the scattering matrix S(k). The nonuniqueness problem encountered in the previous two examples can also occur in the nonscalar case, as shown in the following example. This new example is the generalization of Example 8.2 from the scalar case to the n × n matrix case for any positive integer n. For the relevant scattering theory for the matrix Schrödinger equation on the full line, we refer the reader to [3]. 266 Tuncay Aktosun and Ricardo Weder Example 8.3. In this example we assume that n is any positive integer and not necessarily restricted to n = 1. Let us use (8.12) as the definition of the scattering matrix, instead of using (3.3). Let us again use (2.1) to describe an input data set D and use (3.12) to describe a scattering data set S on the half line. Consider the class of n × n matrix-valued potentials V (x) on the full line satisfying V (x) = V (x)†, x ∈ R, (8.18)∫ ∞ −∞ dx (1 + |x|) |V (x)| < +∞, (8.19) where we recall that the dagger denotes the matrix adjoint and |V (x)| denotes the matrix operator norm. The reader is referred to [3] for the matrix-valued scattering coefficients for the full-line matrix Schrödinger equation with such po- tentials. Associated with V (x) we have the full-line scattering coefficients Tl(k), R(k), and L(k), each of which is an n× n matrix. These matrix-valued scatter- ing coefficients are the matrix generalizations of the scalar scattering coefficients T (k), R(k), and L(k) considered in Example 8.2. Let us further restrict the full- line potentials V (x) so that they vanish when x < 0, they do not possess any bound states on the full line, and they correspond to the purely exceptional case. We refer the reader to [3] for the details on the bound states and the purely exceptional case on the full line. The absence of bound states on the full line is equivalent to having the determinant of the matrix inverse of Tl(k) not vanishing on the positive imaginary axis in the complex k-plane. The purely exceptional case on the full line is equivalent to having the limit of k Tl(k)−1 as k → 0 equal to the n×n zero matrix. For such potentials Tl(0)−1 is well defined and we have det[I±L(0)] 6= 0, where we recall that I denotes the n×n identity matrix. Since we only consider the full-line potentials V (x) vanishing when x < 0, we can view their restrictions on x ∈ R+ as half-line potentials V (x). From (8.18) and (8.19) we see that their restrictions on x ∈ R+ belong to the Faddeev class. When x ∈ R+, the full-line Jost solution from the left fl(k, x) coincides [3] with the half-line Jost solution f(k, x) appearing in (3.1). Furthermore, we have [3] fl(k, 0) = [I+L(k)]Tl(k)−1, f ′l (k, 0) = ik [I−L(k)]Tl(k)−1, k ∈ R. (8.20) Let V1(x) be a specific full-line matrix potential satisfying (8.18) and (8.19) such that it vanishes when x < 0, does not contain any bound states on the full line, and corresponds to the purely exceptional case on the full line. Let {Tl(k), R(k), L(k)} be the corresponding full-line scattering data. Let V2(x) be the full-line matrix potential corresponding to the full-line scattering data {Tl(k),−R(k),−L(k)}, where the signs of the matrix-valued reflection coefficients are changed. The matrix potential V2(x) also vanishes for x < 0, satisfies (8.18) and (8.19), does not possess any bound states on the full line, and corresponds to a purely exceptional case on the full line. The restrictions of V1(x) and V2(x) on x ∈ R+ can be viewed as half-line potentials. Let A1, B1, A2, B2 be four n× n constant matrices in such a way that A1 = 0, B2 = 0, A2 is an arbitrary in- vertible matrix, and B1 is an arbitrary invertible matrix. Let D1 := {V1, A1, B1} Inverse Scattering on the Half Line for the Matrix Schrödinger Equation 267 and D2 := {V2, A2, B2} be the half-line input data sets as in (2.1), with the un- derstanding that the domains of V1(x) and V2(x) are restricted to x ∈ R+. Let f1(k, x) and f2(k, x) be the half-line Jost solutions corresponding to D1 and D2, respectively. From (8.20) we see that f1(k, 0) = [I + L(k)]Tl(k)−1, f ′1(k, 0) = ik [I − L(k)]Tl(k)−1, k ∈ R, (8.21) f2(k, 0) = [I − L(k)]Tl(k)−1, f ′2(k, 0) = ik [I + L(k)]Tl(k)−1, k ∈ R. (8.22) Using (8.21) and (8.22), because of the purely exceptional case on the full line [3], it follows that neither of the determinants of f1(k, 0) and f ′2(k, 0) vanish. Let J1(k), S1(k), S1 be the respective Jost matrix, scattering matrix, and scattering data set corresponding to D1. Similarly, let J2(k), S2(k), S2 be the respective Jost matrix, scattering matrix, and scattering data set corresponding to D2. Using (8.21) and (8.22) in (3.2) we obtain J1(k) = f1(−k, 0)†B1 = [Tl(−k)†]−1 [I + L(−k)†]B1, (8.23) J2(k) = −f ′2(−k, 0)†A2 = −ik [Tl(−k)†]−1 [I + L(−k)†]A2. (8.24) Using (8.23) in the first line of (8.12) we obtain S1(k) = J1(−k) J1(k)−1 = f1(k, 0)† [ f1(−k, 0)† ]−1 , which yields S1(k) = [Tl(k)†]−1 [I + L(k)†] [I + L(−k)†]−1 Tl(−k)†. (8.25) Using (8.24) in the second line of (8.12) we obtain S2(k) = −J2(−k) J2(k)−1 = −f ′2(k, 0)† [ f ′2(−k, 0)† ]−1 , yielding S2(k) = [Tl(k)†]−1 [I + L(k)†] [I + L(−k)†]−1 Tl(−k)†. (8.26) There are no half-line bound states associated with either of the scattering data sets corresponding to S1(k) and S2(k). Hence we have S1 = {S1} and S2 = {S2}. From (8.25) and (8.26) it follows that S1(k) ≡ S2(k) and hence we have S1 = S2 even though D1 6= D2. This nonuniqueness would not occur if we used (3.3) as the definition of the scattering matrix S(k). We would then get S1(k) ≡ −S2(k), and hence S1 6= S2. Supports. The research leading to this article was supported in part by CONACYT under project CB2015, 254062, Project PAPIIT-DGAPA-UNAM IN103918, and by Coordinación de la Investigación Cient́ıfica, UNAM. 268 Tuncay Aktosun and Ricardo Weder References [1] Z.S. Agranovich and V.A. Marchenko, The Inverse Problem of Scattering Theory, Gordon and Breach, New York, 1963. [2] T. Aktosun and M. Klaus, Chapter 2.2.4: Inverse theory: problem on the line, In: E. R. Pike and P. C. Sabatier (eds.), Scattering, Academic Press, London, 2001, 770–785. [3] T. Aktosun, M. Klaus, and C. van der Mee, Small-energy asymptotics for the Schrödinger equation on the line, J. Math. Phys. 17 (2001), 619–632. [4] T. Aktosun, M. Klaus, and R. Weder, Small-energy analysis for the self-adjoint matrix Schrödinger operator on the half line, J. Math. Phys. 52 (2011), 102101. [5] T. Aktosun, M. Klaus, and R. Weder, Small-energy analysis for the self-adjoint matrix Schrödinger operator on the half line. II, J. Math. Phys. 55 (2014), 032103; arXiv: 1310.4809. [6] T. Aktosun, P. Sacks, and M. Unlu, Inverse problems for selfadjoint Schrödinger operators on the half line with compactly supported potentials, J. Math. Phys. 56 (2015), 022106; arXiv: 1409.5819. [7] T. Aktosun and R. Weder, Inverse spectral-scattering problem with two sets of discrete spectra for the radial Schrödinger equation, Inverse Problems 22 (2006), 89–114; arXiv: math-ph/0402019. [8] T. Aktosun and R. Weder, High-energy analysis and Levinson’s theorem for the self-adjoint matrix Schrödinger operator on the half line, J. Math. Phys. 54 (2013), 112108; arXiv: 1206.2986. [9] T. Aktosun and R. Weder, Inverse scattering for the matrix Schrödinger equation, preprint, 2018; arXiv: 1708.03837. [10] T. Aktosun and R. Weder, Direct and Inverse Scattering for the Matrix Schrödinger Equation, the monograph to be published by Springer-Verlag. [11] K. Chadan and P. C. Sabatier, Inverse Problems in Quantum Scattering Theory, 2nd ed., Springer, New York, 1989. [12] P. Deift and E. Trubowitz, Inverse scattering on the line, Commun. Pure Appl. Math. 32 (1979), 121–251. [13] L.D. Faddeev, Properties of the S-matrix of the one-dimensional Schrödinger equa- tion, Amer. Math. Soc. Transl., Ser. 2 65 (1967), 139–166. [14] M.S. Harmer, Inverse scattering for the matrix Schrödinger operator and Schrö- dinger operator on graphs with general self-adjoint boundary conditions, ANZIAM J. 44 (2002), 161–168. [15] M.S. Harmer, The Matrix Schrödinger Operator and Schrödinger Operator on Graphs, Ph.D. thesis, University of Auckland, New Zealand, 2004. [16] M. Harmer, Inverse scattering on matrices with boundary conditions, J. Phys. A 38 (2005), 4875–4885. [17] V. Kostrykin and R. Schrader, Kirchhoff’s rule for quantum wires, J. Phys. A 32 (1999), 595–630. https://arxiv.org/abs/1310.4809 https://arxiv.org/abs/1409.5819 https://arxiv.org/abs/math-ph/0402019 https://arxiv.org/abs/1206.2986 https://arxiv.org/abs/1708.03837 Inverse Scattering on the Half Line for the Matrix Schrödinger Equation 269 [18] V. Kostrykin and R. Schrader, Kirchhoff’s rule for quantum wires. II: The in- verse problem with possible applications to quantum computers, Fortschr. Phys. 48 (2000), 703–716. [19] B.M. Levitan, Inverse Sturm–Liouville Problems, VNU Science Press, Utrecht, 1987. [20] V.A. Marchenko, Sturm–Liouville Operators and Applications, revised ed., Amer. Math. Soc. Chelsea Publ., Providence, R.I., 2011. [21] R.G. Newton and R. Jost, The construction of potentials from the S-matrix for systems of differential equations, Nuovo Cim. 1 (1955), 590–622. [22] R. Weder, Scattering theory for the matrix Schrödinger operator on the half line with general boundary conditions, J. Math. Phys. 56 (2015), 092103; arXiv: 1505.01879. [23] R. Weder, Trace formulas for the matrix Schrödinger operator on the half- line with general boundary conditions, J. Math. Phys. 57 (2016), 112101; arXiv: 1603.09432. [24] R. Weder, The number of eigenvalues of the matrix Schrödinger operator on the half line with general boundary conditions, J. Math. Phys. 58 (2017), 102107; arXiv: 1705.03157. Received March 1, 2018. Tuncay Aktosun, University of Texas at Arlington, Arlington, TX 76019-0408, USA, E-mail: aktosun@uta.edu Ricardo Weder, Instituto de Investigaciones en Matemáticas Aplicadas y en Sistemas, Universidad Na- cional Autónoma de México, Apartado Postal 20-126, IIMAS-UNAM, México DF 01000, México, E-mail: weder@unam.mx Обернене розсiювання на пiвпрямiй для матричного рiвняння Шредiнгера Tuncay Aktosun and Ricardo Weder На пiвпрямiй розглянуто матричне рiвняння Шредiнгера iз загаль- ною самоспряженою крайовою умовою в нулi, яка задається двома мат- рицями, що задовольняють певнi умови. Вважається, що матричний потенцiал є самоспряженим, iнтегровним та має скiнченний перший момент. Побудовано вiдповiдну множину даних розсiяння. Цю множи- ну даних розсiювання характеризовано набором необхiдних i достатнiх умов, якi гарантують єдинiсть та взаємно однозначну вiдповiднiсть мiж множиною даних розсiяння та множиною вхiдних даних, яка мiстить потенцiал та крайовi матрицi. Ця робота надає узагальнення з крайової умови Дiрiхле на загальну самоспряжену крайову умову для класичного результату Аграновича та Марченка. Ключовi слова: матричне рiвняння Шредингера, самоспряжена гра- нична умова, метод Марченка, матричний метод Марченка, матриця Йоста, матриця розсiяння, обернене розсiювання, характеризацiя. https://arxiv.org/abs/1505.01879 https://arxiv.org/abs/1603.09432 https://arxiv.org/abs/1705.03157 mailto:aktosun@uta.edu mailto:weder@unam.mx Introduction The matrix Schrödinger equation The solution to the direct problem The solution to the inverse problem The characterization of the scattering data An alternate characterization of the scattering data Another characterization of the scattering data Some elaborations