310
Distributions with given Marginals and Moment Problems

Distributions with given Marginals and Moment Problems

  • Upload
    others

  • View
    7

  • Download
    0

Embed Size (px)

Citation preview

Distributions with given Marginals and Moment Problems Edited by
Viktor BeneS Department of Mathematics, Czech Technical University, Prague, Czech Republic
and
SPRINGER SCIENCE+BUSINESS MEDIA, B.V.
A CLP. Catalogue record for this book is available from the Library of Congress.
ISBN 978-94-010-6329-6 ISBN 978-94-011-5532-8 (eBook) DOI 10.1007/978-94-011-5532-8
Printed on acid-free paper
All Rights Reserved © 1997 Springer Science+Business Media Dordrecht Originally published by Kluwer Academic Publishers in 1997 Softcover reprint of the hardcover 1st edition 1997
No part of the material protected by this copyright notice may be reproduced or utilized in any form or by any means, electronic or mechanical, including photocopying, recording or by any information storage and retrieval system, without written permission from the copyright owner.
TABLE OF CONTENTS
vii PREFACE ACKNOWLEDGEMENT x G. A. ANASTASSIOU / Optimal bounds on the average of a rounded
off observation in the presence of a single moment condition 1 T. RYCHLIK / The complete solution of a rounding problem under two
moment conditions 15 V. GIRARDIN / Methods of realization of moment problems with en-
tropy maximization 21 E. KAARIK / Matrices of higher moments: some problems of represen-
tation 27 L. B. KLEBANOV, S. T. RACHEV / The method of moments in to-
mography and quantum mechanics 35 V. BENES / Moment problems in stochastic geometry 53 M. SCARSINI, M. SHAKED / Frechet classes and nonmonotone de-
pendence 59 A. CHATEAUNEUF, M. COHEN, 1. MEILIJSON / Comonotonicity,
rank-dependent utilities and a search problem 73 P. CAPERAA, A. L. FOUGERES, C. GENEST / A stochastic ordering
based on a decomposition of Kendall's tau 81 M. J. W. JANSEN / Maximum entropy distributions with prescribed
marginals and normal score correlations 87 . V. E. PIPERIGOU / On bivariate distributions with Polya-Aeppli or L iiders-Delaporte marginals 93
E.-M. TIlT, H.-L. HELEMAE / Boundary distributions with fixed mar- ~~ ~
X. LI , P . MIKUSINSKI, H. SHERWOOD, M. D. TAYLOR / On ap- proximations of copulas 107
G. DALL'AGLIO / Joint distributions of two uniform random variables when the sum and difference are independent 117
R. B. NELSEN, G. A. FREDRICKS / Diagonal copulas 121 G. A. FREDRICKS, R. B. NELSEN / Copulas constructed from diag-
onal sections 129 C. M. CUADRAS, J. FORTIANA / Continuous scaling on a bivariate
copula 137 H. G. KELLERER / Representation of Markov kernels by random map-
pings under order conditions 143 J. STEP AN / How to construct a two dimensional random vector with
a given conditional structure 161 A. HIRSHBERG, R. M. SHORTT / Strassen 's theorem for group-valued
charges 173
vi
G. LETAC / The Lancaster's probabilities on R2 and their extreme points 179
M. STUDENY / On marginalization, collapsibility and precollapsibili- ty 191
J. DUPACOV A / Moment bounds for stochastic programs in particular for recourse problems 199
T. SZANTAI / Probabilistic constrained programming and distributions with given marginals 205
V. KANKOV A / On an £-solution of minimax problem in stochastic programming 211
T. ViSEK / Bounds for stochastic programs - nonconvex case 217 R. JIROUSEK / Artificial intelligence, the marginal problem and in-
consistency 223 O. KIlii / Inconsistent marginal problem on finite sets 235 V. L. LEVIN / Topics in the duality for mass transfer problems 243 C. S. SMITH, M. KNOTT / Generalising monotonicity 253 L. RUSCHENDORF, 1. UCKELMANN / On optimal multivariate cou-
plings 261 L. UCKELMANN / Optimal couplings between one-dimensional distri-
butions 275 D. RAMACHANDRAN, 1. RUSCHENDORF / Duality theorems for
assignments with upper bounds 283 J. H. B. KEMPERMAN / Bounding the moments of an order statistics
if each k-tuple is independent 291 SUBJECT INDEX 305
PREFACE
The last decade has seen a remarkable development of the "Marginal and Moment Problems" as a research area in Probability and Statistics. Its attractiveness stemmed from its lasting ability to provide a researcher with difficult theoretical problems that have direct consequences for appli­ cations outside of mathematics. The relevant research aims centered mainly along the following lines that very frequently met each other to provide sur­ prizing and useful results :
-To construct a probability distribution (to prove its existence, at least) with a given support and with some additional inner stochastic property defined typically either by moments or by marginal distributions.
-To study the geometrical and topological structure of the set of prob­ ability distributions generated by such a property mostly with the aim to propose a procedure that would result in a stochastic model with some optimal properties within the set of probability distributions.
These research aims characterize also, though only very generally, the scientific program of the 1996 conference "Distributions with given marginals and moment problems" held at the beginning of September in Prague, Czech Republic, to perpetuate the tradition and achievements of the closely related 1990 Roma symposium "On Frechet Classes" 1 and 1993 Seattle" AMS Summer Conference on Marginal Problem".
A careful selection of the contributions presented to the 1996 conference is published in these proceedings. It has been an aim of the editors not only to reflect the recent developments in the field, but also, with an indispens­ able assistance of numerous referees, to publish new and important results. These are the reasons that make the editors believe that the book will prove itself useful even for the probabilists and statisticians working outside its specialized domain and will surely stimulate the research in the "Moments & Marginals" area itself.
1 For the proceedings see "Advances in Probability Distributions with Given Ma.rginals" . Kluwer Academic Publishers, Dodrecht 1991.
vii
viii
Roughly speaking, the contributions of the proceedings belong to one of the following topics centered around the invited papers written be carefully selected scholars :
- Moment Problems and Applications. G.A.Anastassiou applies Geometric Moment Theory to a statistical problem over observations sub­ jected to one moment condition. L.B.Klebanov and S. T.Rachev introduce Computer Tomography and Quantum Mechanics as research fields where the moment and marginals technologies may be applied to get new intrigu­ ing results - a solution of the computer tomography paradox, for example. J.H.B.Kemperman shows how to calculate lower and upper bounds of the moments of an order statistics if each k-tuple in the sample is formed be independent random variables with a known probability distribution.
- Marginal Problems and Stochastic Order. M.Scarsini and M.Sha­ ked examine in their paper non monotone dependence bivariate stochastic models in the framework of a Frechet class. By nonmonotone dependence is meant the tendency of a bivariate distribution to concentrate its probability mass around the graph of a (not necessarily monotone) function.
- Copulas. This well established topic with a remarkable history in the field is represented by "On Approximation of Copulas" by X.Li, P.Mi­ kusinski, H.Sherwood, and M.D. Taylor. The authors observe that uniform convergence of copulas is not completely satisfactory when trying to verify some of their properties that could be derived first on a suitable class of "simple" copulas and then extended to all copulas by a limiting process. Two different ways of approximating copulas are proposed (Checkerboard and Bernstein approximations) to meet the needs mentioned above.
- Marginal and Moment Problems - Measure Theory Approach. G.Letac comes back to a problem with the long history namely that of find­ ing extremal measures in the convex set of probability measures with a fixed pair of marginals. The problem is attacked, this time, in the framework of Lanchester R 2-probabilities. H.G.Kellerer treats the problem of represent­ ing a Markov kernel that connects a pair of ordered spaces by an order preserving random map. The problem is a natural counterpart to the pre­ vious studies on Markov kernels represented by continuous random maps.
- Stochastic Programming and Artificial Intelligence. J.Dupa­ cova discusses in her paper different methods for bounding the optimal values of stochastic programs, in particular for recourse problems. R.Jirou­ sek studies the problem of constructing a distribution with given marginals as an inherent part of artificial intelligence approaches to applied problems. In this context, only the constructions that can be handled efficiently by computer programs are acceptable.
- Optimizations in Marginal Problem The paper by L.Levin con­ cerns the duality theory for nontopological versions of mass transport prob-
ix
lem with the aim to extend the classical Monge-Kantorovich results. L.Rii­ schendorf and 1. Uckelmann treat optimal multivariate problems, prove the existence of optimal Monge solutions in this context and propose an ex­ plicit construction method for the optimal transportation in the case that one mass distribution is discrete. Finally, D.Ramachandran and L.Riischen­ dorfformulate a version of the nonatomic assignment .problem with upper admissible assignments in the form of a dominating measure and obtain a general duality theorem for this model.
The invited papers, the achievements of which have been reviewed above are accompanied by 25 contributed papers that, according to the editors' opinion, increase significantly the scientific value of the proceedings. Cordial thanks are due to all authors and referees.
Prague, January 1997
ACKNOWLEDGEMENTS
The editors desire in the first place to acknowledge that it would never have been possible to organize the 1996 Prague Conference "Distributions with Given Marginals and Moment Problems" and to prepare its Proceed­ ings for the publication if it had not been for the research grant they have received from the Grant Agency of the Czech Republic in 1994 under the contract 201/94/0471. The grant not only encouraged their own sci­ entific work but also provided an invaluable help in keeping and promoting necessary international contacts.
The editors are no less indebted to Faculty of Mathematics and Physics of Charles University in Prague, Faculty of Mechanical Engineering of Czech Technical University in Prague, Institute of Information Theory and Automation in Academy of Sciences of Czech Republic and to Technical Faculty of Czech Agricultural University in Prague which institutions spon­ sored generously the Conference activities.
Among many individuals to whom the organizers of the Conference and the editors of its Proceedings owe for advice, help and encouragement are firstly the members of the Conference scientific and organizing committees: L. Riischendorf (Freiburg i. Br.), H. Sherwood (Orlando), S. T. Rachev (Santa Barbara), J. Dupacova, V. Slavik, M. Studeny, P. Lachout, D. Hlu­ binka, T. Visek.
D. Hlubinka and T. Visek also have the editor's grateful thanks for preparing the manuscript technically. Last but not least cordial thanks are due to H. Jandova for the skill and patience with the editors when mastering their extensive correspondence and paperwork.
Viktor Benes and Josef Stepan
x
OFF OBSERVATION IN THE PRESENCE OF A SINGLE MOMENT
CONDITION
GEORGE A. ANASTASSIOU Department of Mathematical Sciences The University of Memphis Memphis, TN 38152 U.S.A.
Abstract. The moment problem of finding the maximum and minimum expected values of the averages of nonnegative random variables over various sets of observations subject to one simple moment condition is encountered. The solution is given by means of geometric moment theory (see [3]).
1. Background
Here [.j and r'l stand for the integral part (floor) and ceiling of the number, respectively.
We consider probability measures p, on A := [0, aj, a > ° or [0, +00). We would like to find
Ua := sup f ([t] + rtl) p,(dt) p. JA 2
and
subject to
L tT J.L( dt) = dT , r > 0,
where d > ° (here the subscript Q in Ua, La stands for average).
V. Bend and J. Stepan (ew.). Distributions with given Marginals and Momenl Problems. 1-13. @ 1997 Kluwer Academic Publishers.
(1)
(2)
(3)
2 GEORGE A. ANASTASSIOU
In this article, the underlined probability space is assumed to be nonatomic and thus the space of laws of nonnegative random variables coincides with the space of all Borel probability measures on R+ (see [4], pp. 17-19).
The solution of the above moment problems are of importance because they permit comparisons to the optimal expected values of the averages of nonnegative random variables, over various sets of observations, subject to one simple and flexible moment condition.
Similar problems have been solved earlier in [1], [2]. Here again the solutions corne by the use of tools from the Kemperman geometric moment theory, see [3]. Here of interest will be to find
and
subject to the one moment condition (3). Then, obviously,
(4)
(5)
(6)
So we use the method of optimal distance in order to find U, L as follows: Call
M :== cOnvtEAW, ([tl + ftl )). (7)
When A = [0, a] by Lemma 2 of [3J we get that ° < d ~ a, in order to find finite U and L. Then U = sup{z: (dT,z) E M} and L = inf{z : (dT,z) E M}. Thus U is the largest distance between (dT,O) and (dT,z) EM. Also L is the smallest distance between (dT , 0) and (dT , z) EM. SO we are dealing with two-dimensional moment problems over all different variations of A :=
[0, a], a > 0, or A := [0, +00) and T ~ 1 or ° < T < 1. We intend elsewhere to solve the corresponding three-dimensional mo­
ment problems, where we include one more moment condition satisfied by JL.
Description of the graph (XT, ([xJ + f xl)), T > 0, a ~ +00. This is a stairway made out of the following steps and points, the next step is two units higher than the previous one with a point included in between, namely we have: the lowest part of the graph is Po = (0,0), then open interval {PI = (0,1), Qo = (1, I)}, then included point (1,2), open interval {P2 = (1,3), QI = (2T,3)}, included point (2T,4), open interval {P3 = (2T,5), Q2 = (3T,5)}, included point (3T,6), ... , {open intervals {PHI := (kT,2k + 1), Qk := ((k+lY, 2k+ I)}, included points (kT, 2k); k = 0,1,2,3, ... }, ... , keep going like this if a = +00, or if ° < a ¢ N then the last top part, segment
OPTIMAL BOUNDS ON THE AVERAGE 3
ofthe graph, will be the open closed interval {([ar, 2[aJ + 1), (aT, 2[aJ + I)}, or if a E N then the last top part, included point, will be (aT, 2a). Here Pk+2 = ((k + 1Y, 2k + 3). Then
which is decreasing in k if r ~ 1, and increasing in k if ° < r < 1. The closed convex hull of the above graph M changes according to values of r and a.
Here X stands for a random variable and E for the expected value.
2. Results
Our first result follows
Theorem 1. Let d > 0, r ~ 1 and 0< a < +00. Consider
U := sup{E([XJ + rXl): ° ~ X ~ a a.s., EXT = dT }. (8)
1) If k ~ d ~ k + 1, k E {O, 1,2, ... , [ aJ- I}, a ¢ N, then
2dT + (2k + 1)(k + 1y - (2k + 3)kT U= .
(k + ly - kT (9)
U = 2[a] + 1. (10)
3) Let A = [0, +00), (i.e., a = +00) and k ~ d ~ k + 1, k = 0,1,2,3, ... , then U is given by (9).
4) Let a E N - {I, 2} and k ~ d ~ k + 1, where k E {O, 1,2, ... , a - 2}, then U is given by (9).
5) Let a E N and a-I ~ d ~ a, then
U = dT + aT(2a - 1) - (a - lY2a. aT - (a - ly (11)
Proof. We are working on the (XT, y) plane. Here the line through the points Pk+I, Pk+2 is given by
2xT + (2k + 1)(k + It - (2k + 3)P y = (k + 1 Y - kT •
(12)
4 GEORGE A. ANASTASSIOU
In the case of a E N let Pa := ((a - l)r, 2a - 1). Then the line through Pa
and (ar , 2a) is given by
(13)
The above lines describe the closed convex hull !VI in the various cases mentioned in the statement of the theorem. Then we apply the method of optimal distance described earlier in §1. 0
Its counterpart comes next.
Theorem 2. Let d > 0, 0 < r < 1 and 0 < a < +00. Consider
U := sup{E([X] + rXl):O::; X::; a a.s.,EXT = dT }. (8)
1) Let 0 < a ¢ N and [a] ::; d::; a, then
U = 2[a] + 1. (14)
2) Let 0 < a ¢ Nand 0 < d::; [a], then
2[a] U = [a]T ~ + 1. (15)
2*) Let a = 1 and 0 < d :::; 1, then U = dr + 1. 3) Let a E N - {I}, 0 < d :::; a and 0 < r ::; !, then
( 2a - 1) U= -;r- ~+l. (16)
4) Let a E N - {I}, 0 < d::; a-I, ! < r < 1 and a ~ (2Tt ,then (2r) -r-l
U = 2(a _1)I-T~ + 1. (17)
1 (2Tt h 5) Let a E N - {I}, a-I::; d::; a, '2 < r < 1 and a ~ ,t en (2T) -r-l
(18)
U = +00. (19)
OPTIMAL BOUNDS ON THE AVERAGE 5
Proof. We apply again the method of optimal distance. From the graph of (XT, ([X] + r xl)) we have that the slope (PHI Pk+2) = (k+1)r-kr is increasing in k (see §1).
Let 0 < a ¢ N, then the upper envelope of the closed convex hull if is formed by the line segments {PI = (0,1), P[aJ+I := ([aY, 2[a] + I)} and {p[a)+l> (aT, 2[a] + I)}. Then case (1) is clear ..
The line (PI p[a)+1) in the (x r , y) plane is given by
y = 2[a]xT + 1. [aV
Therefore case (2) follows immediately. Case (6) is obvious by the nature of M when 0 < r < 1. N ext let a E N - {I} and 0 < r ~ !. I.e.,
( ~ )l-T 2r~1< a-I ,forany~E(a-1,a).
(20)
That is 2(a - l)I-T < re!-i' any ~ E (a - 1, a). Hence by the mean value theorem we obtain
2 a-I I-T < 1 ( ) aT - (a - 1Y (21)
Therefore slope(mo) < slope(mt), where mo is the line through the points PI = (0,1), Pa = ((a - 1Y, 2a - 1), and ml is the line through the points Pa , (aT, 2a) . In that case the upper envelope of if is the line segment with endpoints PI = (0,1) and (aT, 2a) . This has equation in the (XT, y) plane
( 2a - 1) Y = --;;;- xT + 1. (22)
The last implies case (3). Finally let a EN - {I} and! < r < 1, along with
1 (2r)r=T
Hence
(_a_) ~ (2r) l':r . a-I
6 GEORGE A. ANASTASSIOU
( ) I-T I.e., a~I ~ 2r and e-T < aI- T ~ (a - I)I-T2r, any ~ E (a - 1, a).
Thqs e-T ~ 2r(a _1)I-T, and T~!-l ~ 2(a - l)l-T, any ~ E (a - 1, a). And by the mean value theorem we obtain that
1 slope(mI)= ( ) ~slope(mo)=2(a-l)I-T,
aT - a-I T
where again ml is the line through Pa = ((a-lY, 2a-l) and (aT, 2a), while mo is the line through PI = (0,1) and Pa. Therefore the upper envelope of M is made out of the line segments PI Pa and Pa (aT, 2a). The line mo in the (x T , y) plane has equation
Now case (4) is obvious. The line ml in the (XT, y) plane has equation
xT + 2a( aT - (a - 1 Y) - aT y=
aT - (a - lY A t last case (5) is clear.
(23)
(24)
o Next we determine L in different cases. We start with the less compli­
cated case where 0 < r < 1.
Theorem 3. Let d > 0, 0 < r < 1 and 0 < a < +00. Consider
L:= inf{E([X] + rXl):O ~ X ~ a a.s., EXT = dT }. (25)
1) Let 1 < a ¢ Nand 0 < d ~ 1, then
(26)
1 *) Let a < 1, 0 < d ~ a, then L = a -T dT •
2) Let 2 < a ¢ Nand k + 1 ~ d ~ k + 2 for k E {O, 1, ... , [a] - 2}, then
L = 2dT + (2k + l)(k + 2Y - (2k + 3)(k + lY. (27) (k+2Y-(k+1Y
3) Let 1 < a ¢ N and [a] ~ d ~ a, then
L _ 2dT + aT (2[a] - 1) - [aY(2[aJ + 1) - ~_~]T . (28)
4) Let a E Nand 0 < d ~ 1, then
L = dT. (29)
OPTIMAL BOUNDS ON THE AVERAGE 7
5) Let a E N - {I} and k + 1 :::; d :::; k + 2, where k E {O, 1, ... , a - 2}, then
L is given by (27).
6) Let A:= [0,+00) (i.e., a = +00) and 0 < d:::; 1, then
7) Let A := [0, +00) and k + 1 :::; d:::; k + 2, where k E {O, 1,2, ... }, then L is given by (27).
(30)
Proof. The lower-envelope of the closed convex hull M is formed by the points Po:= (0,0), {Qk := ((k + lY, 2k + 1), k = 0,1,2,3, ... }, Q[al-l := ([ar,2[a]-I), F:= (ar,2[a] + 1). Here
which is increasing in k since 0 < r < 1, and slope(PoQo) = 1. Notice that slope(QoQl) = 2r:l > 1. When ° < a ¢ N we observe that
Therefore the lower envelope of M is made out of the line segments PoQo, QOQb QIQ2' .. ·' QkQk+b ... , Q[a]-2Q[a]-1 and Q[al-I F . In the cases of a E N or a = +00 we have a very similar situation. Cases (1), (4) and (6) are obvious.
The line (QkQk+d has equation in the (xr,y) plane
2xr + (2k + 1)(k + 2Y - (2k + 3)(k + lY (31) y= (k+2Y-(k+l)r .
Now the cases (2), (5) and (7) follow easily by the principle of optimal distance.
Finally the line (Q[al-lF) in the (xr,y) plane has equation
2xr + ar(2[a]- 1) - [an2[a] + 1) y = ar - [a]r . (32)
Thus, case (3) now is clear by applying again the principle of optimal (min­ imum) distance. 0
The more complicated case of L for r ~ 1 follows in several steps. Here the lower envelope of the closed convex hull M is formed again by the points Po = (0,0), Qo = (1,1), ... ,Qk = ((k+ ly,2k+ 1),.·.,Q[al-l = ([a]r,2[a] -1), F = (ar,2[a] + 1) or F* = (ar,2a) if a E N. In the case of
8 GEORGE A. ANASTASSIOU
A = [0, +00), the lower envelope of if is formed just by Po, Q k; all k E Z+. Observe that the
is decreasing in k since r 2: 1.
Theorem 4. Let A = [0, +00), (i.e., a = +00) r 2: 1 and d > O. Consider
L:= inf{E([X] + rXl):X 2: 0 a.s., EXT = dT}. (33)
We meet the cases: 1) If r > 1 we get that L = O. 2) If r = 1 and 0 < d ~ 1, then L = d. 3) If r = 1 and 1 ~ d < +00, then L = 2d - 1.
Proof. We apply again the principle of optimal (minimum) distance. When r > 1 the lower envelope of if is the x T -axis. Hence case (1) is
clear. When r = 1, then slope(QkQk+d = 2 any k E Z+ and slope(PoQo) = 1.
Therefore the lower envelope of if here is made by the line segment (PoQo) with equation y = x, and the line (QoQ 00) with equation y = 2x - 1. Thus cases (2) and (3) are established. 0
Theorem 5. Let a E N - {l}, d> 0 and r 2: 1. Consider L as in (25).
1) If 2~~:::~) 2: 1 and 0 < d ~ 1, then
2) If 2!~:::~) 2: 1 and 1 ~ d ~ a, then
L= 2(a-l)dT+(aT -2a+l). aT -1
3) If 2i~=~) ~ 1 and 0 < d ~ a, then
(34)
(35)
(36)
Proof. Here mo := line(PoQo) has equation y = x T and slope(mo) = 1. Also ml := line(QoQa-l), where Qa-l = (aT, 2a - 1), has equation
2( a-I )XT + (aT - 2a + 1) y = T 1 ' a -
(37)
. h I ( ) 2(a-l) WIt s ope ml = ar-l .
Notice also that slope(QkQk+d is decreasing in k. If 2i~=~) ? 1, then
the lower envelope of M is made out of the line segments PoQo and QOQa-l. Hence the cases (1) and (2) are now clear by application of the principle of minimum distance.
If 2~~=~) ~ 1, then the lower envelope of M is the line segment (POQa-l) with associated line equation
( 2a - 1) T y= -- x. aT
Thus case (3) is established similarly.
Theorem 6. Let 2 < a ¢ N, n := [a], d> 0 and r ? 1. Assume that
1 < 2(n - 1) < 2
Consider L as in (25). 1) If 0 < d ~ 1, then
2) If 1 ~ d ~ [a], then
- nT - 1 - aT - nT
L = 2([a]- l)dT + ([a]T - 2[a] + 1). [ay - 1
3) If [a] ~ d ~ a, then
L = 2dT + aT(2[a] - 1) - [a]T(2[a] + 1). aT - [a]T
(38)
o
(39)
(40)
(41)
(42)
Proof. The lower envelope of M is made out of the line segments: {Po = (0,0), Qo = (1, I)} with equation y = XT, {Qo = (1,1), Q[a]-l = ([ar, 2[a]­ I)} with equation
2([a]- l)xT + ([a]T - 2[a] + 1) Y = [a]T - 1 '
and {Q[a]-l, F = (aT, 2[a] + 1)} with equation
2xT + aT(2[a] - 1) - [aY(2[a] + 1) Y = aT - [a]T .
Then cases (1), (2), (3) follow immediately by application of the principle of minimum distance as before. 0
10 GEORGE A. ANASTASSIOU
Theorem 7. Let 2 < a rJ. N, n := [aJ, d > 0 and r 2: 1. Assume that
2(n - 1) 2 1> > . - nT - 1 - aT - nT (43)
Consider L as in (25). If 0 < d::; a, then
(44)
Proof. Here the lower envelope of M is made out of the line segment {Po = (0,0), F = (aT, 2[a] + I)} with equation
y = (2[al T + 1) XT,
Etc,
Theorem 8. Let 2 < a rJ. N, n := [aJ, d> 0 and r 2: 1. Assume that
and
Consider L as in (25). 1) If 0 < d $ [aJ, then
2) If [a] $ d $ a, then
2(n - 1) < 2 < 1 nT - 1 - aT - nT -
2n - 1 2 --<--nT aT - nT
L = (2[a] - 1) dT, [a]T
L _ 2dT + aT (2[a]- 1) - [ar(2[a] + 1) - aT - [a]T '
o
(45)
(46)
( 47)
(48)
Proof. Here the lower envelope of M is given by the line segments: {Po = (0,0), Q[a]-l = ([ar, 2[a]- I)} with equation
_ (2[a]- 1) T
Y - [a]T x
and {Q[a]-l, F = (aT, 2[a] + I)} with equation
2xT + aT (2[a]- 1) - [a]T(2[a] + 1) Y = aT - [a]T ' (49)
OPTIMAL BOUNDS ON THE AVERAGE 11
o
Theorem 9. Let 2 < a ¢ N, n := [aJ, d > 0 and r 2: 1. Assume that
2(n - 1) < 1 < 2 nT - 1 - - aT - nT (50)
Consider L as in (25). Then L is given exactly as in cases (1) and (2) of Theorem 8.
Proof. The same as in Theorem 8.
Theorem 10. Let 2 < a ¢ N, n := [aJ, d > 0 and r 2: 1. Assume that
and
Consider L as in (25). 1) If 0 < d ~ 1, then
2) If 1 ~ d ~ a, then
2 < 1 < 2(n - 1) aT - nT - - nT - 1
2n -->1. aT - 1
L = 2[a]dT + (aT - 2[a]- 1). aT - 1
o
(51)
(52)
(53)
(54)
Proof. Here the lower envelope of M is made out of the following line segments: {Po = (0,0), Qo = (1, I)} with equation y = xr, and {Qo, F = (a T ,2[a] + I)} with equation
2[a]xT + (aT - 2[a]- 1) y = aT - 1 . (55)
Etc. o
Theorem 11. Let 2 < a ¢ N, n := [aJ, d> 0 and r 2: 1. Assume that
2 2(n- 1) 1< < .
- aT - nT - nT - 1 (56)
Consider L as in (25). Then L is given exactly as in cases (1) and (2) of Theorem 10.
Proof. Notice that the lower envelope of M is the same as in Theorem 10, especially here a;~l > 1. Etc. 0
12 GEORGE A. ANASTASSIOU
Theorem 12. Let 2 < a ~ N, n := [a], d> 0 and r ~ 1. Assume that
2(n -1) < 2 < 1 nT - 1 - aT - nT - (57)
and 2n -1 2 -->---nT - aT - nT (58)
Consider L as in (25). If 0 < d ~ a, then
(59)
Proof. Here the lower envelope of Sf is the line segment PoF with equation
(60)
Etc. o
Theorem 13. Let 2 < a ~ N, n := [a], d > 0 and r ~ 1. Assume that
2 < 1 < 2(n - 1) aT - nT - - nT - 1
(61)
(63)
Proof. Same as in Theorem 12. o
Proposition 1. Let a = 1, r ~ 1 and 0 < d ~ 1. Then L = dT.
Proposition 2. Let 0 < a < 1, r ~ 1 and 0 < d ~ a. Then L = a-TdT.
Proposition 3. Let a = 2, r ~ ~ and 0 < d ~ 2. Then
Proposition 4. Let a = 2, 1 ~ r ~ ~~. 1) If 0 < d ~ 1, then L = dT.
2dr 2) If 1 ~ d ~ 2, then L = 2r-l'
OPTIMAL BOUNDS ON THE AVERAGE 13
Proposition 5. Let 1 < a < 2, r ~ 1 and 2(ar - 1)-1 > 1. 1) If 0 < d ~ 1, then L = dr. 2) 111 < d < a then L = 2dr tar -3.
- - , ar-l
Proposition 6. Let 1 < a < 2, r ~ 1 and 2( ar - 1)-1 ~ 1. IfO < d ~ a, then L = 3a-r dr •
Open Problem. In (1) and (2) replace (ltl~[tl) by (.x[t] + (1- .x)ftl),
where.x E (0,1), .x -:f ~. Then find the general Ua, La .
Acknowledgement. The author wants to thank the referee for making interesting comments.
References
[1] G.A. ANASTASSIOU, Moments in probability and approximation theory, Longman Sci. & Tech., Harlow, UK, 1993.
[2] G.A. ANASTASSIOU and S.T. RACHEV, Moment problems and their applications to characterization of stochastic processes, queueing theory, and rounding problem, in Proc. 6th S.E.A. Meeting on Approximation Theory, pp. 1-77, Marcel Dekker, New York, 1992.
[3] J .H.B. KEMPERMAN, The general moment problem, a geometric approach, Ann. Math. Statist. 39(1968), 93-122.
[4] S.T. RACHEV, Probability metrics and the stability of stochastic models, John Wiley & Sons, New York, 1991.
THE COMPLETE SOLUTION OF A ROUNDING PROBLEM
UNDER TWO MOMENT CONDITIONS
"
TOMASZ RYCHLIK Institute of Mathematics, Polish Academy of Sciences, Chopina 12, 87100 Torun, Poland
Abstract. We complete a partially known solution of the problem of deter­ mining the extreme expectations of arbitrarily rounded nonnegative (pos­ sibly bounded) random variables with given two moments: the expectation and another moment of a positive order.
1. STATING THE PROBLEM AND INTRODUCTION
We consider the family of nonnegative random variables with a given, pos­ sibly infinite range and two fixed moments
o ~ X ~ a, a.s., 0 < a ~ +00, EX = mI,
O<r=J.l.
Our problem consists in determining the accurate bounds for the expec­ tation of c-rounded variable E[X]e, 0 ~ c ~ 1, where the integer valued c-rounding function [']e is defined by
[x]e = inf {k E Z: x ~ k + c}, 0 ~ c < 1,
and [x h = inf {k E Z: x < k + 1}.
The cases c = 0 and c = 1 describe the popular Adams and Jefferson rules of rounding, or the ceiling and floor of a number, respectively. We shall write
L(mI' mr )
inf E[X]e, sup E[X]e,
V. Benes and J. ~tipcln (eels.), Distributions with given Margina/s and Moment Problems, 15-20. © 1997 Kluwer Academic Publishers.
16 TOMASZ RYCHLIK
ignoring in notation the dependence of extremes on the other parameters: a, rand c.
For a number of subcases, the problem was solved in Anastassiou and Rachev (1992) and Anastassiou (1993, Chap. 4) by means of a geometric approach to moment problems due to Kemperman (1968). The crucial tool of the geometric moment theory, proved independently by Richter (1957) and Rogosinsky (1958), is the fact that for every probabilistic measure with given number of moment conditions, there exists a probability distribution with the same moments as the original one which is supported on a finite number of points, exceeding the number of conditions by no more than one.
To adopt the theory for our problem, we introduce some useful notation. Given 0 < r 'II and 0 ~ c ~ 1, put t = (t,tT) E R2 and T = (t,tT,[t]e) E R 3 , 0 ~ t ~ a. We shall further adhere to the convention that points of the plane and space are written in bold lower case and capital letters, respectively (e.g., 0 = (0,0) and 0 = (0,0,0)). We denote S]S2 = {t E R2 : SI ~ t ~ S2}. Also, let 8182 and 6818283 stand for the line segment and triangle, respectively, determined by their extreme points.
The problem is well posed iff m = (ml' mT ) E cooo, the region between the curve tT and the line aT-It, 0 ~ t ~ a. The solution consists in deter­ mining the endpoints ML = (m, L(m)) and Mu = (m, U(m)) of the line segment arising by intersecting the moment space co{T E R3: 0 ~ t ~ a} by the vertical line t = m. Since t t-t t is continuous and t t-t [t]e, 0 ~ c < 1, is lower semi continuous , and t t-t [tlt is upper semicontinuous, inf E[X]e, o ~ c < 1, and sup E[Xh are attained (see Kemperman (1968, Remark on p. 99)). To make possible determining the unattainable extremes, we should add Lk = (tk, k) = (k+c, (k+ct, k) and Uk = (tk' k+ 1), 0 ~ k < [ale = n, say, to the moment space (see Kemperman (1968, Theorem 6)).
2. THE SOLUTION
We proceed to review the results of Anastassiou and Rachev (1992) and then to present new ones. We have
L(m)
(1) (2)
if m E CO{tk: 0 ~ k < n}. Note that these trivial bounds, implied by X ~ [X]e + C ~ X + 1, are the best ones here. For c < 1, (1) is attained by any distribution on k + c with probabilities o.k, 0 ~ k < n, satisfying L o.ktk = m. Bound (2) is merely approximated by the sequences supported on k + c + €kj, 0 ~ k < n, €kj ! 0 as j -+ 00, and obeying the moment constraints. (For the Jefferson rule, it is the supremum that
ROUNDING UNDER TWO MOMENT CONDITIONS 17
is attained). If m E coot;;, then L(m) = 0, being attained by any pair o ~ SI,S2 ~ c such that m E 81,82.
Suppose that a < +00 and m to the tetragon with the vertices 0, to, tn- l and a. If A = (a, n) lies above the plane containing 0, Lo and Ln-I, then
and
alO + a2LO + a3Ln-l, for mE 6.ototn_I, alO + a2Ln-l + a3A, for m E 6.otn- l a,
ML - { alO + a2 LO + a3A , for mE 6.otoa, - alLo + a2Ln-l + a3A, for mE 6.tOtn_l a,
(3)
(4)
otherwise. In each formula of (3) and (4), the vertices of triangles define the support of the (limiting) extreme distribution, the first and second equa­ tions enable us to calculate the coefficients of the convex combination, and the last one provides L( m). To derive U( m), it suffices to replace Lo, Ln- l and ML in (3)-(4) and the preceding condition by Uo, Un- l and Mu, re­ spectively. The explicit expressions for the extremes are complicated and we do not present them here (see Anastassiou (1993, pp. 91-100)). More­ over, for m E cot;:;a, we have U(m) = n, attained by a combination of some SI, S2 E [n, a].
If a = +00 and r < 1, the for all m from the halfstrip 0 ~ mr ~ cr ,
ml ~ c1-rmn we have L(m) = ml - c1-rmr and U(m) = c-rmr • For a = +00, and r > 1, and m such that 0 ~ ml ~ c, mr ~ cr-lml, it follows that L(m) = 0 and U(m) = C-lml. Except the cases that the moments belong to the joint border of the two regions, the extremes are unattainable and approached by measures with a support point running to infinity and the respective probability vanishing at an appropriate rate.
Neither of extremes is known, when m E cott:;tk, 1 ~ k < n. Also, there are lacking the upper one for m E coot;;, and the lower one for bounded random variables with m E cot;:; a. Such pairs of moments come from variables highly concentrated around the expectation.
Proposition 1. Suppose that m E cotk-:;tk for some 1 ~ k < n. Let VL, vu E [k - 1 + c, k + c] and a, f3 E [0,1] be uniquely defined by the equations
Then
18 TOMASZ RYCHLIK
and the limiting distributions of the sequences approximating (7) and (8) are the two-point distributions:
respectively.
Pr(X = k + c) = 1- Pr(X = vu) = {3,
Proof We examine the infimum problem, only. The other one may be handled in much the same way and we leave it to the reader to adapt the following reasoning for it. Using arguments of the geometric moment theory, we can confine ourselves to the random variables distributed on three point, say v, wand z, at most. Due to the two moment conditions, the distributions are determined by the support points. The idea of proof is to provide other values such that the respective moment point M = (ml' mT , m(c)) E R 3 ,
m(c) = E[X]c, has a smaller third coordinate than the assumed distribution for all possible positions of v, w, z. Consecutive improvements lead to the assertion. It is evident that some numbers among v, wand z must belong to [k - 1 + c, k + c], since otherwise m ¢ cotk~tk' Because there are random variables with assumed moments ml and mT , and m(c) < k (see the statement of the Proposition), we can conclude that the candidates for the infimum must have at least one support point less than or equal to k - 1 + c.
We first consider the case 0 5 v 5 k - 1 + c < w < z 5 k + c. The first step consists in replacing wand z by a single point. Let M = (m, m(c)) E 6.VWZ and V' be the element of the line segment WZ such
that M E VV'. Let v" denote the point at which t';;tk and the half­ line '0'0' ...... , beginning at v and passing through v', cross each other. Observe that the halfline VV" ...... runs beneath VV' ...... , because they reach level k at v" and v', respectively. Therefore, for M' = (m,m(c)) E VV", we have
m(c) 5 m(c) ' If v = k - 1 + c, this is the final claim. Otherwise we proceed to the second step, in which we present a way of
eliminating points lying outside of [k-l +c, k+c] . Suppose that v < k-l + c < z 5 k + c and M En. Define v' = vz n tk-ltk, V' = (v',v(c)) E VZ,
and V" = (v',v(~)) ELk-ILk. By (1) , v(~) = L(v') = v~ - c 5 v(c) so that
the whole V"Z is below V'Z . In particular, for M' = (m,m(c)) E V"Z yields m(c) 5 m(c)' On the other hand, M' E 6.Lk-lZLk, and applying the arguments of the first step, we can reduce m(c) by taking a single point instead of Z and Lk .
It remains to treat the case of two support points, say wand z, situated apart from (k - 1 + c, k + c]. As in the second step of the previous case, we replace both Wand Z by W', Z' ELk-ILk so that 6. VW' Z' with
ROUNDING UNDER TWO MOMENT CONDITIONS 19
the modified M' lie under L VW Z and M, respectively. Since M' may be represented as a combination of Lk-b Z and Lk, it suffices to take the first step again to derive the desired conclusion. 0
The method of proof carries over to the remaining open problems.
Proposition 2. If m E co~, then U(m) = {3 for some 0 ~ f3 ~ 1 and o ~ Vu ~ c satisfying
(1 - (3)vu + (3to = m. (9)
Proposition 3. If a < +00 and m E cot~a, then L(m) = n - a for o ~ a ~ 1 and n - 1 + c ~ VL ~ a such that
(1 - a)vL + atn - I = m. (10)
In contrast with the previously known solutions for which the points of the lower and upper envelopes of the moment space compose pieces of several plane~we here obtain curved surfaces. E.g., the part of the ~ce for m E cotk-Itk consists of two cones with the common base COUk-lLk and vertices Lk-I and Uk. This is the implication of the fact that each extreme distribution has two support points. One is in common for various moment conditions and yields either Lk-I (the infimum case) or Uk in the moment space. The moments of the other vary along the arc Uk~k.
Only for a few specific r we can solve equations (5), (6), (9) and (10) analytically and derive explicit expressions for Land U in Propositions 1-3. For r = 2 and r = ! we have
L(m)
U(m)
and
k - 2" 2 ' (m! - vlk + 1- c)2 + ml - m 1
2 2"
U(m)
respectively. The bounds hold for 1 ~ k < n if m E cotk~tk. The upper ones are also true with k = 0 as m E co04;, and the lower ones apply to k = nand mE cot:=;a (for finite a). They become more meaningful when
20 TOMASZ RYCHLIK
we realize that m2 - m~ =VarX and ml - m~ = Van/X. In the case r = 3, 2"
we obtain (7) and (8) with a and f3 being expressed by huge formulas as the greater and smaller roots of the following quadratic equations:
(m3-a~k-l+c)3
(1 - f3)2[ml - a(k + c)],
respectively. The ranges of application coincide with those given above. Similar formulas can be concluded for r = ~ .
We finally point out that simply putting c = 0 and 1 we derive the extreme expectations for the Adams and Jefferson rounding proportions, respectively.
References
Anastassiou, G .A. and Ra.chev, S.T . (1992) Moment problems and their applications to characterization of stochastic processes, queueing theory, and rounding problem. In: Proc. 6th S.E.A . Meeting "Approximation Theory". Marcel Dekker, New York, pp. 1-77.
Kemperman, J .H.B. (1968) The genera.! moment problem, a geometric approach, Ann. Math. Statist., Vol. no. 39, pp. 93-122.
Richter, H. (1957) Para.meterfreie Abschatzung und Rea.lisierung von Erwartungswer­ ten, Blatter der Deutschen Gesellschaft for Versicherungmathematik, Vol. no. 3, pp. 147-161.
Rogosinsky, W.W. (1958) Moments of non-negative mass, Proc. Roy. Soc. London Ser. A, Vol. no. 245, pp. 1-27.
METHODS OF REALIZATION OF MOMENT PROBLEMS
WITH ENTROPY MAXIMIZATION
VALERIE GIRARDIN UFR Sciences, Universite de Caen, BP5186, 14032 Caen, Statistique, Batiment 425, UPS, 91405 Orsay Cedex, France [email protected]·fr
Let K be a finite subset of 71/ including 0 = (0, ... ,0). Let c) = (¢>k)kEK be a family of measurable and linearly independent complex or real func­ tions defined over a compact I C lRr , with ¢>o == 1. Let p be a reference measure over I.
We construct here h maximizing an entropy functional in both the fol­ lowing problems
(1)
where C = (Ck)kEK is positive in the sense of Krein & Nudelman [5] (if (ak)kEK is such that LkEK ak¢>k ~ 0 over I then Lk akCk ~ 0), and
(2)
where D = (dkd(k,I)EK2 is a positive definite matrix. We call PE the subspace of polynomials of L;(I) spanned by (¢>k)kEE.
Problem (1) is a generalized moment problem. Relation (2) defines a d-scalar product over PK by
Hence Problem (2) is a problem of realization of this d-scalar product. It may also be considered as a moment problem for the family (¢>k¢>d(k,I)EK2.
This has a probabilistic interpretation. Let D be the autocovariance ma­ trix of a purely nondeterministic multidimensional process X = (Xk)kE,Zr.
21
V. Benes and 1. Stipdn (eds.), Distributions with given Marginals and Moment Problems, 21-26. © 1997 Kluwer Academic Publishers.
22 VALERIE GIRARDIN
It admits a stationary realization if there exists a nonnegative bounded measure J.L over the torus nr such that
The process admits then the representation
where Z( >.) is a process with orthogonal increments and energy J.L. If J.L is absolutely continuous with respect to the Lebesgue measure A over nr , its density h is the spectral density of the process. Note that the coefficients dkl of D are functions of k - I only so that Problem (2) becomes here a problem (1) for a set of indices of the form K - K.
By extension, see Priestley [7], if D admits a representation
dkl = 14>k(>')<M>')h(>')dP(>'), (k,l) E yzr x yzr
for a given family ~, then X admits the spectral representation
(3)
The functions 4>k may have a precise physical meaning, real exponentials for dampings, algebraic monomials for generalized power spectra, product of a function by complex exponentials for the output process of a linear system ...
So, (2) is the problem of representing the covariance matrix of (Xk)kEK by a measure of density hand (1) is a moment problem for the process X.
We are interested here in solutions maximizing an entropy functional. Given W E C2(JR+), a concave function, the entropy is defined for non­ negative functions h by
SIJ!(h) = 1 W[h(>.)]d>..
The spectral density h* which maximizes an entropy among the solu­ tions of (1) or (2) corresponds to a process X* with particular properties. For the Burg entropy, where W(h(>')) = logh(>'), the distribution of X* is the most regular possible, for the Kullback (or Shanon) entropy, where W (h(;\)) = h(;\) log h(;\), it is the closest to a given one ...
Methods of realization of moment problems with entropy maximization 23
This probabilistic point of view leads to results such as developed in Part 1 for the entropy Slog.
In Part 2, we use a Hilbertian setting to construct a rational function solution of (2).
Its denominator is the prediction polynomial. If X admits a spectral representation (3), the prediction polynomial gives the linear predictor of Xo given (XkheK\{O}, optimal in quadratic mean.
The construction is based on the projection of 4>0 onto PK\{O} for thed­ scalar product and on its connection with the usual scalar product of L~(I). We are interested here in showing that its practical application comes down simply to solve linear systems. Note that it also yields a fixed point method to construct a rational function solution of (1), with a chosen numerator. When it is chosen to be constant, the solution maximizes the Burg entropy.
1. Probabilistic point of view
We gave in [4] the form of the density h* maximizing an entropy functional. We shall give here some of its probabilistic consequences. Let us first
recall it.
Theorem 1 Let 1J~ be the set of nonnegative solutions of (1). Assume there exists a positive h * E V~. Then
SIJ/(h*) = max SIJ/(h) if and only if ho = !(qT/)-l 0 g, g E PK. he1J'J, P
In certain cases, it is possible to get rid of the existence assumption. Theorem 2 Let r = 1 or 2. Let (> be a family of continuous functions. Then a polynomial 9 = EkEK gk4>k in PK exists such that h* = 1/g E V~. Slog is maximum over 1J~ at h*.
Proof Dacunha-Castelle & Gamboa [3] proves that agE PK exists such that 1/g is solution of (1) and is locally integrable. But if r =1 or 2, this is true only if 9 never takes the value 0 over I. And Theorem 1 yields the result. 0
This theorem has a probabilistic consequence which we give here for r = 1.
Corollary 1 Among all stationary processes with given autocovariance co­ efficients (co, ... ,cn), there exits an AR process. This process gives the max-
imum of Slog and its spectral density is h*(>.) = En 1 k)". k=O ak cos
This can be extended to processes having a representation of the form (3). It can also be generalized to problems combining (1) and (2) such as the determination of processes with given autocovariances and impulse response
24 VALERIE GIRARDIN
coefficients. For r = 1, this problem is equivalent to the determination of h E L1 (11) analytic such that
{ 2~ lr eikA lh(A)1 2dA = ci, 0 ~ j ~ n,
2~ Z eikAh(A)dA = dk, 1 ~ k ~ m.
Franke [2] proved that Slog is maximum among the solutions at the den­ sity of the ARMA process, that is for a trigonometric rational function. Extensions of this result for multidimensional processes admitting a spec­ tral representation (3) will be given in a future paper, Castro & Girardin [1]. They lead to the construction of generalized rational functions such as made in Part 2.
Optimization results too can be developed, such as the following which is a direct consequence of the preceding propositions.
Theorem 3 Let tP E L1(I). Let c E [-1,1]. Let 1'(1,4» be the set of non negative functions h E LHI) such that II h(A)dA = 1 and II htP(A)dA = c. Then,
sup sup Slog(h) = 1Iogh*(A)dA, forh*(A) = l+a1cosA 4>eL~ (I). he'D(l .• ) 1
This means that the maximum of the Burg entropy for densities under the constraint II h4>(A)dA = c is obtained for the density of the AR process.
2. Construction of a rational function
We construct here first a rational function solution of (2). Its form is given in the following theorem, which characterizes also the prediction polynomial
Pmin' We ca.ll rational functions the quotients of elements ofPKePK. We have
the orthogonal decomposition
Let (Pk)keL be an orthonormal system for the usual scalar product of L~(I) suited to this decomposition such that (Pk)keK is deduced from (tPk)keK starting by Po constant as 4>0. Let AK be the change matrix from (tPk)kEK to (PklkeK' Theorem 4 Let D = (dkl)(k,l)eKxK be a positive definite matrix. Then,
the projection - P min for the d-scalar product of Po onto PK\{O} is
Methods of realization of moment problems with entropy maximization 25
Moreover, i!(Po+Pmin)(>'):f 0 over I, there exists Q E (PKePK)8PK such that h = (lIPo + Q)/(Po + Pmin) satisfies (2). The proof, given fully in [4], is based on the following idea: lIPo is the pro­ jection for the d-scalar product of Po + Pmin onto PK. Set! E PKePK such that < <Pk,<Pt >d=< !<Pk,<P/ >q(I) for (k,l) E K x K. Then lIPo is the projection for the usual scalar product of !(Po + Pmin) onto PK. Thus 1I and Pmin are determined. Then h is taken such that its projection for the usual scalar product onto (PKePK)8PK is equal to the same projection of ! - lIPo/(Po + Pmin).
Let us show its practical form in the algebraic case. The product of two elements of PK belongs to PK+K, where K+K={l + 1'/1 E K, l' E K}. Therefore, a measure realizes the d-scalar product (2) if its density h is so­ lution ofthe moment problem (1) for <Pk(>') = >.k, k E K+K and Ck = d/+11. Note that here D is a definite positive matrix if (Ck)kEK is a positive defi­ nite sequence and that h is here a rational function in the usual sense.
Take for example r = 1, p equal to A over I = [-1,1] and K = {0,1}. Then! = c(jPo + ci PI + ciP2, where c* = Akc and
bO = 0 ~ 0 . 00) (~ 00)
o e2 ;jl 0 ~
Here, Pmin = b1Ph and the projection of !(Po + Pmin) onto PK is
( < ! Po, Po > <! Po, PI > ) ( 1) (1I) < ! Po, PI > <! Ph PI > b1 0'
so that b1 = -acI/bc2 and 1I = a2(co - CUC2). Then h = (lIPo + Q)/(Po + b1Pt}, where Q = b2P2 is given by
11 h(Po + Pm in)P2(>')d>' = c2. -1
Hence b2 = [aK2(coK - lIH)]/[(eoK2 + e2)H - 2e2K], where K = -Cl/C2
and H = In[(1 + K)/(1 - K)], and finally
h(>') = (all + b2eo + b2e2>.2). (a + aK >.)
If 0 < leI I < c2, h is well defined and its denominator is nonnegative over I. Its numerator too is nonnegative over I if f32/1l belongs to the in­ terval [-a(eo + e2)-1,-aeo1] or [1s,ts]. If Co = 1 (corresponding to a
26 VALERIE GIRARDIN
probability measure), this is generally true for C2 - leI I et 1- C2 both small enough or both large enough.
This construction yields a fixed point method for the determination of
a rational function h = I/fo°: j solution of the moment problem (1)
whose numerator is chosen and whose denominator is a polynomial of PK. Indeed, Q is above implied by the process of construction from 1/ and Pmin '
If the moments are only given for k E J(, this can be reversed, Q implying 1/ and P. Hence, a linear system may be written whose unknowns are the coefficients of P. Its matrix involves the known moments c of the solution h and its now unknown moments a for ti>kti>/. The determination of the fixed point of the function connecting a to the moments of h constitutes an iterative process for the determination of h.
If Q is chosen equal to 0 over I and p equal to A, then the rational function obtained by this fixed point method is of the form
1 h= ,
EkEK akti>k
and maximizes the entropy Slog among the solutions of (1). Note that in the trigonometric case (i.e. for stationary processes), the
denominator of h can be expressed as the square of a polynomial, a more usual form. A Hilbertian setting was used in Landau [6] too to construct solutions of the trigonometric moment problem for r = 1.
References
1. G. Castro & V. Girardin Multidimen!ional nondationary proceues and entropy maximization In Preparation
2. J. Franke ARMA proceue! have maximal entropy among time series with prescribed autocovariances and impulse re!pon!e! Adv. Appl. Prob., V17, p810, 1985
3. D. Dacunha-Castelle & F. Gamboa Maximum d'entropie et probleme de! moments Ann. Inst .H. Poincare, Prob. Stat., V26, p567, 1990
4. V. Girardin Methodes de realisation de produit !calaire et de probleme de moments avec maximi!ation d'entropie Stud. Mat., 1996
5. M. Krein et A. Nudelman The Markov moment problem and extremal problems Amer. Math. Soc., 1977
6. H. Landau Maximum entropy and the moment problem Bulletin of the American Mathematical Society, V16, 1987
7. M. Priestley Non linear and nonstationary time series analysis Academic Press, London, 1988
MATRICES OF HIGHER MOMENTS:
SOME PROBLEMS OF REPRESENTATION
E. KAARIK Institute of Mathematical Statistics University of Tartu 2 J.Liivi, EE2400 Tartu, ESTONIA
1. Introd uction
There exist different representations of matrices - via tensors, sets, block­ matrices, vectors. Each of these representations has some preferencies and some shortages in different procedures. In multivariate statistics some addi­ tional problems arise due some kind of symmetry of the arrays, causing the identity of some groups of elements. It is reasonable to find a parsimonious representation of multivariate matrices, avoiding repeated elements caused by symmetry of the array. In the paper the compressed representation for symmetrical multi-dimen­ sional arrays (Tiit, Tammet, 1994) will be introduced. The results will be applied for the special case, when a matrix of empirical moments as the array is regarded, and some estimation algorithms will be given (Kaarik, 1994; Kaarik, Tiit 1995), using the classification of moments by their types.
2. Index vectors, index-sets and arrays
Let us start with the definitions of the basic concepts. Definition 1. The vector I = (il' ... , ik) is said to be an index-vector, if
all its components are integers, ij ~ Pj, j = 1, ... , k, where (Pl,'" ,Pk) is a fixed vector. The natural number k is the cardinality of I. If the condition ij < ij+l or ij ~ ij+l is satisfied for j = 1, ... , k - 1, then the index-vector I is said to be ordered (partially ordered). In the case of opposite inequalities the index-vector is D-ordered or partially D-ordered (decreasingly). If all parameters Pj are equal, Pj = p, j = 1, ... , k, then the index-vector I is
27
V. Benes and J. 5tepan (eds.), Distributions with given Marginals and Moment Problems, 27-33. © 1997 Kluwer Academic Publishers.
28 E. KAARIK
said to be symmetrical and p is said to be the order of index-vector. The set of all symmetrical index-vectors of order p and cardinality k will be denoted by I(p, k).
Definition 2. Let us have a set of elements ail ... ik' depending on an index-vector I = (il, ... ,ik), where 1 ~ ij ~ Pj, j = 1, ... ,k. Then we say that we have a k-dimensional array A. In some cases we assume that the symmetricity conditions are fulfilled, as well:
Definition 3. The array A with elements ail ... ik is said to be symmetric, if the following conditions are fulfilled: 1) All indices have the same set of values, Pj = p, j = 1, ... , k,
i.e, Ie I(p,k). 2) The elements ail ... i k are exchangeable by the indices. Almost all useable in practice statistical functions (empirical or theoreti­ cal) of k random arguments, calculated by p-variate random vector (e.g. moments, central moments, cumulants and other statistics) satisfy the con­ ditions 1 - 2.
3. Compressing transformations
3.1. EQUIVALENCE CLASSES OF INDEX-VECTORS
Let us regard the symmetric index-vectors IE I(p, k), where the parameter p and the dimension k are fixed. Let us define the equivalence classes of index-vectors by permutations of indices in the following way:
Definition 4. Let I = (ill" .,ik) and J = (jl, ... ,jk) be two index­ vectors from the class I(p, k). If there exists a permutation P so, that the equality I = P J holds, then we say that the index-vectors I and J are equivalent. Shortly, the index-vectors, corresponding to the same index-set, are equiv­ alent. As a representative of a class of equivalent index-vectors we will take the partially D-ordered index-vector l' = (i~, ... ,i~J from this class, where
., ., . 2 k tj ~ tj_l, J = , ... , . (1)
The index-vectors that belong to different classes are said to be essentially different. In the future we will regard the set ili(p, k) of equivalence classes (essentially different index-vectors).
3.2. ORDERING OF INDEX-VECTORS
Let us define the position f of a component if of an index-vector I as its successive number in the index-vector (ib"" ik)' Every symmetric index­ vector (i1 , ... , ik) can be regarded as an k-figure natural number in the
MATRICES OF HIGHER MOMENTS 29
positional number system having the basis p. Notice that for getting the usual form of natural number we should take the transformation ij = ij - 1, j = 1, ... ,p. Let us order all index-vectors ofthe set 'II(p, k) lexicographically; as a result we receive the sequence £(p, k) with terms h (here h is the h-th element ofthe sequence). The sequence £(p,k) = {h} does not include all possible k-figure p-numbers, but only these which satisfy the conditions (1).
Example 1. Let us regard the set of fourth moments of a five-variate random vector i = (Xl,X2,X3,X4,XS)' The usual way to present the set of moments is to use the moment's matrix (of order 5 x 5 x 5 x 5). In this matrix, which is usually written with the help of 5 x 5 blocks, there are 625 entries. But most of the elements of the matrix are superfluous, as almost all moments are represented in this matrix repeatedly. The situation is similar to that of correlation matrix, where all correlation coefficients are represented twice: in the upper and in the lower triangle. The main practical problem is to eliminate the repeated copies of the mo­ ments from the matrix. The order of moments k is dimensionality of the index-vector and the dimension of random vector p is the order of index­ vector. It is possible to get all different fourth moments of the random vector, identified by the indices, as a finite sequence £(p, k). In Table 1 all different fourth moments of a five-variate random vector are represented as 4-dimensional index-vectors, identifying the indices of random variables (the sequence £(5,4)).
TABLE 1.
h(I) h(I) h(I) I h(I) h(I)
1 1111 15 3333 29 4431 43 5331 57 5521 2 2111 16 4111 30 4432 44 5332 58 5522 3 2211 17 4211 31 4433 45 5333 59 5531 4 2221 18 4221 32 4441 46 5411 60 5532 5 2222 19 4222 33 4442 47 5421 61 5533 6 3111 20 4311 34 4443 48 5422 62 5541 7 3211 21 4321 35 4444 49 5431 63 5542 8 3221 22 4322 36 5111 50 5432 64 5543 9 3222 23 4331 37 5211 51 5433 65 5544 10 3311 24 4332 38 5221 52 5441 66 5551 11 3321 25 4333 39 5222 53 5442 67 5552 12 3322 26 4411 40 5311 54 5443 68 5553 13 3331 27 4421 41 5321 55 5444 69 5554 14 3332 28 4422 42 5322 56 5511 70 5555
30 E. KAARIK
Let us assign to every index-vector h its rank h = h(I). Our task is to find the ordering rule </>, describing the correspondence h ~ h(I) and </>-1: h(I) ~ h. The ordering rule will be given by the following theorem (Tiit, Tammet, 1994)
THEOREM 1. Let W(k,p) be the set of essentially different index­ vectors (i1, ... ik), where the dimensionality k and the parameter pare fixed. The rule </> for the lexicographical ordering of index-vectors in the sequence £(p, k) is the following:
k
h(I) = 1 + L A~g+-Lg, (2) g=l
h I ( . .) d AP (p+k-1)(p+k-2)· ... ·p were = t}, ... , tk an k = k! .
The rule </> is a generalization of the well-known operator vech, transforming the upper triangle of a symmetric matrix into the vector. From Theorem 1 immediately two corollaries follow.
COROLLARY 1. The cardinality of the sequence £(p,k) is A~. COROLLARY 2. The transformation </> has a unique inverse transfor­
mation </>-1. The rule </> and its inverse </>-1 are both easily programmable.
4. Classification of higher moments of a random vector
In the set W(k,p) there are different types of fourth moments, depending on the number of variables, included into the moment and the degrees of them. To describe the higher moments of a random vector it is convenient to use the concept of the partition of a number, see e.g. Andrews (1976). Let us regard the set of moments of order k of a p-dimensional random vector, k ~ p, and introduce the following classification, using the concept of the partition of a number.
Definition 5. Let n be a natural number. A set of natural numbers Q = {Q1, ... , qv}, satisfying the condition
is a partition of the natural number n, the addendae qj are said to be the parts of n, v = K,( Q) is the cardinality of the partition Q.
Usually the set Q is represented as a non-decreasingly ordered set of parts. As the parts have the values 1,2, ... , n, so there exists an equivalent representation ofthe partition via its characteristic numbers aj, i = 1, ... , n
in the following sense.
MATRICES OF HIGHER MOMENTS 31
Definition 6. Let Q = {ql, ... , qv} be a partition of a natural number n. Let aj denote the number of repetitions of the integer j in the set Q, i.d., the following equation must be valid:
n
n=I:>.aj. j=l
The numbers aj are said to be the characteristic numbers of the partition Q, and the vector a = (aI, ... an) is its characteristic vector.
Definition 7. Let (Xl,"" Xp) be a given random vector and Q = (ql, .. . , qv) a partition of a natural number k, k ~ p. Then we say that the moment E(X?ll ..... X iq:) belongs to the type Q, if ib"" iv are differ­ ent integers from the set {I, ... , p}. The following problem is to clear up the number K,k(Q) of moments of a type Q in the sequence £(p, k) . The answer to this problem will be given in the following theorem.
THEOREM 2. Let £(p, k) be a sequence of all moments of the order k of a p-variate random vector, and let Q be a partition of the number k, having the characteristic vector a = (al, ... ,ak), see Andrew (1976). Then the number of moments of the type Q equals to the following expression:
k
K,k(Q) = II C;;, (3) j=l
where: dJ = p - bj - 1 , bj = 2::::~=1 aj, bo = 0; j = 1, ... , k and C: = n!j(m!· (n - m)!).
Example 2. Let us regard the set of the fourth moments of a five-variate index-vector, given in Example 1. As the number 4 (the order of moments k) has 5 different partitions ([4], [3,1], [2,2], [2,1,1], [1,1,1,1]), so there are five different types of the fourth moments, see Table 1. To the partition [4] all marginal moments correspond, to the partition [2,2] the moments E(X? . Xi), etc. Let us calculate the cardinalities of different fourth moments of all types K,4(Q). At first, let us write the characteristic vectors of all partitions:
[4]::} (0,0,0,1); [3,1]::} (1,0,1,0); [2,2]::} (0,2,0,0); [2,1,1]::} (2,1,0,0); [1,1,1,1]::} (4,0,0,0);
The cardinalities are calculated using the formula (3): K,4{[4]} = CJ = 5; K,4{[3][1]} = CJ . Cl = 20; 1'1,4 {[2][2]} = Cg = 10; 1'1,4 {[2][1][1]} = Cg . C~ = 30; K,4{[1][1][1][1]} = ct = 5;
It is easy to check from the Table 1 the correctness of results.
32 E. KAARIK
For some types of moments it is quite easy to find the rule for determining their ranks in the sequence £(p, k). E.g., the marginal moment, correspond­ ing to index j, has the rank h(f) = At, j = 1, ... ,p.
4.1. THE RATIO OF NUMBER OF DIFFERENT MOMENTS AND NUMBER OF ALL MOMENTS
THEOREM 3. The ratio of number of different moments and number of all moments of the order k of p-dimensional random vector approaches to the constant b as p tends to infinity, i.e. the following equality is valid
1. A~ 1 1m -k = k' . p-oo p .
(4)
Example 3. Let us see the number of all4-order moments (p4), the num­
ber of different moments (A~) and the ratio (#) by different dimensions p p of random vector
TABLE 2.
AP 4 5 15 35 70 126 210 330
p4 16 81 256 625 1296 2401 4096 ratio 0.312 0.185 0.136 0.112 0.097 0.087 0.080
By connection (4) the limit of the ratio is ~ = 0.0417. Example 4. Construction of a matrix of fourth moments of a random
vector having normal distribution and constant correlations. Let us see the set of fourth moments of a bivariate random vector X = (XI, X 2 ), X rv N2(0, R), DXi = 1. Different fourth moments of the type Q, representative index-vectors and all possible index-vectors are: Q = {4}, 1-'4 = 3; (1,1,1,1),(2,2,2,2); Q = {3,1}, 1-'31 = 3p; (2,1,1,1,) -+ (1,2,1,1),(1,1,2,1),(1,1,1,2);
(2,2,2,1) -+ (2,2,1,2),(2,1,2,2),(1,2,2,2);
Q = {2,2}, J.l22 = 1 + 2p2; (2,2,1,1) --+ (2,1,2,1),(2,1,1,2),(1,2,2,1), (1,2,1,2),(1,1,2,2).
Compressing representation (representative index-vectors) of fourth mo­ ments of a two dimensional random vector is: 1. (1,1,1,1) 3 2. (2,1,1,1) 3p 3. (2,2,1,1) 1 + 2p2 4. (2,2,2,1) 3p 5. (2,2,2,2) 3 As we can see, the compressing representation is a good tool for simpli­ fying the representation procedures and calculation of matrices of higher moments.
References
Andrews, T.W. (1976) The Theory of Partitions. Kiiiirik, E. (1994) Representations of index-functions of random vector. Ma&ter-thesis. Kiiii.rik, E., Tiit E.-M. (1995) Arrays of Multivariate statistics and their representation,
New Trends in Probability and Statistics, Vol. 3, Multivariate Statistics and Matrices in Statistics, pp 313-324, Vilnius, Utrecht, Tokyo.
Tiit, E.-M., Tammet, S.{1994} Using Index-Vectors and Partitions in Multivariate Anal­ ysis, Acta et Comment UT, Vol. 968, pp 55-75
THE METHOD OF MOMENTS IN TOMOGRAPHY AND
QU ANTUM MECHANICS
AND
RACHEV S.T. University of California Santa Barbara Santa Barbara, CA 93106-3110, USA§
1. Radon transform and its applications to computer tomogra­ phy
Suppose an x-ray examination of an object (a body) is performed, and the intensities of the x-ray at both input (1d and output (10) are measured. It is known, that (under some conditions imposed on the object and the intensity of the input x-ray) the following is true:
L j 10g(...2.) = p(x)dL, fo L
where the integral is taken along L, the straight line which the x-ray follows through the body. The problem is to reconstruct the density p of the body based on the line integrals calculated on (some or all) of the straight lines.
The transformation from the density p( x) to the set of all its line in­ tegrals (considered these as functions of the parameters of the line) is the so-called Radon transform. This was introduced by Radon in 1917 [13], who also provided formulae for the inversion of the transform and proved the uniqueness of the reconstruction of the density from its Radon transform.
IThe research was supported by RFFI Grants No 96-01-00852 and 95-01-01260 §The research was conducted while S.T. Rachev was visiting the University of Freiburg
and Kiel. His visits were supported by a Research Award for U.S. Senior Scientists from the Alexander-von-Humboldt Foundation.
35
V. BeMI and J. Stlptin (eds.), Distributions with given Margi1llJis and Moment Problems, 35-52. @ 1997 Kluwer Academic Publishers.
36 KLEBANOV L.B. AND RACHEV S.T.
In fact, by the Radon transform any probability measure in the n - dimensional Euclidean space can be reconstructed in a unique way in terms of probabilities on half spaces. Probabilists refer to this result as the Wold­ Cramer principle; has been obtained independently by Wold and Cramer in 1932, using characteristic functions.
In the medical sciences applications of Radon's result began in the six­ ties with the first tomography machine of A. Cormack and the first use of the commercial tomography machine by G . Hounsfield; they were both awarded the Nobel prize in Medicine in 1979 for this work.
Computer tomography (CT) is based on the inversion of the Radon transform, which allows us to reconstruct in a unique way the density of a measure.1 However such a unique reconstruction is possible only if one knows an infinite number of marginals of the underlying distribution. In practice one can only have a finite number of marginals, and a unique reconstruction of the density is impossible in this case. One can easily see this from the following "CT paradox".
Let I( x) be a probability density function given on Euclidean plane R2 and having a compact support D C R2. Let 01 , ••• , ON be N directions (N unit vectors) in the plane R2. It is obvious, that a measure on R2 of compact support is uniquelly determined by infinitely many marginals.
It was proved in the recent work [3], that for any density I: 0 ~ I(x) ~ 1, xED and for any finite number of directions 0], ... , ON (N < 00) there exists another density lo{ x), xED such that 10 has the same marginals in the directions 01 , • •• ,ON as 1 and such that 10 has only two values 0 or 1. This result gives the following CT Paradox (see (3)):
It implies that for any human object and corresponding projection data there exist many different reconstructions, in particular, a reconstruc­ tion, consisting only of bone and air (density 1 or 0), but still having the same projection data as the original object. Related non uniqueness results are familiar in tomography and are usually ignored because CT machines seem to produce useful images. It is likely that the "explana­ tion" of this apparent paradox is that point reconstruction in tomogra­ phy is impossible.
In the first part of this paper we give a survey of some recent already published results, conencted with CT Paradox. The second part is con­ cerned with an analogous in quantum mechanics for wich new results are obtained.
IThe methods of computer tomography CT (see for example the survey [14], and books [11], [12]) have been used in science, medicine and technology.
THE METHOD OF MOMENTS IN TOMOGRAPHY... 37
2. The problem of "stability" in computer tomography with par­ tially known data
The CT paradox shows that it is impossible (in general situation) to re­ construct the density basing on partially known tomography data. The purpose here is to show that, under moment-type conditions, measures having a "large" number of coinciding marginals are close to each other in metrics metrizing the weak topology (so-called here "weak metrics"). Our method is based on techniques used in classical moment problem. The key idea in showing that measures with large numbers of common marginals are close to each other in the weak metrics is best understood by compar­ ing the following three results. The first is the Theorem of Guttman et al. (1991) [3] mentioned above. The second (see [6, page 265]) states that if a finite number of moments /-LI. .•. ,/-Ln of a function f, 0 ::; f ::; 1 are given, then there exists a function 9 which takes the values 0 or 1 only, and pos­ sesses the moments /-Ll, .•. ,/-Ln. It is clear that these two results are similar; however, the condition of equality of the marginals is more complex than the condition of coincidences of the moments. Finally the third result (see [5, p.170-197]) gives estimates of the closeness in the A - metric on RI (a weak metric) for measures having common moments ILl, ••. , /-Ln (n < (0). These estimates are expressed in terms of the truncated Carleman's series 13m = L~=1 IL~]I(2j), (2m < n), and show that the closeness is of order
f3~1/4. Of course, since the condition of common marginals seems to be more restrictive than the condition of equal moments, one should be able to construct a similar estimate expressed in terms of the common marginals only.
First we shall derive estimates for closeness of measures in R2 having coinciding marginals in n directions. We consider the case when one of the measures has a compact support: in this case, the A-closeness of measures has order lIn. Further, the compactness assumption will be relaxed by Carleman's assumption in the problem of moments. Here the A-closeness of measures is of order f3:N4. We also display estimates of the closeness of measures with f-coinciding marginals, (0 < f < 1). These estimates differ from the corresponding ones with equal marginals by an additional term of order 1Jlog(1/f). We conclude by applying our results to the problems of computer tomography. In particular we offer a solution of the computer tomography paradox mentioned above and compare this solution with anal­ ogous results in [7J. To highlight the basic ideas, let us consider only the 2-dimensional case in full.
Let (h, ... , On be n unit vectors on the plane and PI, P2 be two prob­ abilities on R2, having the same marginals in the directions 01 , ••• ,On. To estimate the distance between PI and P2, different weak metrics can be
38 KLEBANOV L.B. AND RACHEV S.T.
used; however it seems, that the .x-metric is the most convenient for our purposes. This metric is defined as
(see, e.g., [17]), where (.,.) is the inner product and II." is the Euclidean norm. Clearly, .x metricize the weak convergence.
Our first result concerns the important case where one of the probability measures considered has compact support (see [8]). Theorem 2.1 Let (h, ... , On be n ~ 2 unit vectors in R2, no two of which are collinear. Let the support of the probability PI be a subset of the unit disc, and let the probability P2 have the same marginals as PI in the direc­ tions OJ, ... , On. Set
S=2[n;1], (2.2)
where [r] denotes the integer part of the number r, then
(2.3)
Remark 2.1 Let 01 , ••• , On be n ~ 2 directions in R2 no two of which are collinear. Suppose that the marginals of the probabilities PI and P2 along the directions 01 , •.• ,On have the same moments up to the even order k :::; n - 1. Then the marginals of PI and P2 with respect to any direction have the same moments up to order k.
Remark 2.2 In view of Remark 2.1 the proof of Theorem 2.1 even implies that Theorem 2.1 still holds if we replace the assumption that PI and P2 have coinciding marginals along the directions OJ (j = 1, ... , n) with the assumption that these marginals have the same moments up to order n - 1.
We next relax the condition of compactness for the support of PI, assum­ ing only the existence of all moments together with Carleman's conditions for the moment problem.
Set J.Lk=SUP f (x,O)kPI (dx),k=O,l, ... ,
9E81 iR2 (6-2)/2 1
f36 = L J.L;j2), j=1
where the number s is determined from equation 2.2 and SI is the unit circle.
THE METHOD OF MOMENTS IN TOMOGRAPHY.. . 39
Theorem 2.2 [8} Let (h, ... ,8n be n 2: 2 directions in R2 no two of which are collinear. Suppose that the measure PI has moments of any order. Sup­ pose also that the marginals of the measures PI and P2 in the directions 81 , .•. , 8n have the same moments up to order n - 1. Then there exists an absolute constant C such that
Let us give another type of results (see [7]). Let FD be the set of all probabilistic densities with compact support in
the square D = {x = (X},X2) :1 Xl I~ 1,1 X2 I~ I}.
Let i.p(1 be the Gaussian density:
and let f * i.p(1( x) (J E F D) be the corresponding Gaussian convolution.
Theorem 2.3 Let n 2: 2 be a natural number. There exist N = 2n di­ rections 81 , ..• ,8N on the Euclidean R2 plane such that if the densities f, g E FD have the same marginals on the directions 81, ... , 8N then
sup 1 f * i.p(1(x) - g * i.p(1(x) I~ 1/(y1ran+22(3n-4)/2r«n+ 1)/2)), (2.4) xER2
where r( z) is the Gamma-function. The vectors 81 , ... , 8N can be chosen on the following definition:
8j = (vj,-I)/(v] + 1)1/2,j = 1, ... ,n;
8j = (l,vj-n)/(V]_n + 1)1/2,j = n+ 1, ... ,2n,
where Vk = cos(1I'(2k-l)/(2n)),k = 1, ... ,n.
Let us now consider a more realistic situation where the marginals of PI and P2 in the directions 81 , ... ,8n are not the same but are close in the metric A. We use the same notation as that introduced in Theorem 1.
Theorem 2.4 Let 81 , ... ,8n be n 2: 2 directions in R2, no two of which are collinear. Suppose that the supports of the measures PI and P2 lie in the unit disc, where they have f - coinciding marginals with respect to the directions 8j (j = 1, ... , n) i.e.,
(2.5)
40 KLEBANOV L.B. AND RACHEV S.T.
= min max ( max I 1t'1( Tj 8j) - 1t'2( Tj 8j) I, l/T) $ E, T>O I"I~T
j = 1, ... , n. Then there exists a constant C depending on the directions 8j (j = 1, ... , n) such that for sufficiently small E > 0, we have
1 A(Pb P2) $ G(l/log( -) + l/s),
E (2.6)
where, as before, s is determined from equation {2}.
Remark 2.3 The statement in Theorem 2.4 still holds if instead of the E-coincidence of the marginals we require the E-coincidence of the mo­ ments up to order s of these marginals. The latter condition is defined by the inequalities
Let us demonstrate another application of the method of mo­ ments to problems with incomplete data.
Consider a body emitting x-rays. This body can move along horizontal axis in intervals [bo, (0) and (-00, -bo]. These x-rays go through a body situated in the strip {(x, y) : Ixl $ bo,lyl < oo}. H the body has density p(x,y) then we know all integrals of the form J p(x,a(x - b»dx where Ibl ~ bo and a is an arbitrary number. Let us denote
t/J(a, b) = j p(x,a(x - b»dx.
Actually, t/J( a, b) is well defined for all real a and b. Only the case under condition Ibl ~ bo is of interest. More precisely, one is studying the situation where t/J( a, b) is observed for all real values a and b such that Ibl ~ boo So, for any positive integer n we have
100 bnt/J(a,b)db = [00 bndbjp(x,a(x _ b»dx. bo Jbo
After substituting y = a( x - b) we get
100 j 100 y 1 bnt/J(a,b)db = dx (x - - )np(x, y)-dy bo a(z-bo) a a
~ (n) (_l)n j d 100 k n-k ( ) =L.J k n-k+t X xy px,ydy. k=O a a(z-bo}
THE METHOD OF MOMENTS IN TOMOGRAPHY.. . 41
We can represent the summands in the right hand side of previous equality in the form
But
Therefore
= J dx 1000 xkyn-kp(x, y)dy + O(an- k)
and
[00 bn1/1(a, b)db = t (:) (:':~l J dx [00 xkyn-kp(x, y)dy + O(a-1). Ao k=O a h
Hence, knowing the function 1/1 implies that we also know all integrals
J dx 100 xkyn-kp(x, y)dy
for all n = 1, ... and k = 0, ... , n - 1. In other words, we know all moments of the density (as the coefficients of expansion in powers of ~)
Pl(X, y) = y2p(x, y).
In case when the density p(x, y) has compact support (or in case when the measure is defined by its moments in unique way) we can reconstruct the density Pl(X, y) in unique way by its moments. It means that we can reconstruct the density p( x, y) in unique way for y i: O.
3. Quantum mechanics and computer tomography
Consider a physical system with one continuous degree of freedom, with a generalized coordinate operator x and a conjugate operator p. Suppose that "the rotated quadrature" operators xe, pe are defined by
Xe=xcos8+psin8, pe=-xsin8+pcos8 (3.1)
for all angles 8 (xo, Po are related with X, p (9 = 0) by unitary trans­ formations). Let Po( xo) be the probability distribution of the observable
42 KLEBANOV L.B. AND RACHEV S.T.
Xo. It was shown in [16] that if an uncountable set of such distributions {Po(xo) is known, then the corresponding Wigner distribution W(x,p) for pure state
W(x,p) = ~ J 1fJ(x + x')1fJ*(x - x')ei2PX'dx' (3.2)
can be determined uniquely by the Radon transform (see, for example [11]). W(x,p) is in one-to-one relation to the wave function:
1fJ( x + x')1fJ* (x - x') = J W (x, p )ei2pX' dp. (3.3)
An experimental method in [15] for measuring of corresponding distribu­ tions Po(xo; W) is given by
Po(xo; W) = J W(xo cos 8 - Po sin 8, Xo sin 8 + Po cos 8)dpo; (3.4)
Po(xo; W) represent the set of marginals of W(x,p)2. So to "measure" of the wave functions one needs to solve the inverse problems: (A) Po(xo; W) - W(x,p) and (B) W(x,p) - 1fJ(x) . To solve (A) via Radon transform the filtered back-projection numerical algorithm (FBPNA) of the usual tomography [11] was used in the orig­ inal works [15]. Both inverse problems (A) and (B) are ill-posed (in the Hadamard sense, see for example [2]) mathematical problems, what means that very small errors in Po( Xo; W) may cause very large (or even infinite) errors in W(x,p), and correspondingly, in wave functions. Moreover, in fact we can only measure a finite number (N) of marginals Po(xo; W), 8 E {81 , ... ,8N}; however for a finite number of marginals the Radon Theo­ rem on the uniqueness of solution of the inverse problem (A) fails. As it was recently derived [3], this non-uniqueness produces the so-called tomog­ raphy paradox. In our terminology this means that two different Wigner distributions W1(x,p) and W2(x,p), which have the same N marginals, may differ essentially. Another disadvantage is that Wigner distributions W(x,p) cannot have bounded support. In fact,
11fJ(x)1 2 = J W(x,p)dp, 1..j;(p)12 = J W(x,p)dx (3.5)
where ..j;(p) is the wave function in the momentum (impulse) representation
-¢;(p) = J 1fJ( x )e -ipx dx. (3.6)
2In general, the Wigner distributions W(x,p) may not be measurable in physical experiment since W(x,p) is not non-negative in general. It was proved in [4] that W(x,p) are non-negative iff the wave functions are Gaussian.
THE METHOD OF MOMENTS IN TOMOGRAPHY.. . 43
If we assume that W(x,p) has bounded support, then (3.5) implies that tP( x) and ;j,(p) both must have bounded support, but this is impossible due to the Paley - Wiener Theorem. Therefore in reality only a finite number of truncated marginals F9(X9j W) are available, say
(3.7)
w here the finite X 01, X 92 are determined by the possibilities of the exper­ imental device. It follows from the finiteness of the number of truncated marginals F9( X9j W), (J E {(JI, .. . ,(IN}, and necessary discretization of the integral Radon transform [11] that the solution of problem (A) produced by FBPN A [15] has the so-called artifacts. Artifacts are non-existing details in the exact W(x,p) and tP(x) which originate from the non-accuracy of mathematical part of the original scans. Such artifacts are easy to detect even in the simplest case of the Gaussian W(x,p), see [15].
Using the methods in [7] and [8] we obtain the following result. Its proof goes beyond the scope of this paper and will be published elsewhere.
Theorem 3.1 Let tPI, tP2 be two wave functions and W}, W 2 be the corre­ sponding Wigner distributions. Suppose that tPl (I = 1,2) has the modulus of continuity w( u) and
ItPI(x)1 S; c}, 1 ItPI(x)ldx s; c2,
gl(a,(jO') = 271'10'21 1 tPI(a-(-6 -6)-l/Ji(a-6)
exp( -(~i + ~i)/(20'2))d6d6, (l = 1,2), (3.8)
where C 1, C 2 are positive constants. For any f > 0 and any integer N ~ 2 there exist N + 1 directions (Jo, ... ,(IN such that if the marginals of WI and W2 in these directions are f-identical, i.e.
(3.9)
suplg1(a,(jO')-g2(a,(jO')1 S; (3.10) a
44 KLEBANOV L.B. AND RACHEV S.T.
where C is absolute constant. Moreover, the directions 80 , •.. , 8N may be chosen so that the points
1 1 ao = - tan 80 ' •.• ,aN = - tan 8N
are dividing the interval [-A, A] into N equal parts.
Remark 3.1 The condition (3.9) may be regarded as a measure of the deviation of the truncated from the exact marginals.
Remark 3.2 In (3.8) the Gaussian kernel was used only as an example of a smoothing kernel, but it is possible, indeed, to use other kernels which would lead to similar estimates.
Remark 3.3 For mixed states the estimates for the density matrix can be derived by the same method. The corresponding results will be published separately.
4. A sufficient condition for the solution uniqueness of the moment­ problem for quasi-probability Wigner distribution
For a given a sequence of constants J.'j,k where j, k are multi-indices r
) = (il'''',)n), k =\(kb ... ,kn), (4.1)
and n > 0 is some fixed integer, Narcowich [9] (see also [10]) provided necessary and sufficient conditions for the existence of a quasi-probability Wigner density W( q, p), (q = (ql, ... , qn), p = (PI, ... , Pn)) such that
(4.2)
(4.3)
The conditions for uniqueness of the problem (4.2) are unknown,3 and the essence of this section is to -give a sufficient condition for uniqueness of the problem (4.2) under assumption of its solvability (that is under conditions derived in [9]). For simplicity we sha.ll consider only the case n = 1, but our results are valid for arbitrary n ~ 1. Note that the problem of moments for W will be equivalent to a system of "probabilistic" problems of moments.
3Narcowich [10] provided an example of non-uniqueness of problem {4.2}.
THE METHOD OF MOMENTS IN TOMOGRAPHY... 45
The Wigner function (more precisely, the quasi-probability Wigner den­ sity function) of a quantum system in a mixed state is given by the formula:
W(q,p) = ~ f [.l "pj(q + q')"pj(q - q')dF(j)] e2inpql dq'. (4.4)
Complex-valued functions "pi are supposed to be continuous and square module integrable over the whole real line. The symbol * is stated for complex conjugation. The functions "pi are indexed by j from an abstract set J, and F is a measure on J.
Let us formulate some results by F.J.Narcowich. Firstly give the defi­ nition of Ii-positive definiteness. Let r = Rn x Rn be phase space. Denote z = (q1, ... ,qniP1, ... ,Pn) E r and define
n
= z'Jz,
where
J - (On In) - -In On . Definition 4.1 A continuous function F given on r is called Ii-positive definite if for any integer m 2: 1, for any set {a1, ... , am} C r and any set of complex numbers {All ... , ).m} holds the inequality
m
L F(ai - ak)ei~(7(a",aj)Xi).k 2: o. j,k=l
Let us note that for the case Ii = 0 we have well known definition of ordinary positive definiteness.
Theorem 4.1 (Narcowich) A continuous function F given on r is a Wigner function (4.4) if and only if it is Ii-positive definite.
This Theorem is thus very similar to the well known Bochner Theorem. We suppose here that the Plank constant Ii equals to 1 in the chosen
units system. Now we give some notations (due to Narcowich) connected with the conditions of solvability of the problem of moments for Wigner distribution. Suppose that A and B are two (sufficiently smooth and quickly decreasing at infinity) functions given on the phase space r. Define the operation 0 by
46 KLEBANOV L.B. AND RACHEV S.T.
Set also
< A, B >= J A(z)B(z)dz.
Definition 4.2 We shall say that a function F given on r is Ii-positive if
< F,AoA >~ 0
for all A E S(Rn), where S(Rn) is Schwartz space on Rn. For any given polynomial P(z) = L,j,k Cj,kpiqk we define a "moment func­ tional" J.t:
J.t( P) = ~ Cj,kJ.tj,k j,k
Pj,k = J J piqkW(q, p)dnq dnp,
and then
J.t(P) = J P(z)W(z)dz =< P, W > .
Theorem 4.2 (Narcowich) For the existence of a Wigner function W hav­ ing prescribed moments J.tj,k it is necessary and sufficient that J.t( P) ~ 0 for any Ii-positive polynomial P. Of course, this Theorem is similar to the well known necessary and sufficient condition for the solvability of classical moment problem.
Suppose the Wigner function has all moments
J.tm,n = J J qmpnW(q,p)dqdp, m ~ 0, n ~ O. (4.5)
Conversely, let {J.tm,n, m ~ 0, n ~ O} be a double sequence of numbers sat­ isfying the Narcowich conditions. Then there exists at least one Wigner function W for which (4.5) holds. We are interested under what additional conditions posed on the sequence J.tm,n this Wigner function is unique. A partial answer is given by the following theorem.
Theorem 4.3 Given {J.tm,n, m ~ 0, n ~ O} let
and suppose that the series
ji,k = m~ lJ.tk-j,jl J
(4.6)
(4.7)
diverges. Then there exists no more than one Junction W oj the Jorm (4.4) with continuous and square module integrable functions"pj satisfying (~.5).
THE METHOD OF MOMENTS IN TOMOGRAPHY... 47
Proof. Suppose that there exists a function W satisfying (4.5). For arbi­ trary real a and integer k ~ 0 let us consider the quantity
j j W(q,aq + b)bkdqdb.
Denote
On the other hand in view of (4.4),
J j W(q, aq + b)bkdqdb
( 4.8)
(4.9)
= ~ j j j bk [.l 'lj;j(q + ql)'Ij;j(q - ql)dF(j)] e2i(aq+b)q'dqdbdq'
= 2~ J J J bk [.l 'lj;j(U)'Ij;j(V)dFU)] ei(a;2+bu)e-W~2+bv)dudvdb
where
(f;j( b; a) = j 'lj;j( U )ei('1~2 +bu) duo
From (4.8) and (4.10) we get
Pk(a) = 2~ J bk [J; l{f;j(b;a)12dF(j)] db.
( 4.10)
(4.11 )
48 KLEBANOV L.B. AND RACHEV S.T.
Therefore for arbitrary fixed a the sequence {Jlk(a), k ~ O} is a sequence of moments corresponding to a non-negative density
(4.12)
The Carleman's criterion (see, for example,