## Access

You are not currently logged in.

Access your personal account or get JSTOR access through your library or other institution:

## If You Use a Screen Reader

This content is available through Read Online (Free) program, which relies on page scans. Since scans are not currently available to screen readers, please contact JSTOR User Support for access. We'll provide a PDF copy for your screen reader.

# Bounds on the Efficiency of Linear Predictions Using an Incorrect Covariance Function

Michael L. Stein
The Annals of Statistics
Vol. 18, No. 3 (Sep., 1990), pp. 1116-1138
Stable URL: http://www.jstor.org/stable/2242045
Page Count: 23
Since scans are not currently available to screen readers, please contact JSTOR User Support for access. We'll provide a PDF copy for your screen reader.
Preview not available

## Abstract

Suppose z(·) is a random process defined on a bounded set \$R \subset \mathbb{R}^1\$ with finite second moments. Consider the behavior of linear predictions based on z(t1), ..., z(tn), where t1, t2, ⋯ is a dense sequence of points in R. Stein showed that if the second-order structure used to generate the predictions is incorrect but compatible with the correct second-order structure, the obtained predictions are uniformly asymptotically optimal as n → ∞. In the present paper, a general method is described for obtaining rates of convergence when the covariance function is misspecified but compatible with the correct covariance function. When z(·) is Gaussian, these bounds are related to the entropy distance (the symmetrized Kullback divergence) between the measures for the random field under the actual and presumed covariance functions. Explicit bounds are given when R = [ 0, 1] and z(·) is stationary with spectral density of the form f(λ) = (a2 + λ2)-p, where p is a known positive integer and a is the parameter that is misspecified. More precise results are given in the case p = 1. An application of this result implies that equally spaced observations are asymptotically optimal in the sense used by Sacks and Ylvisaker in terms of maximizing the Kullback divergence between the actual and presumed models when z(·) is Gaussian.

• 1116
• 1117
• 1118
• 1119
• 1120
• 1121
• 1122
• 1123
• 1124
• 1125
• 1126
• 1127
• 1128
• 1129
• 1130
• 1131
• 1132
• 1133
• 1134
• 1135
• 1136
• 1137
• 1138