Regret (decision theory)

From Wikipedia, the free encyclopedia

Regret (often also called opportunity loss) is defined as the difference between one's actual payoff and the payoff in a better position that he could have got if a different course of action had been chosen. This is also called difference regret. One can define the ratio regret, namely the ratio between the actual payoff and the best one.

Contents

[edit] The minimax regret concept

The minimax regret approach is minimizing the worst-case regret. The rational behind this is performing uniformly close to the optimal course. Since the minimax criterion applied here to the regret (difference or ratio of the payoffs) rather that to the payoff itself it's not so pessimistic as the ordinary minimax approach. Similar approaches have been used in variety of areas such as:

[edit] Example: Linear estimation setting

Here we illustrate how the concept of regret can be used to design a linear estimator. Now the regret is the difference between mean-squared error (MSE) of the linear estimator that doesn't know the parameter x and the mean-squared error (MSE) of the linear estimator that knows x. Since the estimator is restricted to be linear the zero MSE cannot be achieved in the latter case also.

Consider the problem of estimating the unknown deterministic parameter vector x from the noisy measurements in the linear model

y = Hx + w

where H is a known n \times m matrix with full column rank m, and w is a zero mean random vector with covariance matrix Cw, which models the noise.

Let

\hat{x}=Gy

be a linear estimate of x from y, where G is some m \times n matrix. The MSE of this estimator is given by

MSE = E\left(||\hat{x}-x||^2\right) = Tr(GC_wG^*) + x^*(I-GH)^*(I-GH)x.

Since the MSE depends explicitly on x it cannot be minimized directly. Instead we can use the concept of regret in order to define a linear estimator with good MSE performance. To define the regret in this setup consider a linear estimator that knows the value of the parameter x, i.e. the matrix G can explicitly depend on x:

\hat{x}^o=G(x)y.

The MSE of \hat{x}^o is

MSE^o=E\left(||\hat{x}^o-x||^2\right) = Tr(G(x)C_wG(x)^*) + x^*(I-G(x)H)^*(I-G(x)H)x.

To find the optimal G(x) we differentiate it with respect to G and equate to 0 getting

G(x) = xx * H * (Cw + Hxx * H * ) − 1

and using the Matrix Inversion Lemma

G(x)=\frac{1}{1+x^*H^*C_w^{-1}Hx}xx^*H^*C_w^{-1}.

Substituting this G(x) back into MSEo

MSE^o=\frac{x^*x}{1+x^*H^*C_w^{-1}Hx}.

This is the smallest MSE achievable with a linear estimate that knows x. Of course, in practice this MSE cannot be achieved but it serves as a bound on the optimal MSE. Now the regret is defined by

R(x,G)=MSE-MSE^o=Tr(GC_wG^*) + x^*(I-GH)^*(I-GH)x-\frac{x^*x}{1+x^*H^*C_w^{-1}Hx}.

The minimax regret approach here is to minimize the worst-case regret as defined above. This will allow us to perform as close as possible to the best achievable performance in the worst case of the parameter x. Although this problem appears difficult, it can be formulated as a convex optimization problem and solved definitely. For details see [1]. Similar ideas can be used when x is random with uncertainness in the covariance matrix. For details see [2][3].

[edit] References

  1. ^ Y. C. Eldar, A. Ben-Tal, and A. Nemirovski, "Linear Minimax regret estimation of deterministic parameters with bounded data uncertainties," IEEE Trans. Signal Process., vol. 52, no. 8, pp. 2177–2188, Aug. 2004.
  2. ^ Y. C. Eldar and Neri Merhav, "A Competitive Minimax Approach to Robust Estimation of Random Parameters," IEEE Trans. Signal Processing, vol. 52, pp. 1931-1946, July 2004.
  3. ^ Y. C. Eldar and Neri Merhav, "Minimax MSE-Ratio Estimation with Signal Covariance Uncertainties," IEEE Trans. Signal Processing, vol. 53, no. 4, pp. 1335-1347, Apr. 2005.

[edit] See also

[edit] External links