Z-Estimator

Recall that an M-Estimator seeks the minimizer of a function . Suppose the function is differentiable and convex, its minimizer equals the zero of its derivative

where in this case, is the derivative of in the definition of M-estimator. For example, MLE is an estimator w.r.t the Score Function .

However, can be more general without necessarily corresponding to an optimization problem. For example, recall that for a Moment Estimator, we solve a system

Thus, Moment Estimators are also a special case of Z-estimators.

As we can see, Z-estimators are a class of more general estimators that solve the zero point of a system of estimating equations .

  • Table. Comparison of optimization and system solution.
OptimizationSystem Solution
M-estimatorsZ-estimators
Optimizing an objective functionSolving an equation system
Utilize optimization landscape, e.g., gradientUtilize system dynamics, e.g., contraction
optimizationsystem

As we discussed earlier, for convex/concave and differentiable objective functions, optimization is equivalent to solving a system regarding the gradient. Conversely, we can also define an objective function for solving a system of equations. For example, for a linear system , we can define the squared cost , whose minimizer is the solution of the system.

Equivalence of optimization and system solution.|300
Equivalence of optimization and system solution.

However, different problem formulations offer different insights and solution methods.

  • Optimization is more suitable if you have a clear and well-motivated objective function;
  • System solution is more suitable when you know how the solution determines the system dynamics.
  • When optimizing a function, we usually care more about how the local landscape, e.g., gradient, carries the decision variable to the optimum;
  • When solving a system, we usually want to follow some system dynamics, e.g., a contractive operator, to reach the solution.

Properties

Asymptotic Normality

Let solve the system . Suppose the consistency holds: . Then, under some regularity conditions:

  • is twice differentiable for all and for some integrable function ;1
    • or, there exists an function such that for any in a neighborhood of , we have ;
  • exists and is non-singular in a neighborhood of ;
  • exists;

we have

where .

Relation to Asymptotic Normality of M-Estimators

If , where is the objective function for an M-Estimator, we can see that the asymptotic normality of Z-estimators implies the asymptotic normality of M-estimators.

However, the regularity conditions for Z-estimators are stronger than those for M-estimators (see Asymptotic Normality). For example, we can show that quantile regression satisfies the asymptotic normality conditions for M-estimators; but it does not satisfy the conditions for Z-estimators.

Proof Sketch

Denote ; recall that . By Taylor expansion,

Since we assume the consistency, we get

By LLN, ; by CLT,

Thus, by Slutsky’s theorem,

Asymptotic Normality of Least Squares

We cast Ordinary Least Squares as a Z-estimator and verify the asymptotic normality conditions. The cost function is , which gives the z-function .

For the first condition, we verify its alternative Lipschitz condition:

Suppose has finite moments, then the first condition is met.

For the second condition, we have

For the third condition, we have

Therefore, by the asymptotic normality of Z-estimators, we have

Footnotes

  1. The measure is always the probabilistic measure.