Z-Estimator
Recall that an M-Estimator seeks the minimizer of a function . Suppose the function is differentiable and convex, its minimizer equals the zero of its derivative
where in this case, is the derivative of in the definition of M-estimator. For example, MLE is an estimator w.r.t the Score Function .
However, can be more general without necessarily corresponding to an optimization problem. For example, recall that for a Moment Estimator, we solve a system
Thus, Moment Estimators are also a special case of Z-estimators.
As we can see, Z-estimators are a class of more general estimators that solve the zero point of a system of estimating equations .
- Table. Comparison of optimization and system solution.
Optimization | System Solution |
---|---|
M-estimators | Z-estimators |
Optimizing an objective function | Solving an equation system |
Utilize optimization landscape, e.g., gradient | Utilize system dynamics, e.g., contraction |
As we discussed earlier, for convex/concave and differentiable objective functions, optimization is equivalent to solving a system regarding the gradient. Conversely, we can also define an objective function for solving a system of equations. For example, for a linear system , we can define the squared cost , whose minimizer is the solution of the system.
However, different problem formulations offer different insights and solution methods.
- Optimization is more suitable if you have a clear and well-motivated objective function;
- System solution is more suitable when you know how the solution determines the system dynamics.
- When optimizing a function, we usually care more about how the local landscape, e.g., gradient, carries the decision variable to the optimum;
- When solving a system, we usually want to follow some system dynamics, e.g., a contractive operator, to reach the solution.
Properties
Asymptotic Normality
Let solve the system . Suppose the consistency holds: . Then, under some regularity conditions:
- is twice differentiable for all and for some integrable function ;1
- or, there exists an function such that for any in a neighborhood of , we have ;
- exists and is non-singular in a neighborhood of ;
- exists;
we have
where .
Relation to Asymptotic Normality of M-Estimators
If , where is the objective function for an M-Estimator, we can see that the asymptotic normality of Z-estimators implies the asymptotic normality of M-estimators.
However, the regularity conditions for Z-estimators are stronger than those for M-estimators (see Asymptotic Normality). For example, we can show that quantile regression satisfies the asymptotic normality conditions for M-estimators; but it does not satisfy the conditions for Z-estimators.
Proof Sketch
Denote ; recall that . By Taylor expansion,
Since we assume the consistency, we get
Thus, by Slutsky’s theorem,
Asymptotic Normality of Least Squares
We cast Ordinary Least Squares as a Z-estimator and verify the asymptotic normality conditions. The cost function is , which gives the z-function .
For the first condition, we verify its alternative Lipschitz condition:
Suppose has finite moments, then the first condition is met.
For the second condition, we have
For the third condition, we have
Therefore, by the asymptotic normality of Z-estimators, we have
Footnotes
-
The measure is always the probabilistic measure. ↩