High-Dimensional Probability

High-dimensional probability theory studies a high-dim random vector and its transformation (high-dim functions). Without any additional structure, there is nothing special about this random vector. In high-dim probability, we always assume has independent (or weakly dependent) coordinates, i.e., where are independent random variables. Suppose are iid, the very first result we have is regarding its variance:

Such a factorization into the dimension and one-dim property is called tensorization, a key technique in high-dimensional probability.

Applications:

  • Statistical Learning
  • Compressed sensing
  • random matrices
    • covariance matrix
    • random graphs
  • Sampling
  • Optimal transport
  • Gaussian approximation

Concentration of Measure

Consider the simplest form: . If is IID and has a finite mean, by the strong Central Limit Theorem, we know . We ask

Qn

  1. How about general functions?
  2. How fast is this convergence? (non-asymptotic)

Informal Principle

If is IID, then provided that is “smooth” enough, i.e., does not depend too heavily on any of its coordinates.

Suprema of Stochastic Processes

Qn

How large is ?

Ex

  • L2 error: .
  • Convex conjugate:

Informal Principle

If is “smooth”, then can be controlled by the complexity of .