We propose dual regression as an alternative to the quantile regression process for the global estimation of conditional distribution functions under minimal assumptions. Dual regression provides all the interpretational power of the quantile regression process while avoiding the need for repairing the intersecting conditional quantile surfaces that quantile regression often produces in practice. Our approach introduces a mathematical programming characterization of conditional distribution functions which, in its simplest form, is the dual program of a simultaneous estimator for linear location-scale models. We apply our general characterization to the specification and estimation of a flexible class of conditional distribution functions, and present asymptotic theory for the corresponding empirical dual regression process.Now suppose that U ∼ U (0, 1) does not hold. Because X includes an intercept, any random variable U such that E{X ⊗ m J ( U )} = 0 for all J must also satisfy E{m J ( U )} = 0 for all J, and therefore U ∼ U (0, 1). It follows that E{X ⊗ m J (U )} = 0 in the large J limit.Therefore, E{X ⊗ m J (U )} = 0 for all J if and only if E(X | U ) = E(X) and U ∼ U (0, 1), and the result follows.Proof of part (ii). Let e be a random variable with mean 0 and variance 1 satisfying E( X | e) = 0. Then E{ X c ⊗ h J (e)} = E{E( X c | e) ⊗ h J (e)} = 0, for all J, by iterated expectations and mean independence.In order to show the converse statement, suppose that E( X c | e) = 0. Letting ϕ(e) = E( X c | e) and Ψ J such that E{||ϕ(e) − Ψ J h J (e)|| 2 } → 0, and following steps similar to the proof of Lemma 2.1 in Donald et al. (2003),as J → ∞, which implies E{ X c ⊗ h J (e)} = 0 as J → ∞.Therefore, a random variable e with mean 0 and variance 1 satisfies E{ X c ⊗ h J (e)} = 0 for all J large enough if and only if E( X c | e) = 0, and the result follows.