Flip Probabilities for Random Projections of Θ-separated Vectors

نویسنده

  • ROBERT J. DURRANT
چکیده

We give the probability that two vectors in d-dimensional Euclidean space m,n ∈ R which are separated in R by an angle θ ∈ [0, π/2] have angular separation θR > π/2 following random projection into a k-dimensional subspace of R, k < d. This probability, which we call the ‘flip probability’, has several interesting properties: It is polynomial of order k in θ; it is independent of the original dimensionality d depending only on the projection dimension k and the original separation θ of the vectors; it recovers the existing result for flip probability under random projection when k = 1 as a special case, and has a geometric interpretation as the quotient of the surface area of a hyperspherical cap by the area of the corresponding hypersphere which is a natural generalisation of the k = 1 case. We also prove the useful fact that, for all k ∈ N, the flip probability when projecting to dimension k is greater than the flip probability when projecting to dimension k + 1. 1. Statement of Theorem and Proof Theorem 1 (Flip Probability). Let n, m ∈ R with angular separation θ ∈ [0, π/2]. Let R ∈ Mk×d, k < d, be a random projection matrix with entries rij iid ∼ N (0, 1/d) and let Rn, Rm ∈ R be the projections of n, m into R with angular separation θR. (1) The ‘flip probability’ PrR[θR > π/2|θ] = PrR[(Rn)Rm < 0|nm > 0] is given by: PrR[(Rn) Rm < 0|nm > 0] = Γ(k) (Γ(k/2))2 ∫ ψ 0 z(k−2)/2 (1 + z)k dz (1.1) where ψ = (1− cos(θ))/(1 + cos(θ)). (2) The expression above can be shown to be of the form of the quotient of the surface area of a hyperspherical cap subtending an 1991 Mathematics Subject Classification. Primary 60D05; Secondary 15B52.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

ثبت نام

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Berry-esseen Bounds for Projections of Coordinate Symmetric Random Vectors

For a coordinate symmetric random vector (Y1, . . . ,Yn) = Y ∈ R, that is, one satisfying (Y1, . . . ,Yn) =d (e1Y1, . . . , enYn) for all (e1, . . . , en) ∈ {−1,1}, for which P(Yi = 0) = 0 for all i = 1,2, . . . ,n, the following Berry Esseen bound to the cumulative standard normal Φ for the standardized projection Wθ = Yθ/vθ of Y holds: sup x∈R |P(Wθ ≤ x)−Φ(x)| ≤ 2 n ∑ i=1 |θi |E|X i | + 8.4E(...

متن کامل

Near Optimal Sketching of Low-Rank Tensor Regression

We study the least squares regression problem min Θ∈S D,R ‖AΘ− b‖2, where S D,R is the set of Θ for which Θ = ∑R r=1θ (r) 1 ◦ · · · ◦ θ (r) D for vectors θ (r) d ∈ Rpd for all r ∈ [R] and d ∈ [D], and ◦ denotes the outer product of vectors. That is, Θ is a low-dimensional, lowrank tensor. This is motivated by the fact that the number of parameters in Θ is only R ·Dd=1pd , which is significantly...

متن کامل

Stochastic Comparisons of Random Vectors with Conditionally Independent Components

In this paper we consider the following model. We consider n nonnegative random variables T1, T2, . . . , Tn such that they are conditionally independent given some random vector Θ, where Θ takes values on some set χ ⊆ R. If Fi(·|θ) denotes the marginal distribution of Ti(θ) ≡ (Ti|Θ = θ), and G(θ) denotes the joint distribution function of Θ, then the joint distribution of (T1, T2, . . . , Tn) ...

متن کامل

Supplement to “Adaptive Confidence Bands for Nonparametric Regression Functions”

This supplement contains the proofs of Theorem 2, Propositions 2 and 3, Lemma 1 and Eq.(14). 7 Proof of Theorem 2 Rather than prove Theorem 2 directly it is convenient to first prove an analogue of the Theorem in the context of multivariate Normal random vectors. This is done in section 7.1. The proof of Theorem 2 is then given in section 7.2 7.1 Confidence Bound For Multivariate Normal Vectors...

متن کامل

A note on the Lovász theta number of random graphs

We show a strong concentration bound for the Lovász θ function on G(n, p) random graphs. For p = 1/2, for instance, our result implies that the θ function is concentrated in an interval of length polylog(n) w.h.p. The best known bound previously was roughly n. The general idea is to prove that all the vectors in an optimal solution have “roughly equal lengths” w.h.p.

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

عنوان ژورنال:

دوره   شماره 

صفحات  -

تاریخ انتشار 2010