where a is a positive constant.
Note that the zero norm,∣∣ ∙ ∣∣o, in Equation 2.3 measures the number of non-
zero elements of the vector. This objective function is not convex which means
solving the optimization problem in Equation 2.3 is difficult. Instead, Danaho
et al. showed [10,15,24] that one can use the following optimization problem to
approximate the sparse vector X.
min∣∣X∣∣ι + α∣∣e∣∣2
such that Y = UX + e.
(2.4)
They proved that, for a Gaussian measurement matrix, an «-sparse vector can
be retrieved via ^ɪ-norm optimization if
S<C K ■
<Ulog(X/K)’
where C is a constant. Moreover, for a general measurement matrix U, Restricted
Isometry Property (IRP) should be satisfied [15].
Most of the real world vectors have an approximately sparse representation.
A vector Xv ×ι is called approximately «-sparse if it has s large elements and
N — s very small elements. It is also shown that the optimization problem in
Equation 2.4 can be used to recover approximately sparse vectors that fie in weak
lp ball of radius r [15]. i.e.,
I I — 1
(2.5)
k∣(i) ≤ ri P, 1 ≤ i ≤ N
21
More intriguing information
1. Ein pragmatisierter Kalkul des naturlichen Schlieβens nebst Metatheorie2. Manufacturing Earnings and Cycles: New Evidence
3. MULTIPLE COMPARISONS WITH THE BEST: BAYESIAN PRECISION MEASURES OF EFFICIENCY RANKINGS
4. Analyse des verbraucherorientierten Qualitätsurteils mittels assoziativer Verfahren am Beispiel von Schweinefleisch und Kartoffeln
5. The name is absent
6. The name is absent
7. Segmentación en la era de la globalización: ¿Cómo encontrar un segmento nuevo de mercado?
8. The name is absent
9. The name is absent
10. Accurate, fast and stable denoising source separation algorithms