Tamkang Journal of Science and Engineering, Vol. 14, No. 3, pp. 217-224 (2011) 217 Signal Sparse Representation Based on the Peak Transform and Modulus Maximum of Wavelet Coefficients Yi-Gang Cen1,2, Rui Xin1,2, Xiao-Fang Chen3*, Li-Hui Cen3, Ming-Xing Hu4 and Shi-Ming Chen5 1 Institute of Information Science, Beijing Jiaotong University, Beijing, 100044, P.R. China 2 Beijing Key Laboratory of Advanced Information Science and Network Technology, Beijing 100044, P.R. China 3 School of Information Science and Engineering, Central South University, Hunan, Changsha 410083, P.R. China 4 Centre for Medical Image Computing, University College London, UK 5 School of Electrical and Electronic Engineering, East China Jiao Tong University, Jiangxi, Nanchang 330013, P.R. China Abstract It is now well-known that one can reconstruct sparse or compressible signals accurately from a very limited number of measurements. This technique is known as “compressed sensing” or “compressive sampling” (CS). A basic requirement of CS is that a signal should be sparse or it can be sparsely represented in some orthogonal bases. Based on the Peak Transform (PT) and modulus maximum of wavelet coefficients, a new algorithm was proposed for the signals that are non-sparse themselves and can not be sparsely represented by wavelet transform such as the Linear Frequency Modulated signal. According to this algorithm, K-sparse wavelet coefficients can be obtained. For the peak sequence produced by the Peak Transform, value expansion approach of reversible watermarking is exploited such that the peak sequence can be embedded into the measurements of the signal, which avoids increasing additional points for transmission. By using the Peak Transform and modulus maximum, non-sparse wavelet coefficients can be transformed into K-sparse coefficients, which improves the reconstruction result of CS. Simulation results showed that our proposed algorithm achieved better performance comparing with the original CS algorithm. Key Words: Compressed Sensing, Sparse Representation, Peak Transform, Modulus Maximum, Wavelet Transform 1. Introduction Conventional approaches to sampling signals or images follow Shannon’s celebrated theorem: the sampling rate must be at least twice the maximum frequency present in the signal (the so-called Nyquist rate). In the field of data conversion, for example, standard analog-to*Corresponding author. E-mail: [email protected] digital converter (ADC) technology implements the usual quantized Shannon representation: the signal is uniformly sampled at or above the Nyquist rate. In the recent years, a new theory known as Compressed Sensing or Compressive Sampling [1,2] has been proposed, which is a novel sensing/sampling paradigm that goes against the common wisdom in data acquisition. CS theory asserts that one can recover certain signals and images from far fewer samples or measurements than traditional 218 Yi-Gang Cen et al. methods use, which significantly reduces the cost for sampling, processing, transmission and storage of signals. By [3], if a signal f is sparse or compressible in the sense that they have concise representations when expressed in a proper basis Y, then by using few measurements, the signal can be recovered with high accuracy. In [4], seven types of reconstruction problem and the corresponding solutions were given. It can be seen that a most important premise of CS is that the signal can be sparsely represented. In this paper, a novel approach for the Ksparse representation is proposed. Then the interior point method is employed for solving the Min-l1 with equality constraints problem as that presented in [4] for the reconstruction of signals. The method of Min-l1 with equality constraints is very powerful for the reconstruction of sparse signals. But in general, we almost can not obtain strict K-sparse coefficients by the wavelet transform for the real signals, which will lead to large errors for the recovered signals. Peak Transform (PT) [5] is a nonlinear geometric transform that can transfer the high frequency components of a signal into the low frequency components. This property has been successfully applied in the image compression [5]. Combining PT and wavelet transform, the sparsity of wavelet coefficients can be greatly increased. Moreover, modulus maximum can be utilized to obtain sparser wavelet coefficients. Thus, an algorithm based on the PT and modulus maximum of wavelet coefficients is proposed in this paper such that K-sparse wavelet coefficients can be obtained. Especially for the signals that are non-sparse and can not be sparsely represented by the wavelet bases, our proposed algorithm achieves much better performance for the reconstruction signals. 2. Compressed Sensing and Peak Transform 2.1 Compressed Sensing Suppose that x Î RN´1 is a one dimension signal and it can be expanded by an orthogonal basis (such as a wavelet basis) Y = {y1, ¼, yN}, i.e., (1) where, yi = < x, yi >, Y Î CN´N and YYH = YHY = IN (the N ´ N identity matrix). If there are only K (K << N) non- zero coefficients in y, then Y is called as the sparse basis of x. In general, most signals are not sparse themselves. But their wavelet coefficients can be considered to be sparse or compressible. For example, one can consider only keeping the largest K coefficients and discarding the N - K small coefficients without much perceptual loss. Then the coefficient vector is sparse in a strict sense since all but a few of its entries are zero. We will call K-sparse for such objects with at most K nonzero entries. Multiplying the signal x by an M ´ N sensing matrix F = {f1, ¼, fM} to obtain M measurements s = Fx (2) Then the signal can be recovered by the M measurements and F. By substituting (1) into (2), we have s = FYy = Qy (3) Here Q = FY is an M ´ N matrix. According to the above process, the N dimension signal x is reduced into an M dimension signal s. Since the unknown variable number is greater than the number of the equations, (2) has not a determinate solution. But the y in (3) is K-sparse, i.e., there are only K nonzero components in y and K < M £ N. Thus, y can be reconstructed from s by solving the convex optimization problem. Finally, x can be recovered by (1). Reference [6] pointed out that Q should obey the Restricted Isometry Property (RIP), i.e., (4) where d > 0. An equivalent condition of RIP is that F and Y satisfy the incoherence requirement [6,7]. In order to reconstruct the signal x, we need to solve the following convex optimization problem (5) This is in general difficult to implement because it is an NP-hard problem in combinatorial mathematics. But this problem can be converted into solving [8] (6) Signal Sparse Representation Based on the Peak Transform and Modulus Maximum of Wavelet Coefficients There were some iterative greedy algorithms for solving (6) such as the Primal-Dual Interior-Point Algorithm [4], Matching Pursuit (MP) algorithm [9], Orthogonal Matching Pursuit (OMP) algorithm [7], Gradient Projection for Sparse Reconstruction (GPSR) algorithm [10] etc. 219 3. Sparse Representation based on the PT and Modulus Maximum 2.2 Peak Transform Peak transform (PT) [5] is able to convert highfrequency signals into low-frequency ones. Coupled with wavelet transform and subband decomposition, PT is able to significantly reduce signal energy in highfrequency subbands and achieve a significant transform coding gain. This has important applications in efficient data representation and compression. In the following, we will illustrate the PT by a simple example. As shown in the Figure 1(a), suppose that the original signal is f (x). We select seven local minimum and maximum points, i.e., x0, ¼, x6, which is called as peaks. Then f (x) can be partitioned into six curve segments by the peaks, denoted as f1(x), ¼, f6(x). We first cascade all odd-numbered curve segments then all even-numbered curve segments and form a new curve, denoted as PT{f (x) | xi}, as shown in the Figure 1(b). The above transform process is called as the forward PT. It can be seen that the PT only changes the order of curve segments and is reversible. The backward transform can be done by simply recascading the curve segments according to their original order. The backward PT operation is denoted by PT-1{f (x) | xi}. 3.1 Peak Transform for Signals According to the theory of CS, a basic requirement of CS is that the signal should be sparse itself or it can be sparsely represented in an orthogonal basis such as the wavelet bases. The wavelet decomposition layer should be as large as possible. This is because that one can obtain a low-pass approximation sub-band and a high-pass detail sub-band (both the low-pass and high-pass coefficient numbers are respectively only half of the original signal’s point number) for one dimension signal by one layer wavelet transform. The high-pass coefficients can be considered as sparse or compressible. But the lowpass sub-band is an approximation of the original signal. If the original signal is not sparse, then the low-pass coefficients must not be sparse. Thus, we need to decompose the signal as possible as we can so that there are only several points in the maximal low-pass layer J. Then they can be omitted in the future process and all the wavelet coefficients can be considered sparse or compressible, which just meet the requirements of CS. But for some signals such as the Linear Frequency Modulation (LFM) signal, its wavelet coefficients are not sparse. As shown in the Figure 2, a LFM signal is decomposed by the 4-layer wavelet transform. It can be seen that the coefficients in each layer are not sparse. Thus, if we simply apply the CS process on the wavelet coefficients of LFM signal, the reconstruction error will be increased. Figure 1. Peak transform. Figure 2. LFM signal and its coefficients of 4 layers wavelet decomposition. 220 Yi-Gang Cen et al. According to section 2.2, a signal will be much smoother after PT. The zero point (or the points with very small magnitudes) number will dramatically increase, which means that the sparsity of the coefficients is improved greatly. Figure 3 is a toy example. We can see that after the PT, the nonzero high-pass wavelet coefficients of the PT output signal are significantly reduced. Based on the above discussions, although the wavelet coefficients of some signals are not sparse, but we can perform PT on the non-sparse signals and pass the PT output into the wavelet filter banks. Then sparse wavelet coefficients can be obtained. Experiment results show that the energy of high-frequency sub-bands with PT is only about 40% of that without PT. 3.2 Embedment of Peak Sequence into Measurements In the forward PT, some points are selected as peaks so that curve segments can be determined, i.e., we need to know the truncated positions of the original signal. In the backward PT, the peaks also need to be used. Thus, the information about the peak positions need to be encoded with the measurements and transferred to the decoder. For a discrete signal f (x) Î RN, the peak sequence is defined as follows. the first N/q measurement points is left shifted for q bits, respectively. Then the N bits of B{f (x)}k are embedded into the last q bits of these N/q measurements. When the decoder receives the signal, it only needs to pick up the last q bits from the first N/q measurements to obtain the peak sequence and then right shift the first N/q measurements for q bits to obtain the original measurements. 3.3 Modulus Maximum and Hard-Threshold for Sparse Representation In the practice, signals are more complex than the triangle wave shown in the Figure 3. The PT output signal will not as smooth as that shown in Figure 3(c), which leads to the fact that there are a lot of nonzero wavelet coefficients with small magnitude. This deteriorates the sparsity of the wavelet coefficients. In the wavelet theory, de-noising algorithm based on the modulus maximum has been successfully exploited. Here, it will be used for the sparse representation of signals. We only keep the modulus maximum of the wavelet coefficients for sensing. Then, a hard-threshold t is used by (8) (7) Since all elements in B{f (x)}k are either 0 or 1, B{f (x)}k is a binary sequence with N bits. In order to reduce the information for transmission, we can encode B{f (x)}k by using some algorithms such as run-length coding, Huffman coding and arithmetic coding etc. But these encoding algorithms will greatly increase the computational cost for the encoder. For example, for a LFM signal with 256 points, its peak sequence also has 256 bits. If the arithmetic coding is applied, the encoding result still has 216 bits. But the run-time of arithmetic encoder and decoder is almost the same as the CS algorithm. Thus, we give up encoding the peak sequence by these coding algorithms. For the sake of not increasing the measurements number for transmission, value expansion approach of reversible watermarking is used directly such that B{f (x)}k can be embedded into the measurements of the signal, i.e., each of the binary values of where Max | WPT{f (x) | xi}| denotes the modulus maximum of the wavelet coefficients. By the hard-threshold, Figure 3. Wavelet coefficients of triangle wave and the coefficients of PT output. Signal Sparse Representation Based on the Peak Transform and Modulus Maximum of Wavelet Coefficients the sparseness of the modulus maximum sequence will be significantly improved. The selection of t is important, if t is too large, many modulus maximum points are set to zeros. Although the ~ sequence Max WPT { f ( x ) x i is sparse enough so that CS ~ algorithm can recover Max WPT { f ( x ) x i accurately. The error of the reconstructed signal will increase since a lot of information of the signal was lost by thresholding. If t is too small, the sparseness of the modulus maximum sequence will decrease. Then the CS algorithm may not re~ cover Max WPT { f ( x ) x i accurately. This also increases the error of the reconstructed signal. ~ In the decoder, when Max WPT { f ( x ) x i has been recovered by a convex optimization algorithm, it needs to re~ construct the wavelet coefficients from Max WPT { f ( x ) x i . Alternate Projection (AP) [11] gave a method to reconstruct the wavelet coefficients but with high computational cost. In [12], a Piecewise Cubic Spline Interpolation (PCSI) algorithm was proposed, which was simple and fast. More important, the quality of the reconstructed signal could be improved. Thus, we will use PCSI in the decoder for the reconstruction of wavelet coefficients. Finally, we will discuss the property of wavelet decomposition. It is well-known that the low-pass coefficients are the best approximation of the original signal in the wavelet subspace. But in general, low-pass coefficients are not sparse. Thus, the error of the reconstructed signal will increase if the CS algorithm is performed on both the high-pass and low-pass coefficients. According to the Mallat algorithm of wavelet transform, after onelayer wavelet decomposition, the low-pass coefficient number is only half of the original one. Thus, in the jth wavelet decomposition layer, the low-pass coefficient number is only N / 2j, where N is the length of the original signal. For example, if N = 1024, then the low-pass coefficients only have 16 points after 6 wavelet decompositions. In order to improve the quality of the reconstructed signal, we can transmit these 16 points to the decoder directly and only perform the CS algorithm on the high-pass coefficients. When the decoder received the low-pass coefficients and recovered the high-pass coefficients, the signal will be better reconstructed. 221 3.4 Compressive Sensing Algorithm Based on the Peak Transform and Modulus Maximum Based on the above discussion, the CS algorithm based on the PT and modulus maximum proposed in this paper is as follows. Encoder Step 1. Find the peaks of the original signal f (x) Î RN and obtain B{f (x)}k. Perform PT on f (x), the output signal is denoted as PT{f (x) | xi}. Step 2. Perform wavelet transform on PT{f (x) | xi}. The wavelet coefficients are denoted as aJ, dJ, dJ-1, ¼, d1, where aJ is the low-pass coefficients in the maximal decomposition layer, dj, 1 £ j £ J, is the high-pass coefficients of the jth layer. Define WPT{f (x)} = {dJ, dJ-1, ¼, d1}. Step 3. Find out the modulus maximum points of WPT{f (x)} and apply the threshold on them. Only the largest K modulus maximum points are kept and the resultant sequence is denoted as ~ Max WPT { f ( x ) x i . Step 4. Construct the sensing random matrix FM´N, M < N. The entries of F are drawn from the Gaussian ~ distribution (0,1/N). Then Max WPT { f ( x ) x i is sensed by F and M measurements can be obtained. Finally, B{f (x)}k is embedded into the first N/q measurements and transmitted to the decoder together with the low-pass coefficients aJ. Decoder ~ Step 1. Extract M measurements of Max WPT { f ( x ) x i , the low-pass coefficients aJ and the peak sequence B{f (x)}k from the received signal. Step 2. Recover the modulus maximum sequence ~ Max r WPT { f ( x ) x i } by the convex optimization algorithm. ~ Step 3. According to the Max r WPT { f ( x ) x i }, the highpass wavelet coefficients can be reconstructed by the PCSI algorithm. Then the PT output signal can be recovered by the inverse wavelet transform. The recovered PT output signal is de~ noted as WPT { f ( x )}. 222 Yi-Gang Cen et al. Step 4. Perform the backward Peak Transform on ~ WPT { f ( x )} by using B{f (x)}k to obtain the re~ covered signal f ( x ). 4. Simulation Results In order to evaluate the similarity of the original and the reconstructed signals, the error between two vectors is defined as follows. (9) ~ Here, f is the original signal and f is the recon~ structed signal. The smaller the m( f , f ) is, the more accurate the reconstructed signal is. We used the LFM signal with 512 points in the simulation. The initial normalized frequency and the final normalized frequency are 0.1 and 0.3, respectively. The magnitude was normalized to 1. The selection of wavelet basis is very important. According to the properties of wavelet functions, we should select wavelet basis with orthogonality, compactly support and near symmetry. By the experiments, we find that coif3~coif5, sym8, db4~ db10 bases achieve better reconstruction results. The LFM signal was decomposed into 5 layers. Figure 4 illustrates the relations between the measurement number and the reconstruction error for different threshold values t. Because of the randomness of the sensing matrix, the reconstruction result of each simulation is different. Thus in Figure 4, for each M, we conducted the experiment 100 independent trials and calculate the average reconstruction error. According to the figure, it can be seen that if the threshold value t is too small (for example, t = 0.1), the wavelet coefficients after thresholding are not sparse enough so that the reconstruction error is large. But if the threshold value is too large (for example, t = 0.3), too many points with useful information were set to zeroes. Thus, the error between the recovered signal and the original signal can not been decreased significantly. Moreover, in this case, when M > 320, the reconstruction error will not decrease even the measurement number is increased. This is because the threshold is too large, the information contained in the coefficients is limited. After experiments, we found that t = 0.2 is better in this simulation, as shown by the dotted line with diamond points in Figure 4. When M is chosen between 250 and 350 and t = 0.2, our proposed algorithm achieved better performance. Figure 5 was a simulation result. The algorithm of Min-l1 with equality constraints presented in [4] was used for comparison. The wavelet decomposition layer was J = 5. The wavelet function was chosen as the sym8. There were total 496 high-pass coefficients in the 5 decomposition layers and 16 low-pass coefficients in the 5th decomposition layer. Let t = 0.2, after thresholding, there were only 167 nonzero points in the modulus maximum sequence. M was chosen as 300. Thus, in our proposed algorithm, we need to transmit 300 measurements and 16 low-pass coefficients together to the decoder. For the decoder, Primal-Dual Interior-Point Algorithm described in [4] was used to recover the modulus maximum sequence. The reconstruction result was shown in Figure 5(a) and the reconstruction error computed by (7) is m = 6.2551. In order to obtain a fair comparison, for the simulation by using the CS algorithm presented in [4], M was set to be 316. The reconstruction result was shown in Figure 5(b) and the reconstruction error computed by (9) is m = 7.1342. In this simulation, 158 points were selected as peaks, i.e., the corresponding values of these 158 points of B{f (x)}k are 1. The size of the sensing matrix was 300 ´ 512, then 300 measurements were obtained. Thus, each of the binary values of the first 256 points was left shift for 2 bits (q = 2). Then the 512 bits of B{f (x)}k could be embedded into these 256 measurements sequentially for transmission. This embedment method is non-adaptive Figure 4. Reconstruction error comparison for different thresholds. Signal Sparse Representation Based on the Peak Transform and Modulus Maximum of Wavelet Coefficients 223 Figure 5. Reconstruction results of our proposed algorithm and the CS algorithm described in [4]. (a) Comparison of the recovered signal of our algorithm and the original signal, m = 6.2551. (b) Comparison of the recovered signal of the CS algorithm described in [4] and the original signal, m = 7.1342. since if the size of the sensing matrix is changed, the points used for embedment also need to be changed. But in the practice, the sensing matrix should be selected optimally in prior. Thus the measurements number M can be determined. Then the points used for embedding can also be determined according to the length of B{f (x)}k. Figure 6 shows the performance of our proposed algorithm comparing with the algorithm of Min-l1 with equality constraints presented in [4], which illustrated the relations between the reconstruction errors and the values of measurement number M. For each M we have conducted the experiment 100 independent trials. By Figure 6, the performance of our proposed algorithm is better than the CS algorithm presented in [4]. Especially when M is small, the reconstruction error computed by (9) of our algorithm outperforms the algorithm presented in [4] by up to 2. 5. Conclusion Although the theory and applications of CS is very attractive, there are a lot of problems need to be solved. Except the sparse representation problem, the choosing of the sensing matrix is also a hot issue. Since the random matrices generated in each experiment are different, the reconstruction performance will be greatly influenced. Figure 6. Average reconstruction error of our proposed algorithm and the algorithm of Min-l1 with equality constraints presented in [4]. On the other hand, the recover algorithm is another issue for the future development. In this paper, an algorithm based on the Peak Transform and modulus maximum was proposed for sparse representation of signals, which was especially effective for the signals that are not sparse themselves and have no sparse coefficients after wavelet transform. Combining with the original CS algorithm, the quality of reconstructed signal has been improved. 224 Yi-Gang Cen et al. Acknowledgment This work was supported by National Natural Science Foundation of China (No. 60802045, 60804066, 61073079); the Scientific Research Foundation for Returned Overseas Chinese Scholars ([2009] 1001), the Fundamental Research Funds for the Central Universities of China (2009JBM028, 201012200159, 20101220 0110); Beijing Municipal Natural Science Foundation (4113075) and New Teacher Foundation of State Education Ministry (20100162120019). References [1] Donoho, D., “Compressed Sensing,” IEEE Trans. Inform. Theory, Vol. 52, pp. 1289-1306 (2006). [2] Candes, E., “Compressive Sampling,” Proceedings of the International Congress of Mathmaticians, Madrid, Spain (2006). [3] Candes, E. and Tao, T., “Near Optimal Signal Recovery From Random Projections: Universal Encoding Strategies?” IEEE Trans. Inform. Theory, Vol. 52, pp. 5406-5425 (2006). [4] Candes, E. and Caltech J. R., l1-MAGIC: Recovery of Sparse Signals via Convex Programming, Oct., 2005; http://dsp.rice.edu/cs. [5] He, Z. H., “Peak Transform for Efficient Image Representation and Coding,” IEEE Transactions on Image Processing, Vol. 16, pp. 1741-1754 (2007). [6] Candes, E., Romberg, J. and Tao, T., “Robust Uncer- tainty Principles: Exact Signal Reconstruction from Highly Incomplete Frequency Information,” IEEE Trans. Inform. Theory, Vol. 52, pp. 489-509 (2006). [7] Tropp, J. A. and Gilbert, A. C., “Signal Recovery from Random Measurements via Orthogonal Matching Pursuit,” IEEE Trans. Inform. Theory, Vol. 53, pp. 46554666 (2007). [8] Donoho, D. and Tsaic, Y., “Extensions of Compressed Sensing,” Signal Processing, Vol. 86, pp. 533-548 (2006). [9] Tropp, J. A., “Greed is Good: Algorithmic Results for Sparse Approximation,” IEEE Trans. Inform. Theory, Vol. 50, pp. 2231-2242 (2004). [10] Figueiredo, M. A. T., Nowak, R. D. and Wright, S. J., “Gradient Projection for Sparse Reconstruction: Application to Compressed Sensing and Other Inverse Problem,” Journal of Selected Topics in Signal Processing: Special Issue on Convex Optimization Methods for Signal Processing, Vol. 1, pp. 586-598 (2007). [11] Mallat, S. and Zhong, S., “Characterization of Signals from Multiscale Edges,” IEEE Trans. on PAMI, Vol. 14, pp. 710-732 (1992). [12] Zhao, R. Z., Song, G. X. and Qu, H. Z., “A New Piecewise Cubic Spline Interpolating Algorithm for the Reconstruction of Wavelet Coefficients,” Signal Processing (Chinese), Vol. 17, pp. 242-277 (2001). Manuscript Received: Dec. 17, 2010 Accepted: Feb. 15, 2011
© Copyright 2026 Paperzz