Adaptive Reconstruction Algorithm Based on Compressed Sensing Broadband Receiver

Existing greedy reconstruction algorithms require signal sparsity, and the remaining sparsity adaptive algorithms can be reconstructed but cannot achieve accurate sparsity estimation. To address this problem, a blind sparsity reconstruction algorithm is proposed in this paper, which is applied to compressed sensing radar receiver system. The proposed algorithm can realize the estimation of signal sparsity and channel position estimation, which mainly consists of two parts. The ﬁ rst part is to use fast search based on dichotomy search, which is based on the high probability reconstruction of greedy algorithm, and uses dichotomy search to cover the number of sparsity. The second part is the signal matching and tracking algorithm, which is mainly used to judge the signal position and reconstruct the signal. Combine the two parts together to realize the blind estimation of the sparsity and the accurate estimation of the number of signals when the number of signals is unknown. The experimental analyses are carried out to evaluate the performance of the reconstruction probability, the accuracy of sparsity estimation, the running time of the algorithm, and the signal-to-noise ratio.


Introduction
With the continuous development of radar technology, the electronic countermeasure environment has become increasingly complex, and it will become more complex in the future [1].First, the analog-to-digital converter (ADC) sampling rate limits the instantaneous bandwidth of the receiver [2].The use of frequency agile signals requires radar receivers to have increasingly wider instantaneous bandwidths.At present, the development of high-rate ADCs is relatively slow [3][4][5].Second, in order to achieve full-probability reception, radar receivers require increasingly hardware, the system is more and more complex, and the volume is getting larger.In order to meet the needs of modern electronic warfare, it is necessary to solve the problems faced by digital channelized receivers [6].Donoho, Candes et al., Wan et al., and Candes and Tao proved the theory of compressed sensing in 2006 and proposed specific implementation algorithms [7][8][9][10].Compressed sensing is a new sampling method, which is highly concerned in image processing, microwave imaging, wireless communication, radar system, and biomedical engineering [11][12][13][14][15][16][17][18].
In recent years, an analog-to-information converter [19] (AIC) has been proposed based on the compressed sampling theory.The methods to achieve AIC include random demodulation (RD) [20], multicoset sampling (MCS) [21], quadrature compressed sampling (QuadCS) [22], and modulated wideband converter (MWC) [23] structure.The under-Nyquist sampling of the sparse signal can be realized at a sampling rate much lower than the Nyquist theorem, and the compressed sampled data of the signal can be obtained.The modulation broadband converter is currently the most mature multibranch compression sampling structure at.Cohen et al. and Mishali et al. have completed the realization of the hardware structure [24,25].
Compressed sensing theory is mainly divided into three parts: signal sparse representation, observation matrix, and reconstruction algorithm.Signal sparse representation means that the signal can be sparsely represented in a certain transform domain.The observation matrix reduces the dimensionality of the signal through domain conversion.Finally, it restores the original signal through a reconstruction algorithm [26].The reconstruction algorithm is the most important step in compressed sensing.The algorithm can be divided into the following three categories: combination algorithm, convex optimization algorithm, and greedy matching pursuit algorithm [27].The greedy matching pursuit algorithm has been widely studied due to its fast reconstruction speed and small calculation amount.
Existing greedy matching pursuit algorithms mainly include two types.The first category is greedy matching pursuit algorithms that require sparsity as a priori information, mainly including matching pursuit (MP) [28], orthogonal matching pursuit (OMP) [29], regularized orthogonal matching pursuit (ROMP) [30], compressed sampling matching pursuit (CoSaMP) [31], subspace pursuit (SP) [32], and generalized orthogonal matching pursuit (gOMP) [33] algorithms.The OMP algorithm is one of the simplest matching pursuit algorithms.In each iteration, the atom with the largest inner product of the observation matrix and the residual is selected and added to the support set.Use the least square method to update the residuals.Make the residual continuously approach a certain value to achieve convergence.The OMP algorithm guarantees the orthogonality of iterations, but cannot guarantee the correctness of atoms.So once the wrong atom is introduced, it will affect subsequent iterations.The CoSaMP algorithm adds backtracking ideas when selecting atoms.Each time an atom is selected, not only the atom is added to the support set, but also on the wrong atom to delete.Although the CoSaMP algorithm can improve the reconstruction probability and quickly complete the iterations, as the algorithm increases in sparsity, the reconstruction probability will drop sharply.The gOMP algorithm finds multiple support sets each time, the method of finding at most K times improves the probability of reconstruction.However, the sparsity of the algorithm is greater than K after the iteration.
The other is the adaptive algorithm that does not require sparsity, mainly including stagewise orthogonal matching pursuit (StOMP) [34], stagewise weak orthogonal matching pursuit (SWOMP) [35], and sparsity adaptive matching pursuit (SAMP) [36] algorithms.There are three problems in the aforementioned sparsity adaptive matching pursuit algorithm.The first problem is that the StOMP and SWOMP algorithms' support concentration does not consider the selection of the wrong atom, which leads to the limitation of the reconstruction probability of the algorithm [37].The second problem is that although the SAMP algorithm has a high reconstruction probability, it cannot accurately estimate the sparsity of the signal during adaptive estimation due to the limitation of the step size [38].The third problem is that SWOMP algorithm and SAMP algorithm are also susceptible to noise in the signal.Therefore, it is necessary to design a sparsity adaptive matching tracking algorithm that can accurately estimate the signal sparsity and high reconstruction probability in noisy radar signals.
This paper proposes a sparseness adaptive matching tracking algorithm for radar receiver system based on compressed sensing.Aiming at the problem of low reconstruction probability of greedy matching pursuit algorithm.A backtracking-based orthogonal matching pursuit BOMP algorithm is designed using backtracking ideas.In each iteration, the wrong atoms selected in the support set will be eliminated.Compared with other greedy algorithms that require sparsity, it has a higher reconstruction probability.Aiming at the problem of sparsity estimation error.Utilizing the principle of binary search, we devised a method sparsity search.First, set an estimated value of sparsity continuously and change the estimated value of sparsity through judgment criteria.Finally, get an accurate estimated sparsity.The estimated sparsity values are all based on high probability reconstruction.At the same time, due to the characteristics of binary search, it can quickly find the true value within a limited number of steps.For the noise problem in the signal, joining residuals associated with the threshold, the ability dynamically adapts to noise.The matching pursuit algorithm proposed in this paper is compared with similar sparsity adaptive algorithms, which improved the accuracy of output signal-to-noise ratio, reconstruction probability, and sparsity estimation.
The arrangement of this article is as follows.In the second section, the system model of the compressed sensing broadband receiver based on MWC is introduced.In the third section, an orthogonal matching pursuit algorithm based on backtracking is designed, which improves the reconstruction probability.In the fourth section, a sparsity adaptive reconstruction algorithm based on compressed sensing is proposed.In the fifth section, the performance of the algorithm is verified by simulation.Compare the reconstruction probability, sparsity, and running time of several adaptive algorithms under one-dimensional signal.Under the radar signal, the algorithm SNR and sparsity estimation and reconstruction probability are also analyzed.Finally, our conclusions are given in the sixth section.

System Model
In this section, the system model of MWC compressed sampling structure and sparse reconstruction algorithm will be introduced.The structure design and working principle of MWC compressed sampling wideband digital receiver are introduced, including the introduction of mathematical expressions and a brief description of the compressed sensing reconstruction algorithm.
2.1.MWC Structure.The radar signal in space is often continuous.In this article, for the convenience of introduction, the signal is set as a discrete signal under the Nyquist sampling rate.This setting will not affect the result.The received signal x½n is first mixed with the pseudorandom sequence pm ½n to obtain the mixed signal xm ½n, and the mixed signal is then passed through the corresponding low-pass filter h½n to obtain the filtered signal w m ½n.Finally, the signal is extracted by M p time and sent to the low-speed ADC for samples to get y m ½k.The discrete compressed sampling structure based on MWC receiver is shown in Figure 1.

Wireless Communications and Mobile Computing
The discrete time domain of the radar received signal can be expressed as follows: Among them, s½n is the effective signal component, η½n is the Gaussian white noise component in the signal, and N represents the length of the signal.x½n is a complex discrete signal; in the Nyquist sampling rate f NYQ = 1/T NYQ , T NYQ represents the period of Nyquist sampling, and the bandwidth range is F NYQ ≜ ½−f NYQ /2, f NYQ /2.The discrete-time Fourier transform (DTFT) is x½n and expressed as follows: x n ½ e −j2πf nT NYQ : ð2Þ It can be seen from Figure 1 that the structure contains M parallel branches.Taking the mth path as an example, the period of a periodic pseudorandom sequence p m ½n is T p , each period contains M p = T p f NYQ elements.The sequence p m ½n can be expressed as follows: The discrete Fourier series of the sequence pm ½n can be expressed as follows: l represents the index number and 0 ≤ l ≤ M p − 1, and P m ðlÞ is the discrete Fourier transform coefficient of the sequence p m ½n.The mixing rate of the signal and the sequence is f p = 1/T p = f NYQ /M p .The sequence p m ½n will be divided F NYQ into M p segments.The interval of each segment is the same f p .The position of each segment is called an index.The range of the baseband spectrum segment is In order to avoid edge effects, there is a need to meet f p ≥ B, where B represents the bandwidth of s½ n.The discrete-time Fourier transform (DTFT) of the mixed signal xm ½n is expressed as follows: The mixed signal xm ½n is filtered by a low-pass filter h½n.The frequency response is Hðe −j2πf T NYQ Þ.The cutoff frequency of the low-pass filter is f p /2.The discrete-time Fourier transform (DTFT) filtered signal w m ½n is expressed as follows: After the signal passes through the low-pass filter, the spectral components are within F p ≜ ½−f p /2, f p /2.But the data rate of the signal is still f NYQ .The data rate can be reduced by decimating the filtered signal.Use low-speed ADC sampling to complete the decimation of the signal.Decimation factor is M p = f NYQ /f p .The discrete-time Fourier transform (DTFT) of the sampled signal y m ½k is expressed as follows: The output can be written in the form of a matrix; the expression is as follows: and ~Low-pass filter Low-pass filter Low-pass filter 3 Wireless Communications and Mobile Computing the lth element is Xðe j2πf T NYQ ð f −lpÞ Þ.The linear combination can be expressed as follows: The observation matrix C is M × M p dimensional and can be expressed as follows: where P represents M × M p dimensional random matrix, F represents M p × M p dimensional discrete Fourier transform matrix, F l = ½1, e j2πl/M p , ⋯, e j2πðM p −1Þl/M p T is a column vector, and ð•Þ T represents transpose.The expanded form of the formula can be obtained.
The above formula is the frequency domain model of MWC compressed sensing receiver, and the feasibility of this structure is proved through mathematical deduction.Therefore, the compressed sensing of signals can be realized through the MWC structure.

Reconstruction Algorithm.
The sampled signal model obtained by using the MWC structure is the infinite measurement vector (IMV) model.In theory, it takes countless iterations to recover the original signal.Xu et al. proposed continuous finiteness in literature [20].The continuous-tofinite (CTF) algorithm can convert the IMV model of the MWC compressed sample signal into a multiple measurement vector (MMV) model.The signal under this model can be reconstructed from the original signal through the tra-ditional reconstruction algorithm structure.The block diagram of CTF algorithm implementation is shown in the figure.
As can be seen from Figure 2, first, use the sampled signal y½k to get the matrix Q, and the expression is as follows: which ð•Þ H represents the conjugate transpose.Then, decompose Q to get frame V.The expression is as follows: Obtain the joint support set S according to the sparse reconstruction of the vector V.The greedy matching pursuit algorithm can find the index set S where the signal exists.Using the sampling signal and the observation matrix, the time-domain waveform z s ½n of each channel in the baseband can be obtained.The expression is as follows: where ð•Þ † represents the pseudoinverse.The reconstructed signal x½n can be finally obtained.The time domain expression is as follows: h½n is an ideal interpolation filter, and the rate is the Nyquist sampling rate f NYQ .The greedy algorithm part will be studied below.

BOMP
This section proposes a backtracking orthogonal match pursuit (BOMP) algorithm.The algorithm mainly improves the OMP algorithm by three parts: threshold filtering, maximum selection, and backtracking idea.First, threshold filtering is used to calculate the inner product of the residual and the matrix.Filter out all atoms whose inner product is greater than the threshold u t and add these atoms to the support set.Secondly, the backtracking idea is used to find the least square solution using the support set.Through backtracking idea, only some atoms in the support set are selected to form a new support set.Finally, use the maximum selection.In order to ensure that the number of atoms' final output is the same as the sparsity, t atoms are selected to form a new support set during the tth backtracking.It is equivalent to only adding one atom for each cycle of support concentration.After the cycle ends, there are only t atoms in the support set.Specific steps are as follows: As seen in Algorithm 1, we need to enter A, y, and K. Find the atom whose inner product is greater than the threshold u t each time.Combine these atoms with the support set Λ t−1 obtained in the previous cycle to obtain a new support set β t .Use the matrix and residuals corresponding to the support set to find the least squares solution of b δ t .Replace b θ t , A t , and Λ t and find the residual r t according to b θ t and A t .Finally, judge whether the iteration is completed through t and r t .Set ε=10 −6 .Set the threshold th = t s kr s k 2 / kyk 2 , where r s is the residual calculated last time, k•k 2 is the 2 norms.t s ≥ 2 because the found atom needs to be greater than twice the average.
In order to verify the reconstruction probability of the BOMP algorithm, choose OMP, CoSaMP, and gOMP algorithms to compare with the algorithm proposed in this paper.The original signal x uses a Gaussian random signal, and the signal length is N = 256.The observation matrix Φ is a Gaussian random matrix, and the observation value M = 128.The interval of sparsity K is [35,70], and the step size is 5.Each sparsity is measured 1000 times.The same signal is reconstructed by four algorithms.The reconstruction probability under different sparsity is shown in Figure 3.
As shown in Figure 3, when the sparsity is greater than 30, the reconstruction probability of the OMP algorithm begins to gradually decrease.The reconstruction probability of CoSaMP algorithm, gOMP algorithm, and BOMP algorithm is still good.When the sparsity is greater than 50, the reconstruction probability of CoSaMP algorithm, gOMP algorithm, and BOMP algorithm begins to decrease.The reconstruction probability of BOMP algorithm is about 15% higher than that of gOMP algorithm.It can be seen that the BOMP algorithm improves the reconstruction probability.Regarding the reconstruction probability, the CoSaMP algorithm eliminates the wrong atom every time the atom is selected to prevent the wrong atom from affecting the selection of subsequent atoms, so the reconstruction probability is better than the OMP algorithm.The gOMP algorithm selects multiple atoms each time.According to the conditions for stopping iteration, it can be known that theoretically at most SK atoms can be selected.The correct atom (1) Compute inner product u t = r t−1 , a j , find greater than th from u t , and add the index values j to index set J 0 ; (2) Algorithm 1: BOMP algorithm.5 Wireless Communications and Mobile Computing set is a subset of the supporting set, thus further improving the probability of reconstruction.For the BOMP algorithm, it is finally achieved by selecting the best atom while continuously removing the wrong atom.Improve the probability of reconstruction.

DSAMP
This paper proposes a dichotomy-based sparsity adaptive matching pursuit (DSAMP) algorithm for compressed sensing broadband receivers.The algorithm mainly contains two parts.The first part is a fast atom search based on dichotomy.First, select the number of sparsity is 2 n−1 , in order to achieve sparse coverage under high reconstruction probability.The number of atoms selected for the first time is n.After the preestimated sparsity, the unknown sparsity problem becomes a known sparsity problem.Perform the residual calculation on the basis of this sparsity.Use criteria to determine the calculation results, then use the dichotomy to change the preestimated sparsity according to the evaluation results.The iteration is repeated until the dichotomy satisfies the stop iteration condition or the residual calculation of the nth atomic selection is completed.The sparsity when the iteration is stopped is the sparsity finally estimated by the algorithm.Find the true sparsity through this method.The second part is the BOMP algorithm, which guarantees that the reconstruction algorithm under a given sparsity has a high reconstruction probability.At the same time, it can ensure that the sparsity is the same as the number of atoms in the support set.Proposition 1.For the greedy matching pursuit algorithm, the estimated sparsity K 0 , the real sparsity K, when K 0 < K and K 0 ∈ K, the residual r t > ε; when K 0 ≥ K and K ∈ K 0 , the residual r t ≤ ε.
Proof.According to the principle of the greedy matching pursuit algorithm, r t is monotonically decreasing when the atom selection is correct in the iteration process, and then according to the iteration stop condition, it is obvious that the proposition is correct.
In Algorithm 2, for the recovery algorithm, the number of channels M must be at least twice the sparsity K to recover the original signal with high probability.Therefore, limit the sparsity range of the algorithm ½0, 2 n , where n = <log 2 M − 1 > ð<•>is roundÞ.In the first cycle, the number of atoms is set to the midpoint of the entire range, which is 2 n−1 .Steps (2)~( 6) are BOMP algorithm, and it is mainly to calculate the residual under the current sparsity and iterate under the set sparsity.Steps (7)~(9) are to set K 0 , and use Proposition 1 to set K 0 .t < K 0 indicates that the algorithm has completed convergence ahead of time.The real sparsity is less than the sparsity; at this time, the number of cycles is the estimated sparsity.When i = n+1, it means that the last calculation is completed, the estimated sparsity and support set are found, and the iteration can be stopped.When i = n, the sparsity K 0 has been selected for the last time.Satisfying r t < ε indicates that the sparsity is the estimated sparsity.It can directly output all residuals, sparsity, and support set.When r t < ε is not satisfied, K 0 plus 1 is the estimated sparsity.It is necessary to recalculate the residual, sparsity, and support set under the new sparsity.When t = K 0 , the algorithm has iterated K 0 times.r t > ε indicates that the algorithm has not reached the condition to stop iteration.It needs to update the estimated sparsity K 0 = K 0 + 2 n−1−i and continue to iterate.In other cases, the iteration continues after updating the estimated sparsity K 0 = K 0 − 2 n−1−i .

Simulation Results
In this section, the performance of the algorithm is verified through simulation of the algorithm.This article simulates the algorithm in two situations.The first is to compare the performance of the algorithm proposed in this paper with several other algorithms in the case of one-dimensional Input:M × N measurement matrix A N × 1 sampled measurement vector y Output: sparsity level t the estimated support set Λ t the nonzero values b θ t (2) Compute inner product u t = r t−1 , a j , find greater than th from u t , and add the index values j to index set J 0 ; (3)  As shown in Figure 4, it can be seen that the error between the reconstructed signal and the original signal is ε r < 1 × 10 −13 .It can be seen that the algorithm can accurately reconstruct the original signal.
Figure 5 shows the position and estimated position of the sparse signal when the sparsity is K = 51, respectively.It can be seen that the estimated sparsity is the same as the real sparsity of the signal, and the estimated position and the real position are also the same.The algorithm can accurately estimate the sparsity and the location of sparsity.

Sparsity Estimation Probability Diagram.
In this experiment, it is mainly compared with the sparsity adaptive algorithms such as StOMP, SWOMP, and SAMP to illustrate the advantages of the algorithm proposed in this paper.The signal x is a Gaussian random signal, the signal length is N = 256, the observation value is M = 128, the interval of sparsity K interval is [1,66], and the step size is 5.The sparse locations are randomly selected.The observation matrix Φ is a Gaussian random matrix.The sparse position is randomly selected and performs 1000 Monte Carlo experiments at each sparsity.The matrix is a Gaussian random matrix.Figure 6 is the reconstruction probability diagram of 1000 Monte Carlo experiments under different sparsity.Figure 7 is the average sparsity estimation diagram of 1000 Monte Carlo experiments under different sparsity.Figure 8 is the average running time diagram of 1000 Monte Carlo experiments under different sparsity.
As shown in Figure 6, when the sparsity is greater than 40, the reconstruction probability of the StOMP algorithm and the SWOMP algorithm begins to decrease.The reconstruction probability of SAMP algorithm and DSAMP algorithm is still 100%.When the sparsity is greater than 56, the StOMP and SWOMP algorithms cannot be reconstructed.The reconstruction probability of SAMP algorithm and DSAMP algorithm begins to decline, but the reconstruction probability of SAMP algorithm and DSAMP algorithm is similar.

Wireless Communications and Mobile Computing
As shown in Figure 7, the real K represents the real sparsity set, and the remaining four lines represent the corresponding algorithms.When the sparsity is less than 45, the SWOMP algorithm can accurately estimate the sparsity when the sparsity is small, and as the sparsity increases, the error of the estimated sparsity number gradually increases.The number of sparsities estimated by the StOMP algorithm changes with the number of real sparsities, but there is an error with the real sparsity.When the sparsity is greater than 45, the sparsity estimated by the SWOMP algorithm has exceeded 100, which is relatively close to the observed value M. When the sparsity is greater than 45, the amount of spar-sities estimated by the StOMP algorithm continues to decrease.From Figure 6 can be seen that the reconstruction probability of these two algorithms is decreasing under this sparsity.When the sparsity is less than 51, due to the setting of the step size of the SAMP algorithm, the amount of sparsity estimates each time is an integer multiple of 10.The DSAMP algorithm can accurately estimate the number of signal sparsity.When the sparsity is greater than 51, the number of sparsity estimated by SAMP algorithm is not changed, and the sparsity estimated by DSAMP algorithm also has errors.It can be seen from Figure 6 that the reconstruction probability of these two algorithms is rapidly decreasing under this sparsity.
As shown in Figure 8, the average running time of the StOMP algorithm and the SWOMP algorithm is relatively short, the average running time of the DSAMP algorithm is the longest, and the average running time of the SAMP algorithm is between several algorithms.This is because the StOMP algorithm and the SWOMP algorithm can select multiple atoms in each iteration and can quickly complete the iteration.The SAMP algorithm can select multiple atoms in each iteration, but the iterative atoms are repeatedly selected in the iteration, so the average running time is longer than the StOMP algorithm and the SWOMP algorithm.The DSAMP algorithm only adds one atom per iteration, so the average running time is the longest.Wireless Communications and Mobile Computing us, the noise is Gaussian white noise, the signal-to-noise ratio SNR = 10 dB, and the signal amplitude is A 1 = 0:7, A 2 = 0:5, and A 3 = 0:6.The parameter settings of the MWC compressed sampling broadband digital receiver: the number of sampling channels is M = 128, the periodic pseudorandom sequence adopts a Bernoulli pseudorandom sequence of ±1, the period length is M P = 255, the signal sampling time is T = 10:2 us, and the original sampling point number is N = 204000 points.Figure 9 is the radar signal reconstruction diagram of the compressed sampling broadband receiver of MWC by the DSAMP algorithm, including the original signal, the noise signal, and the reconstructed signal.Figure 10 is the corresponding spectrogram.
As shown in Figure 9, the algorithm can estimate 3 signals and reduce the noise of the output signal in the reconstruction process.As shown in Figure 10, the algorithm can accurately estimate the position of the 3 signals.Since the real signal is used in the simulation, the same spectrum will appear in the position of the symmetrical spectrum.It can be seen that the spectrum on the far left and right is the LFM signal.The noise of the four reconstructed signals in the middle is higher than that of the noise-added signal, and all noises beyond the sparsity are eliminated.Therefore, the algorithm has a certain effect on the improvement of signal-to-noise ratio.
In order to prevent multiple signals from appearing in a channel at the same time, the frequencies of the signals are equally spaced.The initial frequency of the signal is, respectively, f 0 = 1:1555 GHz, the interval between each signal is 0:315 GHz, the interval of the number of signals is [1,20], and the step is one.The signal amplitude of is all set to A 1 = 0:7.Perform 1000 Monte Carlo experiments under each signal-to-noise ratio.Other conditions are the same as in Figure 9.
where N is the times of independent  9 Wireless Communications and Mobile Computing algorithm in Figure 11 is the intersection of the estimated sparsity position and the real position.If the intersection is the same as the real position, the reconstruction is considered successful.It can be seen from Figure 11 that the reconstruction probability of SWOMP algorithm and SAMP algorithm is better than DSAMP.The reconstruction probability of StOMP algorithm drops sharply after the number of signals is greater than 10.It can be seen from Figure 12 that the RMSE of the SWOMP algorithm and the SAMP algorithm is very large.It can be seen from Figure 12 that the RMSE of the SWOMP algorithm and the SAMP algorithm is very large.After the number of signals is greater than 3, it decreases.The sparsity estimated by the SWOMP algorithm is close to the observed value M. When the number of signals in the SAMP algorithm is greater than 2, the estimated sparsity is all 60.The RMSE of the StOMP algorithm rises sharply after the number of signals is greater than 10.The RMSE of the DSAMP algorithm has been kept within 5. The SWOMP algorithm and SAMP algorithm can still maintain a reconstruction probability above 0.7 when the number of signals is 20.This is because the two algorithms continue to add atoms to the atomic set until the stop iteration condition is met.Although it can be reconstructed, it introduces a large number of error atoms.When the number of signals in the StOMP algorithm is greater than 10, the reconstruction probability drops rapidly and the RMSE rises rapidly, indicating that the algorithm begins to fail at this time.The DSAMP algorithm maintains a small RMSE when the reconstruction probability decreases.This is because as the signal sparsity increases, the accuracy of the algorithm sparsity estimation decreases, but the estimated sparsity is still close to the true sparsity.
Set the number of signals to one signal, the frequency of the signal is selected randomly between [0, 10 GHz].The signal-to-noise ratio SNR interval is [-15, 20], and the step size is 5. Perform 1000 Monte Carlo experiments under each signal-to-noise ratio.Other conditions are the same as in Figure 9. Figure 13 is a simulation diagram of the algorithm SNR.
As shown in Figure 13, the SWOMP algorithm and the SAMP algorithm are not effective in improving the output signal-to-noise ratio under low signal-to-noise ratio conditions, while the DSAMP algorithm and StOMP algorithm have a greater improvement in the output signal-to-noise ratio under low signal-to-noise ratio conditions.It can be seen from Figures 11 and 12 that the SWOMP algorithm and the SAMP algorithm are in the case of low signal-tonoise ratio.When the observation matrix is mixed, the noise of all frequency bands is mixed into the baseband, which causes the noise of the baseband signal to be strengthened, and the number of signals recovered during reconstruction is greater than the sparsity.Therefore, the output SNR is lower than the input SNR under the condition of low SNR.For DSAMP algorithm and StOMP algorithm, the sparsity estimation is more accurate at low SNR.It is equivalent to filtering out all the out-of-band noise of the signal, so the signal-to-noise ratio is greatly improved.For a high signalto-noise ratio, the accuracy of the sparsity estimation of the four algorithms increases, so the output signal-to-noise ratio is improved.The DSAMP algorithm and the StOMP algorithm have a relatively close output signal-to-noise ratio.This is because the two algorithms have the same threshold setting when selecting the threshold.The DSAMP algorithm can have the advantage of improving the signal-to-noise ratio.

Conclusion
Adaptive sparsity estimation is necessary for wideband receivers based on MWC compressed sampling in radar systems.This paper proposes a sparsity adaptive matching pursuit algorithm based on dichotomy.In the algorithm proposed in this paper, there is no need to know the sparsity of the signal in advance.We estimate the sparsity of the signal 10 Wireless Communications and Mobile Computing in a limited number of iterations through the dichotomy.It can be seen through simulation that an accurate and adaptive estimation of the sparsity of the signal is achieved.The orthogonal matching pursuit based on backtracking improves the accuracy of supporting concentrated atoms and finally improves the reconstruction probability of the algorithm.Due to the algorithm's accurate estimation of a single signal, the noise introduced by error atoms is reduced during reconstruction.Improve the signal-to-noise ratio of the algorithm to the signal.As the sparsity increases, the algorithm can still maintain a better reconstruction probability and more accurate sparsity estimation.In the future, the running time of the algorithm can be improved to reduce the calculation time of the algorithm.

Algorithm 2 :
DSAMP algorithm.6Wireless Communications and Mobile Computing signals, mainly to verify the accuracy and running time of the algorithm for the estimation of reconstruction probability and sparsity.The other is to simulate under radar signals, which mainly verifies that the algorithm can adapt to multiple sparse signals under low signal-to-noise ratio and finally simulates the signal-to-noise ratio of several algorithms.5.1.One-Dimensional Signal Simulation.This experiment mainly verifies the correctness of the one-dimensional signal, including signal reconstruction error and sparsity estimation.The signal x is a Gaussian random signal, the signal length is N = 256, the observation value is M = 128, and the sparsity is K = 51.The sparse locations are randomly selected.The observation matrix Φ is a Gaussian random matrix.Figure 4 is a one-dimensional signal reconstruction diagram of the DSAMP algorithm, including the original signal, reconstructed signal, and reconstruction error.Figure 5 is a diagram of the true position and estimated position of the sparse signal in Figure 4.

8
Figure 11 is a reconstruction probability diagram of different numbers of signals.Figure 12 is an RMSE diagram of sparsity.The simulations use root mean square error (RMSE) to analyze the sparsity estimation performance, expressed as follows:
and Λ t and find t maximum values from b δ t denoted by b θ t , α t number t columns denoted by A t , and β t number t columns denoted by Λ t ; (5) Update residual r t and Λ t and find t maximum values from b δ t denoted by b θ t , α t number t columns denoted by A t , and β t number t columns denoted by Λ t ; (5) Update residual r t Monte Carlo simulations, K i denotes the sparsity estimation, and K denotes real sparsity.Figures11 and 12are the results of the same experiment.The basis for judging the successful reconstruction of the