SymmetricQuantileQuantizerParameterization for theLaplacian Source: Qualification for Contemporary Quantization Solutions

In this paper, we consider the opportunities and constraints, which rest on quantization as a guiding principle for data representation and compression. In particular, we propose a novel model of Symmetric Quantile Quantizer (SQQ) and we describe in detail its parameterization. We suggest a simple method for offline precalculation of its parameters and we examine the inevitable loss of information introduced by SQQ, as an important part of bit optimization task at the traditional network level, which can be globally mapped out in many contemporary solutions. Our anticipation is that such precalculated values can be leveraged in deterministic quantization process. We highlight that this notice heavily relies on the fact that the values of interest are distributed according to the Laplacian distribution, which we consider in the paper. ,e basic difference of our SQQ and the previously established asymptotically optimal quantizer model, that is, Scalar Companding Quantizer (SCQ), is reflected in the fact that, in SCQ model, both decision thresholds and representation levels are determined in accordance with the specified compressor function, whereas in our SQQ model, a precedence of SCQ model for the straightforward decision thresholds calculation is used, while the representation levels are optimally determined for the specified decision thresholds and assumed Laplacian distribution. As a result, our SQQ outperforms SCQ in terms of signal-to-quantization noise ratio (SQNR). As stated in this paper, there are numerous indications to make us believe that appropriate quantizer parameterization will move us closer to an optimization in the amount of the transferred data in bits, which is strongly dependent on the amount of SQNR.


Introduction
Humans are unique in their ability to weigh perceptions and make decisions based on them. Apparently, human observation is inherently very subjective and thus varies across individuals. Additionally, emotional, cognitive, and social processes can override basic instincts and deductive or inductive conclusion and guide behavioral choices. In this respect, the human decision to infer is a consequence of an interaction of all these factors. In practice, it is not realistic to identify and adequately quantify all the factors that play a role in a certain phenomenon, even if the phenomenon works simply. As the amount of data we need to generalize grows, the number of features or dimensions also grows. Assuming that the explanatory part of many problems is characterized as highly dimensional, in order to fix the data in a linear space or on some space of a significantly smaller dimension, it is desirable to involve a prediction model or model ensembles [1][2][3].
When analyzing the effectiveness of predictions, it is preferable to make the prediction models close to the actual data. Although it is possible to achieve high accuracy on the actual training set, the real challenge is to develop a model that generalizes well to data not known in advance. e outlined criterion to the predictive model logically leads to a solution, which lets some amounts of data to dictate algorithms and solutions that penalize underfitting and overfitting. More generally, the presence of overfitting is a common problem of many neural network (NN) models [4]. Additionally, it is not possible to partially prevent from occurring possible extrapolation errors, commonly appearing in the uncovered regions, with no data. e fundamental question is imposed about how to build robust systems that perform reasonably and safely in the real world. Achieving robustness is certainly a goal not only for the fullprecision NN, but also for quantized neural network (QNN) [5]. Inspired jointly by bit optimization and robustness, the exploration and interpretation of the different quantization principles behind the QNN remain as intriguing directions for future research. A growing interest in deep neural network is directed towards the efficient inferencing and training using quantization [6]. In [7], it has been recently proved that with appropriately trained NNs using binary weights and activations, quite well performance could be achieved even by extremely low-precision networks, such as binarized neural network. It turns out that quantization is a desirable mechanism that can dictate the entire NN performance. A promising deployment of QNN in many contemporary approaches and artificial intelligence applications reduces the overall computational and memory cost [5]. As shown in [8], this is of great importance since the amount of data to be processed constantly grows so that more powerful hardware (CPU/GPU) is required. Moreover, the QNN approach addressed in [5] is particularly beneficial for implementing in models with the extreme memory requirements, such as mobile devices and edge devices.
Whereas many contemporary applications are inherently decentralized or heterogeneous, data are distributed throughout a network without an aggregation, due to various communication constraints [9]. It is implied that compressed or quantized, rather than raw, data are transferred [10][11][12][13]. In other words, quantization is seen as a preferred mechanism for typically indispensable bit reduction. In brief, quantization plays a prominent role in optimizing various data transmission models in the existing network solutions. ere are numerous indications to make us believe that appropriate quantizer parameterization will move us closer to an optimization in the amount of the transferred data in bits. Often, the same quantizer is applied to a variety of signals with distinct statistical features. For this reason, it is desirable to design the quantizer to be robust to the mentioned features of the input signal. In numerous literature studies, it has been assumed that the underlying distributions are known and with a specified parametric form [10,[14][15][16][17][18][19][20][21][22][23][24][25][26]. However, in many situations, typically, there is no prior information about the distribution of the signal to be processed, so that one can accept widespread opinion that the Laplacian distribution fits many natural, economical, and social phenomena [27]. Broadly speaking, as a consequence of the partial or total lack of information about the distribution, accepting some of distributions, such as the Laplacian distribution, offers the opportunity for a more complete view of the input attributes and the relations among their representatives.
is explains our choice of considering the Laplacian distribution in the analysis presented in the paper.
Numerous papers have given a general framework to stimulate discussion on many different possible interpretations of the impact of the number of quantization levels and the support region width on both the granular and the overload distortion of the quantizer designed for the assumed Laplacian source [16-18, 21, 22, 25]. Having in mind the definition of the support region, as the area where distortion is small or at least appropriately bounded [11,22], the main trade-off in quantizer design is done between an accommodation to the signal's amplitude dynamic and minimization of the overall distortion. In this regard, as shown in [22], the support region threshold is probably the most striking feature of the quantizer. To this end, we address the problem of determining this striking feature of our SQQ, as well as how to perform its parameterization in order to qualify our novel SQQ model for sources having Laplacian probability density function (pdf ) with improved overall quantization efficiency.
One of the main goals of a quantizer designing or parameterization is to provide minimal possible distortion, that is, maximal possible SQNR, for a given bit rate, or equally, for a given number of quantization levels N. Designing an optimal N-level scalar quantizer consists of determining the set of N + 1 decision thresholds and the set of N representation levels minimizing distortion that measures quantizer performance [11,12]. In general, as we highlighted in [20], it is extremely hard to determine the global minimum of distortion function simultaneously with respect to all the decision thresholds and representation levels. However, Lloyd and Max found that it is possible to solve two partial problems [11]. e first problem is to determine the optimal representation levels for a given set of the decision thresholds. e second problem is to determine the optimal decision thresholds given the set of representation levels. In doing so, Lloyd and Max developed an iterative algorithm for designing an optimal quantizer for a source with known pdf. As stated in [11], the necessary and sufficient conditions for optimal quantizers are that each decision threshold is the midpoint of the adjacent representation levels and that each representation level is the centroid of its respective cell with respect to the given pdf. Due to the mutual dependence of these key quantizer parameters, the optimal quantizer has to be designed iteratively by applying the Lloyd-Max algorithm [11]. However, this algorithm is too complex and time-consuming and its application is due to the prominent problem with the algorithm initialization mainly limited to very low bit rates. Accordingly, novel quantizers with lower design complexities are very beneficial. Taking into account the fact that SCQ is significantly simpler to design than Lloyd-Max's quantizer [11] and that SCQ has worse performance, especially for small and medium bit rates, in this paper we came up with the idea to determine the set of decision thresholds according to the SCQ model and to determine the optimal set of N representation levels for the given set of decision thresholds and the assumed Laplacian pdf, in a similar manner as specified by Lloyd-Max's quantizer. In brief, the outcome of this paper is one completely novel model of scalar quantizer, called SQQ, which is not only simple to design, since it does not require iterative parameterization, but also outperforms asymptotically optimal SCQ in terms of SQNR. Unlike with the SCQ model, where representation levels are determined from the compressor function, in our SQQ each representation level is determined as the centroid of the corresponding quantization cell with respect to the assumed Laplacian pdf. An important aspect of our interest in SQQ design is that it dictates the assumptions one can make on the statistics of the input signal, emerging from maximal SQNR. In particular, recasting the bit optimization problem, as the mean squared error (MSE) prediction problem, we offer the manner for direct computation of the support region threshold and SQNR from the formulas derived in the paper with the goal of characterizing the performance of our SQQ and providing a fair comparison with other well-known quantizer models. Moreover, as we have recently highlighted in [16] that, in general case, the clipping process formulation can be directly related to the support region determination, our anticipation in this paper is that such an offline precalculated value of support region can be stored in a lookup table for a fast retrieval, which can be indeed beneficial in numerous applications. e paper is organized as follows. Section 2 provides an insight into the Laplacian pdf and briefly summarizes known facts from the literature about it. e main results of the paper are given in Sections 3 and 4. In particular, parameterization of our SQQ is presented in Section 3, whereas in Section 4, the discussion on our results and the comparison with the up-to-date related results are provided. Finally, in Section 5, we summarize the paper goals and we conclude with our research results.
In order to improve the readability of the rest of the paper, we opt to summarize here the list of abbreviations and notations we utilize: Symmetric Quantile Quantizer (SQQ); Scalar Companding Quantizer (SCQ); signal-to-quantization noise ratio (SQNR); neural network (NN); quantized neural network (QNN); probability density function (pdf ); mean squared error (

Why Laplacian pdf?
Legacy voice is mature, robust, and high performing. Voice service continues to be globally forecasted as the challenge, which traditionally has a relatively predictable and small use of bandwidth but requires real-time transmission. It is well known that the distinctive attributes of speech signals are well modelled by the Laplacian pdf or the Gaussian one [10,11,27]. In addition, the most common assumption is that the pixel difference values can be described by either the Laplacian or the generalized Gaussian pdf [12]. However, the Laplacian pdf is more tailed than a Gaussian one. Moreover, the Laplacian pdf provides typically tractable closed-form formulas necessary for the design and performance assessment of various transmission systems. is tractability is often impossible in the case where one or more Gaussian pdfs are assumed [28], so that approximations of the special functions, such as erfc function and Q-function, or approximation of the compressor function, are required [29,30]. For that reason, Laplacian pdf is a more preferable pdf, which has been recently used to model not only weights and activations in NNs, but also speech and audio signals [15-20, 25, 31]. Namely, according to [10], the Laplacian pdf (see Figure 1) is arguably the most suitable form for modelling statistical properties of both audio and speech signals. Since audio signals may exhibit varying degrees of stationarity, it is often advantageous to allow for flexible coders backed by the flexible transmission rates, able to adjust in real time to continuously time-varying signal attributes. As known, Laplacian and Gaussian pdfs belong to the family of the log-concave functions. e concept of logconcavity was widely studied in the literature for many symmetric distributions such as the Laplacian one, for instance, in [14]. e log-concavity can facilitate the accurate anticipation of the quantizer design from the standpoint of its symmetry. According to [14], as the Laplacian pdf satisfies the log-concavity condition, the optimal quantizer is itself symmetric. More generally, the Laplacian pdf can be used to model symmetric shaped data, such as weights used in NN [15]. Rather than striving for a full precision, it has been shown in [15] that, by quantizing the NN into low-bit representation and with an analytical clipping assessment, negligible accuracy degradation can be incurred. As we have already mentioned that, in general case, the clipping process formulation can be directly related to the support region determination, we believe that such an offline precalculated value of the support region threshold can be stored in a lookup table for a fast retrieval, which can be indeed beneficial. Given the acceptable quantization error, the question arises whether the approximated quantized network still has the same ability to extract features from the input data. e outlined clipping assessment task is clearly possible by minimizing the MSE. e main reason why we deal with the MSE is because the error function for the observed Laplacian pdf is a convex function [11]. To our knowledge, the convexity of the MSE distortion has recently received a lot of attention in [17]. Essentially, one of the preferences of convexity is that the error of averaging two guesses is always lower than the average error from the same two guesses. e lack of convexity has been seen and proved for odd numbers of quantizer representation level, while for any symmetric density, such as the Laplacian, the MSE is convex for even numbers of representation levels [17].
Before moving on to the quantizer modelling, we should highlight here that some of the bases in our further analysis are straight forwarded by the rigorously stated partial distortion theorem in the case of the Laplacian pdf [18]. We should also highlight that some novel insights in the results of the abovementioned approach applied in our SQQ design are provided in this paper, which could be of great Mathematical Problems in Engineering significance in the further analysis of quantization and in qualification for contemporary quantization solutions.

Symmetric Quantile Quantizer Designed for the Laplacian pdf
Signal coding generally refers to the process of extracting and transforming of the raw analog information from the signal, into digitally coded representations, to facilitate the access to descriptive and invariant attributes of this signal [11]. Signal decoding, as the reciprocal process of signal coding, on the receiver end, has the main task to obtain the closest possible signal to an original signal, using the extracted attributes. Well-designed coder or quantizer can reduce the potential influence of unpredictable statistical characteristics of the signals to the quality, thereby contributing to increased robustness. Following the main aspect of data coding and compression and a constant intention to decrease the number of bits necessary to deliver original analog signal, companding quantizers, based on some compression laws [11,[19][20][21]32], have been receiving much research interest, especially due to the fact that they are well suited in terms of complexity of designing and implementation, as shown in [33]. In general, an input signal is divided by the quantization procedure into a granular region and an overload region, or alternatively, in an inner and an outer region, which are for the symmetric quantizer separated by the support region thresholds denoted by − x max and x max , respectively [22] (see Figure 2). Both analyses for the simplest quantizer model, that is for uniform quantizer, emerged recently, are authored by Sangsin Na et al. [18,25]. Paper [25] deals with the monotonicity-by-two for the symmetric uniform quantizers in the case of the four most common pdfs, including Laplacian pdf, we observe here. Another interesting fact herein stated and proved for the Laplacian pdf and sufficiently large number of quantization levels N is that the optimal step size for the uniform quantizer decreases, as bit rate increases by one. Initially derived from "partial distortion theorem," an analysis given in [18] uncovers several insights about the relationship between the MSE distortion, number and position of quantization levels in inner and outer region, or granular and overload region.
One of the conclusions from [18] is that microscopical contribution to the total distortion is the same from each cell, whereas microscopically is more from outer cells. In what follows, we will ascertain some novel insights in the results of the abovementioned approach applied in our SQQ design.
Let us consider the variance-matched case of our interest in which the Laplacian pdf of zero mean and unit variance is assumed: Referring to the aforementioned log-concavity and the MSE convexity proved for the Laplacian pdf [14], of particular interest, in this paper we propose one simple manner for the SQQ parameterization. Our N-level SQQ is defined by mapping Q: R ⟶ Y [22], where Q(•) is a symmetric characteristic of our quantizer, R is a set of real numbers, and Y ≡ − y N/2 , . . . − y 1 , y 1 , . . . , y N/2 ⊂ R is a set of representation levels (see Figure 2), which makes the code book of size │Y│ � N � 2K. Subsequently, one can accept that, for every symmetric quantizer with an even number of nonoverlapping cells, quantization levels are symmetrically placed about the mean. Without losing of generality, we are focusing only on the K positive counterparts, as shown in Figure 3, where some of the positive counterparts for the odd compressor function c SQQ (x) are presented. e necessary parameters of our N-level SQQ are given as follows: {quantization level density function; compressor function; decision thresholds; representation levels; support region threshold; distortion}.
Quantization level density function of our novel model of quantizer named SQQ is given by (2) Foremost, the substitution of equation (1) in (2) yields Our odd compressor function is c SQQ (x): [0,+∞)⟶ [0, 1) Nonnegative decision thresholds of our SQQ are determined from the compressor function as whereas the negative counterparts explicitly follow from the symmetry. Unlike with the SCQ model, where representation levels are also determined from the compressor function, in our SQQ, each representation level is determined as the centroid of the corresponding quantization cell where X and E p {·} denote the continuous random variable with pdf p(x) and expectation with respect to p(x): For the assumed pdf and for i � 1,2, . . ., K− 1, we derive where Δ SQQ i is a distance between neighbouring decision thresholds of our SQQ.
Let us substitute x SQQ K � +∞ in equations (3) and (4) so that the corresponding quantization level density function and compressor function of our SQQ are derived as follows: Accepting such a defined c SQQ (x), for the nonnegative decision thresholds in our SQQ defined by equation (5), we obtain and for i � K − 1, from equations (5) and (12), we directly derive the closed-form formula for clipping: By applying Bennett's integral [23] to approximate the granular distortion of our novel SQQ where in our model it holds Δ � 2/N � 1/K, we derive Further substituting equation (13) in (15) gives us the closed-form formula for the granular distortion: To calculate the overload distortion, we use y SQQ K obtained from equation (8):

Granular region
Overload region -y 4 Direct substitution of equation (13) in (18) then yields Eventually, by summing equations (16) and (19) for the total distortion, composed of the granular and the overload distortion, we obtain where subscript Asym refers to the application of Bennett's integral in our asymptotic analysis. In order to ascertain the accuracy of the derived asymptotic formula for the overall distortion in what follows, we derive one beneficial exact formula for the overall distortionD SQQ Exact , for the Laplacian pdf of zero mean and unit variance.
Let us start our derivation with the general formula for the exact distortion for any even N value: which can be rewritten as so that we derive By using equation (12), we finally derive To conclude, the necessary parameters of our N-level SQQ are given as follows:  Mathematical Problems in Engineering Quantizer parameterization dictates its entire performance. If decision thresholds and representation levels are chosen more appropriately, the overall distortion is smaller; that is, SQNR is higher, which translates to a reduction in the number of bits required from the quantizer for achieving a certain SQNR. Let us highlight here that SQQ is a modified form of nonuniform quantizer named as compandor. Namely, compandor or SCQ is conceptually realized as cascade of three intrinsic functional blocks: compressor, uniform quantizer, and expandor [11]. In particular, in the considered SQQ model, a precedence of compandor model for the decision thresholds calculation is used, while the representation levels are optimally defined, so that it makes the basic difference with SCQ model where both decision thresholds and representation levels are determined in accordance with the specified compressor function. erefore, we highlight here that our SQQ and SCQ models are characterized with the similar design and implementation complexity level so that both models can be used for the efficient Lloyd-Max algorithm initialization [24]. Moreover, SQQ and SCQ are characterized by the elegant determining of asymptotic quantizer's performances for the Laplacian pdf.
Towards determining the cases of interest where the proposed SQQ outperforms SCQ for the same support region threshold value x SQQ K− 1 � x SCQ K− 1 , we define the necessary parameters of N-level SCQ as We can highlight here that the above described novel SQQ model and the appropriate formulas derived for this model are main results of this paper. In what follows, we show that the overload distortion of the SCQ is not equal to the one of the SQQ. Specifically, for the support region threshold x SCQ K− 1 of SCQ, designed for the Laplacian pdf of zero mean and unit variance, in the following, we determine the overall distortion and we show that it differs from the distortion of SQQ given by equation (20).
Briefly, we omit the complete proof and only highlight its main steps: (i) Recalling equations (14), (16), and (17) and discussions about the granular distortion of the SQQ, g . Paying in-depth attention to these equations, we can intuitively propose how to calculate the granular distortion D SCQ g � (9/2N 2 )(1 − (2/N)).  (17) gives us D SCQ over in the following form:

Mathematical Problems in Engineering
Eventually, the total distortion and SQNR of SCQ are

Numerical Results and Analysis
As earlier outlined in Section 3, we analyze the Laplacian SQQ with an even number of quantization levels, where each level is the centroid of the respective cell. Considering both the overload and the granular region, we can adopt that the large, compressed signal values occur very rarely compared to the small values. For our SQQ analysis refinement, it seems reasonable to accept the monotonicity of the neighbouring cells, which stems from the cell size determination, as illustrated in Figure 3. Note that, from equation (10), we can conclude that Δ SQQ i decreases strictly monotonically to Δ SQQ 0 as i decreases, since the source pdf has a finite support x SQQ K− 1 and Δ SQQ K− 1 is the cell of a finite length; i.e., this is the cell that lies the most outwardly in the granular region.
With the increase of N (N ⟶ ∞) and with the neglection of the overload distortion due to narrowing the overload region, one can notice that equation (20) approaches the PD high-resolution formula D PD � 9/(2 N 2 ) [26]: We can explicitly compute the overload distortion by applying Bennett's integral [20] and adopting x SQQ K− 1 from equation (13): From equations (30) and (32), one can conclude that it is not cumbersome or demanding to derive the following formula for the granular distortion of our SQQ: which is expectedly equal to equation (16). For the sake of the overload distortion analysis, we can compare the formulas given in equations (19) and (32), with the one in equation (28) that we derived for the companding quantizer. Obviously, the following inequality holds: D SQQ over < D SCQ over < D B.I. over . is is indeed one of the beneficial features of the proposed SQQ model since it introduces smaller distortion compared to the asymptotically optimal quantizer model, that is, compared to SCQ. Having in mind that relative competitive relations defined as do not depend on N, we can highlight here that this interesting notice deserves further attention, especially when the importance of the overload distortion is more pronounced. We can also highlight here that, as long as the overload distortion is not predominantly negligible to the granular distortion, D SQQ over provides significant advantage over D SCQ over and D B.I. over . An important aspect of our interest in SQQ design, as a means of data conversion or compression is, that it dictates the assumptions one can make on the statistics of the input signal, emerging from maximal SQNR. In particular, by recasting the bit optimization problem, as the MSE prediction problem, we can derive SQNR directly from the affordable formulas characterizing its performance, given in equations (20) and (21): SQNR SQQ Exact � − 10 log 10 D SQQ Exact .
In order to provide fare performance comparison, we further define relative errors: e values of SQNR presented in Table 1 show that SQNR SQQ Asym provides significant performance gain when compared to SQNR PD . It is worthy to notice that, for N ≥ 128, SQNR SQQ Asym achieves roughly the same level of accuracy as that of SQNR SQQ Exact or SQNR PD , for the same number of levels N, which we have intuitively expected. Our approach of determining distortion, or directly related SQNR, SQNR � − 10 log 10 (D)) [11] provides a more complete insight into the impact of both the granular and the overload distortion to the total distortion.
In what follows, we provide the following observations about some of the results of our analysis and the performances of the proposed SQQ model: 8 Mathematical Problems in Engineering (1) SQQ model introduces smaller distortion compared to the observed asymptotically optimal quantizer model, that is, compared to SCQ. (2) SQQ is very simple to design since it does not require iterative parameterization as it is the case with the optimal Lloyd-Max quantizer. (3) Although outputting approximate values, the derived asymptotic formula for SQNR SQQ Asym is very accurate, even for a small N value, which make it very beneficial. (4) Since δ A and δ PD amount approximately less than or equal to 1% for N � 10 and N � 30, respectively, the correctness of the analysis performed for our SQQ model for small and medium bit rates is confirmed. is is of great importance since the accuracy of the asymptotic theory decreases noticeably for the classes of small or medium numbers of the quantization cells N. To perceive the model of SQQ and SCQ, we perform an analysis for a wide-ranged set of the quantization levels, where for SQQ we specify a relative competitive benefit over the high-resolution theory (asymptotic theory), that is, the performance gain: whereas to provide the comparison with SCQ we specify the following performance gain: G SQQ− SCQ � 10 log 10 D SCQ D SQQ � 10 log 10 By observing In what follows, we compare SQNR values and we analyze the SQNR curves overlapping along the finite part of their length. Note that curve overlapping is an illustrative indicator and hence, more specifically, our goal is to emphasize strong overlapping for SQNR SQQ Asym and SQNR SQQ Exact , as well as SQNR PD and SQNR SCQ when R ranges from 2 bits/ sample to 5 bits/sample (Figure 4(a)). We can also observe that, as shown in Figure 2(b), for R ranging from 3 bits/ sample almost to 5 bits/sample, the derived SQNR SQQ Asym overperforms other asymptotic solutions as SQNR PD and SQNR PD � − 10 log 10 D PD � 10 log 10 while for R � 2 bit/sample, SQNR Hui outperforms SQNR of our SQQ. is can be explained by the fact that, for such low bit rate, the value of the support region threshold, optimized in the asymptotic analysis for N >> 1, is more inaccurate in the case of a nonuniform quantizer than in the case of a uniform one, which is consequently reflected in the SQNR values of these two quantizer models. From the analysis presented in the paper, one can conclude that the granular distortion can be applicable as the total distortion, as long as the overload distortion does not occur, or it is predominantly negligible to the granular distortion. While in the most cases where quantization is traditionally used, for instance, in [16,[19][20][21]24], the highresolution theory (N ⟶ ∞) is well justified, it totally breaks down for the classes of small or medium numbers of the quantization cells N. Large values of N lead to more fineleveled quantization, but they result in a less efficient computation and storage consumption. Many quantizers often operate at a very modest bit rate, i.e., at medium bit rates and, for that reason, they indeed make sense of imposing various constraints on the number of used bits. Accordingly, we believe that our analysis derived not only for higher bit rates (R ≥ 5 bit/sample) but also for small and medium bit rates could be of great significance.
In brief, the model of SQQ that we have proposed is fully specified with the necessary tunable parameter set given by equation (26) and it has very beneficial properties: in pursuit of the reduced total distortion, the number of quantization levels can be gradually incremented; i.e., some compression gain can be provided. Having recast the total distortion asymptotic formula given by equation (20) and by specifying apparently simple solution for the SQQ support region (see equation (13)), in this paper we intended to deal with an opportunity in considering a wide usage of SQQ in NN   training process. However, this has been left for the future research. Eventually, we can anticipate that our results for the successful deployment of clipping in neural networks [15] are very encouraging.

Summary and Conclusions
As a main contribution in this paper, we have tackled SQQ parameterization by suggesting one simple method for offline calculation of fully qualified SQQ parameter set, which can be stored for fast retrieval. Our anticipation has been based on the fact that such precalculated values can be leveraged in the deterministic quantization process. We have highlighted that this notice heavily relies on the fact that the values of interest are distributed according to the Laplacian pdf. Namely, we have shown that the practical constraint that rests on quantization, as a guiding principle for data compression, imposes an inevitable information loss, which with the proposed SQQ model we have made smaller compared to the asymptotically optimal SCQ. We have not only analyzed the particular use cases where formidable convenience of SQQ could be applied, but we have also addressed explicit SQQ parameterization, emphasizing the wide area of quantizers usage, as well as the possibility for its qualification for contemporary quantization solutions. Although we intended to consider the application of SQQ in NN training process, this has been left for our future research since in this paper we opted to describe one general quantizer solution applicable for processing any data modelled by the Laplacian pdf. However, from the analysis presented in this paper, we can anticipate that our results for the successful deployment of clipping in NNs are very encouraging and, for that reason, our future work will be directed toward the research on quantization benefits in compression of NN parameters.

Data Availability
No data were used to support this study.

Conflicts of Interest
e authors declare that there are no conflicts of interest regarding the publication of this paper.