Design of the Poster Image System Based on Human Vision

At present, the human visual perception system is the most effective, accurate, and fast image processing system in the world.)is is because human eyes have some special visual features, but the features closely related to image enhancement include color constancy and brightness constancy. )is paper presents a new image enhancement framework and computational model which can better simulate human visual features. It is based on the analysis of color constancy and luminance constancy and Retinex theory. And, this is a new image enhancement method in the compressed domain based on Retinex theory. In Retinex theory, DCTcoefficients consist of incident components (DC coefficients) and reflection components (AC coefficients). By adjusting the dynamic range of DC coefficients, carefully adjusting AC coefficients, and using the threshold method for block suppression, the compressed domain image can be enhanced. On the basis of Retinex theory, the incident light and reflected light components are considered synthetically, the dynamic range (DC coefficient) of the incident light component and the details of the reflected light component (AC coefficient) are adjusted, and then the incident light component is reexamined. Moreover, it achieves a better image enhancement effect and avoids the blocking effect.


Introduction
With the development of modern society, posters have gradually become an important medium of cultural communication. Poster design conveys its message through its unique expression. Posters have a natural and moving charm. e development of poster design is not stagnant but keeps pace with the times, keeping a high degree of concern and forward-looking thinking on society and things. Nowadays, the poster design industry has embarked on a broader road of exploration under the guidance of new art aesthetics and has experimental characteristics. In the future, the design forms of image media in poster design will be richer and more diverse, and the development trend will be better and better.
is may be a design experience with audience participation, or it may be "out of nothing" that advocates simplicity first. With the progress of the times, the carrier of posters is constantly changing, and the new media is constantly developing. Poster design has also moved from plane to space full of greater development possibilities, which can better communicate and interact with the audience. e study in [1] defined a human visual system model which describes the monochrome features of each spatial location image as encoding the luminance and chrominance of the corresponding three colors of each spatial location image. ere are many requirements for human perception [2], but the functions of fast analysis and selective coding of information in scenes in image compression are very suitable for human perception requirements. e basic information types of image processing are human brain perception, processing, and interpretation [3]. e study in [4] improved by building edge-enhanced images and original sketches. Spatial filters used in the human visual system seem to be a low flux of color stimulation and band pass of photometric stimulation [5]. Image processing technology focuses on two main tasks [6], such as improving image information for human interpretation and processing of image data and for storage, transmission, and representation of automatic machine perception. Disjointed information [7] is used as an indicator of image comparison and applied to perceptual image quality assessment, image registration, and video tracking. Recent developments in digital technologies utilizing large-scale image and video data such as the Internet, multimedia systems, digital television, digital cinemas, and mobile/wireless devices have increased the demand for efficient processing [8], storage, and transmission of visual information. Image processing apparatus [9] can quickly detect objects. In the image processing apparatus of [10], an input unit is used for inputting image information. e purpose of the image relocation algorithm of [11] is to adjust the image to a display screen that is different from the aspect ratio or size of the image by uneven and perceptible content image size adjustment. Algorithms are capable of computing and clustering types of photos and videos for precise and nearly repetitive media items [12]. e study in [13] relied on more and more high-resolution multispectral digital art images (mainly painting and painting) and complex computer vision methods. is paper proposes a new research method [14], which takes the physical illumination model as the core, that is, dynamically calculating HDR illumination in the virtual environment. Our method allows the reuse of existing virtual environments as input and computes HDR images in photometric units. e study in [15] had some common difficulties in many visual learning tasks. One of them is the lack of supervision information because labels can be boring, expensive, or even impossible. Another difficulty is the high dimension of visual data. However, these difficulties can be alleviated by mixing labeled and unlabeled training data.
Based on the analysis of color constancy, brightness constancy, and Retinex theory, an enhancement framework and computational model that can better simulate human visual characteristics are proposed in this paper, and it is applied to poster processing to verify its effectiveness. e threshold method is used to suppress the block effect, thus realizing image enhancement in the compressed domain. Simultaneous processing of incident and reflection components can make full use of the low-frequency and highfrequency information of the original image to enhance the details, preserve the color, and suppress the block effect.

e Primary Function of Posters.
In the basic theory of social symbol theory, symbol theory holds that all symbol systems are social, which is reflected in the social purpose of symbol systems. erefore, posters also have their social purposes. Information dissemination is one of the most obvious social functions of posters, especially in the era of underdeveloped information technology. However, great changes have taken place after the industrial revolution. As we all know, the industrial revolution has greatly improved the level of productivity development, and for the first time in human history, products exceeded demand. In this context, various disciplines are discussing the methods of influencing consumers' decision-making, which directly lead to the development of new disciplines including marketing and advertising. erefore, the social function of posters has also changed greatly, from the initial information dissemination to the persuasion of the audience. It can be said that persuading the audience to accept the information conveyed by posters is the primary role of posters in today's society. Based on the above understanding of the main social functions of posters, it is the key to explain how posters convey information and persuade audiences through the image system.

Human Visual Properties.
e human visual system is a complex concept of biology and related disciplines. People initially studied and utilized the visual characteristics of HVS to manufacture products or put forward standards. For example, in the process of transmitting TV signals, the color signal is compressed to a smaller bandwidth, which reduces the overall signal transmission bandwidth, so that color TV can be popularized under the condition of limited bandwidth at that time. Another example of applying human visual features is image compression, such as the JPEG compression standard, which takes advantage of the insensitivity of human eyes to the details of high-frequency signals. In the compression process, most high-frequency coefficients are quantized to zero, so reducing the required storage space will only bring imperceptible image quality loss. e human visual system is a comprehensive threedimensional world perception system. e human eye receives reflected light from surrounding objects through retinal cells. HVS processing starts with the eyes and is mostly processed by the brain. At present, how the brain processes this information is still unknown. e research on HVS mainly focuses on the visual characteristics of HVS, which are called human visual characteristics. rough computer science, they try to simulate these characteristics and functions of human eyes and apply human visual characteristics to information processing fields such as image enhancement.

Human Eye Structure and Visual Processing Process.
e eye is the starting position of HVS information processing. Light passes through the lens and falls on the retina behind the eyeball. e eye structure diagram is shown in Figure 1. Six muscles located on the outside of the eye are connected with the eye so that the eye can focus on any target and adjust the shape of the lens. e shape of the lens determines the focal length. By adjusting the shape of the lens, the scene captured by the eye can be focused behind the retina. e discoid vascular membrane composed of two annular muscles is called an iris, and the gap in the center of the iris is called a pupil, which is located in front of lens. e diameter of the pupil determines the amount of light entering the eye. One of the two annular muscles is used to enlarge the diameter of the pupil, and the other is used to reduce the diameter of the pupil. e retina is essentially composed of a group of receptors that can sense the light entering the eye. e structure is divided into three main layers: the receptors are located at the bottom of the retina, the central cells of the retina are located at the top of the retina, and the polar cells are located in the middle of these two layers. e light entering the eye passes through the upper two layers before reaching the receptor located behind the retina. e light-absorbing material is attached to the lower part of the receptor layer, which can effectively prevent the scattering of light received by the retina due to astigmatism. Rod cells and cone cells are two very important photoreceptor cells in the eyes. Cone cells are used to distinguish colors under strong light conditions. ey are the organs of bright vision. Rod cells are dark vision organs, which work under the condition of less light.
Rod cells are generally about 25 times more sensitive than cone cells because of their larger diameter and longer length. Generally, cones can be divided into three categories. e first category responds to the red part of the visible spectrum, the second category mainly responds to the green part, and the third category mainly responds to the blue part.
is constitutes the physiological basis of the tricolor theory of color vision. e tricolor theory explains why the three primary colors must be red, green, and blue, while other colors can be composed of these three colors. e receptor is the main organ for measuring the amount of light entering the retina. e amount of light measured by a single receptor (expressed in energy) can be calculated by integrating all wavelengths λ. It is known that three types of cones are sensitive to the red, green, and blue parts of the spectrum. Record S i (λ) as the response curve of these three receptors, where i ∈ r, g, b . E(λ) is used to denote the light that falls on a receptor and has a given wavelength λ, so that the light energy Q i measured by a single receptor can be obtained by the following formula: When observing an object with an unsmooth surface, the light ray E perceived by the receiver can be considered as the common product of the reflected light R of the object (light reflected from the object surface into the eye) and the incident light L irradiated on the object surface; that is to say, the light ray E perceived by the human eye is composed of the reflected light E and the incident light L, as shown by the following formula: (2) en, the complete calculation formula of energy Q i is as follows: e receptors transmit the measured energy information to the polar cells located in the second layer, and then the polar cells use the central retinal cells to form some chromosome binding connections to stimulate nerve endings to transmit the information to the cerebral cortex. Of the 126 million receptors in the retina, rod cells account for the vast majority of about 120 million, and the remaining 6 million are cone cells. Most pyramidal cells are located in a very small area called the fovea. is area is located in the center of the retina. e resolution of the center of the retina is much higher than that of its edge. Although there are about 126 million receptors in the retina, only about 1 million central retinal cells carry this information to the cortex through nerve endings, so this information needs to be compressed in some way to pass through the axons of the central cells. Processing begins with the retina. A human visual system called the luminance visual model has been developed to describe the process of light entering the eye and reaching the cerebral cortex through the retina, as shown in Figure 2.
In order to establish the model quickly, this paper simplifies the visual model of human eyes into a linear system. e light reflection in Figure 2 refers to the light response system of human eyes. Lamant, Rohler, and others measured the spatial frequency characteristics of the human eye system, which are also called low-pass characteristics. Spectral response refers to the spectral response system of Scientific Programming human eyes. Internal imaging f(x, y) is formed by light passing through two reaction systems: optical reaction and spectral reaction. Imaging signal f(x, y) acts on receptors in the retina, including rod cells and cone cells. Corner compression refers to the lateral inhibition effect of the visual neural network, because the number of receptors is much larger than the number of retinal central cells, so this information must be compressed before it can continue to be transmitted.
e lateral inhibition effect shows that the visual signal perceived by human eyes is formed by the weighted sum of the outputs of other adjacent receptors, not just a single receptor. e internal imaging signal f(x, y) produces signal c(x, y) through cell reaction, and then the signal o(x, y) produced by corner compression is the final visual subjective luminance signal, which is processed by the cerebral cortex and finally causes visual perception.

Dynamic Range of Vision.
e dynamic range of vision is the range of light intensity that human eyes can perceive. e whole dynamic range of human eyes ranges from 10 to 2 cm/m 2 to 106 cd/m 2 , which can be said to be relatively wide. Such a wide dynamic range is beyond the capabilities of machinery, displays, and even photosensors. Take the gray image that can be displayed by the display as an example. Generally, the dynamic range of an 8-bit gray image is only 256 gray levels from 0 to 255. e smaller the gap in each level, the smoother, more delicate, and natural the brightness change, which indicates the wider the dynamic range.

Spatial Resolution.
Spatial resolution is the ability of human eyes to observe and distinguish details in scenes. e inherent properties of human eyes determine that the resolution of human eyes depends on the physiological structure of eyes.

Image Contrast.
Contrast refers to the measurement of different brightness levels between the brightest white and the darkest black in an image, which is also called the gray contrast of an image. e larger the difference range of image gray contrast, the greater the contrast. e smaller the difference range of image gray contrast, the smaller the contrast. When the contrast reaches 120: 1, it is good, and vivid and rich colors can be easily displayed at this time. When the contrast reaches 300: 1, it can support all levels of colors.

Relative Visual Acuity Curve.
e human eye has different sensitivity to different wavelengths of light. Light with different wavelengths but the same radiation power can be seen in different brightness and colors in human eyes. erefore, the general object's perception of the brightness of the spatial and wavelength distribution I(x, y, λ) of optical radiation is where V (λ) is called the relative luminous efficiency function of the visual system. For human eyes, V (λ) is a bellshaped curve, and the relative visual acuity curves of rod cells and cone cells are different.

Contrast Sensitivity.
Under the uniform background with brightness I, there is an area with brightness ΔI higher than the background brightness; that is, the brightness of this area is I + ΔI, and the brightness difference I that human eyes can distinguish is a function of I. With the increase of background brightness I, the two-degree difference ΔI must also increase, so that human eyes can distinguish the area with brightness I + ΔI. Experiments show that brightness I can change in a considerable range, ΔI/I can still keep a fixed value, and the ratio is constant, which is called the Weber ratio. Weberby also believed that the visual characteristics of human eyes have a certain degree of logarithmic characteristics because if the brightness is logarithmic and then differentiated, the following formula can be obtained: 3.3. Retinex eory and Algorithm. Land believed that color constancy is the result of the joint action of retina and cerebral cortex. So, he put forward the Retinex theory, that is, retina + cortex. Land proposed the application of the Retinex theory in the field of digital poster images, where long-wave, medium-wave, and short-wave wavelengths correspond to three color channels of digital poster RGB images, respectively. e image to be enhanced in R, G, and B channels is considered to be composed of two parts: the reflected light component and the incident light component. Estimating incident light components by comparing pixel values between pixels is called illuminance estimation. e illuminance component is removed from the image, and the reflected light component is obtained to restore the original appearance of the object, thus realizing the enhancement of the original image. e specific process is as follows (the three channels of color image are the same, taking singlechannel processing as an example). ink of the image as the product of the incident light component and the reflected light component: where I(x, y) is the input image, L(x, y) is the reflected light component, and R(x, y) is the reflected light component. en, the incident light component of the scene is estimated by various methods to obtain the illuminance estimation L(x, y), and the reflection estimation value R(x, y) is obtained by the following formula: where θ is a positive number infinitely close to 0 to ensure that the divisor is not 0. How to estimate the incident light component of the scene to get the illuminance estimation L(x, y) is the core step of the Retinex algorithm. In the process of comparing the pixel values between pixels, path selection, that is, how to select the compared pixels, is the most important link, and it is also one of the common bases for distinguishing different Retinex algorithms. Path selection determines which pixels in the image to compare and how to compare, which directly affects the accuracy of estimating the incident light component of the scene and further determines the enhancement effect.
Provenzi's RSR method is an improvement of the Random Path Retinex method proposed by Land, which changes the path selection from a one-dimensional path to a two-dimensional region and can better reflect the influence of surrounding pixels on target pixels. MSR and MSRCR algorithms are proposed by Jobson. ey were first applied to space pictures taken by NASA and achieved a good enhancement effect, which attracted wide attention. e implementation process of MSR is as follows (take singlechannel processing as an example).
Step 1. Divide the image into two light quantities: incident light component and reflected light component, as shown in formula (6).
Step 2. Take logarithms on both sides of formula (6) and separate the incident and reflected components as follows: Step 3. Filter the original image using Gaussian smoothing filter as follows: where the symbol * is a convolution operation, a low-pass filtered image L(x, y) obtained after smoothing processing; that is, an illuminance estimate of an incident component. F(x, y) is a Gaussian filter, and the expression is as follows: Step 4. Subtract the original image from L(x, y) in the logarithmic field as follows: where R ′ (x, y) is the logarithmic domain reflected light component.
Step 5. MSR algorithm selects several standard deviations of Gaussian functions for linear weighted average: where k is the number of different standard deviations σ and generally three are taken; that is, w i , i � 1, 2, 3 are weights, σ i is three different standard deviations, and F i (i, j) is a Gaussian function with parameter σ i .
Step 6. Take an opposition number to R ′ (x, y) to obtain an enhanced image, namely, In equation (10), the standard deviation σ parameter of Gaussian function is called scale parameter, and the enhancement effect σ is determined by its size. e stronger the detail enhancement ability, the smaller the value of σ, and the worse the color fidelity ability and vice versa. MSR algorithm selects three scales from small to large for weighted fusion to achieve a better balance effect. However, the MSR algorithm does not consider the local information of the image and uses three fixed scale parameters to fuse the whole image, so it cannot enhance the local information of the image pertinently. I(i, j) cos (2i + 1)uπ 16 cos (2j + 1)vπ 16 i, j � 0, 1, · · · · · · , 7.

Image Compression in DCT
Among them, Scientific Programming After quantizing the DCT coefficients, each coefficient is divided by the corresponding quantization parameter and the result is rounded to the nearest integer. e initial value of the quantization parameter depends on the compression standard. e quantized DCT coefficients are stored in a row or column matrix in a zigzag order for average information encoding. Coefficients will be compressed to zero after quantization, leaving only a few nonzero coefficients in subblocks. When stored in zigzag order, these nonzero coefficients will be in front of the queue. To reconstruct the original image from the decompression process, the compressed image can be decoded first and multiplied by the corresponding quantization coefficients; then the quantification is inversed, and finally the image is transformed by DCT. e inverse DCT transform IDCT is defined as

YC b C r Color Space. YC b C r color space is widely used in digital poster images in which luminance information is represented by a single component Y and color information is stored by two chromatic aberration components Cb and
Cr. e difference between the blue component and the reference value is component Cb and component Cr is the difference between the red component and a reference value. YCbCr, HSI (hue, saturation, brightness) color space is more in line with the human description and interpretation of color. At the same time, to simplify the processing and reduce the amount of computation, this algorithm transforms the image to YC b C r color space for processing. Assuming that each color component of YC b C r space and RGB space images is stored in 8 bits, the relationship between the two spaces is as follows: Although the relationship between YC b C r space and RGB space is not linear, from the point of view of algorithm implementation, we can define a YC b C r space, where Cb ′ � Cb − 128, Cr ′ � Cr − 128 has a linear relationship with RGB space. After conversion to the Y, C b , C r space, the contrast of the image can be enhanced only by processing the luminance component Y of the image and the C b and C r component operations are also required to preserve the color information.
It is assumed that, in the YC b C r color space, the DCT coefficient of the luminance component Y of an image I of length and width N × N is Y � Y(i, j)|0 ≤ i, j ≤ N − 1 , k dc is an adjustment scaling factor for quantizing DC coefficients, and k ac is an adjustment scaling factor for quantizing AC coefficients. e DCT coefficients of the adjusted component Y are given by e contrast of the enhanced image is k ac /k dc times that of the original image. However, if only the parameters of the Y component are carried out, the color of the original image cannot be preserved. In RGB color space, the three components R, G, B of each pixel in the original image have the same effect on the color, and image enhancement algorithms based on RGB color space often process these three channels in the same way. is algorithm adopts a similar strategy: while adjusting the parameters of the Y component, the parameters of the Cb ′ and Cr ′ components are simpler than those of the Y component, and DC coefficients and AC coefficients are treated separately in the process. Assuming that the DCT coefficients of the Cb ′ and Cr ′ components of the image are D, respectively, and the adjusted scaling factors of the Y component are k, the adjusted DCT coefficients U and V of the Cb ′ and Cr ′ components are e function of U and V is to calculate DCT coefficients, realize compression function in calculating image processing, reduce data amount, and improve transmission efficiency.

Experimental Flow.
e Retinex theory defines image I(x, y) as the product of incident light component L(x, y) and reflected light component R(x, y) of the scene: namely, en, various methods are used to estimate the incident light component of the scene to obtain the illumination estimation L(x, y), the illumination estimation value is removed by the following formula, and then the reflection estimation R(x, y) is obtained as the final output result: 6 Scientific Programming

I(x, y) L(x, y) . (21)
To improve the overall contrast of the poster image, the dynamic range adjustment of the incident light component must be carried out to obtain L out (x, y): (L(x, y)). (

22)
To increase the clarity of the poster image, the reflected light component must be enhanced in detail, and the following results are obtained R out (x, y): (R(x, y)). (23) Finally, the output image O(x, y) is obtained by multiplying L out (x, y) and R out (x, y), and the poster image is enhanced: (24)

Comparison with Traditional Image Enhancement Algorithms in Compressed Domain.
Comparing the algorithm with the traditional compression domain image enhancement algorithm, Jinshan Tang proposed a compression domain image enhancement algorithm with DCT domain frequency band adjustment as the core. e algorithm transforms the image into the compressed domain and uses the same coefficients for each frequency band to adjust. e processing results are compared as shown in  e quadratic function τ (x) has no parameters and is easy to realize. Experiments on different images show that τ (x) has a good enhancement effect on dark areas with low contrast. e parametric function η (x) has an adjustable parameter c near 1, which has a good dynamic range compression effect for both dark and overbright areas. e mean value, standard deviation, and entropy of each processing result in Figures 3-5 are given below, and the R channel of each image is selected for comparative explanation in Tables 1-3.
All the algorithms involved in the comparison are image enhancement algorithms in the compressed domain. It is more targeted to choose the JPEG image quality standard proposed by Wang as the evaluation standard. As shown in Table 4, the higher the value, the better the image quality.
As can be seen from Figure 6, compared with the traditional compressed domain image enhancement algorithm, the image quality enhanced by this algorithm has been significantly improved.
e results show that the image quality of this algorithm is better than that of the traditional algorithm, which is more suitable for the visual characteristics of human eyes and accords with the subjective evaluation of human eyes.

Comparison with the Traditional Retinex Algorithm.
e processing effect of this algorithm is compared with the processing effect of MSR, as shown in Figures 7-9.
In the above experimental results, (a) of each image is the original image, (b) is the processing result of MSR, (c) is the processing result of mapping function τ(x) in the dynamic range adjustment process of illumination component, and (d) is the processing result of mapping function η(x) in this algorithm. As can be seen from Figures 7-9, the color of MSR processing results is darker, and the brightness and color of the processing results are considered in Y, C b , C r color space. e processing effect is better than MSR, and the color is more natural. e details are also richer.
Objective evaluation of image quality, such as average value, standard deviation, and entropy of processing results, is shown in Tables 5-7. e mean value reflects the average brightness of the image, the standard deviation reflects the contrast of the image, and the entropy reflects the amount of information contained in the image.
It can be seen from Tables 5-7 that the brightness, contrast, and information content of the images processed by this algorithm are relatively good. e following image quality evaluation method based on statistics proposed by Jobson is used to further compare the processing results. e statistical results show that when the average gray level is between 90 and 190 and the standard variance is between 30 and 85, the image quality is better, as shown in Figure 10.
e specific method of statistics is to divide the image into nonoverlapping subblocks with the same size (generally 50 × 50 or 60 × 60); then calculate the standard deviation of each subblock, respectively, and then average the standard deviation of the subblocks; and multiply the average value of standard deviation σ with the gray value I of the image to get the evaluation result. e larger the result value, the better the image quality. e statistical results of the images in this chapter are shown in Table 8 and Figure 11.
As can be seen from Table 8, the quality evaluation results of the images processed by this algorithm are generally higher than those of the traditional MSR algorithm. e original image I * σ value is relatively small, and the brightness and contrast of the image are not good. e quality evaluation data processed by this algorithm is improved compared with that processed by the traditional MSR algorithm.

Comparison with HE and MSR Algorithms.
e poster image enhanced by this algorithm is compared with HE and MSR algorithms, and the results are as follows: Compared with the existing HE and MSR methods, the proposed algorithm has clearer processing results. e calculation of the incident component is more accurate than the traditional MSR algorithm, and the constancy of brightness and color is considered. e combination of the two methods further shows that the adaptive filter selection strategy and weighted fusion computing framework adopted in this algorithm are superior to the cumulative distribution equalization of the existing HE algorithm and the calibration filter of the MSR algorithm.
At the same time, it can be seen from Figures 12-14   the dynamic range of the gray scale, but the gray scale distribution of the results processed by this algorithm is concentrated between 120 and 200, which is more in line with the average range of gray scale required by the best human vision. e gray mean and variance of the statistical image are used to verify the gray distribution, as shown in Tables 9 and 10.
As shown in Table 9 and Table 10, the average gray level of the algorithm processing results in this chapter is generally higher than that of the original image, HE, and MSR processing results, and most of them are concentrated between 140 and 180 gray levels, which is more in line with the observation range of human eyes. However, the standard deviation of the gray level of the algorithm processing results     in this chapter is reduced, which shows that the brightness difference between pixels is reduced, and the gray level is more concentrated near the median value, which is more suitable for human vision to observe. Experimental results show that the contrast and texture of low illumination images with different degradation degrees can be effectively enhanced, and better visual effects can be achieved in brightness, color, and detail. Finally, the output image is obtained by multiplying the incident component and the reflection component, which enhances the contrast while keeping the color information of the original image, and makes the enhanced image more natural in color. In the future, the quality evaluation of the processed image is often a relatively subjective view in enhancement processing. According to different application requirements, the evaluation criteria may vary greatly. Most of the existing image evaluation criteria are calculated according to the gray level of pixels, which to some extent ignores that pixels with the same gray level may have different meanings in the image. In addition, color information often reflects the quality of an      image better than brightness information. erefore, how to put forward a quality evaluation standard with more visual significance and taking into account color information will become an important part of the follow-up research.

Conclusion
Aiming at the problem that the existing image enhancement algorithms are not ideal in the spatial domain, the image contrast is enhanced in the compressed domain, but the details and color information of the image cannot be preserved. An image enhancement algorithm in DCT compressed domain based on retina theory is proposed. Different from the traditional Retinex algorithm, which discards the incident light component and only enhances the reflection component, this algorithm processes the incident and reflection components at the same time, fully considers the image details, and defines a new spectral content ratio. By manipulating the spectral content ratio to process the reflection component, the details of the image can be effectively enhanced. Finally, the output image is obtained by multiplying the incident component and the reflection component, which enhances the contrast while keeping the color information of the original image, and makes the enhanced image more natural in color. In this algorithm, the threshold method is used to suppress the block effect, and the information of adjacent subblocks can be used in the subblock processing in the DCT domain, so that halo artifacts can no longer be produced around the boundary with obvious light and shade jump. Experimental results show that the algorithm can effectively enhance the contrast and texture of low illumination images with different degradation degrees and achieve good visual effects in brightness, color, and detail. With compressed images, especially JPEG images, becoming the popular data transmission standard in the network, a new image enhancement algorithm based on retina theory is proposed in the compressed domain to make full use of the incident light component information of images. Based on the Retinex theory, DCT coefficients are divided into an incident component (DC coefficient) and a reflection component (AC coefficient). By adjusting the dynamic range of DC coefficients, carefully adjusting AC coefficients, and using the threshold method for block suppression, the compressed domain image can be enhanced. To enhance the compressed image better, we must make full use of the information contained in the DC coefficient.   Data Availability e experimental data used to support the findings of this study are available from the author upon request.

Conflicts of Interest
e author declares that no conflicts of interest regarding this work.