One of the main objectives of watermarking is to achieve a better tradeoff between robustness and high visual quality of a host image. In recent years, there has been a significant development in gray-level image watermarking using fractal-based method. This paper presents a human visual system (HVS) based fractal watermarking method for color images. In the proposed method, a color pixel is considered as a 3-D vector in RGB space. And a general form of 3 × 3 matrix is utilized as the scaling operator. Meanwhile, the luminance offset vector is substituted by the range block mean vector. Then an orthogonalization fractal color coding method is achieved to obtain very high image quality. We also show that the orthogonalization fractal color decoding is a mean vector-invariant iteration. So, the range block mean vector is a good place for hiding watermark. Furthermore, for consistency with the characteristics of the HVS, we carry out the embedding process in the CIELa*b* space and incorporate a just noticeable difference (JND) profile to ensure the watermark invisibility. Experimental results show that the proposed method has good robustness against various typical attacks, at the same time, with an imperceptible change in image quality.
1. Introduction
With the rapid development of internet, digital multimedia data such as image, audio, and video are readily reproduced and distributed with ease. Many watermarking methods have been exploited for copyright protection due to the properties of digital media and the popularity of the internet. Typically, a watermark can be a random signal, a meaningful message, or a logo. It is viewed as an effective way to prevent user’s content from illegal distribution [1]. Figure 1 shows a general scheme for digital watermarking. Firstly, an encrypted watermark is inserted into an original image by using an embedding algorithm. Then, the watermarked image will pass through the transmission channel, which may include some attacks, such as lossy compression, low-pass filtering, noising, and geometric distortions. When the watermark is to be extracted, the corresponding extraction algorithm will be implemented, usually with aid of a secret key and the original image. Only the owner of the image knows the key and no other person can identify the watermark without the knowledge of the key. A good watermarking scheme should possess many important properties, for example, invisibility, robustness to various types of attacks, and accurate detection.
Generic watermark insertion and extraction procedure.
In general, digital image watermarking algorithms addressing a wide variety of applications can be classified into two categories depending on the domain in which the watermark is embedded. The first group is mainly implemented in spatial domain [3, 4] while the second group of algorithms is achieved in transform domain, such as discrete cosine transform (DCT) [5, 6], discrete Fourier transform (DFT) [7] and discrete wavelet transform (DWT) [8, 9]. The watermark in spatial domain methods is inserted into the image by directly modifying a preselected set of pixels in the image. On the other hand, in transform domain watermarking methods, the watermark is embedded into the image by modifying the transformed frequency coefficients. The advantage of watermarking in the frequency domain is that the characteristics of the human visual system (HVS) are better captured by the spectral coefficients, since the HVS is more sensitive to low-frequency coefficients and less sensitive to high-frequency coefficients [10, 11]. When focusing on color image watermarking, many methods are realized by embedding watermarks in the image luminance or by processing each color channel separately. By observing the fact that human eyes are insensitive to changes in the blue component, Kutter et al. [12] proposed a method for embedding a watermark in the blue component. Barni et al. [13] introduced a color image watermarking scheme based on the cross-correlation of RGB channels. In Barni’s scheme, a watermark is embedded within the host image by modifying a subset of full frame DCT coefficients of each color channel. And the existence of the watermark is verified by calculating the correlation of RGB channels. Tsai et al. [14] provided a scheme of embedding a watermark on a quantized color image. Tsui et al. [15] introduced a solution of embedding a watermark in the frequency domain of the chromatic channels by using the spatiochromatic discrete fourier transform (SCDFT). In Tsui’s method, the chromatic content of a color image is encoded into CIEa*b* chromaticity coordinates while the achromatic content is encoded as CIEL tristimulus value. Kuo and Cheng [16] presented a color watermarking method that combines color edge detection and color quantization using principal axes analysis in 3-D color space. Chou and Liu [17] introduced a color image watermarking scheme that hides watermark signals in most distortion-tolerable signals within three color channels of the host image without resulting in perceivable distortion in wavelet domain. Also, Vahedi et al. [18] proposed a watermarking scheme of color images based on wavelet transformation. In Vahedi’s method, the strength of the embedded watermark is controlled locally with the aid of the visual characteristics of the host image and so forth.
The idea of fractal image coding (FIC) was originally introduced by Barnsley and Demko [19]. And in 1992, Jacquin [20] achieved the first practical fractal image coding method. In the past decades, fractal image coding has mainly been developed for the purpose of image compression [21–33] as well as some for image denoising [34–37] and image indexing [38]. Recently, it is also well studied for gray-level image watermarking [39–43]. Davern and Scott [39] described a steganographic method for inserting secret information into images using fractal image coding. The method allowed a user to specify a visual key when hiding the secret information. The visual key must then be used when retrieving the hidden data. Čandik et al. [40] investigated two possible approaches of embedding digital watermarks into fractal codes of images, which are embedding digital watermarks into parameters for position of similar blocks and coefficients of block similarity. Kiani and Moghaddam [41] presented a watermarking method to embed a binary watermark into an image using a special type of fractal coding that its parameters are contrast scaling and the mean of range block. And it utilized the fuzzy C-mean clustering to address the watermark bits. Pi et al. [42] studied a novel watermarking method utilizing a special type of orthogonalization fractal coding. In the method, a permutated pseudorandom binary sequence used as watermark was embedded into the range block means. And the detecting procedure was carried out by computing the correlation coefficient between the original and the extracted watermark. Shahraeini and Yaghoobi [43] proposed a blind watermarking algorithm based on fractal model in discrete wavelet domain and so forth.
However, few known papers considered fractal coding for watermarking of color images. In this paper, motivated by [42], we advance a HVS-based fractal color image watermarking algorithm. Meanwhile, an analogous RGB version of orthogonalization fractal coding is introduced. Just like other watermarking methods, good robustness and high image quality are our goals. In our proposed method, based on the observation that the r, g, and b components for a natural image are correlated, we consider a pixel as a 3-D vector in RGB space. Then, instead of using three independent linear predictions for the pixel intensity of the three planes (namely, r, g, and b planes) in the classic fractal color coding [2], we adopt a general form by using a 3 × 3 matrix as the scaling factor. In addition, the range block mean vector substitutes the luminance offset vector. In terms of the quality of the fractal representation measured by PSNR, the proposed fractal coding significantly performs better than the traditional fractal color image coding. We also show that orthogonalization fractal color decoding is a mean-invariant iteration. So, we use the range block mean vectors to hide the watermark. On the other hand, in order to accord with the characteristics of the HVS, we convert the range mean vectors from RGB space to CIELa*b* space before the inserting process (with the aid of the JND model [44]) is carried out. The experimental results show that the proposed method is very robust against different attack, as well as with retaining very high image quality.
2. Foundations of Fractal Image Coding
The theory of fractal-based image coding using iterated function system (IFS) and collage theorem was proposed by Barnsley and Demko [19]. Let (X,d) be a complete metric space. An iterated function system (IFS) is a finite set of contraction mappings wi:X→X, where i=1,2,…,N with contraction factor si on (X,d). In general, the system is denoted by {X;w1,w2,…,wN} and the contraction factor is s=max{s1,s2,…,sN}. Furthermore, we define W:ℋ(X)→ℋ(X) as W(B)=⋃i=1Nwi(B), ∀B∈ℋ(X). Then, the transformation W is a contraction mapping on the complete metric space (ℋ(X),h), where h is the Hausdorff metric. So, according to the contractive mapping fixed-point theorem, there exists a unique fixed point A∈ℋ(X) satisfying
(1)A~=W(A~)=⋃i=1Nwi(A~)
which is known as an attractor of the IFS. The fixed point can be obtained by following iteration
(2)A~=limn→∞W(n)(B),∀B∈ℋ(X),
where W(n)(B)=W(W(n-1)(B)) and W(0)(B)=B. Let ϵ≥0 be a small given real number. With the above hypothesis, collage theorem says that, suppose L∈ℋ(X) satisfying h(L,⋃i=1Nwi(L))≤ϵ; then,
(3)h(L,A~)≤ϵ1-s.
The collage theorem implies that if a set can approximate to contracted copies of itself as well as possible by a set of contractive mappings, then the corresponding attractor is close enough to the set. Thus, the IFS code for the set can be stored since the approximator can be obtained by iterating the IFS on any initial set.
However, in practice it is unlikely to find “collages” for most natural images by reasonably simple transforms. The well-known practical method for fractal image coding, namely partitioned iterated function system (PIFS), was proposed by Jacquin [20]. Let I be a given gray-level image, and hence I is a subset in three-dimensional Euclidian space that is, I∈ℝ3. The image I is first segmented into n nonoverlapping range blocks of size, say b×b, and denoted by ℛ={Ri}i=1n. Note that I=⋃i=1nRi and Ri∩Rj=Ø for any i≠j. Let 𝒟={Dj}j=1m be the pool of all domain blocks which are of size 2b×2b and all are extracted from I that is, Dj⊂I for j=1,…,m. Let Φj={ϕ:ϕ(Dj)→ℝ3}, where ϕ is usually an affine contractive transformation. For each range block Ri, let d be distance measure. Our goal is to find a domain block Dk as well as an appropriate transformation ϕi|k such that
(4)d(Ri,ϕi∣k(Dk))=minj{d(Ri,ϕi∣j(Dj))},
where ϕi∣j∈Φj and ϕi∣j satisfies
(5)d(Ri,ϕi∣j(Dj))≤d(Ri,ϕ(Dj)),∀ϕ∈Φj.
When all the range blocks are processed, the set of maps Φ={ϕi∣·}i=1n thus is obtained and called the PIFS code of the image I.
In classic fractal coding, the distance d always uses l2-norm, ∥·∥2, as the so-called collage distance measure. The affine contractive map ϕ consists of three parts defined as
(6)ϕ=φ∘τ∘ω.
Here, ω is a uniform scaling from domain block to range block dimension; τ is one of the eight isometries that map a block into itself (identity, 90°, 180°, and 270° rotations around the center and the four reflections over the symmetry axes); and φ is a linear transform in the gray level, defined as
(7)φ(x)=αx+β,
where α and β are known as the contrast and luminance offset parameters, respectively, and α is usually restricted to satisfy |α|<1 for contractivity. (Some authors, e.g., [21], have restricted |α|<2 with no noticeable effects on the contractivity.)
3. Fractal Color Image Coding
The traditional fractal color image coding method in RGB space discussed in [2] for the purpose of compression was performed by using three gray-level linear predictions to encode the r, g, and b image components independently. It was defined as
(8)φ(x)=diag([α1α2α3])x+b,
where x=[x1x2x3]T represents (r,g,b) values of a pixel and b=[b1b2b3]T denotes the corresponding offset values. However, for most natural images, the components r, g and b are always strongly correlate. For instance, when testing Lena image, the correlations between r and g, r and b, g and b are 0.8786, 0.6764 and 0.9106, respectively. The correlation of r, g and b is 0.9881, which is estimated by [45]
(9)ρ=1-|Σ|∏i|Σii|,
where Σ is the covariance matrix of vectors r, g and b. Likewise for Barbara image, the correlations between r and g, r and b, g and b are 0.8860, 0.8061 and 0.9582, respectively; the correlation of r, g and b is 0.9921. Based on this observation, in this work, with above notations, we consider the r, g, b components as a whole and adopt a more general functional form:
(10)φ(x)=Ax+b,
where the linear operator A is defined by a 3 × 3 matrix:
(11)A=[a11a12a13a21a22a23a31a32a33].
In general, the distance between the pixels x and y is defined by using l2-norm (Euclidean distance), that is,
(12)d2(x,y)=∥x-y∥2=∑i=13(xi-yi)2.
Let xk≐[x1kx2kx3k]T and yk≐[y1ky2ky3k]T, k=1,…,K, where K=#Ri denoting the pixel number of Ri, be the (r, g, and b) values on Ri and transformed Dj, respectively. Then the distance of {xk} and {yk} can be determined by
(13)Δi,j2=∑k=1Kd22(Axk+b,yk)Δi,j2=∑k=1K∥Axk+b-yk∥22Δi,j2=∑k=1K∑u=13(∑v=13au,vxv,k+bu-yu,k)2.
The purpose of using squared l2norm in function (13) is to facilitate the estimation of the matrix A. According to the matrix theory, formula (13) can be represented using the well known (squared) Frobenius norm; that is,
(14)Δi,j2=∥AX+bI-Y∥F2,
where bI≐[b1Ib2Ib3I]T and I=[11⋯1]T being a K×1 vector. X≐[X1X2X3]T, Xv=[xv1xv2⋯xvK]T, and v=1,2,3 denote the respective r, g, b values of the pixels in the domain block Dj. Likewise, Y≐[Y1Y2Y3]T, Yv=[yv1yv2⋯yvK]T, and v=1,2,3 denote the respective r, g, and b values of the pixels of Ri.
In order to minimize the residual Δi,j2, we rewrite (13) as
(15)Δi,j2=∑u=13∥∑v=13au,vXvT+buI-YuT∥22
and differentiate it with respect to bu, au,v and set them to zero; that is, ∂Δi,j2/∂bu=0 and ∂Δi,j2/∂au,v=0 yielding
(16)bu=1K∑k=1Kyu,k-au,11K∑k=1Kx1,kbu=-au,21K∑k=1Kx2,k-au,31K∑k=1Kx3,kbu≐Y¯u-au,1X¯1-au,2X¯2-au,3X¯3u=1,2,3,(17)au,1∑k=1Kx1,kxv,k+au,2∑k=1Kx2,kxv,k+au,3∑k=1Kx3,kxv,k+bu∑k=1Kxv,k=∑k=1Kyu,kxv,k,
that is,
(18)au,1X1TXv+au,2X2TXv+au,3X3TXv+buKX¯v=YuTXv,000000000000000000000000000000000u,v=1,2,3,
where Y¯u=(1/K)∑k=1Kyu,k and X¯v=(1/K)∑k=1KXu,k, v=1,2,3. Inserting (16) into (18), we have
(19)[au,1au,2au,3]·[X1TXv-KX¯1X¯vX2TXv-KX¯2X¯vX3TXv-KX¯3X¯v]=YuTXv-KY¯uX¯vhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhhu,v=1,2,3.
For a test image, let the r, g, b values Xu={xu,k}k=1K and Yu={yu,k}k=1K, u=1,2,3 be random samples of the random variables Xu and Yu representing the pixel value distribution of the parent subblock and the child subblock, respectively. Then, the formula (19) can be comprehensively written as follows:
(20)AC=M,
where C is a covariance matrix
(21)C=[D(X1)Cov(X1,X2)Cov(X1,X3)Cov(X1,X2)D(X2)Cov(X2,X3)Cov(X1,X3)Cov(X2,X3)D(X3)],M=[Cov(Y1,X1)Cov(Y1,X2)Cov(Y1,X3)Cov(Y2,X1)Cov(Y2,X2)Cov(Y2,X3)Cov(Y3,X1)Cov(Y3,X2)Cov(Y3,X3)].
As a result, if C is an invertible matrix, which can be easily satisfied by randomly adding extremely small positive real numbers to each element of C, then
(22)A=M·C-1.
Moreover, summarizing (16) gives
(23)b=Y¯-AX¯,
where Y¯=[Y1¯Y2¯Y3¯]T and X¯=[X1¯X2¯X3¯]T.
According to the Banach’s fixed point theorem, in order to guarantee the contractivity of the fractal transformation with respect to l2norm, A should satisfy the restriction ∥A∥2<1, which implies that the maximum eigenvalue λ of ATA must be smaller than one. There is no simple relationship between the coefficients of A and λ. However, in the l∞ norm, contractivity is guaranteed if all A satisfy the condition ∑v=13|au,v|<1 for each u∈{1,2,3}. The resulting fractal transforms satisfying this condition are always contractive in l∞, hence in l2, due to the equivalence of the norms in finite pixel space. For this reason, we will “clamp” the coefficients of A by checking each row of A if it satisfies ∑v=13|au,v|<1. If ∑v=13|au,v|≥1 for some u∈{1,2,3}, we re-encoded the u-component Yu of the child subblock Ri by “discarding” one or two unimportant coefficient(s) in {au,v}v=13. Firstly, the correlation coefficients between Yu and X1, X2, X3 of the corresponding parent subblock Dj are calculated and denoted by ρ1, ρ2, ρ3, respectively. Secondly, the magnitudes of the three correlations are sorted in ascending order, for example, |ρv1|≥|ρv2|≥|ρv3|, which means that Xv1 has maximum correlativity. Finally, we “clamp” the coefficients au,v,v=1,2,3 by following way:
If ρv2 has the same sign as ρv1, then set au,v3=0 and
(24)[au,v1au,v2]T=[Cov(Yu,Xv1)Cov(Yu,Xv2)]T[D(Xv1)Cov(Xv1,Xv2)Cov(Xv1,Xv2)D(Xv2)]-1.
If |av1|+|av2|≥1, goto (iv).
If ρv2 and ρv1 have different sign whereas ρv3 and ρv1 take the same sign, then set au,v2=0 and av1, av3 are computed with the similar equation as (24) by replacing Xv2 with Xv3. If |av1|+|av3|≥1, goto (iv).
If both ρv2 and ρv3 have different sign with ρv1, then goto (iv).
Set au,v2=au,v3=0 and au,v1=Cov(Yu,Xv1)/D(Xv1). If |au,v1|>1, then au,v1 is “clamped” to au,v1=sign(au,v1)*0.99.
When the “clamped” A is determined, b can be calculated naturally by (23). Accordingly, the (nonorthogonalization) fractal decoding algorithm with A and b is given by
(25)R(n)=Aω(D(n-1))+bI.
If A and R¯ are taken as fractal parameters, the orthogonalization decoding algorithm is
(26)R(n)=Aω(D(n-1))+bIR(n)=R¯I+Aω(D(n-1)-D(n-1)¯I),
where R¯ denotes the mean of r,g, and b components of the range block R; that is, R¯≐[R¯1R¯2R¯3]T; D(n-1) is the matched domain block from the (n-1)th iteration and D(n-1)¯≐[D1(n-1)¯D2(n-1)-D3(n-1)¯]T. From (26), it can be readily verified that R(n) is a mean vector-invariant iteration; that is,
(27)R(n)¯=R(n-1)¯=⋯=R¯.
In Figure 2, we test the proposed method on the 512 × 512 color Lena (Figure 2(a)), Barbara (Figure 2(d)), and Peppers (Figure 2(g)) images using full search scheme. We set the size of range blocks to be 4 × 4 and the search step size to be 2. We can see that, in terms of PSNR qualitative measures, the image quality in Figures 2(b), 2(e), and 2(h) with PSNRs 34.71 dB, 30.37 dB, and 32.01 dB encoded by the method proposed here is significantly better than the quality of Figures 2(c), 2(f), and 2(i) encoded by the traditional fractal image coding in [2], around 2.0 dB.
Comparisons of the two fractal color image coding methods: (a), (d) and (g) are original images; (b), (e) and (h) are the fractal representations of our proposed methods; (c), (f) and (i) are the fractal representations of the classic method [2].
Lena
PSNR 34.71 dB
PSNR 32.54 dB
Barbara
PSNR 30.37 dB
PSNR 28.57 dB
Peppers
PSNR 32.01 dB
PSNR 30.35 dB
4. Fractal Color Watermarking
In general, there are two types of techniques for embedding a predefined watermark into an image. One is pseudorandom sequence used for objective detection; the other is binary image or gray image used for subjective detection [1]. In this paper, in order to subjectively verify the ownership of an image with the aid of extracting a watermark, a binary logo image is used. And the watermark bit is either 1 or 0. One of the main challenges of the watermarking is to achieve a tradeoff between robustness and perceptivity. In general, increasing the strength of the embedded watermark can achieve robustness, but it would lead to an increase in the visible distortion as well, and vice versa. Since the orthogonalized fractal decoding is a mean-invariant iteration, the range block mean is a good robust place to hide a watermark, as discussed in [42]. After fractal decoding, the embedded watermark diffuses throughout the reconstructed image.
In order to gain high robustness as well as low sensitivity in color image watermarking, the knowledge of human visual perception of color stimuli must be well utilized in designing embedding algorithms. It is well known that the human visual system (HVS) is not perfect sensor of perceiving color visual information. Although the HVS is particularly sensitive to changes in image hue, it is less sensitive to the yellow-blue component. CIELa*b* is the most complete color model conventionally used to describe all of the colors that are visible to the human eye [10]. The parameter L represents the lightness of the color, whereas a*, b* are chromatic information. And a* is always named magenta-green axis and b* is named yellow-blue axis. Based on the observations, in the method here, we will convert the r, g, and b values of the range block means {R¯u}u=1n to La*b* space before they are inserted watermark {wu}u=1n. The transformation from RGB space to the CIELa*b* space is through the XYZ space. The RGB space is first converted to the XYZ space through a linear transformation
(28)[XYZ]=[0.4900.3100.2000.1770.8130.0110.0000.0100.990][RGB]
and the XYZ space is converted to the CIE-La*b* space through a nonlinear transformation
(29)L=166f(YY0)-16,a*=500[f(XX0)-f(YY0)],b*=200[f(YY0)-f(ZZ0)],
where
(30)f(t)={t1/3,ift>0.0088567.787x+16116,otherwise
and X0, Y0, Z0 represent the reference white. Conversely, the reverse transformation is easily expressed by using the inverse of the function f above:
(31)Y=Y0f-1((L+16)116),X=X0f-1((L+16)116+a*500),Z=Z0f-1((L+16)116-b*200),
where
(32)f-1(t)={t3,ift>0.20690.1284(t-429),otherwise.
The color difference ΔE at each pixel is defined as
(33)ΔE=(ΔL)2+(Δa*)2+(Δb*)2.
Under a specific viewing condition, there exits a certain amount of information that is not perceivable by the human eyes, or the so-called perceptual redundancy. Experiments have shown that ΔE<1 is not detectable by the HVS, and ΔE<3 is not apparent [46]. ΔE=3 is called the uniform just-noticeable color difference (UJNCD). Since the sensitivity of the HVS to the yellow-blue component is approximately 1/5 compared to the luminance component [10], we can set ΔL=Δa*=1/5Δb*. So, if we set Δb*=1.5, then ΔL=Δa*=0.3 and ΔE≈1.56≪3. Assume a randomly scrambled watermark to be {wu}u=1n the watermark insertion procedure is illustrated in Figure 3 and includes the following steps.
Fractally encode the original image (using full search scheme) to produce the fractal codes {R¯u,Au,x→u}u=1n in RGB space, where x→u≐(xu,yu) is the position of the optimal domain block.
Convert the range block means {R¯u}u=1n from RGB space to La*b* space and denote them by {R¯u′}u=1n.
Embed the permutated watermark {wu}u=1n into {R¯u′}u=1n.
If the message to be embedded is wu=1, replace R¯u′ with R¯u*≐R¯u′+ζ, where ζ=[0.3,0.3,1.5]T.
Otherwise, if wu=0, replace R¯u′ with R¯u*≐R¯u′-ζ.
Convert {R¯u*}u=1n back to RGB space and denote them by {R~u*}u=1n.
Hide the watermark by performing fractal decoding using {R~u,Au,x→u}u=1n.
Watermark insertion procedure using our fractal color image coding.
To extract the embedded watermark, the original image and watermarked (possibly attacked) image are needed. R¯u′+ζ and R¯u′-ζ are calculated from the original image in La*b* space, followed by a distance comparison to the R~u′ from the watermarked image. If the extracted R~u′ is closer to R¯u′+ζ, we regard “1” as embedded; otherwise, “0” was inserted. The flow diagram is illustrated in Figure 4. The watermark extraction is very quick since it does not need fractal encoding and decoding.
Watermark extraction procedure.
5. Experimental Results
In order to verify that the proposed algorithm described before indeed increases the robustness of the watermarked images against attacks, a series of experiments has been conducted by using the attacks to the 512 × 512 color Lena, Peppers, and Barbara images, as shown in Figures 2(a), 2(d), and 2(g). The size of the color range block is 4 × 4, so the original image can be partitioned into 128 × 128 (i.e., 16384) color range blocks. For convenience, the size of binary logo image to be embedded is 128 × 128, and the logo image as shown in Figure 5(a), a lovely panda, is encrypted by using the block shifting method [47]. That is, we first divide the logo image into 32 × 32 small blocks, each with 4 × 4 pixels. Successively, the positions of the small blocks are randomly shifted by a random matrix, such as
(34)M=[30201457⋯131⋯]
which is generated by randomly scrambling the integers 1,2,…,32. As an example used here, the resulting image is shown in Figure 5(b). If people want to recover the original information, the same random matrix should be used,; otherwise people could not obtain the correct information.
The attacked image quality is measured by peak signal-to-noise ratio (PSNR) defined as
(35)PSNR=10log102552MSE,
where
(36)MSE=13fH×fW∑i=1fH∑j=1fW∥f(i,j)-f~(i,j)∥22.
Here, fH and fW are the height and width of the image, respectively. f(i,j) and f~(i,j) are the values located at coordinates (i,j) of the original image and the attacked image, respectively.
When extracting the watermark, the normalized correlation coefficient (NC) is calculated using the original watermark and the extracted watermark to judge the existence of the watermark and to measure the correctness of the extracted watermark. It is defined as
(37)NC=1wh×ww∑i=1wh∑j=1www(i,j)w~(i,j),
where wh and ww are the height and width of the watermark, respectively. w(i,j) and w~(i,j) are the watermark bits located at (i,j) of the original watermark and the extracted watermark. And w(i,j) is set to 1 if it is watermark bit 1; otherwise, it is set to −1. Likewise for w~(i,j). The PSNRs of the three watermarked images are 33.30 dB, 29.05 dB, and 30.59 dB, respectively. For brevity, only the Lena image is shown. Figure 6(a) shows that in the unattacked watermarked image, no visual difference exists at all compared with Figure 2(b). The extracted logo image shown in Figure 6(b) with NC=1.0 implies that the extracted watermark is exactly the same as the original watermark.
Now, various classic image-processing operations, including JPEG compression, median filtering, mean filtering, sharping, blurring, noising, cropping, and scaling, are simulated to investigate the robustness of the proposed watermarking method. JPEG is one of the most frequently used image compression formats. The JPEG quality factor is a number between 0 and 100. When decreasing the quality factor, the image compression ratio increases, whereas the quality of the resulting image is significantly reduced. Table 1 lists the results of applying various JPEG quality factors to the test images. The proposed method can detect the existence of a watermark even for quality factor equal to 15. And the values of NC still exceed 0.49 for the three images. For other nongeometric attacks, such as median filters, mean filters, sharpening, and blurring, the resulting images are blurred or sharpened at the edge and so forth. The extracted watermarks are illustrated in Table 2. Here, both the median and mean filters are applied with masks of size 3 × 3, 5 × 5 and 7 × 7, respectively. As we can see, the proposed method is very robust against these non-geometric attacks and most of the extracted logo images are only slightly contaminated.
Extracted watermarks (EW) and NCs after attacked by JPEG compression with the quality factors (QF) 15, 25, 35, 45, 55, 65, 75, and 85 on Lena, Barbara, and Peppers.
QF
15
25
35
45
55
65
75
85
Lena
EW
NC
0.6837
0.7656
0.8257
0.8969
0.9957
0.9995
1.0000
1.0000
Barbara
EW
NC
0.6320
0.6792
0.7510
0.8060
0.9529
0.9653
0.9882
0.9956
Peppers
EW
NC
0.4950
0.5752
0.6438
0.7336
0.9070
0.9241
0.9442
0.9529
Extracted watermarks (EW) and NCs after attacked by median filter (3×3, 5×5, 7×7), mean filter (3×3, 5×5, 7×7), sharpening, and blurring of the three images: Lena, Barbara, and Peppers.
Median filter
Mean filter
Sharpening
Blurring
3×3
5×5
7×7
3×3
5×5
7×7
Lena
EW
NC
0.9937
0.9421
0.8593
0.9764
0.8832
0.7648
0.9872
0.9847
Barbara
EW
NC
0.9966
0.9502
0.8641
0.9796
0.8837
0.7561
0.9957
0.9900
Peppers
EW
NC
0.9780
0.9288
0.8456
0.9437
0.8392
0.7111
0.9675
0.9629
The results of noising and some common geometric attacks, for example, rotation, scaling and, cropping, are illustrated in Table 3. The Gaussian noise variation is varied from 5 to 15 and the step size is 5. As shown, the proposed method can effectively resist the attacks and all of the extracted watermarks could be clearly recognized. The attack of scaling is performed by downsampling a watermarked color image by factor 2 in both row and column directions. Since the image is shrunk from size 512 × 512 to 256 × 256, all range blocks are then contracted from 4 × 4 to 2 × 2 and {R~u}u=1n should be calculated in a 2 × 2 block. For the clopping attack in our test, three types, namely, Type I, Type II, and Type III, are considered. In Type I, the left-top of an image of quarter size is cut away (i.e., 75% is remained), as shown in Figure 7(a); in Type II, the positions of the left-top and the right-bottom of the remained subimage are (31,471) and (81,461), respectively. In other words, the remaining region is about 64%, as shown in Figure 7(b); in Type III, two rectangular regions are removed and the remaining part is about 79%, as shown in Figure 7(c). In practical use, the position of the clipped block is usually needed to be located, especially for the case of Type II. Similar to the method in [42], it can be easily achieved by moving the clipped block point by point from the top-left of watermarked image and calculating the correlation coefficient. Clearly, the position of the clipped block should be located at the position where the correlation coefficient is maximum (theoretically, it equals 1 if without any other attacks). Then, in order to facilitate obtaining the subwatermark of the clipped block correctly, the vacant region is filled up with 0 (black). Hence, the size of the patched color image is still 512 × 512. Finally, the watermark extraction procedure is performed and the NC value of the watermark is calculated, as described in Section 4. The results are also listed in Table 3. It is worth noticing that the NC value is closely related with the size of clipped block: smaller size of the clipped block generally means smaller NC value. So, a more reasonable method for detecting the existence of a watermark in the clipped block is only computing the NC value of the sub-watermark of the clipped region (it is almost 1 in theory).
Extracted watermarks (EW) and NCs after attacked by Gaussian noise added from σ=5 to 10, and to 15, scaling with factor 2, and various cropping to the original size of the three images.
Gaussian noise
Cropping (Figure 7)
Scaling(factor 2)
σ=5
σ=10
σ=15
I
II
III
Lena
EW
NC
0.9803
0.8225
0.6276
0.6997
0.6075
0.7957
0.9633
Barbara
EW
NC
0.9148
0.7429
0.5657
0.7086
0.5800
0.7463
0.9767
Peppers
EW
NC
0.9010
0.7024
0.5280
0.6959
0.5665
0.7396
0.9480
Attacked watermarked Lena image by various croppings.
Type I
Type II
Type III
We next compare the proposed method to three traditional methods: Tsui et al. [15], Kutter et al. [12] and Cox et al.’s [5] methods using the Lena image. The results are shown in Table 4. In the comparisons, all the methods still use the panda as the test binary logo image, but the size becomes 64 × 64, which can be obtained by downsampling the original 128 × 128 logo image. In our proposed method, the logo is needed to be expanded to 128 × 128 size by filling up with 0 on its three neighboring regions: left, bottom, and left-bottom, before it is embedded to the Lena image during the embedding procedure. Correspondingly, at the end of the extracting process, the logo is obtained by clipping the extracted logo image on the left-top with the size of 64 × 64. As shown in Table 4, the proposed method generally outperforms the other three methods for various attacks.
Comparing the proposed method with Tsui at al. [15], Kutter et al. [12] and Cox at al. [5] methods.
Attacks
Tsui et al. [15]
Kutter et al. [12]
Cox et al. [5]
Proposed
Attack free
1.0
0.9476
0.9957
1.0
Median filter (3×3)
0.9877
0.8750
0.9146
0.9936
JPEG
10
0.4237
N/A
0.6889
0.7001
50
0.9302
0.3504
0.9396
0.9401
Gaussian noise (σ=20)
0.5038
0.4788
0.4516
0.5124
Scaling (factor 2)
0.8743
0.8500
0.8460
0.9633
Sharpening
0.9781
0.8893
0.9405
0.9872
Bluring
0.9650
0.8956
0.9463
0.9847
Cropping
0.6981
0.6330
0.5954
0.6990
6. Conclusion
In this work, we have developed a fractal color image watermarking method and assessed its performance. As we know, the main objectives of watermarking are to achieve a good robustness against attacks and retain high image quality as well. In the proposed method, we consider a pixel as a 3-D vector and an RGB version of orthogonalization fractal coding is performed in RGB space. In order to obtain higher image quality, instead of using three independent linear functions in classic fractal color coding, we utilize a general form of fractal affine transform by using the range block mean vector as the luminance offset and a 3 × 3 matrix as contrast scaling. In terms of the image quality measured by PSNR, the proposed fractal color coding significantly outperforms the traditional one. It is worth noting that, since the compression ratio is not the main point for watermarking, we do not discuss the quantization of the fractal parameters. On the other hand, we also show that the orthogonalization fractal color decoding is a mean vector-invariant iteration. Hence, the range block mean vector is a good place for embedding the watermark. For the sake of consistency with the human visual system and ensuring the watermark invisibility as well as possible, we further implement the hiding procedure in the CIELa*b* space and incorporate a JND scheme. Our experiments show that the proposed watermarking has good robustness against various distortions such as JPEG compression, median filtering, and geometric distortion such as scaling and cropping, as well as with an imperceptible change in image quality.
Conflict of Interests
The authors declare that there is no conflict of interests regarding the publication of this paper.
Acknowledgments
The authors would like to thank the anonymous referees for their helpful comments and suggestions that led to a significant improvement of the paper. This work was supported by the National Natural Science Foundation of China (nos. 61003178, 61070087, 61373087, 11201312, 61272252, and 11071150), and by the Municipal Science and Technology Plan of Shenzhen in China (JC201105170615A, JC201005280508A).
PotdarV. M.HanS.ChangE.A survey of digital image watermarking techniquesProceedings of the 3rd IEEE International Conference on Industrial Informatics (INDIN '05)August 2005Perth, Australia7097162-s2.0-3384528585210.1109/INDIN.2005.1560462ZhangY.PoL.-M.Fractal color image compression using vector distortion measure3Proceedings of the IEEE International Conference on Image Processing (ICIP '95)October 1995Washington, DC, USA2762792-s2.0-0029511976ChenB.WornellG. W.Quantization index modulation methods for digital watermarking and information embedding of multimedia2001271-27332-s2.0-003525178010.1023/A:1008107127819YeR.WuS.Application of chaotic ergodicity in image encryption and watermarkingProceedings of the 2nd International Conference on Mechanic Automation and Control Engineering (MACE '11)July 2011Inner Mongolia, China719671992-s2.0-8005479509510.1109/MACE.2011.5988711CoxI. J.KilianJ.LeightonF. T.ShamoonT.Secure spread spectrum watermarking for multimedia1997612167316872-s2.0-003136211610.1109/83.650120AhmidiN.SafabakhshR.A novel DCT-based approach for secure color image watermarking2Proceedings of the International Conference on Information Technology: Coding Computing (ITCC '04)April 2004Las Vegas, Nev, USA7097132-s2.0-304254500810.1109/ITCC.2004.1286738SolachidisV.PitasI.Circularly symmetric watermark embedding in 2-D DFT domain20011011174117532-s2.0-003551021110.1109/83.967401LinW.-H.WangY.-R.HorngS.-J.KaoT.-W.PanY.A blind watermarking method using maximum wavelet coefficient quantization200936911509115162-s2.0-6734918265310.1016/j.eswa.2009.03.060Al-OtumH. A.Al-Taba'aA. O.Adaptive color image watermarking based on a modified improved pixel-wise masking technique20093556736952-s2.0-6784908870210.1016/j.compeleceng.2009.01.007WandellB. A.1995Sunderland, Mass, USASinauer AssociatesLiC. P.WuY. J.YeR. S.2013SingaporeWorld Scientific PublishingKutterM.JordanF.BossenF.Digital watermarking of color images using amplitude modulation1998723263322-s2.0-0001001980BarniM.BartoliniF.PivaA.Multichannel watermarking of color images20021231421562-s2.0-003649348310.1109/76.993436TsaiP.HuY.-C.ChangC.-C.A color image watermarking scheme based on color quantization2004841951062-s2.0-034535653210.1016/j.sigpro.2003.07.012TsuiT. K.ZhangX.-P.AndroutsosD.Color image watermarking using the spatio-chromatic fourier transformProceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP '06)May 2006Toulouse, France3053082-s2.0-33947655367KuoC.-T.ChengS.-C.Fusion of color edge detection and color quantization for color image watermarking using principal axes analysis20074012369137042-s2.0-3454769950510.1016/j.patcog.2007.03.025ChouC.-H.LiuK.-C.A perceptually tuned watermarking scheme for color images2010191129662982MR28142002-s2.0-7804931039110.1109/TIP.2010.2052261VahediE.ZoroofiR. A.ShivaM.Toward a new wavelet-based watermarking approach for color images using bio-inspired optimization principles20122211531622-s2.0-8155520582510.1016/j.dsp.2011.08.006BarnsleyM. F.DemkoS.Iterated function systems and the global construction of fractals19853991817243275MR79911110.1098/rspa.1985.0057ZBL0588.28002JacquinA. E.Image coding based on a fractal theory of iterated contractive image transformations19921118302-s2.0-0026630755FisherY.1995Berlin, GermanySpringerxviii+34110.1007/978-1-4612-2472-3MR1313035LaiC.-M.LamK.-M.SiuW.-C.A fast fractal image coding based on kick-out and zero contrast conditions2003121113981403MR20267812-s2.0-024272064710.1109/TIP.2003.817246WangX.-Y.LiF.-P.ChenZ.-F.An improved fractal image coding method2009174451457MR2650305ZBL1183.280192-s2.0-7724908369810.1142/S0218348X09004545WangX.LiF.WangS.Fractal image compression based on spatial correlation and hybrid genetic algorithm20092085055102-s2.0-7035013174910.1016/j.jvcir.2009.07.002TsengC.-C.HsiehJ.-G.JengJ.-H.Fractal image compression using visual-based particle swarm optimization2008268115411622-s2.0-4454908489410.1016/j.imavis.2008.01.003ZhangY.WangX.Fractal compression coding based on wavelet transform with diamond search2012131106112MR2846823ZBL1238.940112-s2.0-8115516093210.1016/j.nonrwa.2011.07.017TongC. S.PiM.Fast fractal image encoding based on adaptive search2001109126912772-s2.0-003544185710.1109/83.941851WangX.-Y.ZouL.-X.Fractal image compression based on matching error threshold2009171109115MR2498749ZBL1159.943112-s2.0-6524911339510.1142/S0218348X09004247ChenH.-N.ChungK.-L.HungJ.-E.Novel fractal image encoding algorithm using normalized one-norm and kick-out condition20102835185252-s2.0-7044972650710.1016/j.imavis.2009.08.007WangX. Y.WangY. X.YunJ. J.An improved no-search fractal image coding method based on a fitting plane20102881303130810.1016/j.imavis.2010.01.008WangX.ZhangD.GuoX.Novel hybrid fractal image encoding algorithm using standard deviation and DCT coefficients2013731-234735510.1007/s11071-013-0790-2MR3080672WangX. Y.GuoX.ZhangD. D.An effective fractal image compression algorithm based on plane fitting201221910.1088/1674-1056/21/9/090507090507WangX. Y.WangY. X.YunJ. J.An improved no-search fractal image coding method based on a modified gray-level transform200832444545010.1016/j.cag.2008.02.004GhazelM.FreemanG. H.VrscayE. R.Fractal image denoising20031212156015782-s2.0-1084423654010.1109/TIP.2003.818038LuJ.YeZ.ZouY.YeR.An enhanced fractal image denoising algorithm200838410541064MR2435602ZBL1152.943162-s2.0-4584910837610.1016/j.chaos.2007.06.048JengJ.-H.TsengC.-C.HsiehJ.-G.Study on Huber fractal image compression20091859951003MR30299792-s2.0-6734918017810.1109/TIP.2009.2013080LuJ.YeZ.ZouY.Huber fractal image coding based on a fitting plane201322113414510.1109/TIP.2012.2215619MR3017452PiM.MandalM. K.BasuA.Image retrieval based on histogram of fractal parameters2005745976052-s2.0-2494447753710.1109/TMM.2005.846796DavernP.ScottM.Fractal based image steganography19961174Berlin, GermanySpringer279294Lecture Notes in Ccomputer ScienceČandikM.LevickýD.KlenovičováZ.Fractal image coding with digital watermarks2000942226KianiS.MoghaddamM. E.Fractal based digital image watermarking using fuzzy C-mean clusteringProceedings of the International Conference on Information Management and Engineering (ICIME '09)April 2009Kuala Lumpur, Malaysia6386422-s2.0-7034948818510.1109/ICIME.2009.72PiM. H.LiC. H.LiH.A novel fractal image watermarking2006834884982-s2.0-3364672517410.1109/TMM.2006.870738ShahraeiniS.YaghoobiM.A robust digital image watermarking approach against JPEG compression attack on bybrid fractal-wavelet5Proceedings of the International Conference on Computer Communication and Management (ICCCM '11)2011Wuhan, China616622ChouC.-H.LiY.-C.A perceptually tuned subband image coder based on the measure of just-noticeable-distortion profile1995564674762-s2.0-002951725110.1109/76.475889LiuJ. Z.Some estimates of correlation coefficients for systems of random vectors19991414348MR1691826ChouC. H.LiuK. C.An oblivious and robust watermarking scheme using perceptual modelProceedings of the 4th EURASIP Conference on Video/Image Processing and Multimedia Communications (VIPMC '03)2003Zagreb, Croatia713720HennellyB.SheridanJ. T.Optical image encryption by random shifting in fractional Fourier domains20032842692712-s2.0-0037443119