Reversible Data Hiding
Reversible Data Hiding
Reversible Data Hiding
28, 2021
Abstract—How to predict images is an important issue in the we are noting that the multi receptive fields [18], [19] and global
reversible data hiding (RDH) community. In this letter, we propose optimization capabilities of convolutional neural networks
a novel CNN-based prediction approach by luminously dividing (CNNs), which could be served as a predictor for improvement
a grayscale image into two sets and applying one set to predict
the other set for data embedding. The proposed CNN predictor is a of the prediction accuracy for RDH. In [20], the authors proposed
lightweight and computation-efficient network with the capabilities a CNN-based RDH method for stereo images by using the right
of multi receptive fields and global optimization. This CNN predic- image and the left image to predict each other. Also, CNN-based
tor can be trained quickly and well by using 1000 images randomly data-driven methods have achieved satisfactory performance in
selected from ImageNet. Furthermore, we propose a two stages of global image analysis [21]–[24] and image coding [25], [26].
embedding scheme for this predictor. Experimental results show
that the CNN predictor can make full use of more surrounding In this letter, we propose a new and efficient predictor based on
pixels to promote the prediction performance. Furthermore, in CNN. An 8-bits grayscale image is first divided into two subset
the experimental way we have shown that the CNN predictor images, which can be used to predict each other by using the
with expansion embedding and histogram shifting techniques can proposed CNN-based predictor (CNNP). The proposed CNNP
provide better embedding performance in comparison with those is a lightweight and computation-efficient network with the
classical linear predictors.
capabilities of multi receptive fields and global optimization,
Index Terms—Convolutional neural network, reversible data which can be trained well by using 1000 images randomly
hiding, global optimization capability. selected from ImageNet [27]. Experimental results show that
the CNNP can promote prediction performance due to the use of
I. INTRODUCTION more surrounding pixels as the context. Experimentally we have
shown that the CNNP with classical expansion embedding and
EVERSIBLE data hiding (RDH) technologies are well
R known for the ability to recover the original image and the
information without any loss [1], [2]. Based on the characteristic
histogram shifting techniques can provide better performance by
comparing with several classical linear predictors. To the best
of our knowledge, this work is the first report in detail on how
mentioned above, RDH methods are widely used in military, to use CNN to predict grayscale images for RDH.
medical, and super-resolution processing fields [3]. The rest of this letter is organized as follows. The proposed
So far, many approaches have been developed in RDH. One RDH scheme is described in detail in Section II, and the
type of approaches focus on finding new embedding ways on experiments comparing with other predictors are reported in
how to deal with the prediction errors for reduction of embedding Section III. Finally, we conclude our work in Section IV.
distortion, such as the difference expansion [4], [5], histogram
shifting [6], [7], and prediction-error expansion [8]. The other
II. PROPOSED METHOD
type of approaches focus on how to design advanced predictors
to improve prediction accuracy, including difference predictor A. Pre-Processing Images
(DP) [4], median edge direction predictor (MEDP) [8], gradient At first, the original image I is divided into two subset images:
adaptive predictor (GAP) [9], [10], bilinear interpolation predic- the “Cross” set image IC and the “Dot” set image ID . The image
tor (BIP) [11], [12], and others by using multi-predictors [13] partition pattern is shown in Fig. 1. For the “Cross” set image, the
and adaptive strategies [14]–[17]. pixel values of the positions belong to the “Dot” set are assigned
In the literature, all of the predictors for grayscale images to 0, and so does the “Dot” set image. Such a partition pattern
share a weakness, that is to only apply one or a few adjacent ensures that the two subset images are independent but relevant.
pixels as the context for prediction. That is to say, there still After dividing the original image, the two subset images will be
exists room for RDH by making full use of more neighboring used to train and converge the proposed CNNP.
pixels as the context of a to-be-predicted pixel. To this direction,
B. Architecture Overview
Manuscript received December 9, 2020; revised January 22, 2021; accepted
February 6, 2021. Date of publication February 12, 2021; date of current version Based on the correlation of image pixels and the properties
March 10, 2021. This work was supported in part by the NSFC under Grant of CNN, the “Cross” set image and the “Dot” set image are
61772234. The associate editor coordinating the review of this manuscript, and
approving it for publication was Prof. Mylene Q. Farias. (Corresponding author: designed to predict each other in this letter. Fig. 2(a) illustrates
Shijun Xiang.) how to use the “Cross” set image to predict the “Dot” set image,
The authors are with the College of Information Science and Technol- including two main steps: the feature extraction step (for the
ogy/College of Cyber Security, Jinan University, Guangzhou 510632, China
(e-mail: [email protected]; [email protected]). use of the multi receptive fields) and the image prediction step
Digital Object Identifier 10.1109/LSP.2021.3059202 (for the use of global optimization). After the feature extraction
1070-9908 © 2021 IEEE. Personal use is permitted, but republication/redistribution requires IEEE permission.
See https://www.ieee.org/publications/rights/index.html for more information.
Authorized licensed use limited to: DELHI TECHNICAL UNIV. Downloaded on February 08,2022 at 07:05:22 UTC from IEEE Xplore. Restrictions apply.
HU AND XIANG: CNN PREDICTION BASED REVERSIBLE DATA HIDING 465
Authorized licensed use limited to: DELHI TECHNICAL UNIV. Downloaded on February 08,2022 at 07:05:22 UTC from IEEE Xplore. Restrictions apply.
466 IEEE SIGNAL PROCESSING LETTERS, VOL. 28, 2021
and generate the data hidden “Dot” set image IDW . Similarly, TABLE I
AVERAGE MSE, ABSOLUTE MEAN AND VARIANCE OF THE PREDICTION
IDW is fed into the proposed CNNP to generated the predicted ERRORS IN 100 IMAGES FOR FIVE DIFFERENT PREDICTORS
“Cross” set image I˜C , which is in cooperated with IC to embed
the other part of the information W2 and generate the data hidden
“Cross” set image ICW . After embedding the whole information
W (W = W1 + W2 ), ICW is added with IDW to generate the
data hidden image IW .
The flowchart of extracting the hidden information and re-
covering the original image is shown in Fig. 5. The data hidden
image IW is divided into two subset images (i.e., ICW and of the training image set, 100 images are randomly selected from
IDW ) with the same partition pattern in Fig. 1. After that, ImageNet for statistical experiments. Without special mention,
the data hidden “Dot” set image IDW is firstly fed into the the results in this section are the average of 100 images.
proposed CNNP to generate the predicted “Cross” set image
I˜C , which is in cooperated with ICW to extract the information
W2 and recover the original “Cross” set image IC . The recovered A. Prediction Accuracy
“Cross” set image IC is fed into the proposed CNNP to generate Table I lists the MSE, the absolute mean, and the variance of
the predicted “Dot” set image I˜D , which is used to recover the prediction errors for the proposed CNNP and four classical
the original “Dot” set image ID and extract the information predictors. The MSE of the CNNP is 99.4, which is lower than
W1 . Subsequently, the recovered images ID and IC are added BIP (154.8), MEDP (234.2), GAP (231.9), and DP (230.8).
together in the spatial domain to recover the original image I. Besides, the absolute mean and the variance in the proposed
Meanwhile, W1 and W2 are combined to recover the information CNNP are 4.77 and 66.9, respectively, which are more con-
bits W . centrated and lower than BIP (6.25 and 100.5), MEDP (7.37
and 161.3), GAP (9.86 and 167.6), and DP (5.13 and 196.6).
From Table I, we can conclude that the prediction accuracy of
III. EXPERIMENTAL RESULTS the proposed CNNP is better than several classical predictors.
The performance of the proposed CNNP is evaluated by The basic reason is that the proposed predictor can exploit multi
comparing it with several classical predictors, including the receptive field prediction and global optimization capabilities
BIP, MEDP, GAP, and DP. The detailed description of these of CNN by optimizing the network parameters with the loss
predictors can be seen in [30]. function. Furthermore, we have plotted the histograms of the
Aiming to assess the prediction performance of the proposed prediction error for these five different predictors on the image
CNNP, the mean square error (MSE) is adopted since it can Lena, as shown in Fig. 6, in which the abscissa range is set
reflect the difference between the predicted images and the target to [−30, 30] for better display. We can see that the histogram
images well. For a fair comparison, the predictors mentioned of the prediction errors generated from the proposed predictor
above are combined with the same embedding schemes for RDH, is higher and sharper than the histograms of the other four
in which the performance of these predictors is measured by predictors, and the number of the prediction error values in
computing the peak signal-to-noise ratio (PSNR) values of the the region [−3, 3] of the proposed CNNP is larger than that
watermarked images at the same embedding rates. Independent of other predictors. From these experimental results, we can
Authorized licensed use limited to: DELHI TECHNICAL UNIV. Downloaded on February 08,2022 at 07:05:22 UTC from IEEE Xplore. Restrictions apply.
HU AND XIANG: CNN PREDICTION BASED REVERSIBLE DATA HIDING 467
TABLE II
AVERAGE PSNR (IN dB) OF 100 IMAGES FOR FIVE DIFFERENT PREDICTORS BY
USING EXPANSION EMBEDDING TECHNIQUE IN [8]
Fig. 6. Histograms of the image Lena under five different predictors. TABLE III
AVERAGE PSNR (IN dB) OF 100 IMAGES OF THE PROPOSED CNNP-BASED
METHOD AND THE METHOD [11]
Authorized licensed use limited to: DELHI TECHNICAL UNIV. Downloaded on February 08,2022 at 07:05:22 UTC from IEEE Xplore. Restrictions apply.
468 IEEE SIGNAL PROCESSING LETTERS, VOL. 28, 2021
Authorized licensed use limited to: DELHI TECHNICAL UNIV. Downloaded on February 08,2022 at 07:05:22 UTC from IEEE Xplore. Restrictions apply.