arxiv:2110.05687v1 [cs.cv] 12 oct 2021

7
NO WAY TO CROP: ON ROBUST IMAGE CROP LOCALIZATION Qichao Ying ? , Xiaoxiao Hu ? , Hang Zhou , Xiangyu Zhang ? , Zhengxin You ? and Zhenxing Qian ? ? Fudan University, China. Simon Fraser University, Canada ABSTRACT Previous image forensics schemes for crop detection are only lim- ited on predicting whether an image has been cropped. However, cropping different areas in a same image leads to different semantic changes. This paper presents a novel scheme for image crop local- ization using robust watermarking. We train an anti-crop processor (ACP) that embeds a watermark into a target image. The visually indistinguishable protected image is then posted on the social net- work instead of the original image. During data sharing, the cloud users may tamper, crop or benignly attack the image using JPEG compression, scaling, etc. At the recipient’s side, ACP extracts the watermark from the attacked image, and we conduct feature match- ing on the original and extracted watermark to locate the position of the crop in the original image plane. We further extend our scheme to detect tampering attack on the attacked image. We demonstrate that our scheme is the first to provide high-accuracy and robust im- age crop localization. Besides, the accuracy of tamper detection is comparable to many state-of-the-art methods. Index Termsimage crop localization, image tamper detec- tion, robustness, image forensics 1. INTRODUCTION To fight against daily-world image forgery, researchers have devel- oped a number of schemes which detect various kinds of digital at- tacks, e.g., against tampering [24, 13, 12], DeepFake [18] and crop- ping [22, 21, 19]. Among them, cropping is a simple yet power- ful way to maliciously alter the message of an image. This kind of forgery has historically been less investigated by the forensic com- munity than other image manipulations. The main difficulty is that much less semantic or statistical clue of cropping can be discov- ered. The existing crop detection algorithms [21, 19] mainly focus on predicting whether an image is cropped. They are represented by detecting the exposing evidences of asymmetrical image cropping. However, cropping different areas in a same image will definitely result in different semantic changes, and therefore we need to locate the position of the crop in the original image plane. There is little previous arts for image crop localization. Van et al. [22] proposes to investigate the impact that cropping has on the image distribution, featured by chromatic aberration and vignetting. However, there are certain limitations. First, it requires the images to be untouched and uncompressed so that the camera pipeline artefacts and photography patterns are preserved. Otherwise the scheme can- not be applied. Second, the clues from the camera pipeline artefacts are usually not enough to ensure high-accuracy localization results. Therefore, the real-world application of [22] is rather limited. Mo- tivated by the short-comings of the previous schemes, we propose the first image crop localization scheme using robust watermarking. Qichao Ying and Xiaoxiao Hu contribute equally. Corresponding author: Zhenxing Qian ([email protected]). This work is supported by National Natural Science Foundation of China under Grant U20B2051, U1936214. (a) (b) (c) (d) Fig. 1. Examples of crop localization. (a) Given an image to be pro- tected, (b) we conduct imperceptible data hiding on the targeted im- age. (c) The social network users redistribute, crop, and even modify the image. (d) The recipient identifies the cropped-out area. Since the original clues are weak, we propose to hide crafted clues into a targeted image to protect them in the first place. We further observe that besides cropping, there are also benign attacks, e.g., JPEG compression, scaling, etc. These usually cause a significant drop in performance of many other computer vision tasks such as data hiding [62] or tamper detection [13]. It requires the scheme to be robust against these typical attacks. We further extend our work to localize tamper on the protected image before crop localization. The proposed scheme is efficient in revealing misleading photojournal- ism or copyright violation. Fig. 1 shows two examples of malicious image cropping, where our scheme accurately localizes the cropped area on receiving the ambiguous cropped images. We train a normalizing-flow-based anti-crop processor (ACP) to embed a watermark into a targeted image. The watermark is shared with the recipient. ACP produces a visually indistinguishable pro- tected image, which is then posted on the social network instead of the unprotected version. On receiving the attacked version of the protected image, we use ACP again to extract the watermark, and conduct a feature matching algorithms (SURF [1]) between the original and extracted watermark to determine where the crop was positioned in the original image plane. We further extend our work to detect tampers on the doubted image by introducing a tamper de- tector to predict the tamper mask. The features within the predicted tamper mask is disabled to prevent mismatching. We test our scheme by introducing man-made hybrid attacks. The results demonstrate that our scheme can accurately localize the crooped region. We also show the effectiveness of tamper detection by comparison with some state-of-the-art schemes [12, 13]. The highlights of this paper are three-folded. 1) This paper presents the first high-accuracy robust image crop localization scheme. 2) With the embedded watermark, the proposed scheme can also conduct high-accuracy tamper detection, which is comparable with the state-of-the-art works. 3) We use normalizing flows to build an efficient invertible function for image forensic problems. arXiv:2110.05687v1 [cs.CV] 12 Oct 2021

Upload: others

Post on 08-Dec-2021

5 views

Category:

Documents


0 download

TRANSCRIPT

NO WAY TO CROP: ON ROBUST IMAGE CROP LOCALIZATION

Qichao Ying?, Xiaoxiao Hu?, Hang Zhou†, Xiangyu Zhang?, Zhengxin You? and Zhenxing Qian?

? Fudan University, China. † Simon Fraser University, Canada

ABSTRACT

Previous image forensics schemes for crop detection are only lim-ited on predicting whether an image has been cropped. However,cropping different areas in a same image leads to different semanticchanges. This paper presents a novel scheme for image crop local-ization using robust watermarking. We train an anti-crop processor(ACP) that embeds a watermark into a target image. The visuallyindistinguishable protected image is then posted on the social net-work instead of the original image. During data sharing, the cloudusers may tamper, crop or benignly attack the image using JPEGcompression, scaling, etc. At the recipient’s side, ACP extracts thewatermark from the attacked image, and we conduct feature match-ing on the original and extracted watermark to locate the position ofthe crop in the original image plane. We further extend our schemeto detect tampering attack on the attacked image. We demonstratethat our scheme is the first to provide high-accuracy and robust im-age crop localization. Besides, the accuracy of tamper detection iscomparable to many state-of-the-art methods.

Index Terms— image crop localization, image tamper detec-tion, robustness, image forensics

1. INTRODUCTION

To fight against daily-world image forgery, researchers have devel-oped a number of schemes which detect various kinds of digital at-tacks, e.g., against tampering [24, 13, 12], DeepFake [18] and crop-ping [22, 21, 19]. Among them, cropping is a simple yet power-ful way to maliciously alter the message of an image. This kind offorgery has historically been less investigated by the forensic com-munity than other image manipulations. The main difficulty is thatmuch less semantic or statistical clue of cropping can be discov-ered. The existing crop detection algorithms [21, 19] mainly focuson predicting whether an image is cropped. They are represented bydetecting the exposing evidences of asymmetrical image cropping.However, cropping different areas in a same image will definitelyresult in different semantic changes, and therefore we need to locatethe position of the crop in the original image plane.

There is little previous arts for image crop localization. Van etal. [22] proposes to investigate the impact that cropping has on theimage distribution, featured by chromatic aberration and vignetting.However, there are certain limitations. First, it requires the images tobe untouched and uncompressed so that the camera pipeline artefactsand photography patterns are preserved. Otherwise the scheme can-not be applied. Second, the clues from the camera pipeline artefactsare usually not enough to ensure high-accuracy localization results.Therefore, the real-world application of [22] is rather limited. Mo-tivated by the short-comings of the previous schemes, we proposethe first image crop localization scheme using robust watermarking.

Qichao Ying and Xiaoxiao Hu contribute equally. Corresponding author:Zhenxing Qian ([email protected]). This work is supported by NationalNatural Science Foundation of China under Grant U20B2051, U1936214.

(a) (b) (c) (d)

Fig. 1. Examples of crop localization. (a) Given an image to be pro-tected, (b) we conduct imperceptible data hiding on the targeted im-age. (c) The social network users redistribute, crop, and even modifythe image. (d) The recipient identifies the cropped-out area.

Since the original clues are weak, we propose to hide crafted cluesinto a targeted image to protect them in the first place. We furtherobserve that besides cropping, there are also benign attacks, e.g.,JPEG compression, scaling, etc. These usually cause a significantdrop in performance of many other computer vision tasks such asdata hiding [62] or tamper detection [13]. It requires the scheme tobe robust against these typical attacks. We further extend our work tolocalize tamper on the protected image before crop localization. Theproposed scheme is efficient in revealing misleading photojournal-ism or copyright violation. Fig. 1 shows two examples of maliciousimage cropping, where our scheme accurately localizes the croppedarea on receiving the ambiguous cropped images.

We train a normalizing-flow-based anti-crop processor (ACP) toembed a watermark into a targeted image. The watermark is sharedwith the recipient. ACP produces a visually indistinguishable pro-tected image, which is then posted on the social network insteadof the unprotected version. On receiving the attacked version ofthe protected image, we use ACP again to extract the watermark,and conduct a feature matching algorithms (SURF [1]) between theoriginal and extracted watermark to determine where the crop waspositioned in the original image plane. We further extend our workto detect tampers on the doubted image by introducing a tamper de-tector to predict the tamper mask. The features within the predictedtamper mask is disabled to prevent mismatching. We test our schemeby introducing man-made hybrid attacks. The results demonstratethat our scheme can accurately localize the crooped region. We alsoshow the effectiveness of tamper detection by comparison with somestate-of-the-art schemes [12, 13].

The highlights of this paper are three-folded. 1) This paperpresents the first high-accuracy robust image crop localizationscheme. 2) With the embedded watermark, the proposed scheme canalso conduct high-accuracy tamper detection, which is comparablewith the state-of-the-art works. 3) We use normalizing flows to buildan efficient invertible function for image forensic problems.

arX

iv:2

110.

0568

7v1

[cs

.CV

] 1

2 O

ct 2

021

Fig. 2. Overview of the proposed scheme. The anti-crop processor(ACP) generates the protected image and conversely in the back-propagation recovers the hidden watermark. The attack layer simu-lates several kinds of digital attacks. The tamper detector predicts thetamper mask and the feature detector localizes the cropped region ofthe attacked image.

2. RELATED WORKS

Detection on Image Cropping and Tampering. Fanfani et al. [21]exploits the camera principal point insensitive to image processingoperations. Yerushalmy et al. [19] detects whether there are vanish-ing points and lines on structured image content. Many tamper de-tection schemes [12, 40] are developed upon the classic U-Net archi-tecture. Ying et al. [24] not only detects tampers, but also proposesto conduct image self recovery. Besides, Mantra-Net [13] signifi-cantly improves the detection and localization performance by usingself-supervised learning of the robust image manipulation traces.Invertible Neural Networks (INN). Invertible neural networklearns a stable invertible mapping from the source distribution Ps toa targeted distribution Pt, and the forward and back propagation op-erations are in the same network. Pioneering research on INN-basedmapping can be seen in RealNVP [29], Glow [30], etc. Recently,INNs are also used for various low-level computer vision tasks suchas image colorization [31] invertible data hiding network (ISN) [36],and image rescaling [32].Robust Watermarking and its Application on Forensics. Nowa-days, several robust watermarking schemes [42, 26] are proposedwhere a differentiable attacking layer is proposed for adversarialtraining. Yu [37] is the first to ensure robustness and large capacity atthe same time. Previously, robust watermarking has been introducedto aid preventing images from being inpainted [27] or reconstructedby super-resolution [25].

3. METHOD

3.1. Overview

Fig. 2 shows an overview of our scheme. The pipelone of ourmethod is composed of an Anti-Cropping Processor (ACP) P , anattack layer A, a tamper detector V , a discriminator D and a featurematcher M. The proposed scheme consists of five stages: water-mark embedding, image redistribution, tamper detection, watermarkextraction and crop localization. We regard the embedding and ex-traction of the watermark W as the inverse problem, even if the pro-tected image is cropped and attacked. It comes from the observationthat many data hiding schemes embed the secret image into the hostimage according to the spatial order [62]. We formulate the inverse

problem of the ACP as:

(IM ,R) = P(I,W)

I,W = P−1(A(IM ), R),(1)

where R and R are the pseudo-random output and input to keep theconsistency of the channel number. For the invertibility, W and theground-truth cropped watermark WC should be as close as possible.WC is obtained by cropping out the same region from W as on IM .Besides, IM should be visually indistinguishable from I.

First, we embed the watermark W into the targeted image I us-ing the ACP. The protected image IM is generated and we upload itonto the social cloud instead of the unprotected targeted image. Wesimulate the image redistribution stage and generate the attacked im-age IA by freely adding three kinds of attacks (benign attacks, crop-ping, tampering) on the protected images. On the recipient’s side, thetamper detector V predicts the tamper mask M on the attacked im-age to see which parts of the image are tampered. We inversely runthe ACP and feeds the attacked image to extract the watermark W.Afterwards, we rectify the extracted water by W′ = W · (1 − M)to discard the tampered contents. Finally, with the original water-mark W as reference, we uses the feature matcherM to locate theposition of the crop in the original image plane.

In the training stage, the pipeline is slightly modified in order toproperly train the networks. We let the ACP focus on the learn-ing of robust data hiding and extraction by feeding it with non-tampered attacked images. We further add tampering attacks on thenon-tampered attacked images to train the tamper detector.

3.2. Network Design

Considering the efficiency of the invertible U-Net proposed in [32],we build our ACP on top of this architecture. The network consistsof six invertible blocks each of which contains a Haar wavelet trans-formation and a double-side affine coupling. The network ends witha conditional split layer. The number of the input and output chan-nels are four. The sizes of R and R are of the same as W. R andR are not required to be the same.

In the attacking layer A, we first use a differentiable quantiza-tion layer to transfer the data type of the protected image from floatto 8-bit integer. Then, we build differentiable methods to simulatethe benign attacks B, the cropping attack C and the tampering attackT . In B, we take the implementation from [42] except that we buildour own JPEG simulator J . In C, we randomly crop a portion of theprotected image IM . In T , we first randomly select random areasusing a binary matrix M inside IM , and generate the tampered im-age by IA = Iirr ·M + IM · (1 −MR), where Iirr refers to thesource of the tamper.

For JPEG simulation J , there are already many scheme whichinclude a carefully-designed JPEG simulator, e.g., JPEG-SS [67],JPEG-Mask [42], MBRS [68]. However, the real-world JPEG ro-bustness of these schemes is still limited. We believe it mainly at-tribute to that the networks are over-fitted to a fixed compressionmode. For example, [67, 42] uses a fixed quantization table to mimicJPEG compression while in the real world the quantization tablegreatly varys. Luo [69] proposes a trainable attack network that com-petes with the baseline network during training, but previous workshave reported that it usually makes the training very hard and unsta-ble. In this paper, we propose to apply the Mix-Up strategy [70] toconduct an instance-agnostic interpolation by:

Ijpg = θ · IM + (1− θ) ·∑Jk∈J

∑QFl∈[10,100]

ε · Jk(IM , QFl), (2)

Cover

(shared)

Diff (5x)

(shared)

Watermark

Marked

Fig. 3. Illustration of watermark embedding. Row 1-4: targeted im-ages I, watermark W, protected images IM , Augmented differenceD = 5× abs(I− IM )

where θ ∈ (0, 1),∑Jk∈J

ε = 1 and J ∈{JPEG-SS, JPEG-Mask,MBRS}. QF stands for the quality factor. The first part of ?? isto prevent the image recovery from being too hard under simulatedJPEG compression at the begining of the training stage. Thus, weslowly decline θ to zero.

We accepts the implementation of the edge generator and thePatchGAN-based discriminator proposed in [54] as our tamper de-tector V and discriminator D. We use the SURF algorithm [1] toimplement the feature matcherM. SURF is an efficient local fea-ture detector and descriptor widely used to extract points of interest.The crop localization is realized by applying linear weighted fusionalgorithms on the SURF descriptors of the two images.

3.3. Objective Loss Function

For the ACP, we encourages the protected image IM and the ex-tracted watermark IR to respectively resemble the targeted image Iand the ground-truth cropped watermark WC . Lrec = F(I, IM ) +

F(WC ,W), where F is the `2 distance. Owing to the invertibilityof the ACP, we do not use the perceptual loss. We also need to ran-domize the extra output by Lran = F(R, R). The adversarial lossis to further control the introduced distortion by fooling the discrim-inator. We accept the least squared adversarial loss (LS-GAN) [57].The total loss for ACP is LG = Lrec +α · Ldis +β · Lran, where αand β are hyper-parameters. For the tamper detector, we minimizethe binary cross entropy (BCE) loss between the estimated tampermask M and the ground-truth mask M. LV = BCE(M, M).

4. EXPERIMENTAL RESULTS

4.1. Experiment Setup

We train the scheme on the COCO training/test set [38] with au-tomatically generated attacks. The scheme is tested with human-participated attacks. We resize the images to the size of 256× 256.The hyper-parameters are set as α = 1, β = 8. The batch size isset as 16. We use Adam optimizer [48] with the default parameters.The learning rate is 1× 10−4. We provide the volunteers with somegenerated protected image, which are then cropped and processed bybenign attacks such as lossy compression, scaling, etc. The volun-teers may add, modify or delete some important image contents at

Attacked Feature Matching Cropping LocalizationExtraction

(a)

(b)

(c)

(d)

Fig. 4. Results of watermark extraction and crop localization of theprotected images in Fig. 3. The attacks are respectively (a) crop, (b)crop & scaling, (c) crop & JPEG and (d) crop & JPEG & tamper.

their free will using typical image processing tools like Adobe Pho-toshop. The crop rate is roughly δ ∈ [0.25, 1). During training, wearbitrarily and evenly perform one kind of benign attacks. Some-times the benign attack is skipped to simulate lossless communica-tion. Also, to avoid over-fitting, we sometimes skip the tamperingattack, in which case we force the tamper detector to predict a zeromatrix. We adaptively convert the predicted tamper mask M into abinary matrix.

4.2. Real-World Performance of Crop Localization

Quality of the protected images. In Fig. 3, we randomly sampledifferent pairs of images as the targeted and watermark. In the firsttwo groups, we use a shared watermark. From the figures, we canobserve that the difference D is imperceptible. Little detail of thewatermark can be found. We have conducted more embedding ex-periments over 1000 images from the test set, and the average PSNRbetween the protected images and the targeted images is 36.23dB,and the average SSIM [51] is 0.983.Accuracy of the Crop Localization. Fig. 4 shows the results ofwatermark extraction, feature matching and crop localization on theprotected images in Fig. 3. In the first row, we only crop the pro-tected image. We see that the crop mask is accurately predicted. Asa result, even without the prior knowledge of the original image, weknow the relative position of the attacked image in the original imageplane. We suggest the readers not using images with too much repet-itive patterns as watermark where SURF may find multiple matchingsolutions between the extracted watermark and the original version.If it is difficult to share the whole watermark, we suggest to usea pre-trained StyleGAN [76] generator where the latent codes canserve as keys for watermark generation. We have conducted moreexperiments over 1000 images under different crop rate and differ-ent kinds of benign attacks. Here for space limit, we take QF = 70as an example. The average performances are reported in Table 1.Higher IoU (Intersection over Union) and SSIM [51] indicates moreaccurate localization result. The average IoUs are above 0.8 wherethe protected images do not undergo further attacks except cropping.

Table 1. Average performance of cropping localization measured byIoU and SSIM between the extracted and ground-truth watermark.

Rate Index NoAttack JPEG Scaling Blur

90% IoU 0.919 0.895 0.803 0.838SSIM 0.949 0.896 0.957 0.953

70% IoU 0.858 0.813 0.878 0.915SSIM 0.942 0.878 0.940 0.940

50% IoU 0.821 0.706 0.603 0.540SSIM 0.914 0.7223 0.942 0.840

Table 2. F1 score comparison for tamper detection among ourscheme and the state-of-the-art methods.

Method NoAttack JPEG Blur ScalingProposed 0.773 0.736 0.695 0.745

Mantra-Net [13] 0.566 0.480 0.557 0.540RRU-Net [12] 0.435 0.273 0.244 0.417

The scheme is proven to be agnostic to the crop size in that the per-formance does not degrade significantly with larger crop rate.Robustness in the Real-World Application. In the second andthird row of Fig. 4, we test the robustness of our scheme by con-ducting different benign attack on the cropped protected image. Inthe fourth row, tampering attack is further introduced. We subtractthe extracted contents within the predicted tamperea areas (Fig. 5)and prohibit the feature matcherM from using the features inside.In Table 1, we can observe that our scheme provides high-accuracycrop localization despite the presence of the attacks other than crop-ping. The results promote the practical application of the proposedscheme. Thanks to the robustness of feature matching of SURF, theexperiments show that in most cases our scheme do not require aprecise watermark extraction.Comparison with Crop Dissection. In Fig.7 of [22], Van et al.discusses their application on crop localization. However, it requiresthat the targeted image should not have been cropped and mustmaintain a constant fixed aspect ratio and resolution. Otherwise thescheme cannot be applied. In contrast, the proposed scheme doesnot have any restriction on the targeted image. Therefore, we onlycompare our scheme with [22] on crop classification. The crop clas-sification accuracy [22] is 86% on high-quality untouched images,while the accuracy of our scheme is 91% on normal images, as longas the users avoid using inappropriate images as watermark.

4.3. Accuracy of Tamper Detection.

We compare our tamper detection method with Mantra-Net [13] andRRU-Net [12]. Fig. 5 shows the tamper detection results of the at-tacked image in the last two images in Fig.3. We observe that al-though the images are tampered by a variety of random attacks, wesucceed in localizing the tampered areas. The successful localiza-tion helps the feature matcher to find the cropped region accurately.In Table 2, we provide the average results over 1000 images. The F1score of the proposed scheme is 0.773 on uncompressed images and0.736 on JPEG attacked images (QF = 80). The performance ofMantra-Net on JPEG images is much worse than that on plain-textimages. We believe the reason is that less statistical clue is preservedin the compressed version. In contrast, the embedded watermark

Ground Truth No attack JPEG60 No attack JPEG60 JPEG60

Proposed Scheme Mantra-Net [2] RRU-Net [3]

Fig. 5. We compare the result of tamper detection in the attackedimages shown in Fig.3 with Mantra-Net [13] and RRU-Net [12].

Ground Truth Proposed Ablation Test 1 Ablation Test 3 MBRS [21] JPEG-SS [22]

Fig. 6. Results of the ablation studies where we validate the networkdesign and the Mix-Up [70] strategy in JPEG simulation.

signal serves as the alternative clue for tamper detection, which isdesigned to resist benign attacks. The performance on JPEG imagesdoes not drop too much.

4.4. Ablation Study

We study the effectiveness of the network design. In Test 1, we traintwo individual hiding network and revealing network to replace thenormalizing-flow-based ACP. In Test 2, we implement the JPEG at-tack with that proposed in [68, 42, 67]. The JPEG QF and the cropratio are kept the same for fair comparison. We train the implemen-tations together with the baseline under the same losses and batchsize. Fig. 6 shows the detailed comparison results. First, the base-line results outperform the encoder-decoder network design. Sec-ond, while MBRS [68] can provide decent robustness, the extractionperformance is even better using our JPEG simulator. Specifically,the average SSIM between W and WC using MBRS [68] is 0.797compared to 0.878 reported in Table.1. The Mix-Up strategy pre-vents the networks from being over-fitted to any single JPEG sim-ulator, which helps the scheme significantly improve its real-worldrobustness. In Test 3, we do not use the discriminatorD. The resultsshow that it leads to visible artefacts in the protected image. Theextraction result is also worse than the baseline.

5. CONCLUSION

This paper presents a novel scheme of image crop localization usingrobust watermarking. We produces a visually indistinguishable pro-tected image for a targeted image using the ACP. We simulate typicalattacks in the pipeline where we propose an improved JPEG simula-tor. On receiving the attacked image, we extracts the watermark. Wethen conduct tamper detection on the image and use the SURF algo-rithm to locate the position of the crop. The results proves that ourmethod provides high-accuracy and robust image crop localization.Besides, the accuracy of tamper detection is also promising.

6. REFERENCES

[1] Herbert Bay, Tinne Tuytelaars, and Luc Van Gool, “Surf:Speeded up robust features,” in European conference on com-puter vision. Springer, 2006, pp. 404–417.

[2] David G Lowe, “Object recognition from local scale-invariantfeatures,” in Proceedings of the seventh IEEE internationalconference on computer vision. Ieee, 1999, vol. 2, pp. 1150–1157.

[3] Sheng-Yu Wang, Oliver Wang, Richard Zhang, AndrewOwens, and Alexei A Efros, “Cnn-generated images aresurprisingly easy to spot... for now,” in Proceedings of theIEEE/CVF Conference on Computer Vision and Pattern Recog-nition, 2020, pp. 8695–8704.

[4] Hongtao Lu, Ruiming Shen, and Fu-Lai Chung, “Fragile wa-termarking scheme for image authentication,” Electronics Let-ters, vol. 39, no. 12, pp. 898–900, 2003.

[5] HongJie He, JiaShu Zhang, and Heng-Ming Tai, “A wavelet-based fragile watermarking scheme for secure image authenti-cation,” in International Workshop on Digital Watermarking.Springer, 2006, pp. 422–432.

[6] Xinpeng Zhang, Shuozhong Wang, Zhenxing Qian, andGuorui Feng, “Self-embedding watermark with flexiblerestoration quality,” Multimedia Tools and Applications, vol.54, no. 2, pp. 385–395, 2011.

[7] Xinpeng Zhang and Shuozhong Wang, “Fragile watermarkingwith error-free restoration capability,” IEEE Transactions onMultimedia, vol. 10, no. 8, pp. 1490–1499, 2008.

[8] Xinpeng Zhang and Shuozhong Wang, “Fragile watermarkingscheme using a hierarchical mechanism,” Signal processing,vol. 89, no. 4, pp. 675–679, 2009.

[9] Xinpeng Zhang, Shuozhong Wang, Zhenxing Qian, andGuorui Feng, “Reference sharing mechanism for watermarkself-embedding,” IEEE Transactions on Image Processing,vol. 20, no. 2, pp. 485–495, 2010.

[10] Xinpeng Zhang, Zhenxing Qian, Yanli Ren, and Guorui Feng,“Watermarking with flexible self-recovery quality based oncompressive sensing and compositive reconstruction,” IEEETransactions on Information Forensics and Security, vol. 6, no.4, pp. 1223–1232, 2011.

[11] Luisa Verdoliva, “Media forensics and deepfakes: anoverview,” IEEE Journal of Selected Topics in Signal Process-ing, vol. 14, no. 5, pp. 910–932, 2020.

[12] Xiuli Bi, Yang Wei, Bin Xiao, and Weisheng Li, “Rru-net: Theringed residual u-net for image splicing forgery detection,” inProceedings of the IEEE/CVF Conference on Computer Visionand Pattern Recognition Workshops, 2019, pp. 0–0.

[13] Yue Wu, Wael AbdAlmageed, and Premkumar Natarajan,“Mantra-net: Manipulation tracing network for detection andlocalization of image forgeries with anomalous features,” inProceedings of the IEEE/CVF Conference on Computer Visionand Pattern Recognition, 2019, pp. 9543–9552.

[14] Xiuli Bi, Yanbin Liu, Bin Xiao, Weisheng Li, Chi-Man Pun,Guoyin Wang, and Xinbo Gao, “D-unet: A dual-encoder u-netfor image splicing forgery detection and localization,” arXivpreprint arXiv:2012.01821, 2020.

[15] Fan Chen, Hongjie He, and Yaoran Huo, “Self-embedding wa-termarking scheme against jpeg compression with superior im-perceptibility,” Multimedia Tools and Applications, vol. 76, no.7, pp. 9681–9712, 2017.

[16] RO Preda and DN Vizireanu, “Watermarking-based image au-thentication robust to jpeg compression,” Electronics Letters,vol. 51, no. 23, pp. 1873–1875, 2015.

[17] Min-Jen Tsai and Chih-Cheng Chien, “Authentication and re-covery for wavelet-based semifragile watermarking,” OpticalEngineering, vol. 47, no. 6, pp. 067005, 2008.

[18] Lingzhi Li, Jianmin Bao, Ting Zhang, Hao Yang, Dong Chen,Fang Wen, and Baining Guo, “Face x-ray for more generalface forgery detection,” in Proceedings of the IEEE/CVF Con-ference on Computer Vision and Pattern Recognition, 2020, pp.5001–5010.

[19] Ido Yerushalmy and Hagit Hel-Or, “Digital image forgery de-tection based on lens and sensor aberration,” Internationaljournal of computer vision, vol. 92, no. 1, pp. 71–91, 2011.

[20] Arcangelo Ranieri Bruna, Giuseppe Messina, and SebastianoBattiato, “Crop detection through blocking artefacts analysis,”in International Conference on Image Analysis and Process-ing. Springer, 2011, pp. 650–659.

[21] Marco Fanfani, Massimo Iuliani, Fabio Bellavia, CarloColombo, and Alessandro Piva, “A vision-based fully auto-mated approach to robust image cropping detection,” SignalProcessing: Image Communication, vol. 80, pp. 115629, 2020.

[22] Basile Van Hoorick and Carl Vondrick, “Dissecting imagecrops,” arXiv preprint arXiv:2011.11831, 2020.

[23] Ian J Goodfellow, Jean Pouget-Abadie, Mehdi Mirza, BingXu, David Warde-Farley, Sherjil Ozair, Aaron Courville, andYoshua Bengio, “Generative adversarial networks,” arXivpreprint arXiv:1406.2661, 2014.

[24] Qichao Ying, Zhenxing Qian, Hang Zhou, Haisheng Xu, Xin-peng Zhang, and Siyi Li, “From image to imuge: Immunizedimage generation,” in Proceedings of the 29th ACM interna-tional conference on Multimedia, 2021, pp. 1–9.

[25] Minghao Yin, Yongbing Zhang, Xiu Li, and Shiqi Wang,“When deep fool meets deep prior: Adversarial attack onsuper-resolution network,” in Proceedings of the 26th ACMinternational conference on Multimedia, 2018, pp. 1930–1938.

[26] Seung-Min Mun, Seung-Hun Nam, Haneol Jang, DongkyuKim, and Heung-Kyu Lee, “Finding robust domain from at-tacks: A learning framework for blind watermarking,” Neuro-computing, vol. 337, pp. 191–202, 2019.

[27] David Khachaturov, Ilia Shumailov, Yiren Zhao, NicolasPapernot, and Ross Anderson, “Markpainting: Adver-sarial machine learning meets inpainting,” arXiv preprintarXiv:2106.00660, 2021.

[28] Laurent Dinh, David Krueger, and Yoshua Bengio, “Nice:Non-linear independent components estimation,” arXivpreprint arXiv:1410.8516, 2014.

[29] Laurent Dinh, Jascha Sohl-Dickstein, and Samy Bengio,“Density estimation using real nvp,” arXiv preprintarXiv:1605.08803, 2016.

[30] Diederik P Kingma and Prafulla Dhariwal, “Glow: Gener-ative flow with invertible 1x1 convolutions,” arXiv preprintarXiv:1807.03039, 2018.

[31] Rui Zhao, Tianshan Liu, Jun Xiao, Daniel PK Lun, and Kin-Man Lam, “Invertible image decolorization,” IEEE Transac-tions on Image Processing, vol. 30, pp. 6081–6095, 2021.

[32] Mingqing Xiao, Shuxin Zheng, Chang Liu, Yaolong Wang,Di He, Guolin Ke, Jiang Bian, Zhouchen Lin, and Tie-Yan Liu,“Invertible image rescaling,” in European Conference on Com-puter Vision. Springer, 2020, pp. 126–144.

[33] Yazhou Xing, Zian Qian, and Qifeng Chen, “Invertible imagesignal processing,” in Proceedings of the IEEE/CVF Confer-ence on Computer Vision and Pattern Recognition, 2021, pp.6287–6296.

[34] Xintao Duan, Kai Jia, Baoxia Li, Daidou Guo, En Zhang, andChuan Qin, “Reversible image steganography scheme basedon a u-net structure,” IEEE Access, vol. 7, pp. 9314–9323,2019.

[35] Rafia Rahim, Shahroz Nadeem, et al., “End-to-end trainedcnn encoder-decoder networks for image steganography,” inProceedings of the European Conference on Computer Vision(ECCV), 2018, pp. 0–0.

[36] Shao-Ping Lu, Rong Wang, Tao Zhong, and Paul L Rosin,“Large-capacity image steganography based on invertible neu-ral networks,” in Proceedings of the IEEE/CVF Conference onComputer Vision and Pattern Recognition, 2021, pp. 10816–10825.

[37] Chong Yu, “Attention based data hiding with generative adver-sarial networks,” in Proceedings of the AAAI Conference onArtificial Intelligence, 2020, vol. 34, pp. 1120–1128.

[38] Tsung-Yi Lin, Michael Maire, Serge Belongie, James Hays,Pietro Perona, Deva Ramanan, Piotr Dollar, and C LawrenceZitnick, “Microsoft coco: Common objects in context,” inEuropean conference on computer vision. Springer, 2014, pp.740–755.

[39] Russakovsky O Park E, Liu W, “Ilsvrc-2017,” 2017.[40] Olaf Ronneberger, Philipp Fischer, and Thomas Brox, “U-net:

Convolutional networks for biomedical image segmentation,”in International Conference on Medical image computing andcomputer-assisted intervention. Springer, 2015, pp. 234–241.

[41] Fisher Yu and Vladlen Koltun, “Multi-scale contextaggregation by dilated convolutions,” arXiv preprintarXiv:1511.07122, 2015.

[42] Jiren Zhu, Russell Kaplan, Justin Johnson, and Li Fei-Fei,“Hidden: Hiding data with deep networks,” in Proceedingsof the European conference on computer vision (ECCV), 2018,pp. 657–672.

[43] Tero Karras, Timo Aila, Samuli Laine, and Jaakko Lehtinen,“Progressive growing of gans for improved quality, stability,and variation,” arXiv preprint arXiv:1710.10196, 2017.

[44] Phillip Isola, Jun-Yan Zhu, Tinghui Zhou, and Alexei A Efros,“Image-to-image translation with conditional adversarial net-works,” in Proceedings of the IEEE conference on computervision and pattern recognition, 2017, pp. 1125–1134.

[45] Justin Johnson, Alexandre Alahi, and Li Fei-Fei, “Percep-tual losses for real-time style transfer and super-resolution,” inEuropean conference on computer vision. Springer, 2016, pp.694–711.

[46] Takeru Miyato, Toshiki Kataoka, Masanori Koyama, andYuichi Yoshida, “Spectral normalization for generative adver-sarial networks,” arXiv preprint arXiv:1802.05957, 2018.

[47] Leonid I Rudin, Stanley Osher, and Emad Fatemi, “Nonlineartotal variation based noise removal algorithms,” Physica D:nonlinear phenomena, vol. 60, no. 1-4, pp. 259–268, 1992.

[48] Diederik P Kingma and Jimmy Ba, “Adam: A method forstochastic optimization,” arXiv preprint arXiv:1412.6980,2014.

[49] Sergey Ioffe and Christian Szegedy, “Batch normalization: Ac-celerating deep network training by reducing internal covari-ate shift,” in International conference on machine learning.PMLR, 2015, pp. 448–456.

[50] Djork-Arne Clevert, Thomas Unterthiner, and Sepp Hochreiter,“Fast and accurate deep network learning by exponential linearunits (elus),” arXiv preprint arXiv:1511.07289, 2015.

[51] Zhou Wang, Alan C Bovik, Hamid R Sheikh, and Eero P Si-moncelli, “Image quality assessment: from error visibility tostructural similarity,” IEEE transactions on image processing,vol. 13, no. 4, pp. 600–612, 2004.

[52] Jiahui Yu, Zhe Lin, Jimei Yang, Xiaohui Shen, Xin Lu, andThomas S Huang, “Generative image inpainting with contex-tual attention,” in Proceedings of the IEEE conference on com-puter vision and pattern recognition, 2018, pp. 5505–5514.

[53] Ning Wang, Jingyuan Li, Lefei Zhang, and Bo Du, “Mu-sical: Multi-scale image contextual attention learning for in-painting.,” in IJCAI, 2019, pp. 3748–3754.

[54] Kamyar Nazeri, Eric Ng, Tony Joseph, Faisal Qureshi, andMehran Ebrahimi, “Edgeconnect: Structure guided im-age inpainting using edge prediction,” in Proceedings ofthe IEEE/CVF International Conference on Computer VisionWorkshops, 2019, pp. 0–0.

[55] Satoshi Iizuka, Edgar Simo-Serra, and Hiroshi Ishikawa,“Globally and locally consistent image completion,” ACMTransactions on Graphics (ToG), vol. 36, no. 4, pp. 1–14, 2017.

[56] Nobuyuki Otsu, “A threshold selection method from gray-levelhistograms,” IEEE transactions on systems, man, and cyber-netics, vol. 9, no. 1, pp. 62–66, 1979.

[57] Xudong Mao, Qing Li, Haoran Xie, Raymond YK Lau, ZhenWang, and Stephen Paul Smolley, “Least squares generativeadversarial networks,” in Proceedings of the IEEE interna-tional conference on computer vision, 2017, pp. 2794–2802.

[58] Jessica Fridrich, Tomas Pevny, and Jan Kodovsky, “Statisti-cally undetectable jpeg steganography: dead ends challenges,and opportunities,” in Proceedings of the 9th workshop on Mul-timedia & security, 2007, pp. 3–14.

[59] Tomas Pevny, Tomas Filler, and Patrick Bas, “Using high-dimensional image models to perform highly undetectablesteganography,” in International Workshop on InformationHiding. Springer, 2010, pp. 161–177.

[60] Abbas Cheddad, Joan Condell, Kevin Curran, and Paul Mc Ke-vitt, “Digital image steganography: Survey and analysis of cur-rent methods,” Signal processing, vol. 90, no. 3, pp. 727–752,2010.

[61] Ingemar Cox, Matthew Miller, Jeffrey Bloom, Jessica Fridrich,and Ton Kalker, Digital watermarking and steganography,Morgan kaufmann, 2007.

[62] Shumeet Baluja, “Hiding images in plain sight: Deepsteganography,” Advances in Neural Information ProcessingSystems, vol. 30, pp. 2069–2079, 2017.

[63] Jamie Hayes and George Danezis, “Generating stegano-graphic images via adversarial training,” arXiv preprintarXiv:1703.00371, 2017.

[64] Weixuan Tang, Shunquan Tan, Bin Li, and Jiwu Huang, “Au-tomatic steganographic distortion learning using a generativeadversarial network,” IEEE Signal Processing Letters, vol. 24,no. 10, pp. 1547–1551, 2017.

[65] Pin Wu, Yang Yang, and Xiaoqiang Li, “Stegnet: Mega im-age steganography capacity with deep convolutional network,”Future Internet, vol. 10, no. 6, pp. 54, Jun 2018.

[66] K. A. Zhang, A. Cuesta-Infante, and K. Veeramachaneni,“Steganogan: Pushing the limits of image steganography,”2019.

[67] Kunlin Liu, Dongdong Chen, Jing Liao, Weiming Zhang, HangZhou, Jie Zhang, Wenbo Zhou, and Nenghai Yu, “Jpeg robustinvertible grayscale,” IEEE Transactions on Visualization andComputer Graphics, 2021.

[68] Zhaoyang Jia, Han Fang, and Weiming Zhang, “Mbrs: En-hancing robustness of dnn-based watermarking by mini-batchof real and simulated jpeg compression,” arXiv preprintarXiv:2108.08211, 2021.

[69] Xiyang Luo, Ruohan Zhan, Huiwen Chang, Feng Yang, andPeyman Milanfar, “Distortion agnostic deep watermarking,” inProceedings of the IEEE/CVF Conference on Computer Visionand Pattern Recognition, 2020, pp. 13548–13557.

[70] Hongyi Zhang, Moustapha Cisse, Yann N Dauphin, and DavidLopez-Paz, “mixup: Beyond empirical risk minimization,”arXiv preprint arXiv:1710.09412, 2017.

[71] Eric Wengrowski and Kristin Dana, “Light field messagingwith deep photographic steganography,” in Proceedings of theIEEE/CVF Conference on Computer Vision and Pattern Recog-nition, 2019, pp. 1515–1524.

[72] Cheol Woo Park, Yong Ho Moon, and Il Kyu Eom, “Imagetampering localization using demosaicing patterns and singu-lar value based prediction residue,” IEEE Access, vol. 9, pp.91921–91933, 2021.

[73] Nhan Le and Florent Retraint, “An improved algorithm fordigital image authentication and forgery localization using de-mosaicing artifacts,” IEEE Access, vol. 7, pp. 125038–125053,2019.

[74] Jun-Ho Choi, Huan Zhang, Jun-Hyuk Kim, Cho-Jui Hsieh,and Jong-Seok Lee, “Evaluating robustness of deep imagesuper-resolution against adversarial attacks,” in Proceedings ofthe IEEE/CVF International Conference on Computer Vision,2019, pp. 303–311.

[75] Paarth Neekhara, Brian Dolhansky, Joanna Bitton, and Cris-tian Canton Ferrer, “Adversarial threats to deepfake detection:A practical perspective,” in Proceedings of the IEEE/CVF Con-ference on Computer Vision and Pattern Recognition, 2021, pp.923–932.

[76] Tero Karras, Samuli Laine, and Timo Aila, “A style-basedgenerator architecture for generative adversarial networks,” inProceedings of the IEEE/CVF Conference on Computer Visionand Pattern Recognition, 2019, pp. 4401–4410.