a robust detection algorithm for copy-move forgery in digital images

Post on 19-Jan-2016

52 Views

Category:

Documents

5 Downloads

Preview:

Click to see full reader

DESCRIPTION

A robust detection algorithm for copy-move forgery in digital images. Presented by Issam Laradji. Authors: Yanjun Cao, Tiegang Gao , Li Fan, Qunting Yang Course: COE589-Digital Forensics Date: 18 September, 2012. Outline. Introduction Challenges Background Concepts - PowerPoint PPT Presentation

TRANSCRIPT

A robust detection algorithm for copy-move forgery in digital images

Presented by Issam Laradji

Authors: Yanjun Cao, Tiegang Gao, Li Fan, Qunting Yang

Course: COE589-Digital Forensics

Date: 18 September, 2012

2

Outline

– Introduction– Challenges– Background Concepts– Related Work– Proposed Approach– Experimental Results– Summary

Most definitions were obtained from Wikipedia, others are from online tutorials

Introduction

Some statistics state that around 20% of accepted manuscripts are tempered– 1% are fraudulent manipulations– Innocent people can get framed of crimes they

didn’t commit– Negative social impact– Premature propaganda

3

Challenges

Sophisticated tools– 3D Max, Photoshop– Automated lighting, and processing that

conceal forgery Increase of large size images

– High-definition images– Much more costly to process

4

Background Concepts

Normal Distribution– Used to describe real-valued random variables that cluster around

a single mean value

– The most prominent distribution

– All distributions areas add up to 1, bell-shaped

– Allows for tractable  analysis

– Observational error in an experiment is usually assumed to follow a normal distribution

– Has a symmetric distribution about its mean

5Normal distribution formula

Background Concepts (2)

Energy of the image– Amount of information present:

• High energy: city, lots of details

• Low energy: plain, minimal details

Feature vector– N-dimensional vector of numerical features to represent

some object• Facilitates statistical analysis

• Explanatory “independent” variables used in procedures such as linear regression

6

Background Concepts (3)

Feature vector cont.– Linear regression can be used to model the relationship

between independent variable X (Feature vector) and dependent variable Y

– least square is commonly used for fitting Time Complexity

– The time complexity of an algorithm quantifies the amount of time taken by an algorithm to run as a function of the length of the variables representing the input

7

Background Concepts (4)

Global and local features– Global features represent details about the

whole image• Color distribution, brightness, and sharpness

• Faster to process

– Local features represent more finer details such as the relationship between pixels

• Similarities and differences between pixels

• Much more costly in processing

8

Background Concepts (5)

Eigenvector & Eigenvalue– An eigenvector of a square matrix is a non-zero vector that,

when multiplied by the matrix, yields a vector that is parallel to the original

– The eigenvalue is the scalar value that corresponds to the eigenvector λ 

In this case, [3;-3] is an eigenvector of A, with eigenvalue 1 (MATLAB syntax)

9

Background Concepts (6)

Principal analysis component– Mathematical procedure that uses orthogonal

transformation

– Converts correlated variables to linearly uncorrelated variables called principal components

– Principal components are guaranteed to be independent only if the data set is normally distributed

– Identifies patterns in data• Highlighting their similarities and differences

10

Background Concepts (7)

Principal analysis component cont.– Eigenvalue decomposition of data  correlation

– Eigenvalues obtained can measure weights of different image features

– Main advantage• Data compression without much loss of information

– Applications• Data mining, image processing, marketing, and chemical

research

11

Background Concepts (8)

Scale-invariant feature transform (or SIFT) – An algorithm in computer vision to detect and describe

local features in images

– Local image features helps in object recognition

– Invariant to image scale and rotation

– Robust to changes in illumination, noise, and minor changes in viewpoint

– Applications• Object/face recognition, navigation, gesture recognition, and

tracking

12

Background Concepts (9)

Discrete Cosine transform– Transforms image from spatial to “frequency domain”

in which it can be efficiently encoded

– Discard high frequency “sharp variations” components which refines the details of the image

– Focuses on the low frequency “smooth variations”, holds the base of an image

13Zigzag scanningDCT basis (64)

Background Concepts (10)

Discrete Cosine transform cont.– Removes redundancy between neighboring pixels

– Prepares image for compression / quantization

– Quantization:

• Maps large set of values to smaller set

• Reduces the number of bits needed to store the coefficients by removing less important high frequency.

14

Background Concepts (11)

Why DCT?– Approximates better with fewer coefficients as

compared to other contemporary transforms• However wavelet transforms is the new trend

– Less space required to represent the image features, hence easier to store in memory

– Applications:• Lossy compression for .mp3 and .jpg

15

Related Work

Straightforward approach – Compare each group of pixels with the rest, and check

for similarities!

– Very impractical, exponential time complexity

– False positives could be high Related work

1. Exhaustive search

2. Fridrich used DCT-based features for duplication detection

• Sensitive to variations (additive noise)16

Related Work (2)

3. Haung et al. increased performance by reducing feature vector dimensions

However, none considered multiple copy-move forgery

4. Popescu: PCA-based feature, – Can endure additive noise – Low in accuracy

17

Related Work (3)

5. Luo proposed color features as well as block intensity ratio

6. Bayram et al. applied Fourier-Mellin transform to each block, then projected to one dimension to form the feature vector

7. B. Mahdian, and S. Saic used a method based on blur moments invariants to locate the forgery regions

8. X. Pan, and S. Lyu took the advantages of SIFT features to detect the duplication regions

However, all these are of higher time complexity than the proposed approach!

18

Proposed Approach

Basically, the algorithm divides the original image into overlapping blocks, then similarities between these blocks are calculated, based on some threshold the duplicated regions are highlighted in the output image

19

Proposed approach advantages (contributions)

Improved version of copy-move forgery detection algorithm– Lower feature vector dimension

– Robust to various attacks: multiple copy-move forgery, Gaussian blurring, and noise contamination

– Lower time complexity

20

Step 1- dividing the image into blocks Say we have an input image of size m x n If its not gray scale

The image is converted to Grayscale using the formulae: I=0.228R+0.587G+0.114B

Human eye is most sensitive to green and red That’s why most weights are on green and red

21

Green channel gives the clearest image

Step 1- dividing the image into blocks (2)

The input image is split into overlapping blocks The standard block size is 8 x 8

22

n

B

B

m

B

BGenerates

(m-b+1)(n-b+1) = N Blocks

Let ‘N’, and ‘b’ be the number of blocks obtained, and the height of the block respectively

Each block differ by one row or one column by its preceding block

Step 1- dividing the image into blocks (3)

23

155 155 155 158 158 156 158 159

155 155 155 158 158 156 158 159

155 155 155 158 158 156 158 159

155 155 155 158 158 156 158 159

155 155 155 158 158 156 158 159

151 151 151 154 157 156 156 156

155 155 155 156 157 158 156 153

149 149 149 153 155 154 153 154

Original image …

Block size : 8 x 8

155 155 155 158 158 156 158 159

155 155 155 158 158 156 158 159

155 155 155 158 158 156 158 159

155 155 155 158 158 156 158 159

155 155 155 158 158 156 158 159

151 151 151 154 157 156 156 156

155 155 155 156 157 158 156 153

149 149 149 153 155 154 153 154

Dividing into blocks

Complexity: O(N) where N is the number of blocks

Step 2 – Applying DCT transform

For each block, DCT is applied We get DCT coefficients matrix

24

155 155 155 158

155 155 155 158

155 155 155 158

155 155 155 158

420.75 37.70297 -3.25 4.136577

-2.98619 0.926777 2.1744 -0.32322

-0.25 -5.44081 0.75 -0.72292

2.589912 0.676777 -0.63007 0.573223

DCT Transfor

m

Original Sample block DCT coefficient block

Step 2 – Applying DCT transform (1) The block is compared with its 64 DCT

basis to get the correlation coefficients

25

DCT basis (64)

155 155 155 158 158 156 158 159

155 155 155 158 158 156 158 159

155 155 155 158 158 156 158 159

155 155 155 158 158 156 158 159

155 155 155 158 158 156 158 159

151 151 151 154 157 156 156 156

155 155 155 156 157 158 156 153

149 149 149 153 155 154 153 154

… … … … … … … …

.. … … … … … … …

Discrete Coefficients

Step 2 – Applying DCT transform (2) The transformation allows us to focus on the low

frequency coefficients which hold the basis of the image Zigzag extraction is done so that coefficients are in order

of increasing frequency– Allows for zeroing the high frequency blocks

Time complexity: O(N x b x b)

26(a) the Lena image (b) Zigzag order scanning (c) the reconstruction image of Lena by using 1/4 DCT coefficients.

Step 3 – feature extraction

The coefficients matrix are divided and represented by four parts: C1,C2, C3, and C4

p_ratio=c_area/m_area is approximately 0.79 The circle block represents the low frequency, hence

decreasing the computation cost without affecting efficiency

27

Generate matching feature :

,

, ( , _ , 1,2,3,4)_i i

i

f x yv f x y c area i

c area

1 2 3 4 : , , ,feature vector V v v v v

420.75 37.70297 -3.25 4.136577

-2.98619 0.926777 2.1744 -0.32322

-0.25 -5.44081 0.75 -0.72292

2.589912 0.676777 -0.63007 0.573223

DCT coefficient block

C2C2C1C1

C3C3C4C4

Step 3 – feature extraction (2) Each v is quantized by its corresponding c_area Four features that represent the matrix are obtained vi is the mean of the coefficients value, corresponding to each ci

28

DCT coefficient block

Matching features generated:

2

2

_ 4

_ 4 / 4 1, 2,3,4i

r

c area r

c area for i

1

420.75 37.70297 2.98619 0.926777v

≒ 145.2746

2

( 3.25) 4.136577 2.1744 ( 0.32322)v

≒ 0.8715

3

0.75 ( 0.72292) ( 0.63007) 0.573223v

≒ -0.0095

420.75 37.70297 -3.25 4.136577

-2.98619 0.926777 2.1744 -0.32322

-0.25 -5.44081 0.75 -0.72292

2.589912 0.676777 -0.63007 0.573223

C2C2C1C1

C3C3C4C4

Step 3 – feature extraction (3)

The extracted features are invariant to a lot of processing operations according to the results below

Time complexity of feature extraction: O(N x 4 x b x b)

29

Step 4 – Matching

The extracted feature vectors are arranged in a matrix A

A is then lexicographically sorted , with time complexity of O(N log N)

Each element (vector) of A is compared to each subsequent vector to check if the thresholds Dsimilar, Nd, are satisfied i.e. the equations:

30

1

2

( 1)( 1)

N B N B

V

VA

V

42

1

_ ( , ) ( )k ki i j i i j similar

k

m match A A v v D

2 2( , )i i j i i j i i j dd V V x x y y N

Step 4 – Matching (2)

31

145.2746,0.8715, 0.0095, 0.7716

196.6815,0.7681,0.2534,1.6032

145.1673,0.9771,0.0032, 1.0438

179.2334,4.8015, 0.4968,4.1904

A

Pr :

0.4, 25similar d

edict value

D N

2 2 2 21 116_ ( , ) (145.2746 196.6815) (0.8715 0.7681) ( 0.0095 0.2534) ( 0.7716 1.6032) 51.4625m match A A similarD

Not Similar

Not Similar

2 2 2 21 117_ ( , ) (145.2746 145.1673) (0.8715 0.9771) ( 0.0095 0.0032) ( 0.7716 1.0438) 0.3113m match A A similarD

Step 4 – Matching (3)

32

145.2746,0.8715, 0.0095, 0.7716

196.6815,0.7681,0.2534,1.6032

145.1673,0.9771,0.0032, 1.0438

179.2334,4.8015, 0.4968,4.1904

A

Pr : 0.4, 25similar dedict value D N

2 2 2 21 117_ ( , ) (145.2746 145.1673) (0.8715 0.9771) ( 0.0095 0.0032) ( 0.7716 1.0438) 0.3113m match A A similarD

2 2

1 117( , ) 1 100 1 81d V V ≒ 127.28 5dN

SimilarSimilar

1 (1,1)V

117 (100,81)V

Detected image

Step 5 – Displaying duplicated regions Finally, regions showing the duplicated

regions are expected to be displayed

33

The green rectanglesindicate a duplicated region

The computational complexities of extraction methods are compared

Time complexity analysis

As claimed, the total computation complexity:– O(N)+O(Nxbxb)+O(Nx4xbxb)+O(4NxlogN)

• Where N, b are the number of blocks and the height of the block respectively

– Questionable?• The computation complexity of matching was not

calculated which could be O(NxN)• However, they stated that their computational

complexity is dominated by the matching blocks

34

Experimental results - environment Photoshop 8.0 2.59 GHz, AMD processor Matlab2009a software First dataset

– Gray images of size of 128 x 128– DVMM lab at Columbia University

Second dataset– uncompressed colour PNG images of size 768 x 521– the Kodak Corporation

Third dataset – Internet collection of images of size 1600 x 1000

35

Experimental results - Setting Thresholds

Detection accuracy rate (DAR) and False positive rate (FPR)

– psis & “psis tilde” are set as the copy region, and the detected copy respectively

– psit and “psit tilde” are set as the tampered region and detected tampered respectively

– Questionable?• Vague formulas

• Nothing in the paper have shown what the symbols really mean

• Accuracy check is normally calculated in ratios

36

Experimental results - Setting Thresholds (2) Selecting the circle representation for matching features

extraction can be challenging– Therefore, 200 images are randomly chosen from the three

datasets

– Series of forgeries are applied to them

– Different circle radius ranging from 2 to 6 are used, with 1 increment

– Optimum at r = 4, as shown in the diagram below

37

Experimental results - Setting Thresholds (3) Choosing the threshold parameters, b, Dsimilar, Nd, and Nnumber, is also

challenging Colour images:

– The optimal values: 8, 0.0015, 120 and 5 for b, Dsimilar, Nd, and Nnumber,, respectively

Gray images:– The optimal values: 8, 0.0025, 25 and 5 for b, Dsimilar, Nd, and Nnumber,

respectively

38

Experimental results – Effective testing To test the proposed method, gray images of different sizes are chosen:

– Tempered regions of sizes: 32x32, 64x64, 96x96, 128x128, are tested

39The detection results (from left to right is the original image, tampered image, detection results)

Experimental results – Robustness and accuracy test

40

Signal-to-noise ratio (SNR): level of a desired signal to the level of background noise

(a)–(b) DAR/FPR performance with SNR, and (c)–(d) DAR/FPR performance with Gaussian blurring

Experimental results – Robustness and accuracy test (2)

41

DAR/FPR curves for DCT, DCT-improved, PCA, FMT, and Proposed methods when the duplicated region is 64 pixels 64 pixels. (a)–(b) with different SNR levels, and (c)–(d) with Gaussian blurring

Experimental results – Demonstration

42The detection results for non-regular copy-move forgery

Experimental results – Demonstration (2)

43

The test results for multiple copy-move forgery under a mixed operation

Experimental results – Demonstration (3)

44The top row are tampered images with duplicated region size of 32 pixels × 32 pixels. Shown below are the detection results using our algorithm

Experimental results – Demonstration (4)

45

c) The analyzed image (Python script)•Duplicated regions were detected

b) the manipulated image

a) the original image

Experimental results – Demonstration (5)

46

c) The analyzed image (Python script)• Used --blcoldev=0.05• False positive• Duplicate regions were not detected

b) Manipulated image

a) Original image

Experimental results – Demonstration (6)

47

c) The analyzed image (Python script)• Partial part of the duplicated region was detected

b) the manipulated image

a) the original image

Summary The chart illustrates a summary of how the proposed

algorithm works

48Flowchart of the proposed scheme

Summary(2)

Automatic and efficient detection algorithm for copy-move forgery have been presented

Contributions– Outperforms contemporary algorithms in speed and storage

– Robust to various attacks: multiple copy-move forgery, Gaussian blurring, and noise contamination

– Different way of representing blocks (circles), reducing memory requirements

49

References

A robust detection algorithm for copy-move forgery in digital images; By: Yanjun Cao a,*, Tiegang Gao b, Li Fan a, Qunting Yang

Wikipedia: The Free Encyclopedia. Wikimedia Foundation, Inc. 22 July 2004. Web. 10 Aug. 2004

cao2012-image-forgery-slides.ppt; By: Li-Ting Liao The Discrete Cosine Transform (DCT): Theory and

Application1; By: Syed Ali Khayam A tutorial on Principal Components Analysis; By:

Lindsay I Smith

50

top related