Loading [MathJax]/extensions/TeX/boldsymbol.js
PRNU-Based Content Forgery Localization Augmented With Image Segmentation | IEEE Journals & Magazine | IEEE Xplore

PRNU-Based Content Forgery Localization Augmented With Image Segmentation


The proposed multi-orientation image forgery localization algorithm. An image in question is first segmented into multi-scale irregular superpixels and the forgery probab...

Abstract:

The advances in image editing and retouching technology have enabled an unskilled person to easily produce visually indistinguishable forged images. To detect and localiz...Show More

Abstract:

The advances in image editing and retouching technology have enabled an unskilled person to easily produce visually indistinguishable forged images. To detect and localize such forgeries, many image forensic tools rely on visually imperceptible clues, e.g. the subtle traces or artifacts introduced during image acquisition and processing, and slide a regular, typically square, detection window across the image to search for discrepancies of specific clues. Such a sliding-window paradigm confines the explorable neighborhood to a regular grid and inevitably limits its capability in localizing forgeries in a broad range of shapes. While image segmentation that generates segments adhering to object boundaries might be a promising alternative to the sliding window-based approach, it is generally believed that the potential of the segmentation-based detection scheme is hindered by object removal forgeries where meaningful segments are often unavailable. In this work, we take forgery localization based on photo-response non-uniformity (PRNU) noise as an example and propose a segmentation-based forgery localization scheme that exploits the local homogeneity of visually indiscernible clues to mitigate the limitations of existing segmentation approaches that are merely based on visually perceptible content. We further propose a multi-orientation localization scheme that integrates the forgery probabilities obtained with image segmentation and multi-orientated detection windows. The multi-orientation scheme aggregates the complementary strengths of image segmentation and multi-oriented detection window in localizing the object insert and object removal forgeries. Experimental results on a public realistic tampering image dataset demonstrate that the proposed segmentation-based and multi-orientation forgery localization schemes outperform existing state-of-the-art PRNU-based forgery localizers in terms of both region and boundary F_{1} scores.
The proposed multi-orientation image forgery localization algorithm. An image in question is first segmented into multi-scale irregular superpixels and the forgery probab...
Published in: IEEE Access ( Volume: 8)
Page(s): 222645 - 222659
Date of Publication: 07 December 2020
Electronic ISSN: 2169-3536

Funding Agency:


CCBY - IEEE is not the copyright holder of this material. Please follow the instructions via https://creativecommons.org/licenses/by/4.0/ to obtain full-text articles and stipulations in the API documentation.
SECTION I.

Introduction

The ease of manipulating images with increasingly powerful image editing tools has also led to the growing appearance of digitally altered forgeries, which raise serious concerns about the credibility of digital images. Developing image forgery detection and localization techniques, therefore, becomes essential under the circumstances where digital images serve as critical evidence. A variety of approaches have been proposed to detect and localize image forgeries. Active techniques such as digital watermarking [1], [2] are effective in verifying the authenticity of an image, but the requirement of embedding additional information at the creation of the image limits their widespread use. On the contrary, passive or blind techniques rely on the image data itself without requiring any pre-embedded information and thus have attracted great interests over the past two decades. The passive image forgery detection and localization techniques in the literature mainly fall into five categories:

  1. The first category is based on specific statistical properties of natural images such as higher-order wavelet statistics [3], [4], image texture features [5] and residual-based features [6], [7].

  2. The second category includes the techniques that seek the traces left by specific image manipulations such as resampling [8], [9], contrast enhancement [10], [11], median filtering [12], copy-move manipulations [13] and JPEG compression [14]–​[16].

  3. The third category relies on the regularities or constraints that make images physically plausible. For instance, anomalies in lighting [17], shadows [18], [19], reflections [20] and perspective constraints [21] have been exploited for exposing image manipulations.

  4. The fourth category exploits the artifacts introduced in the image acquisition pipeline of digital camera. These artifacts can be either caused by specific processing components such as demosaicking artifacts [22], camera response function abnormalities [23], [24], optical aberrations [25] and imaging sensor noise [26]–​[28], or caused by the complex interplay of multiple components, such as local noise levels [29].

  5. Inspired by the success of convolutional neural networks (CNN) in various multimedia tasks, the fifth category comprises the techniques that automatically learn the features for image forgery detection or localization using CNN [30]–​[32].

Besides the above-mentioned techniques, recent years have seen some works on fusing the outputs of multiple forensic detectors, e.g. under the framework of fuzzy logic [33], the Dempster-Shafer theory of evidence (DSTE) [34], simple logical rules [35], [36] or Bayesian inference [37], but the performance of the overall system depends on that of each individual forensic detector. To localize the forgeries in an image, a prevailing paradigm is to move a regular, typically square, detection window across the image and analyze the forensic clues in the detection window at different locations. Such a sliding-window paradigm allows for the convenient and efficient processing of the image but also has two inherent limitations: 1) The regular shape of the detection window hinders the detection of forgeries in various possible shapes. For instance, a square detection window is not suitable for detecting tree branches or human limbs. 2) Due to the object boundary unawareness of the sliding-window detection scheme, pixels of different natures, e.g. some pixels are forged and others are pristine, are often contained in the same window. Jointly processing those ‘heterogeneous’ pixels without any distinction makes the detection highly error-prone.

The above limitations can be alleviated by image segmentation. Local structure of visual content is typically exploited by segmentation algorithms to divide an image into a number of ‘segments’ or ‘superpixels’, each of which consists of a group of connected pixels that are perceptually or statistically similar. The resultant superpixels provide informative clues for inferring the boundaries of the forged regions. However, given the variability of possible forgeries, the limitations of image segmentation are also apparent: 1) There might be no distinct local structure information in homogeneous regions to guide the segmentation. 2) The boundaries of the forged regions do not always align with the object boundaries. For instance, it is not uncommon to copy some collateral background pixels along with the forged object to make it blended into the background more naturally. In such cases, pixels of different natures, i.e. the collateral forged pixels and their neighboring authentic pixels, might be grouped into the same superpixel and cause further uncertainties.

In this article, we take forgery localization based on photo-response non-uniformity (PRNU) noise as an example and show that it is possible to mitigate the limitations of segmentation merely based on visual content by incorporating the local homogeneity of visually imperceptible clues. The main contributions of our work can be summarized as follows.

  • We identify the essential criteria that a superpixel segmentation algorithm should satisfy for the task of forgery localization and propose a multi-scale segmentation-based localization scheme in cooperation with some segmentation algorithm that fulfills the identified criteria.

  • We propose a novel superpixel segmentation algorithm that encodes both visual content and visually imperceptible clues for content forgery localization. Different from all previous works that only use visual boundary information for image segmentation, the proposed algorithm incorporates the information from visually imperceptible forensic clues to guide the segmentation for forgery localization.

  • We propose a multi-orientation fusion strategy that aggregates the complementary strengths of image segmentation and multi-oriented detection in localizing various types of image forgeries.

  • We conduct comprehensive experiments on the effects of various segmentation algorithms on different forgery types, i.e. hard-boundary and soft-boundary forgeries, in terms of both region and boundary F_{1} scores.

The rest of this article is organized as follows. In Section II, we will revisit the background of PRNU-based forgery localization and some related work. In Section III, the details of the proposed segmentation-based and multi-orientation forgery localization schemes will be given. Section IV presents the experimental results and analysis on realistic forgeries. Finally, Section V concludes the work.

SECTION II.

Related Work

In this section, we will first revisit the background of the PRNU-based forgery localization. PRNU noise mainly arises from the inhomogeneity of silicon wafers introduced during imaging sensor fabrication and manifests itself as the pixel-to-pixel variation in light sensitivity. It is present in every image and practically unique to the sensor that captures the image. Therefore, its absence can be used to signify the forgery in an image under investigation, provided that the reference PRNU \boldsymbol {z} of the source camera is available for comparison. For this purpose, a binary hypothesis test is typically carried out for each pixel i in the image’s noise residual \boldsymbol {\omega } , i.e. the difference between the original image and its denoised version:\begin{equation*} \begin{cases} h_{0}: \omega _{i} = \upsilon _{i}\\ h_{1}: \omega _{i} = z_{i} + \upsilon _{i} \end{cases}\tag{1}\end{equation*} View SourceRight-click on figure for MathML and additional features. where \upsilon _{i} is Gaussian random noise. If pixel i is forged, z_{i} is supposed to be absent in \omega _{i} , which corresponds to the null hypothesis h_{0} . Otherwise, the alternative hypothesis h_{1} is accepted.

Due to the weak nature of \boldsymbol {z} , the above hypothesis testing usually requires joint processing of a large amount of pixels, so it is typically conducted in a sliding-window fashion by calculating the normalized cross-correlation \rho _{i} over the pixels within a w{\times }w detection window \Omega _{i} centered at each pixel i :\begin{equation*} \rho _{i} = \frac {\sum _{j \in \Omega _{i}}(\omega _{j}-\bar {\omega })(z_{j}-\bar {z})}{\sqrt {\sum _{j \in \Omega _{i}}(\omega _{j}-\bar {\omega })^{2}}\sqrt {\sum _{j \in \Omega _{i}}(z_{j}-\bar {z})^{2}}},\tag{2}\end{equation*} View SourceRight-click on figure for MathML and additional features. where the overbar stands for the respective arithmetic mean within the detection window. Provided that the distributions of the test statistic \rho _{i} under h_{0} and h_{1} , i.e. p(\rho _{i}|h_{0}) and p(\rho _{i}|h_{1}) , are available, either Neyman-Person [26] criterion or Bayes’ rule [27], [28] can be used to make the final decision on the authenticity of pixel i . We will adopt the Bayes’ rule as in [27], [28] for the easier formulation of multi-clue fusion under a uniform probability framework. We model p(\rho _{i}|h_{0}) as a zero-mean Gaussian distribution \mathcal {N}(0, \hat {\sigma }_{0}^{2}) and estimate the variance using the images captured by the cameras that are different from the source camera. While p(\rho _{i}|h_{1}) can also be modeled as a Gaussian distribution \mathcal {N}(\hat {\rho }_{i}, \hat {\sigma }_{1}^{2}) , its estimation is more challenging since the expected correlation \hat {\rho }_{i} is highly dependent on the image content. To address this problem, we adopt the correlation predictor [26] based on local image features, namely the image intensity, texture, signal flattening, texture-intensity features and their second-order terms, to estimate the mean \hat {\rho }_{i} and variance \hat {\sigma }_{1}^{2} of p(\rho _{i}|h_{1}) . Finally, the forgery probability of pixel i is formulated as \begin{align*} P_{i}=&\frac {p(\rho _{i}|h_{0})}{p(\rho _{i}|h_{0})+p(\rho _{i}|h_{1})} \\=&\left ({1+\exp \left ({\frac {\rho _{i}^{2}}{2\hat {\sigma }_{0}^{2}}-\frac {\left ({\rho _{i}-\hat {\rho }_{i}}\right)^{2}}{2\hat {\sigma }_{1}^{2}}-\ln \frac {\hat {\sigma }_{1}}{\hat {\sigma }_{0}}}\right)}\right)^{-1}.\qquad \tag{3}\end{align*} View SourceRight-click on figure for MathML and additional features.

In the above sliding window-based framework, the test statistic \rho _{i} is calculated over all the pixels within the detection window. It becomes problematic if the detection window falls across the boundary between the pristine and the forged regions. In such case, both the pristine and forged pixels contribute to the calculation of \rho _{i} , which gives rise to the chance of the pixels near the boundaries being undetected. This problem can be alleviated by means of hard segmentation, as shown in Chierchia et al. ’s work [38], but its rely on accurate manual segmentation makes it infeasible in most practical scenarios. They further proposed a method [39] based on guided filtering [40] to enforce the calculation of \rho _{i} more aligned with the boundaries of objects. However, its effectiveness relies on the visually perceptible boundary information of the image and thus is largely compromised for object removal forgeries where no meaningful boundary information is available. In light of the fact that meaningful forgeries usually appear in a group of connected pixels, another direction of improvement lies in taking into account pixels’ spatial dependencies in the final decisions. For instance, Chierchia et al. [27] sought to improve the localization performance by modeling the decision statistic as a Markov Random Field (MRF) and recasting the forgery localization as an optimization problem under the Bayesian framework. This method only considers the neighborhood consistency of the final decisions but ignores the content similarity within the neighborhood.

To allow for localizing small-sized forgeries, Korus and Huang [28] proposed a segmentation-based strategy that calculates the test statistic \rho _{i} only using the pixels with similar intensity levels as the central pixel i , as well as two fusion strategies, i.e. multi-scale and adaptive-window, that combine multiple candidate forgery probability maps obtained with detection windows of different sizes. Additionally, they adopted the Conditional Random Fields (CRF) model to incorporate the content-dependent neighborhood interactions into the final decisions. Although only the intensity difference between individual pixels is considered, the incorporation of image content in the simple segmentation-based strategy and the CRF-based model results in significant performance improvement [16], [28]. Further along this path, it poses an interesting question that whether further improvement can be gained if the image content is exploited in a more sophisticated and comprehensive way. This motivates us to resort to explicit image segmentation that fully exploits the local structure and homogeneity of the image for improving the forgery localization performance.

SECTION III.

Proposed Methods

A. Segmentation-Based Forgery Localization Scheme

For PRNU-based forgery localization, a straightforward way to exploit image content would be segmenting the image into a number of superpixels according to the visual content and calculating the forgery probability for each superpixel. However, compared to the methods based on regular detection windows that are able to generate pixel-wise probability maps, this will result in a low-resolution forgery probability map because the pixels belonging to the same superpixel are assigned with the same probability. Consequently, the chance of mis-detection for object removal forgeries will be considerably increased if the size of the superpixel is not appropriately specified. For this reason, we apply image segmentation at different scales and fuse the resultant forgery probability maps to form a single informative probability map. The framework of the proposed segmentation-based multi-scale localization scheme is illustrated in Fig. 1.

FIGURE 1. - The segmentation-based forgery localization scheme. For more details about the Adaptive Window (AW) and Multi-Scale (MS) fusion algorithms, please refer to Section IV-C or [28].
FIGURE 1.

The segmentation-based forgery localization scheme. For more details about the Adaptive Window (AW) and Multi-Scale (MS) fusion algorithms, please refer to Section IV-C or [28].

1) Multi-Scale Image Segmentation

Given the variety of superpixel segmentation algorithms, not all of them suffice for our purpose. We identify a few important criteria that a superpixel segmentation algorithm should satisfy for the task of forgery localization:

  • Boundary adherence. This criterion measures the agreement between the boundaries of the objects and the resultant superpixels. A superpixel algorithm with good boundary adherence effectively avoids segmenting different objects or different parts of an object into the same superpixel, thus reducing the risk of generating heterogeneous superpixels containing both pristine and forged pixels for object insert forgeries.

  • Controllability over superpixel number. Some algorithms do not provide direct control over the number of generated superpixels. As the segmentation needs to be carried out at different scales, easy control over the number of generated superpixels is preferable.

  • Balanced segmentation granularity. Some superpixel algorithms generate superpixels of heavily unbalanced size. The over-sized superpixels substantially increase the chance of generating heterogeneous superpixels while the under-sized superpixels reduce the reliability of the detection. Thus, an algorithm capable of generating superpixels of balanced size is desirable.

Based on these criteria, we shortlist three superpixel algorithms for multi-scale segmentation:
  • Simple Linear Iterative Clustering (SLIC) [41]: SLIC algorithm iteratively updates superpixel centers and assigns pixels to their closest centers in the 5-dimensional pixel color and coordinate space until the algorithm converges. Its simplicity, computational efficiency and high-quality overall segmentation results make it one of the most widely used superpixel algorithms in various applications.

  • Entropy Rate Superpixel segmentation (ERS) [42]: ERS algorithm considers each pixel as a vertex in a graph and formulates the segmentation as an optimization problem of graph topology. It incrementally optimizes an objective function consisting of the entropy rate of random walks on the graph and a balancing term to achieve homogeneous and similar-sized superpixels. ERS exhibits remarkable boundary adherence, which is desirable for localizing object insert forgeries.

  • Extended Topology Preserving Segmentation (ETPS) [43], [44]: ETPS algorithm initially partitions the image into the desired number of regular superpixels and continuously modifies the boundary pixels in a coarse-to-fine manner by optimizing an objective function that encodes information about colors, positions, boundaries and topologies. The diversity of information encoded in the objective function and the efficient coarse-to-fine optimization make ETPS one of the state-of-the-art superpixel algorithms in terms of both segmentation quality and efficiency.

For more details about the above superpixel algorithms, we refer the reader to the original papers. Note that the multi-scale framework in Fig. 1 is similar to the work of Zhang et al. [45]. However, their analysis is only limited to the SLIC algorithm and falls short in the impact of segmentation on different types of image forgeries.

2) Exploiting Homogeneity of PRNU

As mentioned in Section II, only relying on the visual content for image segmentation might become problematic for object removal forgeries where no distinct visual information is available. Although this problem can be mitigated by the multi-scale strategy, it would be beneficial if the additional information from PRNU can be incorporated to guide the segmentation at each scale. The perceptually invisible PRNU not only provides useful clues when salient visual information is unavailable but also serves as a supplement to the visual information to eliminate the ambiguity in regions containing complex patterns or structures. In what follows, we will describe how the homogeneity of PRNU can be integrated to guide the segmentation.

Let N be the number of pixels in the image and K be the desired number of superpixels. The partitioning of the image into superpixels can be represented by a set of random variables \boldsymbol {s}=(s_{1}, \ldots, s_{N}) , where s_{i}\in \{1,..,K\} denotes the superpixel to which pixel i belongs. Following the ETPS algorithm in [43], [44], we formulate the image segmentation as an optimization problem with the following energy function \begin{align*} \mathcal {E}(\boldsymbol {s}, \boldsymbol {\mu }, \boldsymbol {c}, \boldsymbol {P})=&\sum _{i}{\mathcal {E}_{col}(s_{i}, c_{s_{i}})}+\lambda _{pos}\sum _{i}{\mathcal {E}_{pos}(s_{i}, \mu _{s_{i}})} \\&+\,\lambda _{b}\sum _{i}\sum _{j\in \mathcal {N}_{8}(i)}{{\mathcal {E}_{b}(s_{i}, s_{j})}}\!+\!\mathcal {E}_{topo}(\boldsymbol {s})+\mathcal {E}_{size}(\boldsymbol {s}) \\&+\,\lambda _{etp}\sum _{i}{\mathcal {E}_{etp}(s_{i}, P_{s_{i}})},\tag{4}\end{align*} View SourceRight-click on figure for MathML and additional features. where \boldsymbol {c}{=}(c_{1}, \ldots, c_{K}) , \boldsymbol {\mu }{=}(\mu _{1}, \ldots, \mu _{K}) , and \boldsymbol {P}{=}(P_{1}, \ldots, P_{K}) are respectively the set of the mean colors, the mean positions, and the forgery probabilities for K superpixels. The details of different energy terms in Eq. (4) are given below.

Appearance Coherence:\begin{equation*} \mathcal {E}_{col}(s_{i}, c_{s_{i}}) = ||\mathcal {I}(i)-c_{s_{i}}||_{2}^{2}.\tag{5}\end{equation*} View SourceRight-click on figure for MathML and additional features. This term measures the squared Euclidean distance in the CIELAB color space between pixel i and the mean color of the superpixel that i belongs to. It encourages the color homogeneity within superpixels.

Shape Regularity:\begin{equation*} \mathcal {E}_{pos}(s_{i}, \mu _{s_{i}}) = ||\mathcal {L}(i)-\mu _{s_{i}}||_{2}^{2}.\tag{6}\end{equation*} View SourceRight-click on figure for MathML and additional features. It measures the coordinate distance between pixel i and the mean position of the superpixel that i belongs to. This term encourages compact superpixels of regular shapes.

Boundary Length:\begin{equation*} \mathcal {E}_{b}(s_{i},s_{j})= \begin{cases} 1,& \text {if } s_{i} \neq s_{j} \\ 0,& \text {otherwise} \end{cases}\tag{7}\end{equation*} View SourceRight-click on figure for MathML and additional features. This term measures the length of borders between superpixels. Irregular boundaries normally have a length longer than regular boundaries, thus this term penalizes local irregularities in the superpixel boundaries. Note that encouraging the shape and boundary regularity of the generated superpixels does not contradict with our aim to localize forgeries of various shapes because forged objects or regions rarely have wiggly shapes and boundaries.

Topology Preservation: The term \mathcal {E}_{topo}(\boldsymbol {s}) is used to enforce that each superpixel is composed of a single connected component without holes. This is done by checking if changing the assignment of a boundary pixel will violate connectivity.

Minimal Size: The term \mathcal {E}_{size}(\boldsymbol {s}) returns a value of \infty if the size of each resultant superpixel is less than 1/4 of its initial size. It encourages superpixels with similar sizes.

PRNU Homogeneity:\begin{equation*} \mathcal {E}_{etp}(s_{i}, P_{s_{i}})=\left ({255\cdot H(P_{s_{i}})}\right)^{2},\tag{8}\end{equation*} View SourceRight-click on figure for MathML and additional features. where P_{s_{i}} is the forgery probability of superpixel s_{i} and H(P_{s_{i}})=-P_{s_{i}}\log _{2}P_{s_{i}}-(1-P_{s_{i}})\log _{2}(1-P_{s_{i}}) is the entropy of a Bernoulli process characterized by P_{s_{i}} . If a superpixel contains both pristine and forged pixels, the estimated forgery probability P_{s_{i}} will tend to be closer to 0.5, where H(P_{s_{i}}) attains its maximum value. Thus, H(P_{s_{i}}) is an indicator of the homogeneity of PRNU within a superpixel. Actually, a similar measurement has been used in [28], [46] to assess the confidence or reliability of the estimated forgery probability. The factor 255 is used to promote \mathcal {E}_{etp}(s_{i}, P_{s_{i}}) to the comparable level as the color energy term \mathcal {E}_{col}(s_{i}, c_{s_{i}}) . Hereafter, we will refer to this algorithm as entropy-guided ETPS (EG-ETPS) algorithm.

We adopt the coarse-to-fine optimization framework in [43] to minimize the objective function in Eq. (4). As shown in Algorithm 1, the algorithm starts by equally partitioning the image into K non-overlapping square grids of size \lfloor \sqrt {N/K}\rfloor {\times } \lfloor \sqrt {N/K}\rfloor px, where N is the total number of pixels and \lfloor \cdot \rfloor is the floor function. Each grid is considered as an initial superpixel and further divided into a number of macropixels of size M{\times }M pixels. M is initially set to 64 but will be reset to \lfloor \sqrt {N/K}\rfloor /2 if \lfloor \sqrt {N/K}\rfloor < 64 to allow for macropixel division at the beginning of the algorithm. As demonstrated in Fig. 2, a macropixel is a square image block that will be jointly considered to evaluate the objective energy function, i.e. what pixel i is referred to in Eq. (4), and at the finest level, each macropixel only consists of one pixel. Initially, the statistics of mean color \boldsymbol {\mu } , mean position \boldsymbol {c} and forgery probability \boldsymbol {P} are computed for each superpixel based on the corresponding statistics of the macropixels comprising the superpixel. Then, the following coarse-to-fine optimization iteratively updates the segmentation by proposing small local changes at boundaries.

FIGURE 2. - Demonstration of macropixel division for the coarse-to-fine optimization. The regions enclosed by red lines are superpixels and the regular grids bounded by blue dotted lines are macropixels.
FIGURE 2.

Demonstration of macropixel division for the coarse-to-fine optimization. The regions enclosed by red lines are superpixels and the regular grids bounded by blue dotted lines are macropixels.

Algorithm 1 Entropy-Guided ETPS Algorithm

Input:

\boldsymbol {\mathcal {I}} : image to be segmented;

\boldsymbol {r} : camera reference PRNU;

\boldsymbol {n} : noise residual extracted from \boldsymbol {\mathcal {I}} ;

\widetilde {\boldsymbol {I}} , \widetilde {\boldsymbol {T}} : intensity and texture feature maps;

\widetilde {\boldsymbol {S}} , \widetilde {\boldsymbol {\Upsilon }} : signal-flattening and texture-intensity feature maps;

\boldsymbol {\Psi } : camera parameters incl. predictors under h_{0} and h_{1} ;

\boldsymbol {\lambda } : weighting factors in the objective energy function;

K : the desired number of superpixels;

Output:

\boldsymbol {s} : resultant segmentation of the image;

1.

Partition \boldsymbol {\mathcal {I}} into K non-overlapping square superpixels;

2.

Partition each superpixel into regular macropixels of size 64{\times }64 and set current maximal macropixel size m=64 ;

3.

Calculate the initial energy, Eq. (4), for each superpixel;

4.

do

5.

if m < 16 then

6.

set \lambda _{etp}=0

7.

end if

8.

Initialize a FIFO list with all boundary macropixels;

9.

while list is not empty do

10.

Pop out boundary macropixel i from list;

11.

if invalid_connectivity(i ) then

12.

continue

13.

end if

14.

\hat {s}_{i} = \mathop {\mathrm {arg\,min}} _{s_{i}}{\mathcal {E}(\boldsymbol {s},\boldsymbol {\mu },\boldsymbol {c},\boldsymbol {P})}

15.

if \hat {s}_{i} is updated then

16.

Incrementally update \boldsymbol {\mu } , \boldsymbol {c} and \boldsymbol {P} for the

17.

two superpixels involved;

18.

Append any boundary macropixel j in the

19.

4-connected neighborhood of i to the list;

20.

end if

21.

end while

22.

if m>1 then

23.

Bisect or quadrisect any divisible macropixel;

24.

Update maximal macropixel size m\leftarrow \lceil \frac {m}{2}\rceil ;

25.

else

26.

break

27.

end if

28.

while

First, all the boundary macropixels, i.e. those with at least one adjacent macropixel belonging to a different superpixel, are put into a FIFO priority list and popped out one by one to check if changing the label of the popped macropixel i will violate connectivity. If the label change is admissible and results in an energy decrease, we assign macropixel i to one of its neighboring superpixels that gives the lowest energy. Meanwhile, we incrementally update the mean colors, mean positions and forgery probabilities of the two involved superpixels (i.e. the superpixel that macropixel i belonged to and the superpixel that macropixel i is newly assigned to) and append the boundary macropixels adjacent to i at the end of the list. This process is repeated until the list is empty, at which point the optimization will proceed to next finer level by bisecting or quadrisecting each macropixel to a smaller size.

Unlike other terms in Eq. (4), the PRNU homogeneity term can only be reliably evaluated when the size of macropixel is sufficiently large due to the weak nature of PRNU. The above coarse-to-fine framework makes it possible to reliably integrate the PRNU homogeneity information at coarse levels. We set \lambda _{etp}=0 when the maximal macropixel size m{ < }16 when the information from PRNU is no longer reliable. One of the key steps in Algorithm 1 is the incremental update of the mean colors, mean positions and forgery probability for calculating the energy value for each superpixel. Just as other terms, the PRNU homogeneity term can also be efficiently updated in an incremental manner as the assignment of a macropixel changes. The reader is referred to the Appendix for more details.

To see how the PRNU homogeneity term affects the segmentation, we show an example of object removal forgery in Fig. 3. We use the ratio of the average forgery probability \bar {P}_{f} in the forged regions to the average forgery probability \bar {P}_{p} in the pristine regions as the indicator to differentiate the forged regions from the pristine ones. For the image of size 1920\times 1080 px, we vary the superpixel number K from 1000 to 3000 and show the result in Fig. 3f. We can observe a higher ratio for EG-ETPS when the superpixel number K < 2200 , which indicates an easier separation of the forged and the pristine regions. The forgery probability maps when K=1800 are shown in Fig. 3e. It can be observed that EG-ETPS outputs a more homogeneous and coherent probability map.

FIGURE 3. - The effect of PRNU homogeneity term, i.e. Eq. (8), on the segmentation results. For ETPS and EG-ETPS, we used the same parameters except for 
$\lambda _{etp}$
, which is set to 2 for the results shown above.
FIGURE 3.

The effect of PRNU homogeneity term, i.e. Eq. (8), on the segmentation results. For ETPS and EG-ETPS, we used the same parameters except for \lambda _{etp} , which is set to 2 for the results shown above.

B. Multi-Orientation Forgery Localization Scheme

1) Multi-Orientation Forgery Detection

Most existing image forgery detectors apply square detection windows. Such an isotropic detection scheme inherently limits the capability to detect arbitrary-shaped and arbitrary-oriented forgeries. For instance, subtle forgeries such as human body limbs or tree branches might be undetectable with a square detection window. Although this issue can be mitigated by the use of detection windows of smaller size, the reliability will also be compromised at smaller scales. To allow for more accurate forgery localization, the various shapes and orientations of the forged regions need to be taken into consideration in the design of the localization framework. Inspired by the great success of faster R-CNN [47] in detecting objects based on anchor boxes, which are a set of predefined bounding boxes of certain scales and aspect ratios, we extend the multi-scale forgery localization scheme in [28] by adopting detection windows of various aspect ratios and orientations at each scale.

Based on the multi-scale framework in [28], we replace the square detection window at each scale with detection windows of multiple aspect ratios and orientations, as illustrated in Fig. 4. To reduce the computation, we only use 5 scales, i.e. w{\in }\{32,48,64,96,128\} , rather than the 7 scales in [28]. For each scale, we use detection windows of 3 aspect ratios (1:1, 1:2 and 1:3) and rotate the window with a specific aspect ratio by a specific set of orientations. Specifically, we consider 11 orientations: 1 orientation \{0\} for aspect ratio 1:1, 4 orientations \left\{{0,\frac {\pi }{4},\frac {\pi }{2},\frac {3\pi }{4}}\right\} for aspect ratio 1:2, and 6 orientations \left\{{0,\frac {\pi }{6},\frac {\pi }{3},\frac {\pi }{2},\frac {2\pi }{3},\frac {5\pi }{6}}\right\} for aspect ratio 1:3. Such configuration ensures that each window does not have too much overlap with its rotated versions. Note that the numbers of pixels within the detection windows at the same scale are approximately the same. Taking the scale w{=}32 as an example, the sizes of detection windows of the 3 aspect ratios are configured to 32{\times }32 , 22{\times }44 and 18{\times }54 px. For this reason, the offline-trained correlation predictor at one scale can be used to predict the intra-camera correlations, i.e. p(\rho _{i}|h_{1}) in Eq. (3), for detection windows of different orientations at the same scale.

FIGURE 4. - Detection windows used in our multi-orientation forgery localization scheme.
FIGURE 4.

Detection windows used in our multi-orientation forgery localization scheme.

2) Multi-Orientation Fusion

Having obtained 11 candidate forgery probability maps at each scale (corresponding to each of the 11 multi-oriented detection windows), we need a fusion scheme to form a single informative probability map. At first thought, a simple pixel-wise maximum fusion rule, i.e. selecting the largest value of the candidate probabilities for each pixel, will suffice for the task since we aim to detect any possible forgery, but this will also introduce substantial false positives, i.e. mis-labeling pristine pixels as forged. Ideally, the best detection result can be obtained if the detection window is perfectly aligned with the forged region. Thus, it is reasonable to accept the forgery probability calculated with the detection window that agrees best with the segmentation result. Suppose P_{b} , b\in \{1,\ldots,11\} are the candidate forgery probabilities obtained with 11 detection windows centered at pixel i . We adopt the following fusion strategy for pixel i :\begin{equation*} P_{i} = rP_{b^{\star }}+(1-r)P_{s_{i}},\tag{9}\end{equation*} View SourceRight-click on figure for MathML and additional features. where P_{s_{i}} is the forgery probability of the superpixel s_{i} that pixel i belongs to, b^{\star } is the index of the detection window that has the best agreement with s_{i} , i.e.\begin{equation*} b^{\star }= \mathop {\mathrm {arg\,max}} _{b\in \{1,\ldots,11\}}{W_{b}\cap s_{i}}.\tag{10}\end{equation*} View SourceRight-click on figure for MathML and additional features. r is the proportion of the detection window W_{b^{\star }} that overlaps with s_{i} . It measures how much of the probability obtained with detection window W_{b^{\star }} contributes to the final result. r=1 when the detection window W_{b^{\star }} completely falls within a superpixel, which means we only rely on the result obtained with one of the detection windows and is similar to the traditional detector merely based on sliding windows. r < 1 when the detection window W_{b^{\star }} falls across two or more superpixels and (1-r)P_{s_{i}} will be used to compensate for the pixels falling outside of the superpixel that pixel i belongs to. Note that the above fusion is performed at each scale based on the superpixel segmentation and forgery probability maps obtained with 11 detection windows. We introduce a parameter \xi to control the average size of superpixels relative to the size of detection window at each scale. Specifically, if the size of the detection window is w{\times }w px, the number of superpixels K is set to \begin{equation*} K= \lfloor \frac {N}{\xi w^{2}}\rceil,\tag{11}\end{equation*} View SourceRight-click on figure for MathML and additional features. where N is the total number of pixels in the image and \lfloor \cdot \rceil is the rounding operator. To see how the fusion strategy affects the final forgery probability when r < 1 , we analyze the following simplified scenarios as shown in Fig. 5:

  • Scenario I:

    Two forged regions with different color appearances are segmented into two different superpixels s_{i} and s_{j} , which often occurs inside a forged region. Suppose the forgery probability obtained with the detection window W_{b^{\star }} is P_{{b^{\star }}} and the forgery probabilities for s_{i} and s_{j} are P_{s_{i}} and P_{s_{j}} , respectively. The pixels considered in this scenario are all forged pixels, so we can simply assume that P_{s_{i}} \approx P_{s_{j}} \approx P_{{b^{\star }}} since the correlation predictor has been designed to account for different color appearances. Therefore, the final forgery probability P_{i}\approx P_{{b^{\star }}} , which means that the fusion is equivalent to selecting the probability corresponding to the detection window that agrees best with the segmentation results.

  • Scenario II:

    Two neighboring regions, with one forged and the other pristine, have different color appearances and are segmented into two superpixels s_{i} and s_{j} . This case usually occurs for object insert forgeries, where the inserted object usually has a different color appearance from the background. In such case, the pixels within the detection window falling outside the forged region lead to an attenuated P_{{b^{\star }}} . The above fusion strategy compensates the attenuated P_{{b^{\star }}} by adding a term (1-r)P_{s_{i}} . As P_{s_{i}} is calculated over forged pixels and is expected to be >P_{{b^{\star }}} , it results in a P_{i}>P_{{b^{\star }}} compensating for the attenuation caused by the pixels falling outside the forged region.

  • Scenario III:

    In this scenario, the forged and pristine regions have the same or similar color appearance, which often occurs in the case of object removal forgery. Due to the lack of distinguishable color appearance, some parts of the two regions are quite likely to be segmented into the same superpixel. For instance, the segmentation may end up with two superpixels s_{i} and s_{j} separated by the red line. If we assume that s_{i} and window W_{b^{\star }} are equally possible to contain heterogeneous pixels, this scenario is similar to the detection merely based on regular detection windows. In practice, because most superpixel algorithms will try to utilize as much local information as possible to perform the segmentation, the above fusion is expected to deliver comparable or even better performance than the methods merely based on regular detection windows.

Finally, we apply the multi-scale fusion strategies as proposed in [28] for the fused forgery probability maps obtained at different scales. The framework of our proposed multi-orientation forgery localization scheme is demonstrated in Fig. 6.
FIGURE 5. - Three simplified scenarios for multi-orientation forgery probability fusion. The pixel of interest is highlighted in yellow and different color appearances are highlighted in different colors.
FIGURE 5.

Three simplified scenarios for multi-orientation forgery probability fusion. The pixel of interest is highlighted in yellow and different color appearances are highlighted in different colors.

FIGURE 6. - The multi-orientation forgery localization scheme. For more details about the Adaptive Window (AW) and Multi-Scale (MS) fusion algorithms, please refer to Section IV-C or [28].
FIGURE 6.

The multi-orientation forgery localization scheme. For more details about the Adaptive Window (AW) and Multi-Scale (MS) fusion algorithms, please refer to Section IV-C or [28].

SECTION IV.

Experiments

A. Datasets

Our experiments were conducted on the realistic image tampering dataset (RTD) [28], [48], which contains 220 realistic forgeries captured by 4 different cameras: Canon 60D, Nikon D90, Nikon D7000, Sony \alpha 57 (each camera responsible for 55 forgery images). The images in the RTD dataset are 1920{\times }1080 px RGB images stored in the TIFF format and cover various types of forgeries including object insert and object removal forgeries. The RTD dataset provides a good benchmark to evaluate the performance of camera-based content authentication techniques. Considering the nature of our proposed methods, it is reasonable to conduct evaluations separately on the object insert and object removal forgeries. However, the wide variety of forgeries in the RTD dataset makes it hard to simply divide them into two categories. Therefore, we divided the RTD dataset into three subsets:

  1. Hard-boundary Forgeries (HBF): This subset contains the forged images with visually distinguishable boundaries between the pristine and the forged regions. It mainly consists of the forgeries created by object insert, object replacement, color altering, etc.

  2. Soft-boundary Forgeries (SBF): This subset contains the forged images with visually smooth and indistinguishable boundaries between the pristine and the forged regions. It mainly consists of the forgeries created by background texture synthesis or content-aware filling.

  3. Mixed-boundary Forgeries (MBF): This subset contains the images with both hard and soft boundaries between the pristine and the forged regions.

The dataset division was done by visually inspecting each image in the RTD dataset. This results in 100, 80 and 40 images in the HBF, SBF and MBF subsets, respectively. Some examples in these three subsets are shown in Fig. 7. Considering the relatively small size of the MBF subset, we merged MBF with both HBF and SBF to create two subsets, i.e. HBF+MBF (140 images) and SBF+MBF (120 images), for evaluating the localization performance on forgeries of two boundary types.
FIGURE 7. - Examples of forged images in hard-boundary forgeries (HBF), soft-boundary forgeries (SBF) and mixed-boundary forgeries (MBF).
FIGURE 7.

Examples of forged images in hard-boundary forgeries (HBF), soft-boundary forgeries (SBF) and mixed-boundary forgeries (MBF).

B. Evaluation Protocols

The localization performance is commonly evaluated by the region F_{1} score. For an image I_{i}, 1\leq i\leq \mathcal {S} , let G_{i} be its binarized ground truth forgery map and L_{i}(\tau) be the binary forgery map output by a forgery localization algorithm with a threshold \tau . The region F_{1} score is defined as \begin{equation*} F_{r}(G_{i},L_{i}(\tau)) = \frac {2\cdot \mathcal {P}_{r}(G_{i},L_{i}(\tau)) \cdot \mathcal {R}_{r}(G_{i},L_{i}(\tau))}{\mathcal {P}_{r}(G_{i},L_{i}(\tau)) + \mathcal {R}_{r}(G_{i},L_{i}(\tau))},\tag{12}\end{equation*} View SourceRight-click on figure for MathML and additional features. where \mathcal {P}_{r}(G_{i},L_{i}(\tau)) and \mathcal {R}_{r}(G_{i},L_{i}(\tau)) are the precision (the fraction of correctly detected forgery pixels) and the recall (the fraction of ground-truth forgery pixels detected), respectively. Like in [28], we measured the overall performance with a curve of average \bar {F}_{r}(\tau) score (each point in the curve is the F_{r} score averaged over \mathcal {S} images for a given \tau ) and an average peak \hat {F}_{r} score (i.e. the average of the highest achievable F_{r} score over all possible thresholds for \mathcal {S} images):\begin{align*} \begin{cases} \bar {F}_{r}(\tau)=\dfrac {1}{\mathcal {S}}\displaystyle \sum \nolimits _{i=1}^{\mathcal {S}} F_{r}(G_{i},L_{i}(\tau))\\[5pt] \hat {F}_{r}=\dfrac {1}{\mathcal {S}}\displaystyle \sum \nolimits _{i=1}^{\mathcal {S}} \underset {\tau \in (0,1)}{\max }F_{r}(G_{i},L_{i}(\tau)). \end{cases}\tag{13}\end{align*} View SourceRight-click on figure for MathML and additional features.

F_{r} score measures the localization performance by considering the overlapping area (in terms of pixel counts) between G_{i} and L_{i}(\tau) . However, as pointed out in [48], the region F_{1} score is insufficient to accurately assess the contour adherence between the ground-truth and localized forgery regions. In fact, the contour conveys key information about the object shape and provides inferable clues to uncover the potential forged object, which is particularly desirable in the scenario of soft-boundary forgeries. We thus also evaluated the localization performance with boundary F_{1} score [49], which has been widely used to assess the boundary adherence for image segmentation. For an image I_{i}, 1\leq i\leq \mathcal {S} , let B_{i} be the contour map of the binarized ground-truth forgery map and D_{i}(\tau) be the contour map of the binary forgery map output by a forgery localization algorithm given a threshold \tau . The boundary F_{1} score is defined as \begin{equation*} F_{b}\left ({B_{i}, D_{i}(\tau)}\right)=\frac {2\cdot \mathcal {P}_{b}\left ({B_{i}, D_{i}(\tau)}\right) \cdot \mathcal {R}_{b}\left ({B_{i}, D_{i}(\tau)}\right)}{\mathcal {P}_{b}\left ({B_{i}, D_{i}(\tau)}\right) + \mathcal {R}_{b}\left ({B_{i}, D_{i}(\tau)}\right)},\tag{14}\end{equation*} View SourceRight-click on figure for MathML and additional features. where \mathcal {P}_{b}(B_{i}, D_{i}(\tau)) and \mathcal {R}_{b}(B_{i}, D_{i}(\tau)) are the boundary precision and the boundary recall given a distance tolerance \theta :\begin{align*} \begin{cases} \mathcal {P}_{b}(B_{i}, D_{i}(\tau))=\dfrac {1}{|D_{i}(\tau)|}\displaystyle \sum \nolimits _{z\in D_{i}(\tau)}[\![d(z, B_{i}) < \theta]\!] \\[5pt] \mathcal {R}_{b}(B_{i}, D_{i}(\tau))=\dfrac {1}{|B_{i}|}\displaystyle \sum \nolimits _{z\in B_{i}}[\![d(z, D_{i}(\tau)) < \theta]\!]. \end{cases}\tag{15}\end{align*} View SourceRight-click on figure for MathML and additional features. Here, [\![z]\!] is the Iversons bracket notation, i.e. [\![z]\!]=1 if z=\text {true} and 0 otherwise, d(\cdot) is the Euclidean distance, and \theta determines whether a boundary point has a match or not and is set to 0.75% of the image diagonal throughout our evaluation. Similarly to the region F_{1} score, we evaluate the overall boundary quality with a curve of average \bar {F}_{b}(\tau) score and an average peak \hat {F}_{b} score:\begin{align*} \begin{cases} \bar {F}_{b}(\tau)=\dfrac {1}{\mathcal {S}}\sum _{i=1}^{\mathcal {S}} F_{b}(B_{i}, D_{i}(\tau))\\[5pt] \hat {F}_{b}=\dfrac {1}{\mathcal {S}}\sum _{i=1}^{\mathcal {S}} \underset {\tau \in (0,1)}{\max }F_{b}(B_{i}, D_{i}(\tau)). \end{cases}\tag{16}\end{align*} View SourceRight-click on figure for MathML and additional features.

C. Evaluated Algorithms and Parameter Settings

We considered the following single-orientation forgery localization algorithms for the comparison with our proposed schemes:

  • Simple Thresholding (ST) based algorithm [26]: For ST, we first generated a binarized forgery map by comparing the forgery probability map obtained by a detection window of 128{\times }128 px with a threshold \tau \in [{0,1}] . Then we removed the connected forged regions with pixels fewer than 64{\times }64 px and applied image dilation with a disk kernel with a radius of 16 px to generate the final decision map. Note that ST algorithm is a single-scale detector with no image content involved in the final decision-making process.

  • Single-Orientation Multi-Scale (SO+MS) fusion based algorithm [28]: SO+MS formulates the forgery localization as an image labeling problem and solves it by the conditional random field (CRF) model. The data term of the CRF model is the average of the threshold-drifted [16], [28] forgery probability maps obtained with single-oriented detection windows at 7 different scales, and the neighborhood interaction of image content is encoded in the regularization term. The final binary decision map is obtained by optimizing the CRF model.

  • Single-Orientation Adaptive Window (SO+AW) fusion based algorithm [28]: SO+AW aims to fuse the forgery probability maps obtained with single-oriented detection windows of 7 scales. Starting from the smallest scale, it looks for a sufficiently confident decision (i.e. forgery probability that is far from 0.5) for each location in the image. If the decision at a smaller scale is not confident enough, it proceeds to the next larger scale until a sufficiently confident decision and an agreement between two consecutive scales are reached. Finally, the final binary decision map is obtained by applying the threshold drift strategy and optimizing the CRF model.

  • Segmentation-Guided (SG) based algorithm [28]: SG algorithm calculates the forgery probability by only considering the pixels with an intensity value close to the central pixel (the average L_{1} distance in RGB space less than 15) within a detection window of 128\times 128 px. It implements the idea of image segmentation but only exploits the intensity difference between individual pixels. Similarly to SO+AW and SO+MS, the threshold drift strategy and CRF are applied to obtain the final binary decision map.

Note that the notations ‘SO+MS’, ‘SO+AW’, ‘SG’ used in this article correspond to the ‘MSF’, ‘AW+’ and ‘SG+’ algorithms in [28]. We use the notations ‘AW’ and ‘MS’ to denote the adaptive window and multi-scale strategies proposed in [28] for multi-scale fusion. For our proposed segmentation-based schemes, we will use the notation ‘SEG+FUSION’ to represent the FUSION\in {AW, MS} fusion of the probability maps calculated based on the superpixels generated by SEG\in {SLIC, ERS, ETPS, EG-ETPS} algorithm. Similarly, for our proposed multi-orientation forgery localization scheme, we will use the notation ‘SEG+MO+FUSION’ to represent the FUSION\in {AW, MS} fusion of the integrated probability maps obtained with multi-oriented detection windows and the superpixel algorithm SEG\in {SLIC, ERS, ETPS, EG-ETPS}.

For SO+MS, SO+AW and SG, we used exactly the same parameters as summarized in Table 2 of [28]. For the segmentation algorithms used in this work, their parameters are given as follows:

  • SLIC [41]: We set both the compactness parameter and the iteration number of pixel assignment and centroid updating to 10.

  • ERS [42]: As suggested in [42], we set the weighting factor of the balancing term \lambda ^\prime {=}0.5 and the Gaussian kernel bandwidth \sigma {=}5.0 for calculating the pixel similarities.

  • ETPS [43]: We set the weighting factors of both the shape regularity term and the boundary length term to 0.2, i.e. \lambda _{pos}=0.2 and \lambda _{b}=0.2 .

  • EG-ETPS: We used exactly the same parameters as ETPS except for the weighting factor \lambda _{etp} of the PRNU homogeneity term. As can be expected, the setting of \lambda _{etp} will depend on the quality of PRNU. Thus we empirically set \lambda _{etp}=2.5\cdot \exp (-(R^{2}-1)^{2}/0.3) . R^{2} is the adjusted R-squared coefficient for the correlation predictor trained at the scale of 64\times 64 px, which is a good indicator of the quality of PRNU. This results in a \lambda _{etp}\in [{1.8,2.2}] for the four cameras in the RTD dataset.

For our proposed multi-orientation scheme, we set \xi =0.2 in Eq. (11), which controls the average size of superpixels relative to the window size at each scale.
TABLE 1 Segmentation-Based vs. Single-Orientation Forgery Localization Schemes in Terms of the Highest Average Region \bar{F}_{r} Scores and Boundary \bar{F}_{b} Scores Across All Thresholds. ‘SO’ Stands for the Fusion of Probability Maps Obtained With Single-Oriented Detection Windows as Proposed in [28]. The Best Performance on Each Dataset is Highlighted in Bold
Table 1- 
Segmentation-Based vs. Single-Orientation Forgery Localization Schemes in Terms of the Highest Average Region 
$\bar{F}_{r}$
 Scores and Boundary 
$\bar{F}_{b}$
 Scores Across All Thresholds. ‘SO’ Stands for the Fusion of Probability Maps Obtained With Single-Oriented Detection Windows as Proposed in [28]. The Best Performance on Each Dataset is Highlighted in Bold
TABLE 2 Multi-Orientation vs. Single-Orientation Forgery Localization Schemes in Terms of the Highest Average Region \bar{F}_{r} Scores and Boundary \bar{F}_{b} Scores Across All Thresholds
Table 2- 
Multi-Orientation vs. Single-Orientation Forgery Localization Schemes in Terms of the Highest Average Region 
$\bar{F}_{r}$
 Scores and Boundary 
$\bar{F}_{b}$
 Scores Across All Thresholds

1) Results for Segmentation-Based Localization Schemes

In this experiment, we applied the adaptive window and multi-scale fusion strategies directly on the probability maps obtained by calculating the forgery probability on each generated superpixel. To make the average superpixel size consistent with the detection window size at each scale, we specified the desired superpixel number K of each segmentation algorithm as \lfloor \frac {N}{w^{2}}\rceil , where N is the number of pixels in the image, w{\times }w is the size of the detection window at a specific scale and \lfloor \cdot \rceil is the rounding operator. We only used 5 scales, i.e. w\in \{32, 48, 64, 96, 128\} , for our segmentation-based schemes, while for the compared methods SO+AW and SO+MS proposed in [28], we still used 7 scales.

we show the average score curves \bar {F}_{r} and \bar {F}_{b} in Fig. 8 as well as the average peak scores \hat {F}_{r} and \hat {F}_{b} in Fig. 9, respectively. For easy comparison, we also summarized the highest average \bar {F}_{r} , \bar {F}_{b} and the performance gain of the best segmentation-based scheme relative to the corresponding single-orientation scheme in Table 1. Note that ST and SG are not included in Table 1 because they have inferior performance than multi-scale methods as clearly shown in Fig. 8. We can see that even with only 5 scales, the segmentation-based forgery localization schemes are able to deliver better performance than SO+AW and SO+MS on the entire RTD dataset. The advantage of the segmentation-based schemes is more evident if the performance is measured by \bar {F}_{b} (the 3rd and 4th columns of Fig. 8), which is increased by 12.4% and 11.8% for AW and MS multi-scale fusions, respectively. Regarding the localization performance on two different boundary types, it is not surprising to see that segmentation-based schemes outperform the single-orientation schemes by large margins on HBF+MBF, with 9.6% increase in \bar {F}_{r} and 21.8% increase in \bar {F}_{b} when MS is applied. However, due to the lack of distinct local information to guide the segmentation, segmentation-based schemes perform slightly worse than those based on detection windows when localizing the forgeries in SBF+MBF.

FIGURE 8. - Segmentation-based vs. single-orientation forgery localization schemes in terms of average score curves. The 
$\bar {F}_{r}$
 and 
$\bar {F}_{b}$
 on the RTD dataset (220 images), HBS+MBF (140 images) and SBF+MBF (120 images) subsets are shown in the 1st, 2nd and 3rd row, respectively.
FIGURE 8.

Segmentation-based vs. single-orientation forgery localization schemes in terms of average score curves. The \bar {F}_{r} and \bar {F}_{b} on the RTD dataset (220 images), HBS+MBF (140 images) and SBF+MBF (120 images) subsets are shown in the 1st, 2nd and 3rd row, respectively.

FIGURE 9. - Segmentation-based vs. single-orientation forgery localization schemes in terms of average peak scores. The 
$\hat {F}_{r}$
 and 
$\hat {F}_{b}$
 on the RTD dataset (220 images), HBS+MBF (140 images) and SBF+MBF (120 images) subsets are shown in the 1st, 2nd and 3rd column, respectively. The vertical dash lines indicate the best performance of single-orientation schemes, i.e. ST, SG, SO+AW and SO+MS.
FIGURE 9.

Segmentation-based vs. single-orientation forgery localization schemes in terms of average peak scores. The \hat {F}_{r} and \hat {F}_{b} on the RTD dataset (220 images), HBS+MBF (140 images) and SBF+MBF (120 images) subsets are shown in the 1st, 2nd and 3rd column, respectively. The vertical dash lines indicate the best performance of single-orientation schemes, i.e. ST, SG, SO+AW and SO+MS.

As for the comparison between different segmentation algorithms, we can observe that EG-ETPS benefits from the integration of PRNU homogeneity information and slightly outperforms other segmentation algorithms. It is worth mentioning that the benefits of EG-ETPS are mainly reflected in detecting the boundaries of the forgeries, so its overall performance gain over other segmentation algorithms depends on the amount of soft boundaries and is more evident when the performance is measured in terms of boundary measurement F_{b} . For the other three segmentation algorithms, ETPS delivers generally better performance than SLIC and ERS. An interesting observation is that the superpixels generated by ERS exhibit excellent boundary adherence and provide the best \hat {F}_{b} performance among the four segmentation algorithms on HBF+MBF (see Fig. 9), but the trade-off is that its capability in detecting forgeries in SBF+MBF is greatly compromised.

2) Results for Multi-Orientation Localization Schemes

In this experiment, we aim to compare the performance of our proposed multi-orientation localization schemes and the methods based on single-oriented detection windows. The comparison results are shown in Fig. 10 and Fig. 11. Similarly, for easy comparison, we also summarized the highest average measures \bar {F}_{r} and \bar {F}_{b} across all thresholds and the performance gain of the best multi-orientation scheme relative to the corresponding single-orientation scheme in Table 2.

FIGURE 10. - Multi-orientation vs. single-orientation forgery localization schemes. The 
$\bar {F}_{r}$
 and 
$\bar {F}_{b}$
 on the RTD dataset (220 images), HBS+MBF (140 images) and SBF+MBF (120 images) subsets are shown in the 1st, 2nd and 3rd row, respectively.
FIGURE 10.

Multi-orientation vs. single-orientation forgery localization schemes. The \bar {F}_{r} and \bar {F}_{b} on the RTD dataset (220 images), HBS+MBF (140 images) and SBF+MBF (120 images) subsets are shown in the 1st, 2nd and 3rd row, respectively.

FIGURE 11. - Multi-orientation vs. single-orientation forgery localization schemes in terms of average peak scores. 
$\hat {F}_{r}$
 and 
$\hat {F}_{b}$
 on the RTD dataset (220 images), HBS+MBF (140 images) and SBF+MBF (120 images) subsets are shown in the 1st, 2nd and 3rd column, respectively. The vertical dash lines indicate the best performance of single-orientation schemes, i.e. ST, SG, SO+AW and SO+MS.
FIGURE 11.

Multi-orientation vs. single-orientation forgery localization schemes in terms of average peak scores. \hat {F}_{r} and \hat {F}_{b} on the RTD dataset (220 images), HBS+MBF (140 images) and SBF+MBF (120 images) subsets are shown in the 1st, 2nd and 3rd column, respectively. The vertical dash lines indicate the best performance of single-orientation schemes, i.e. ST, SG, SO+AW and SO+MS.

We can see that the MS fusion strategy achieves considerably better performance than the AW fusion for our proposed multi-orientation schemes. A closer inspection revealed that AW is more likely to introduce false positives especially in the regions where PRNU is substantially attenuated. In addition, compared to the single-orientation schemes, significant performance improvement can be observed for the multi-orientation localization schemes when the MS fusion strategy is applied, with the highest \bar {F}_{r} and \bar {F}_{b} increased by 7.5% and 18.7% respectively on the entire RTD dataset. The multi-orientation schemes are even more advantageous on the HBF+MBF subset as evidenced by the performance gains of 10.9% in \bar {F}_{r} and 28.6% in \bar {F}_{b} . While on the SBF+MBF subset, the multi-orientation schemes still deliver comparable performance as the best two single-orientation schemes SO+AW and SO+MS.

Another important observation for the proposed multi-orientation schemes is that the performance gap between the four segmentation algorithms becomes very small. For the segmentation-based localization schemes, the difference between the segmentation algorithms are much more noticeable when localizing the forgeries in the SBF+MBF subset (see the fourth row of Fig. 8). However, for the multi-orientation localization schemes, the capability of localizing soft-boundary forgeries mainly stems from the detection based on multi-oriented detection windows rather than the segmentation algorithms, which narrows the performance gaps between different segmentation algorithms. Some examples of forgery localization can be found in Fig. 12, where we only show the results of EG-ETPS and MS for the proposed multi-orientation forgery localization schemes. Note that for MS, we show the average of the probability maps across different scales to approximate the fused probability map in Fig. 12.

FIGURE 12. - Example forgery localization results. Color coding: green: detected forged regions (true positives); red: detected pristine regions (false positives); blue: mis-detected forged regions (false negatives). More forgery localization results can be found in supplementary materials.
FIGURE 12.

Example forgery localization results. Color coding: green: detected forged regions (true positives); red: detected pristine regions (false positives); blue: mis-detected forged regions (false negatives). More forgery localization results can be found in supplementary materials.

D. Robustness Against Jpeg Compression

One of the main threats to PRNU-based forgery localization is JPEG compression. Thus, another experiment was conducted to evaluate the robustness against JPEG compression. We generated 6 new versions of each image by re-saving the corresponding TIFF images with JPEG quality factors of 100, 95, 90, 85, 80 and 75. We then ran all the localization algorithms on the image set of each version and calculated the performance statistics. To calculate the corresponding forgery probabilities, we used the reference PRNU and correlation predictors trained with TIFF images for JPEG images of different quality levels.

The results on the entire RTD dataset in terms of \hat {F}_{r} and \hat {F}_{b} are illustrated in Fig. 13. For the readability of the figure, we did not show the adaptive-window fusion results for the segmentation-based and multi-orientation localization schemes. As can be observed, when the image quality is above 90, both the segmentation-based and the multi-orientation schemes consistently outperform the single-orientation methods ST, SG, SO+AW and SO+MS on the entire RTD dataset. We can also observe that the \hat {F}_{r} of SG deteriorates more slowly than that of other methods as the compression becomes more severe (see Fig. 13a and 13c), but this phenomenon was not observed if the performance is measured by \hat {F}_{b} .

FIGURE 13. - Impact of JPEG compression on localization performance. ‘
$\infty $
’ in the x-axis corresponds to uncompressed images. (a): segmentation-based vs. single-orientation w.r.t. 
$\hat {F}_{r}$
; (b): segmentation-based vs. single-orientation w.r.t. 
$\hat {F}_{b}$
; (c): multi-orientation vs. single-orientation w.r.t. 
$\hat {F}_{r}$
; (d): multi-orientation vs. single-orientation w.r.t. 
$\hat {F}_{b}$
.
FIGURE 13.

Impact of JPEG compression on localization performance. ‘\infty ’ in the x-axis corresponds to uncompressed images. (a): segmentation-based vs. single-orientation w.r.t. \hat {F}_{r} ; (b): segmentation-based vs. single-orientation w.r.t. \hat {F}_{b} ; (c): multi-orientation vs. single-orientation w.r.t. \hat {F}_{r} ; (d): multi-orientation vs. single-orientation w.r.t. \hat {F}_{b} .

SECTION V.

Conclusion

In this work, we investigated the potential of explicit image segmentation for content forgery localization. We have shown that image segmentation by exploiting the visual content is beneficial for improving the performance of forgery localization based on imperceptible forensic clues, especially for hard-boundary forgeries. While the effectiveness of segmentation merely based on visual content can be compromised for soft-boundary forgeries, such limitation can be mitigated by further integrating the local homogeneity of imperceptible forensic clues to guide the segmentation. To better resolve the issue of detecting soft-boundary forgeries, we also proposed a localization scheme based on the multi-orientation fusion of the forgery probability maps obtained by multi-orientation detection and image segmentation. With the aid of the multi-scale fusion, the multi-orientation detection is effective in detecting soft-boundary forgeries and the segmentation is particularly good at identifying hard-boundary forgeries. Integrating them in a complementary way leads to the superior localization performance for our proposed multi-orientation schemes at the expense of extra computation complexity. Although we used PRNU-based forgery localization as an example in this article, we believe that similar ideas can also apply to forgery detectors based on other forensic clues. Further investigations on the potential of image segmentation in other forensics detectors as well as the combination of them will be conducted in our future work.

Appendix: Incremental Update of Forgery Probability

We calculate the forgery probability P_{s_{i}} of superpixel s_{i} using Eq. (3):\begin{equation*} P_{s_{i}}=\left ({1\!+\!\exp \left ({\frac {\rho _{s_{i}}^{2}}{2\hat {\sigma }_{0}^{2}}-\frac {(\rho _{s_{i}}-\hat {\rho }_{s_{i}})^{2}}{2\hat {\sigma }_{1}^{2}}-\ln \frac {\hat {\sigma }_{1}}{\hat {\sigma }_{0}}}\right)}\right)^{-1},\tag{17}\end{equation*} View SourceRight-click on figure for MathML and additional features. where \rho _{s_{i}} and \hat {\rho }_{s_{i}} are respectively the real and the expected correlations between the reference PRNU and the noise residual of superpixel s_{i} , and \hat {\sigma }_{0} and \hat {\sigma }_{1} are the standard deviations of the correlation distributions under hypotheses h_{0} and h_{1} . Like in [28], we use spline interpolation to dynamically obtain \hat {\sigma }_{0} and \hat {\sigma }_{1} based on the number of pixels in s_{i} and the standard deviations of the correlation distributions for scales \{32,48,\ldots,256\} estimated from training data. By assuming that PRNU noise is locally zero-mean, we simplify Eq. (2) as \begin{equation*} \rho _{s_{i}}=\frac {||\boldsymbol {r}_{s_{i}}\cdot \boldsymbol {n}_{s_{i}}||_{1}}{||\boldsymbol {r}_{s_{i}}||_{2} \cdot ||\boldsymbol {n}_{s_{i}}||_{2}}\tag{18}\end{equation*} View SourceRight-click on figure for MathML and additional features. to allow for the efficient incremental update for each superpixel s_{i} . For the expected correlation \hat {\rho }_{s_{i}} , we adopt the correlation predictor based on least-square estimator [26]:\begin{equation*} \hat {\rho }_{s_{i}}=\boldsymbol {f}_{s_{i}}\hat {\boldsymbol {\theta }}_{s_{i}},\tag{19}\end{equation*} View SourceRight-click on figure for MathML and additional features. where \hat {\boldsymbol {\theta }}_{s_{i}} is a 15{\times }1 least-square parameter vector corresponding to the scale closest to the number of pixels in s_{i} , and \boldsymbol {f}_{s_{i}} is a 1{\times }15 vector composed of four image features (i.e. the intensity, texture, signal-flattening and texture-intensity features) and their second-order terms for superpixel s_{i} . Since the calculation of the image feature is a local operation, e.g. the intensity feature is the mean of the attenuated intensity of an image block, \boldsymbol {f}_{s_{i}} can be updated incrementally if a macropixel j is removed from or added to s_{i} :\begin{equation*} \boldsymbol {f}_{s_{i}}^{(t)} = \boldsymbol {f}_{s_{i}}^{(t-1)} - \frac {N_{j}}{N_{s_{i}}}\left ({\boldsymbol {f}_{s_{i}}^{(t-1)}\pm \boldsymbol {f}_{j} }\right), \tag{20}\end{equation*} View SourceRight-click on figure for MathML and additional features. where \boldsymbol {f}_{s_{i}}^{(t-1)} and \boldsymbol {f}_{s_{i}}^{(t)} are the image features of superpixel s_{i} before and after the update, \boldsymbol {f}_{j} and N_{j} are respectively the image features and the number of pixels in macropixel j , and {N_{s_{i}}} is the number of pixels in s_{i} after the update.

References

References is not available for this document.