Robust Image Forgery Localization for JPEG Artifacts

DOI : 10.17577/IJERTV3IS051666

Download Full-Text PDF Cite this Publication

Text Only Version

Robust Image Forgery Localization for JPEG Artifacts

A. Lakesha

    1. Computer and Communication (II Year) Cape Institute of Technology

      Tirunelveli, India

      1. P. Muthu Priya Assistant Professor-Dept of IT Cape Institute of Technology

        Tirunelveli, India

        Hariharan. B

        Assistant Professor-Dept of IT Cape institute of technology Tirunelveli, India

        AbstractImages are powerful tool for communication and application purpose. Image manipulation is done by splicing techniques for harmless manipulation cases. So identification of true photographs like evidences for real world events. To detect and localize the tampered area into a digital image for forgery purpose new algorithms are used. It is based on machine learning approach and less number of users interaction is required. This technique is applicable to image containing information with respect to no expert interaction for the tampering decision. Statistical based illuminant estimator to extract the texture and edge based features these features are then provided to a machine learning approach for automatic decision making. This Image forgery detection method based on SIFT.A part of an image is taken from one image and they are pasted on a different location in the different image, from that we are finding the forged image. In this approach an improved algorithm based on scale invariant features transform (SIFT) is used to detect such forgery , In this technique Transform is applied to the input image to yield a reduced dimensional representation, After that Apply key point detection and feature descriptor along with a matching over all the key points matching. Furthermore gives output by applying matched points. Although photographers are able to create composites of analog pictures, this process is very time consuming. One of the most common forms of photographic manipulation is known as image composition or splicing.

        Keywords Illuminant; SIFT; tampering; machine learning; texture and edge based; Splicing.

        1. INTRODUCTION

          Digital image processing is a subset which belongs to the electronic domain. In this the image is changed in to an array of small integers, called pixels, which says about the physical quantity that is scene radiance, which is processed by digital hardware and stored in the digital memory. Digital image processing either in the field of improving in human perception or observing their performance leads to the advantage in cost, speed and adaptability. Nowadays it has

          become the dominant method to be used in all the fields like medical, biometrics, forgery etc. The image that is captured by the digital camera is used to identify where the particular person face is located and where the person is present in the picture. In this system we have considered face as a biometric to detect the forgery for the following reasons.

          • The procurement process is non intrusive

          • The procurement process of a face is cheaper simpler and accurate compared to other biometrics such as iris and fingerprints.

          • This technique is secured

There are many steps in face detection methods, first it finds out whether the given input image is human image and checks where the image is present in the database. In the second step, the patches in each face are detected and the output from this step is given to the face extraction phase. In this phase, the face which is in the database will be extracted separately. In order to justify the orientation of the patches we need a robustness to design the system feasibly. Normally, face detection methods focuses mainly on the shape and colour of the face and not on the other features in the face.

Input Image

Face Detection

Face Extraction

Face Recognition

Verification/Identification

Fig.1. Flow diagram of face recognition

In face recognition, the patches in the face may have some drawbacks stated below:

  • Usually, the size of the patches is of more than 1000 pixels, which leads to a non-robustness recognition system.

  • Occlusion and clutter problem arises in the face patch due to different camera alignments, illuminations and face expressions.

After the feature extraction step, the face patches are dimensionally converted in to vectors fiducially based on their locations.

The feature extraction is followed by face Recognition phase where the identities of each face is recognised. If the face database is present, then the automatic recognition is achieved. The two different applications such as face identification and face verification are explained below.

Face identification checks whether the given face image is present in the database among the known individuals that works under one-to- many relationship.

Face Verification verifies whether the unknown individuals identity matches with the database. This performs one-to-one relationship.

    1. RELATED WORKS

      J. F. OBrien [7] et al proposed that a new forensic technique that focuses on geometric inconsistencies that arise when fake reflections are inserted into a photograph or when a photograph containing reflections is manipulated. This analysis employs basic rules of reflective geometry and linear perspective projection, makes minimal assumptions about the scene geometry, and only requires the user to identify corresponding points on an object and its reflection. The analysis is also insensitive to common image editing operations such as resampling, color manipulations, and lossy compression. Images are highly blurry or distorted, highly overlapping occurs by this technique

      Y. Ostrovsky [18] et al report an unexpected failure of this ability in the context of perceiving inconsistencies removed, the visual system is remarkably insensitive to illumination inconsistencies, both in experimental stimuli and in altered images of real scenes. Whether the target is interpreted as oddly illuminated or oddly pigmented, it is very difficult to find if the only cue is deviation from the regularity of illumination or reflectance. Our results allow us to draw inferences about how the visual system encodes illumination distributions across scenes. Specifically, they suggest that the visual system does not verify the global consistency of locally derived estimates of illumination direction. Noise is high; direction of illumination is slow to compute are the major flaws.

      S. Gholap et al [11] proposed a physics-based approach for illuminant color estimation of arbitrary images, which is explicitly designed for handling images with multiple illuminants. The majority of techniques that extract the illuminant color assume that the illumination is constant across the scene. We propose an illuminant-color estimation method which is based on robust local illuminant estimates. An illuminant color estimate is obtained independently from distinct image mini-regions. From these mini-regions a robust

      local illumination color is computed by consensus. These local estimates are then used in deriving the chromaticity of the dominant illuminants. Furthermore, extensive tests on real- world images show that we can reliably process mixed illuminant. Uniform illumination can not accurately recover the illuminant are the major flaws reported.

      M. Johnson et al [3] the availability of sophisticated digital imaging technology has given rise to digital forgeries that are increasing in sophistication and frequency. We describe a technique for exposing such fakes by detecting inconsistecies in lighting. We show how to approximate complex lighting environments with a low-dimensional model and, further, how to estimate the models parameters from a single image. Inconsistencies in the lighting model are then used as evidence of tampering. More accurate results, High dynamic range images are produced.

      C. Riess [24] et al presents a physics-based approach for illuminant color estimation of arbitrary images, which is explicitly designed for handling images with multiple illuminants. The majority of techniques that extract the illuminant color assume that the illumination is constant across the scene. This, however, is not often the case. We propose an illuminant-color estimation method which is based on robust local illuminant estimates. There are no assumptions on the number or type of illuminants. An illuminant color estimate is obtained independently from distinct image mini- regions. From these mini-regions a robust local illumination color is computed by consensus. These local estimates are then used in deriving the chromaticity of the dominant illuminants. Experiments on an established benchmark database of real-world images show that our technique performs comparably to uniform-illuminant estimation methods. Furthermore, extensive tests on real-world images show that we can reliably process mixed illuminant. Uniform illumination can not accurately recover the illuminant.

    2. FORGERY DETECTION FRAMEWORK

      1. Overview

        This project deals with forgery detection. There are different types of forgeries are there, they are mainly used for forgery purpose in courts etc. They are done in faces, signatures, voices etc. In this the persons who are add to original image is detected and say that the image is the forged one. This is calculated by help of SIFT and key point matching. The color image is being used as input for forgery purpose. Several steps are used.

      2. System Module Description

        Face detection:

        In this step from the input image Fig 2.a the noise that are present in the image are being filtered by help of IM filters and produces smooth image or fabric image Fig2.b.From the fabric image it is converted to RGB layers Fig 2.c, Fig2.d, Fig2.e then it is converted to YCbCr image Fig2.f because RGB image is difficult to convert the image for further process that is the pixel value is more, so we are converting to YCbCr image. The YCbCr image is converted to gray image Fig 2.g.Then the segmentation map is done for the image Fig 2.h.

        SIFT features (Scale-invariant feature transform):

        SIFT method, is used to determine the key points, for the entire image by key point detection and feature descriptor. This approach is known as the Scale Invariant Feature Transform (SIFT), related to local features the image data is transformed to scale-invariant coordinates. By converting RGB image to gray scale image Fig 2.g the key point value is calculated.

        Append Images:

        Append is for adding of two or more image with the original image. The image that is being added to the original image will of same pixels and same values. Only joining is done in this step.

        Key point matching:

        Key point is the method, used for matching extracted feature from the image by SIFT algorithm. This reads the input image key point and forged image points and then, it compares the key point image and draws the line indicating the matched points. Then append the two images and draw a line which indicates the matches Fig 2.j.

        Input image

        Face detection

        Sift features

        Append Images

        Key point matching

        Homogeneity calculation

        Fig.2. Flow chart diagram for forgery detection

        Homogeneity Calculation:

        It says the relationship between the original image and forged image by help of key points. The transformation matrix are calculated by help of correspondent points .The input image that is added again with the forged image will have the same value but other persons who are added with the image will

        have the different value, from that it is said to be the forged image Fig 2.k the and the output comes as forgery is detected.

    3. RESULTS AND DISCUSSION

      The image produced is of high quality when compared with forensic algorithm, because the image produced was blur image to avoid that, we used different method. This will increase the overall performance and computation of the system during result execution. Analysis is based on accuracy of image which is compared with forensic the algorithm SIFT, RANSAC gives better result is given below.

      Fig. 2.a. Input image Fig. 2.b.Fabric image

      Fig. 2.c.Red image Fig. 2.d. Green image

      Fig. 2.e. Blue image Fig. 2.f. YCbCr image

      Fig. 2.g. Gray image Fig. 2.h. Segmentation map

      Fig. 2.i.Altered image

      Fig. 2.j. key point matching

      Fig. 2.k. Result

    4. CONCLUSION

In this work, we presented a new method for detecting forged image by using SIFT and RANSAC algorithm this effectively detects the forged image that is being added with the original image. We can produce better quality image than this by help of Surf algorithm in the future.

REFERENCES

    1. H. Farid, A 3-D lighting and shadow analysis of the JFK Zapruderfilm (Frame 317), Dartmouth College, Tech. Rep. TR2010677,2010.

    2. M. Johnson and H. Farid, Exposing digital forgeries by detecting inconsistencies in lighting, in Proc. ACM Workshop on Multimedia and Security, New York, NY, USA, 2005, pp. 110.

    3. M. Johnson and H. Farid, Exposing digital forgeries in complex lighting environments, IEEE Trans. Inf. Forensics Security, vol. 3,no. 2, pp. 450461, Jun. 2007.

    4. M. Johnson and H. Farid, Exposing digital forgeries through specular highlights on the eye, in Proc. Int.Workshop on Inform. Hiding, 2007,pp. 311325.

    5. E. Kee and H. Farid, Exposing digital forgeries from 3-D lighting environments,in Proc. IEEE Int. Workshop on Inform. Forensics and Security (WIFS), Dec. 2010, pp. 16.

    6. W. Fan, K. Wang, F. Cayre, and Z. Xiong, 3D lighting-based image forgery detection using shape-from-shading, in Proc. Eur. Signal Processing Conf. (EUSIPCO), Aug. 2012, pp. 17771781.

    7. J. F. OBrien and H. Farid, Exposing photo manipulation with inconsistent reflections, ACM Trans. Graphics, vol. 31, no. 1, pp. 111, Jan.2012.

    8. S. Gholap and P. K. Bora, Illuminant colour based image forensics,in Proc. IEEE Region 10 Conf., 2008, pp. 15.

    9. X.Wu and Z. Fang, Image splicing detection using illuminant color inconsistency,in Proc. IEEE Int. Conf. Multimedia Inform. Networking and Security, Nov. 2011, pp. 600603.

    10. P. Saboia, T. Carvalho, and A. Rocha, Eye specular highlights telltales for digital forensics: A machine learning approach, in Proc. IEEE Int.Conf. Image Processing (ICIP), 2011, pp. 1937 1940.

    11. C. Riess and E. Angelopoulou, Physics-based illuminant color estimation as an image semantics clue, in Proc. IEEE Int. Conf. Image Processing, Nov. 2009, pp. 689692.

    12. K. Barnard, V. Cardei, and B. Funt, A comparison of computational color constancy algorithmsPart I: Methodology and Experiments With Synthesized Data, IEEE Trans. Image Process., vol. 11, no. 9,pp. 972983, Sep. 2002.

    13. K. Barnard, L. Martin, A. Coath, and B. Funt, A comparison of computational color constancy algorithms Part II: Experiments With Image Data, IEEE Trans. Image Process., vol. 11, no. 9, pp. 985996,Sep. 2002.

    14. A. Gijsenij, T. Gevers, and J. van deWeijer, Computational color constancy:Survey and experiments, IEEE Trans. Image Process., vol. 20,no. 9, pp. 24752489, Sep. 2011.

    15. M. Bleier, C. Riess, S. Beigpour, E. Eibenberger, E. Angelopoulou, T.Tröger, and A. Kaup, Color constancy and non- uniform illumination:Can existing algorithms work?, in Proc. IEEE Color and Photometry in Comput. Vision Wokshop, 2011, pp. 774781.

    16. M. Ebner, Color constancy using local color shifts, in Proc. Eur.Conf. Comput. Vision, 2004, pp. 276287.

    17. A. Gijsenij, R. Lu, and T. Gevers, Color constancy for multiple light sources, IEEE Trans. Image Process., vol. 21, no. 2, pp. 697707, Feb.2012.

    18. Y. Ostrovsky, P. Cavanagh, and P. Sinha, Perceiving illumination inconsistencies in scenes, Perception, vol. 34, no. 11, pp. 1301 1314,2005.

    19. T. Igarashi, K. Nishino, and S. K. Nayar, The appearance of human skin: A survey, Found. Trends Comput. Graph. Vis., vol. 3, no. 1, pp.195, 2007.

    20. P. F. Felzenszwalb and D. P. Huttenlocher, Efficient graph-based image segmentation, Int. J. Comput. Vis., vol. 59, no. 2, pp. 167 181,2004.

    21. G. Buchsbaum, A spatial processor model for color perception,

      J.Franklin Inst., vol. 310, no. 1, pp. 126, Jul. 1980.

    22. A. Gijsenij, T. Gevers, and J. van de Weijer, Improving color constancy by photometric edge weighting, IEEE Trans. Pattern Anal.Mach. Intell., vol. 34, no. 5, pp. 918929, May 2012.

    23. R. Tan, K. Nishino, and K. Ikeuchi, Color constancy through inverse-intensity chromaticity space, J. Opt. Soc. Amer. A, vol. 21, pp.321334, 2004.

    24. C. Riess, E. Eibenberger, and E. Angelopoulou, Illuminant color estimation for real-world mixed-illuminant scenes, in Proc. IEEE Color and Photometry in Comput. Vision Workshop, Barcelona, Spain, Nov.2011.

    25. W. R. Schwartz, A. Kembhavi, D. Harwood, and L. S. Davis, Human detection using partial least squares analysis, in Proc. IEEE Int. Conf. Comput. Vision (ICCV), 2009, pp. 2431.

    26. A. Carkacioglu and F. T. Yarman-Vural, Sasi: A generic texture descriptorfor image retrieval, Pattern Recognit., vol. 36, no. 11, pp.

      26152633, 2003.

    27. O. A. B. Penatti, E. Valle, and R. S. Torres,Comparative study of global color and texture descriptors for web image retrieval, J. Visual Commun. Image Representat., vol. 23, no. 2, pp. 359380, 2012.

    28. J. Canny, A computational approach to edge detection, IEEE Trans.Pattern Anal. Mach. Intell., vol. 8, no. 6, pp. 679698, Jun. 1986.

    29. N. Dalal and B. Triggs, Histograms of oriented gradients for human detection, in Proc. IEEE Conf. Comput. Vision and Pattern Recognition,2005, pp. 886893.

    30. G. Csurka, C. R. Dance, L. Fan, J. Willamowski, and C. Bray, Visualcategorization with bags of keypoints, in Proc. Workshop on StatisticalLearning in Comput. Vision, 2004, pp. 18.

Leave a Reply