November 2015: Yuxing Wu Xinghao Jiang
November 2015: Yuxing Wu Xinghao Jiang
net/publication/284325035
p2693-wu
CITATIONS READS
0 112
4 authors:
Some of the authors of this publication are also working on these related projects:
All content following this page was uploaded by Tanfeng Sun on 22 November 2015.
frame insertion/duplication by the photoresponse non- arg max ReRC u,v (2)
u,v
uniformity noise (PRNU) fingerprinting technique. Chao et al.
[10] proposed to detect frame deletion and insertion through where I i, j, t and I i, j, t 1 are the interrogation windows
at i, j location in t and t 1 frame respectively. F , F 1
are 2-D Fourier transform operator and inverse Fourier trans-
* Corresponding Author is Xinghao Jiang form operator respectively, * is the complex conjugate
2694
3. VIDEO FORGERY IDENTIFICATION to improve the identification accuracy. Let N h and N v de-
note the detected number of the discontinuity peaks in
The discontinuous peaks in the relative factor sequence are horizontal and vertical sequence, respectively. The flowchart
regarded as the evidence of video forgery. The generalized of the identification algorithm is given in Fig. 3.
ESD test has been applied to extract the peaks and identify
the forgery types. The detail of the identification algorithm is Y
Video clip Nh Nv 1 Original
described in this section.
N
3.1. Generalized ESD test Y
Generate Nh Nv 2 Deletion
We find that the probability distribution of the relative factor relative factor sequence
sequence follows an approximate normal distribution. Hence, N
Ri max i xi x / s (6)
Fig. 3 Flowchart of the identification algorithm
where x and s denote the mean and standard deviation of
the n samples respectively. Remove the observation that Finally, the tampered location range is determined based
maximizes xi x / s and then re-compute the above statistic on the relative factor sequence generation process in section
with n 1 observations. Repeat this process until 2. The upper bound of the range is Ru P 1 3 mod LVFI ,3 ,
R1,R2 ,...,Ri have all been computed. Finally pick the corre- where P is the location of the detected peaks in relative fac-
sponding r critical values i at the chosen confidence level tor sequence, “mod” is modulo operation, and LVFI is the
. The number of outliers is determined by finding the larg-
est i such that Ri i . length of the corresponding VFI sequence. Hence, the tam-
In order to determine the exact number of peaks in the pered range is Ru 2,Ru .
relative factor sequence, we have fine-tuned the critical val-
ues i by multiplying a coefficient , and the new definition
is as follows:
t p , n i 1 n i
i ' (7)
n i 1 t n i 1
2
(a) (b) (c) (d)
p , n i 1 Fig. 4. Four source videos with different scenes. (a) scene 1; (b)
scene 2; (c) scene 3; (d) scene 4.
p 1 (8)
2 n i 1
4. EXPERIMENTS
where t p , n i 1 is the pth percentile of a t distribution with
n i 1 degrees of freedom.
4.1. Video database
Some fake peaks might be found in the relative factor
sequence. Comparing with real forgery peaks, these peaks
To the best of our knowledge, there is no open database for
are with relatively low intens ities, which were probably in-
detecting video inter-frame forgery. Therefore, we have in-
troduced by camera noise or video encoding. The fake peaks
vited some volunteers to build one. Our four different scenes
would be determined as outliers with the original i , while
source videos (see in Fig. 4) are downloaded from TRECVID
the fine tuning, which slightly raises the critical values is
surveillance event detection evaluation [14]. Each source
helpful to refuse these fake peaks and pick out the forgery
video split out 10 video clips. Each clip contains about 3000
peaks accurately as well.
frames with 720 576 resolution. Then the 40 video clips
were delicately tampered to generate 40 frame deletion vide-
3.2. Identification algorithm
os and 40 frame duplication videos (defined in section 2.2).
Hence, there are totally 120 video clips in our final inter-frame
According to the description in section 2.2, there are at most
forgery detection database. Note that all the tampered video
two discontinuity peaks in the relative factor sequence,
clips were MPEG-2 re-coded with the same coding standard
hence we set the upper bound number of outliers r 2 .
and parameters as the source videos.
Moreover, the generalized ESD test is carried out on both
horizontal and vertical relative factor sequences , which helps
2695
4.2. Results and analysis compression), 2, 3 are shown in Table IV. When re-
compressing with Qscale=2, the bit rate has averagely de-
The configurations of identification algorithm are as follows. creased by 3%, so the accuracy is the same with the result of
The upper bound number of outliers is r 2 , the signifi- Qscale=2. While when re-compressing with Qscale=3, the bit
cance level is 0.05 and the coefficient of the critical rate dropped a lot (with 30%), the duplication identification
values is . accuracy have slightly decreased. The reason is that the
intensity gap between the two detected peaks enlarged after
4.2.1. Detection accuracy under random deletion re-compression, which makes it easy to be identified as frame
This experiment is to test the sensitivity of our algorithm by deletion forgery. Anyway, the accuracies report the robust-
computing the detection accuracies when frames were ran- ness of our algorithm to some degree of compression.
domly deleted. Table I shows the detection accuracies for
randomly deleting 1 frame, 3 consecutive frames and 5 con- 5. CONCLUSION
secutive frames from original videos. The result illustrates
that our algorithm could have good accuracy when detecting We have proposed a new algorithm to detect video inter-
frame deletion forgery with a few frames removed. frame forgery. This method is based on the consistency of
velocity field. With consecutive frame deletion and frame
4.2.2. Identification accuracy under meaningful forgery duplication forgery operations, some discontinuity peaks
The confusion matrices for the four scenes of video clips can be observed in VFI sequence. And the generalized ESD
and the overall accuracy are given in Table II. The relatively test is applied to extract the peaks and identify the forgery
low accuracies for frame duplication identification are due to type. Experiments show the effectiveness of our algorithm.
the large intensity gaps between their two detected peaks, all
the incorrect identification videos are identified as frame de- ACKNOWLEDGMENT
letion forgery. However, the result demonstrates the
effectiveness of our algorithm with overall 90.0%, 85.0% and The work of this paper is sponsored by the National Natural
80.0% accuracies for identifying original video, frame dele- Science Foundation of China (No.61272439, 61272249), the
tion video and frame duplication video. If we only consider Specialized Research Fund for the Doctoral Program of High-
whether a video is tampered or not, the overall identification er Education (No.20120073110053), and the Fund of State
accuracy for the tampered videos is 96.3%, with 10% false Key Laboratory of Software Engineering, Wuhan University
positives. We did not do comparison experiments because (No.SKLSE2012-09-12). It was also under the Project of In-
no papers were found on identifying consecutive frame dele- ternational Cooperation and Exchanges supported by
tion and duplication forgeries. Shanghai Committee of Science and Technology (No.
12510708500).
4.2.3. Location accuracy under meaningful forgery
Table I Detection accuracies for random deletion.
The location is considered to be incorrectly identified if one
of the detected peaks in both horizontal and vertical VFI s e- Deleted frame number 1 3 5
quences is not in the expected range described in section 3.2. Accuracy 40% 65% 80%
The location accuracies for correctly identified forged videos Table III Location accuracies under meaningful forgery.
are given in Table III. All the locations of detected peaks in Forgery type deletion duplication
forged videos are correctly identified due to the s tatistics-
Accuracy 100%(34/34)a 100%(32/32)
based generalized ESD algorithm. a
(n/m) indicates n of m locations are correctly identified.
4.3. Robustness against compression Table IV Detection accuracies under different Qscales (%).
Qscale 1 2 3
The robustness against lossy compression is tested in this original 90.0 90.0 90.0
experiment. Each video clips was re-compressed by ffmpeg deletion 85.0 85.0 85.0
software with different Qscales (a parameter to control video duplication 80.0 80.0 62.5
quality). The identification results with Qscale=1 (lossless
Table II Confusion matrix for each scene and their overall accuracy (%). – denotes value 0.
Video Scene 1 Scene 2 Scene 3 Scene 4 Overall
Forgery type oria delb dup c ori del dup ori del dup ori del dup ori del dup
ori 80.0 10.0 10.0 90.0 10.0 - 100 - - 90.0 10.0 - 90.0 7.5 2.5
del 20.0 70.0 10.0 - 90.0 10.0 10.0 90.0 - - 90.0 10.0 7.5 85.0 7.5
dup - 30.0 70.0 - 20.0 80.0 - 30.0 70.0 - - 100 - 20.0 80.0
a
Original video type. b Frames deletion video type. c Frames duplication video type.
2696
REFERENCES
[6] Hsu C.C., Hung T.Y., Lin C.W. et al., “Video forgery detection
using correlation of noise residue,” IEEE 10th Workshop on M ulti-
media Signal Processing, pp.170-174, 2008.
[9] Conotter V., O'Brien J.F., Farid H., “Exposing Digital Forgeries
in Ballistic M otion,” IEEE Transactions on Information and Securi-
ty, vol.7, pp.283-296, 2012.
[10] Chao J., Jiang X. H. and Sun T. F., “A novel video inter-frame
forgery model detection scheme based on optical flow consisten-
cy,” International Workshop on Digital Forensics and Watermaking,
pp.267-281, 2012.
[13] Iglewicz B., Hoaglin D.C., “How to Detect and Handle Outli-
ers,” M ilwaukee (Wisconsin): ASQC Quality Press, vol. 16, 1993.