Localization of Copy-Move Forgery in speech signals through watermarking using DCT-QIM

Authors

  • N.V. Lalitha GMR Institute of Technology
  • Ch. Srinivasa Rao JNTU-K University
  • P.V.Y. JayaSree GIT, GITAM University

Abstract

Digital speech copyright protection and forgery identification are the prevalent issues in our advancing digital world. In speech forgery, voiced part of the speech signal is copied and pasted to a specific location which alters the meaning of the speech signal. Watermarking can be used to safe guard the copyrights of the owner. To detect copy-move forgeries a transform domain watermarking method is proposed. In the proposed method, watermarking is achieved through Discrete Cosine Transform (DCT) and Quantization Index Modulation (QIM) rule. Hash bits are also inserted in watermarked voice segments to detect Copy-Move Forgery (CMF) in speech signals. Proposed method is evaluated on two databases and achieved good imperceptibility. It exhibits robustness in detecting the watermark and forgeries against signal processing attacks such as resample, low-pass filtering, jittering, compression and cropping. The proposed work contributes for forensics analysis in speech signals. This proposed work also compared with the some of the state-of-art methods.

Author Biographies

N.V. Lalitha, GMR Institute of Technology

N. V. Lalitha is presently pursuing PhD at GIT, GITAM University. She obtained her M.Tech from Jawaharlal Nehru Technological University, Kakinada and B.Tech from Jawaharlal Nehru Technological University. Presently, she is working as Assistant professor in the Department of Electronics and Communication Engineering at GMR Institute of Technology, Rajam, Srikakulam District. She is having 10 years of teaching experience. Her research interests are Audio and Image Processing. She is a Life Member of IETE.

Ch. Srinivasa Rao, JNTU-K University

Srinivasa Rao Ch is currently working as Professor in the Department of ECE, JNTUK University College of Engineering, Vizianagaram, AP, India. He obtained his PhD in Digital Image Processing area from University College of Engineering, JNTUK, Kakinada, AP, India. He received his M. Tech degree from the same institute. He published more than 50 research papers in international journals and conferences. His research interests are Digital Speech/Image and Video Processing, Communication Engineering and Evolutionary Algorithms. He is a Member of CSI. Dr Rao is a Fellow of IETE.

P.V.Y. JayaSree, GIT, GITAM University

P. V. Y. Jayasree is currently working as Associate Professor in the Department of ECE, GIT, GITAM University. She obtained her PhD from University College of Engineering, JNTUK, Kakinada, AP, India. She received M.E. from Andhra University. She published more than 50 research papers in international journals and conferences. Her research interests are Signal Processing, EMI/EMC, RF & Microwaves.

References

I. J. Cox, J. Kilian, F. T. Leighton, and T. Shamoon, “Secure spread spectrum watermarking for multimedia,” IEEE Trans. Image Process., vol. 6, no. 12, pp. 1673–1687, 1997.

P. Taylor, V. S. Verma, and R. K. Jha, “An Overview of Robust Digital Image Watermarking,” IETE Tech. Rev., no. June, pp. 1–18, 2015.

A. N. Lemma, J. Aprea, W. Oomen, and L. Van de Kerkhof, “A temporal domain audio watermarking technique,” IEEE Trans. Signal Process., vol. 51, no. 4, pp. 1088–1097, 2003.

M. Unoki and R. Miyauchi, “Robust, blindly-detectable, and semi-reversible technique of audio watermarking based on cochlear delay characteristics,” IEICE Trans. Inf. Syst., vol. E98–D, no. 1, pp. 38–48, 2015.

G. Hua, J. Goh, and V. L. L. Thing, “Time-spread echo-based audio watermarking with optimized imperceptibility and robustness,” IEEE/ACM Trans. Speech Lang. Process., vol. 23, no. 2, pp. 227–239, 2015.

A. Merrad, A. Benziane, S. Saadi, and A. Hafaifa, “Robust Blind Approach for Digital Speech Watermarking,” in 2nd International Conference on Natural Language and Speech Processing (ICNLSP), 2018.

A. Al-haj and A. Mohammad, “Digital Audio Watermarking Based on the Discrete Wavelets Transform and Singular Value Decomposition,” Eur. J. Sci. Res., vol. 39, no. 1, pp. 6–21, 2010.

X.-Y. Wang and H. Zhao, “A novel synchronization invariant audio watermarking scheme based on DWT and DCT,” IEEE Trans. Signal Process., vol. 54, no. 12, pp. 4835–4840, 2006.

M. A. Nematollahi, H. Gamboa-Rosales, M. A. Akhaee, and S. A. R. Al-Haddad, “Semi-fragile digital speech watermarking for online speaker recognition,” EURASIP J. Audio, Speech, Music Process., vol. 2015, no. 31, pp. 1–15, 2015.

M. A. Nematollahi, H. Gamboa-rosales, M. A. Akhaee, and S. A. . Al-Haddad, “Robust Digital SpeechWatermarking For Online Speaker Recognition,” Math. Probl. Eng., vol. 2015, pp. 1–12, 2015.

S. Wang, X. Liu, X. Dang, and J. Wang, “A Robust Speech Watermarking Based On Quantization Index Modulation And Double Discrete Cosine Transform,” in 10th International Congress on Image and Signal Processing, BioMedical Engineering and Informatics (CISP-BMEI 2017), 2017.

M. A. Nematollahi, S. A. R. Al-Haddad, and F. Zarafshan, “Blind digital speech watermarking based on Eigen-value quantization in DWT,” J. King Saud Univ. - Comput. Inf. Sci., vol. 27, no. 1, pp. 58–67, 2015.

H. T. Hu, S. J. Lin, and L. Y. Hsu, “Effective blind speech watermarking via adaptive mean modulation and package synchronization in DWT domain,” Eurasip J. Audio, Speech, Music Process., vol. 2017, no. 1, pp. 1–14, 2017.

Subir and A. M. Joshi, “DWT-DCT based blind audio watermarking using Arnold scrambling and Cyclic codes,” in 3rd International Conference on Signal Processing and Integrated Networks, SPIN 2016, 2016, pp. 79–84.

B. Chen and G. W. Wornell, “Quantization index modulation: A class of provably good methods for digital watermarking and information embedding,” IEEE Trans. Inf. Theory, vol. 47, no. 4, pp. 1423–1443, 2001.

P. K. Dhar and T. Shimamura, “Audio Watermarking in Transform Domain Based on Singular Value Decomposition and Cartesian-polar transformation,” Int. J. Speech Technol., vol. 17, pp. 133–144, 2014.

S. Self-recovery, S. Sarreshtedari, S. Member, and M. A. Akhaee, “A Watermarking Method for Digital Speech Self-Recovery,” IEEE/Acm Trans. Audio, Speech, Lang. Process., vol. 23, no. 11, pp. 1917–1925, 2015.

VoxForge speech database “http://www.repository.voxforge1.org/downloads/SpeechCorpus/Trunk/Audio/Original/44.1kHz_16bit/tle.” .

D. Avci, T. Tuncer, and E. Avci, “A New Information Hiding Method for Audio Signals,” in 6th International Symposium on Digital Forensic and Security (ISDFS), 2018, pp. 3–6.

P. K. Dhar and T. Shimamura, “Blind Audio Watermarking in Transform Domain Based on Singular Value Decomposition and Exponential-Log Operations,” Radio Eng., vol. 26, no. 2, pp. 552–561, 2017.

H. Hu and J. Chang, “Efficient and robust frame-synchronized blind audio watermarking by featuring multilevel DWT and DCT,” Cluster Comput., vol. 20, no. 1, pp. 805–816, 2017.

B. Y. Lei, I. Y. Soon, and Z. Li, “Blind and robust audio watermarking scheme based on SVD-DCT,” Signal Processing, vol. 91, no. 8, pp. 1973–1984, 2011.

Z. Liu and H. Wang, “A novel speech content authentication algorithm based on Bessel–Fourier moments,” Digit. Signal Process. A Rev. J., vol. 24, pp. 197–208, 2014.

Downloads

Published

2024-04-19

Issue

Section

Digital Signal Processing