A neural network model for object mask detection in medical images
Abstract
In modern conditions in the field of medicine, raster image analysis systems are becoming more widespread, which allow automating the process of establishing a diagnosis based on the results of instrumental monitoring of a patient. One of the most important stages of such an analysis is the detection of the mask of the object to be recognized on the image. It is shown that under the conditions of a multivariate and multifactorial task of analyzing medical images, the most promising are neural network tools for extracting masks. It has also been determined that the known detection tools are highly specialized and not sufficiently adapted to the variability of the conditions of use, which necessitates the construction of an effective neural network model adapted to the definition of a mask on medical images. An approach is proposed to determine the most effective type of neural network model, which provides for expert evaluation of the effectiveness of acceptable types of models and conducting computer experiments to make a final decision. It is shown that to evaluate the effectiveness of a neural network model, it is possible to use the Intersection over Union and Dice Loss metrics. The proposed solutions were verified by isolating the brachial plexus of nerve fibers on grayscale images presented in the public Ultrasound Nerve Segmentation database. The expediency of using neural network models U-Net, YOLOv4 and PSPNet was determined by expert evaluation, and with the help of computer experiments, it was proved that U-Net is the most effective in terms of Intersection over Union and Dice Loss, which provides a detection accuracy of about 0.89. Also, the analysis of the results of the experiments showed the need to improve the mathematical apparatus, which is used to calculate the mask detection indicators.
References
U. Adithya, C. Nagaraju, "Object Motion Direction Detection and Tracking for Automatic Video Surveillance", International Journal of Education and Management Engineering (IJEME), Vol.11, No.2, pp. 32-39, 2021. DOI: 10.5815/ijeme.2021.02.04.
B. Alpatov, P. Babayan. “Selection of moving objects in a sequence of multispectral images in the presence of geometrically distorted ones.” Herald of RGRTU. 2008. Issue 23. P. 18-25.
V. Badrinarayanan, A. Kendall, R. Cipolla. “SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation.” https://arxiv.org/pdf/1511.00561v2.pdf.
A. Bochkovskiy, C. Wang, H. Liao. “YOLOv4: Optimal Speed and Accuracy.” https://arxiv.org/pdf/2004.10934.pdf.
I. Deepa, A. Sharma, “Multi-Module Convolutional Neural Network Based Optimal Face Recognition with Minibatch Optimization”, International Journal of Image, Graphics and Signal Processing(IJIGSP), Vol.14, No.3, pp. 32-46, 2022. DOI: 10.5815/ijigsp.2022.03.04.
D. Diwakar, D. Raj, “Recent Object Detection Techniques: A Survey”, International Journal of Image, Graphics and Signal Processing (IJIGSP), Vol.14, No.2, pp. 47-60, 2022. DOI: 10.5815/ijigsp.2022.02.05.
Z. Hengshuang, S. Jianping, Q. Xiaojuan, W. Xiaogang, J. Jiaya. “Pyramid Scene Parsing Network.” https://arxiv.org/pdf/1612.01105.pdf.
V. Hoai Viet, H. Nhat Duy, “Object Tracking: An Experimental and Comprehensive Study on Vehicle Object in Video”, International Journal of Image, Graphics and Signal Processing (IJIGSP), Vol.14, No.1, pp. 64-81, 2022.DOI: 10.5815/ijigsp.2022.01.06.
A. Horytov, S. Yakovchenko. “Selection of parametrically defined objects on a low-resolution image.” Management, computing and informatics. 2017. No. 2. P.88-90.
Z. Hu, I. Tereykovskiy, Y. Zorin, L. Tereykovska, A. Zhibek “Optimization of convolutional neural network structure for biometric authentication by face geometry.” Advances in Intelligent Systems and Computing. 2019. Vol. 754. P. 567-577.
T. Kong, F. Sun, H. Liu, Y. Jiang, L. Li, J. Shi, FoveaBox: Beyound Anchor-Based Object Detection, IEEE Trans. Image Process. 29 (2020) 7389–7398. https://doi.org/10.1109/TIP.2020.3002345.
Y. LeCun et al. “Learning Hierarchical Features for Scene Labeling.” URL: http://yann.lecun.com/exdb/publis/pdf/farabet-pami-13.pdf (access date: 02/02/2017).
V. Muraviev “Models and algorithms of image processing and analysis for systems of automatic tracking of aerial objects.” author's review. diss. for the application of scientific degrees of candidate of technical sciences: special. 05.13.01 - system analysis, management and processing. Ryazan 2010. 17 p.
P. Oniskiv, Y. Lytvynenko “Analysis of image segmentation methods”. Theoretical and applied aspects of radio engineering, instrument engineering and computer technologies: materials of IV all-Ukrainian. science and technology conf. 2019. P.48-49.
D. Perfil`ev “Segmentation Object Strategy on Digital Image”. Journal of Siberian Federal University. Engineering & Technologies. 2018. No. 11(2). R. 213-220.
O. Ronneberger, P. Fischer, T. Brox. “U-Net: Convolutional Networks for Biomedical Image Segmentation.” https://arxiv.org/abs/1505.04597.
J. Shen, “Motion detection in color image sequence and shadow elimination.” Visual Communications and Image Processing. 2014. Vol. 5308. P. 731-740.
O. Shkurat “Methods and information technology of processing archival medical images.” dissertation. ... candidate technical Sciences: 05.13.06. K., 2020. 211 p.
N. Stulov “Algorithms for the selection of basic features and methods of formation invariant to rotation, transfer, and rescaling of features of objects.” autoref. diss. for the application of scientific degrees of candidate of technical sciences: special. 05.13.01 - system analysis, management and processing. Vladimir 2006. 16 p.
I. Tereikovskyi, I. Subach, O. Tereikovskyi, L. Tereikovska, S.Toliupa, V. Nakonechnyi “Parameter Definition for Multilayer Perceptron Intended for Speaker Identification.” IEEE International Conference on Advanced Trends in Information Theory. Kyiv, Ukraine. 2019. P. 227-231.
S. Toliupa, Y. Kulakov, I. Tereikovskyi, O. Tereikovskyi, L. Tereikovska, V. Nakonechnyi “Keyboard Dynamic Analysis by Alexnet Type Neural Network.” IEEE 15th International Conference on Advanced Trends in Radioelectronics, Telecommunications and Computer Engineering. 2020. P. 416-420.
S. Toliupa, I. Tereikovskiy, I. Dychka, L. Tereikovska, and A. Trush, "The Method of Using Production Rules in Neural Network Recognition of Emotions by Facial Geometry," 2019 3rd International Conference on Advanced Information and Communications Technologies (AICT), 2019, pp. 323-327, DOI: 10.1109/AIACT.2019.8847847.
H. Wang, X. Wang, L. Yu, and F. Zhong, "Design of Mean Shift Tracking Algorithm Based on Target Position Prediction," 2019 IEEE International Conference on Mechatronics and Automation (ICMA), 2019, pp. 1114-1119, doi:10.1109/ICMA.2019.8816295.
Yudin O., Toliupa S., Korchenko O., Tereikovska L., Tereikovskyi I., Tereikovskyi O. “Determination of Signs of Information and Psychological Influence in the Tone of Sound Sequences”. IEEE 2nd International Conference on Advanced Trends in Information Theory. 2020, pp. 276-280.
S. Zhang, L. Wen, X. Bian, Z. Lei, S.Z. Li, Single-Shot Refinement Neural Network for Object Detection, in 2018: pp. 4203-4212.
Downloads
Published
Issue
Section
License
Copyright (c) 2023 International Journal of Electronics and Telecommunications
This work is licensed under a Creative Commons Attribution-NonCommercial 4.0 International License.
1. License
The non-commercial use of the article will be governed by the Creative Commons Attribution license as currently displayed on https://creativecommons.org/licenses/by/4.0/.
2. Author’s Warranties
The author warrants that the article is original, written by stated author/s, has not been published before, contains no unlawful statements, does not infringe the rights of others, is subject to copyright that is vested exclusively in the author and free of any third party rights, and that any necessary written permissions to quote from other sources have been obtained by the author/s. The undersigned also warrants that the manuscript (or its essential substance) has not been published other than as an abstract or doctorate thesis and has not been submitted for consideration elsewhere, for print, electronic or digital publication.
3. User Rights
Under the Creative Commons Attribution license, the author(s) and users are free to share (copy, distribute and transmit the contribution) under the following conditions: 1. they must attribute the contribution in the manner specified by the author or licensor, 2. they may alter, transform, or build upon this work, 3. they may use this contribution for commercial purposes.
4. Rights of Authors
Authors retain the following rights:
- copyright, and other proprietary rights relating to the article, such as patent rights,
- the right to use the substance of the article in own future works, including lectures and books,
- the right to reproduce the article for own purposes, provided the copies are not offered for sale,
- the right to self-archive the article
- the right to supervision over the integrity of the content of the work and its fair use.
5. Co-Authorship
If the article was prepared jointly with other authors, the signatory of this form warrants that he/she has been authorized by all co-authors to sign this agreement on their behalf, and agrees to inform his/her co-authors of the terms of this agreement.
6. Termination
This agreement can be terminated by the author or the Journal Owner upon two months’ notice where the other party has materially breached this agreement and failed to remedy such breach within a month of being given the terminating party’s notice requesting such breach to be remedied. No breach or violation of this agreement will cause this agreement or any license granted in it to terminate automatically or affect the definition of the Journal Owner. The author and the Journal Owner may agree to terminate this agreement at any time. This agreement or any license granted in it cannot be terminated otherwise than in accordance with this section 6. This License shall remain in effect throughout the term of copyright in the Work and may not be revoked without the express written consent of both parties.
7. Royalties
This agreement entitles the author to no royalties or other fees. To such extent as legally permissible, the author waives his or her right to collect royalties relative to the article in respect of any use of the article by the Journal Owner or its sublicensee.
8. Miscellaneous
The Journal Owner will publish the article (or have it published) in the Journal if the article’s editorial process is successfully completed and the Journal Owner or its sublicensee has become obligated to have the article published. Where such obligation depends on the payment of a fee, it shall not be deemed to exist until such time as that fee is paid. The Journal Owner may conform the article to a style of punctuation, spelling, capitalization and usage that it deems appropriate. The Journal Owner will be allowed to sublicense the rights that are licensed to it under this agreement. This agreement will be governed by the laws of Poland.
By signing this License, Author(s) warrant(s) that they have the full power to enter into this agreement. This License shall remain in effect throughout the term of copyright in the Work and may not be revoked without the express written consent of both parties.