- Faculty of Medicine, University of Buenos Aires, Ciudad Autónoma de Buenos Aires, Argentina
- Department of Neurosurgery, San Fernando Hospital, San Fernando, Argentina
- Department of Neurosurgery, Hospital Padilla de Tucuman, Tucuman, Argentina
Correspondence Address:
Lautaro Ezequiel De Bartolo Villar, Faculty of Medicine, University of Buenos Aires, Ciudad Autónoma de Buenos Aires, Argentina.
DOI:10.25259/SNI_303_2025
Copyright: © 2025 Surgical Neurology International This is an open-access article distributed under the terms of the Creative Commons Attribution-Non Commercial-Share Alike 4.0 License, which allows others to remix, transform, and build upon the work non-commercially, as long as the author is credited and the new creations are licensed under the identical terms.How to cite this article: Lautaro Ezequiel De Bartolo Villar1, Matias Baldoncini2, Alvaro Campero3, Mickaela Echavarria Demichelis2. Development of an artificial intelligence-based convolutional neural network for sellar barrier classification using magnetic resonance imaging. 09-May-2025;16:174
How to cite this URL: Lautaro Ezequiel De Bartolo Villar1, Matias Baldoncini2, Alvaro Campero3, Mickaela Echavarria Demichelis2. Development of an artificial intelligence-based convolutional neural network for sellar barrier classification using magnetic resonance imaging. 09-May-2025;16:174. Available from: https://surgicalneurologyint.com/?post_type=surgicalint_articles&p=13545
Abstract
Background: This study aims to develop an artificial intelligence (AI) model using convolutional neural networks and transfer learning to classify sellar barriers as strong, mixed, or weak based on magnetic resonance imaging (MRI). Accurate classification is essential for surgical planning in endoscopic endonasal approaches for pituitary adenomas, as variations in the sellar barrier can lead to complications such as cerebrospinal fluid leaks.
Methods: The dataset consisted of 600 MRI images of sellar barriers obtained from coronal sections and evenly distributed among the three classes. The EfficientNetB0 architecture was employed, leveraging transfer learning to optimize performance despite the small dataset. The model was implemented and trained on Google Colab using TensorFlow, with techniques such as dropout and batch normalization to improve generalization and reduce overfitting. Performance metrics included accuracy, recall, and F1-score.
Results: The AI model achieved a classification accuracy of 96.33%, with individual class accuracies of 98% for strong barriers, 95% for mixed barriers, and 96% for weak barriers. These results demonstrate the model’s high capacity to accurately classify sellar barriers and identify their specific characteristics, ensuring reliable preoperative assessment.
Conclusion: The proposed AI model significantly enhances the preoperative classification of sellar barriers, contributing to improving surgical planning and reducing complications. While the “black box” nature of AI poses challenges, integrating explainability modules and expanding datasets can further increase clinical trust and applicability. This study underscores the transformative potential of AI in neurosurgical practice, paving the way for precise and reliable diagnostics in managing pituitary lesions.
Keywords: Artificial intelligence, Convolutional neural networks, Medical imaging, Pituitary adenomas, Sellar barrier classification
INTRODUCTION
The endoscopic endonasal approach is one of the most utilized techniques for the resection of pituitary lesions, primarily indicated in patients with endocrinological alterations or neurological deficits.[
Figure 1:
Strong sellar barrier. A 42-year-old male patient with growth hormone produce macroadenoma. (a and b) Preoperative magnetic resonance imaging: the yellow arrows indicate the barrier that captures contrast with a thickness >1 mm. (c-f) Intraoperative images: the barrier constituted by the gland can be seen.
Figure 2:
Mixed sellar barrier. A 59-year-old female patient with a PRL produced macroadenoma. (a and b) Preoperative magnetic resonance imaging: the yellow arrows indicate the barrier that captures contrast with a thickness >1 mm, and the red arrows indicate the barrier that captures contrast with a thickness <1 mm. (c-f) Intraoperative images: the barrier constituted by dura mater and arachnoid can be seen. PRL: Prolactin
Figure 3:
Weak sellar barrier. A 31-year-old female patient with an adrenocorticotropic hormone produced macroadenoma. (a and b) Preoperative magnetic resonance imaging: the red arrows indicate the barrier that captures contrast with a thickness <1 mm. (c-f) Intraoperative images: the barrier constituted only with arachnoid can be seen. The green arrow marks the cerebrospinal fluid leak.
This complication arises due to anatomical variations in the sellar barrier, which may not always provide sufficient resistance to prevent CSF escape during surgical manipulation, regardless of whether the technique used is microscopic, endoscope-assisted, or purely endoscopic.[
In clinical practice, the classification of the sellar barrier is currently based on the evaluation of magnetic resonance imaging (MRI) and intraoperative observations.[
MATERIALS AND METHODS
Code
The complete code is available for review at the following link. This repository provides full access to the implementation, allowing for in-depth analysis and reproducibility of the methods described in this study:
Dataset
Preoperative MRI scans of patients with pituitary adenomas were used. The MRI scans consisted of coronal sections. A total of 200 images were included for each type of sellar barrier (strong, mixed, and weak). The images were standardized to a resolution of 150 × 150 pixels to ensure data uniformity. However, the system also supports the upload of images in other resolutions, as the code includes a module for automatic rescaling to the required format, ensuring compatibility and consistency across the dataset [
CNN architecture
CNNs are a type of deep learning model specifically designed for processing structured grid data, such as images. These networks excel automatically and adaptively, learning spatial hierarchies of features, from low-level edges to high-level patterns, directly from raw input data. This capability makes CNNs particularly well-suited for tasks like image classification and medical imaging analysis[
Figure 5:
Labeled datasets are provided, such as 1,000 images of zebras, 1,000 of lions, and 1,000 of elephants. During training, the network learns to recognize key features that define each category. Once trained, the system can accurately identify an entirely new and different image (one it has never seen before) based solely on the features it has learned. This process is similar to human learning: after observing various examples of tables, we can identify a table we’ve never encountered, even if it differs in size, shape, or design. Similarly, a convolutional neural network processes millions of parameters (over 4 million in this case) to make precise decisions in just milliseconds.
In this study, the EfficientNetB0 architecture was employed, given its optimal balance between performance and computational efficiency. EfficientNetB0 utilizes a compound scaling approach that uniformly scales the depth, width, and resolution of the network, resulting in a model that achieves high accuracy with fewer parameters. This architecture was pretrained on the ImageNet dataset, a widely recognized benchmark comprising millions of labeled images across 1,000 categories, providing a robust starting point for transfer learning.
To adapt EfficientNetB0 for this study’s specific task of classifying sellar barrier types, the top layer of the network – originally designed for 1,000-class classification – was replaced with a fully connected dense layer. This layer includes a SoftMax activation function, which outputs a probability distribution over the predefined categories (strong, mixed, and weak sellar barriers). The replacement process involves fine-tuning the pretrained weights while retraining the final layer to specialize the model for the sellar barrier classification task, leveraging the previously learned hierarchical features from ImageNet.
Key Layers Convolutional layers: Extract local patterns from the images, such as edges, textures, and shapes, using filters that traverse the image. These layers are essential for identifying specific and relevant structures in the visual data. Batch normalization layers: Regulate activations in intermediate layers by normalizing the values, reducing oscillations during training, and enabling faster and more stable convergence Dropout layers: Randomly deactivate a percentage of neurons during training, forcing the model to learn more generalizable features and reducing excessive dependence on specific patterns, thereby minimizing the risk of overfitting.
Transfer learning
Transfer learning is a powerful technique in deep learning where a model pretrained on a large dataset is adapted to perform a related but distinct task. This process capitalizes on the knowledge encoded in the pretrained model’s layers, especially its ability to extract generic features such as edges, textures, and patterns from images, applying it to new data with minimal retraining[
Figure 6:
Transfer Learning is a technique where a pre-trained Convolutional Neural Network (CNN), which has already learned to recognize general features from a large dataset, is reused for a new task. As shown in the image, the initial layers (in dark blue) are kept because they capture basic features like edges and shapes. Then, the final layers (indicated by the orange arrow) are replaced with new layers (highlighted in red), which are trained on a smaller, task-specific dataset. This approach saves time, requires less data, and still achieves high performance.
In this study, this approach enabled excellent results, even with a small dataset. Training a model from scratch with only 600 images, as in this case, would have yielded suboptimal performance due to insufficient data for learning complex patterns. However, using a pretrained model trained on thousands of images and fine-tuning these 600 images not only drastically reduced training time and resource consumption but also significantly improved performance. The lower layers of the EfficientNetB0 architecture, which specialize in capturing fundamental visual features, were frozen to retain their pretrained weights obtained from the ImageNet dataset. This allowed the model to leverage pre-learned representations, minimizing the data and computational resources required for training. The upper layers were replaced and retrained using images of the sellar barrier, fine-tuning the model to identify critical characteristics such as barrier thickness, irregularities, and contrast differences. This process involved optimizing weights in the newly added layers to enable accurate classification into the categories of strong, mixed, or weak sellar barriers.
The model processes over 4 million parameters per image, enabling it to capture complex anatomical nuances critical for high-precision classification. By leveraging transfer learning, this study achieved a substantial reduction in training time while maintaining robust performance, even with limited data. This strategy ensured that the pretrained model’s generalization capabilities were effectively adapted to the specific task of sellar barrier analysis [
Implementation
The programming and training were conducted on Google Colab, a cloud-based platform that allows users to write and execute Python code in a notebook environment. Google Colab is particularly suitable for deep learning tasks, as it provides free access to graphics processing unit (GPUs) and tensor processing unit (TPU), significantly accelerating processing.
The main libraries used included:
TensorFlow: Utilized for building, training, and deploying deep learning models, offering advanced tools for CNNs NumPy: Essential for numerical calculations, matrix management, and structured data manipulation Pandas: Used for organizing, analyzing, and cleaning data, facilitating dataset preparation for the model Matplotlib: Used for visualizing data and training results, such as loss curves and accuracy metrics, to monitor the model’s performance during training Scikit-learn: Provided tools for preprocessing data, splitting datasets, and evaluating model performance with metrics such as precision, recall, and F1 score OpenCV: Assisted in image preprocessing tasks such as resizing, normalization, and augmentation to enhance the robustness of the model.
RESULTS
The model was trained and evaluated over 20 epochs to optimize learning. In deep learning, an epoch refers to one complete pass through the entire training dataset by the model. Multiple epochs are typically used to refine the model’s performance as it iteratively adjusts its internal parameters to minimize error. In this study, the core of the pretrained EfficientNetB0 architecture, which contains the convolutional layers responsible for extracting fundamental visual features, was left unaltered. Only the final dense layers were retrained, allowing the model to specialize in classifying sellar barrier types. This approach drastically reduced training time and computational resources while retaining the rich feature representations learned from the large ImageNet dataset [
Figure 8:
The graph shows the relationship between epochs and the model’s performance. Accuracy represents the correctness of predictions, while loss indicates the prediction error. Training metrics reflect how well the model fits the training data, while validation metrics assess its generalization on unseen data.
Twenty epochs provided sufficient iterations to achieve convergence without overfitting, balancing computational efficiency and learning quality.
In this context, accuracy measures the percentage of correct predictions among all made, while recall reflects the proportion of correctly identified positive cases among all real positive cases, representing the model’s sensitivity to detect all relevant instances. The F1-score combines both metrics, providing a harmonic meaning that is especially useful in scenarios with imbalanced data.
The model achieved an overall accuracy of 96.33%. In detailed class analysis, strong barriers reached 98% accuracy, 97% recall, and a 98% F1-score. Mixed barriers achieved 95% accuracy, 94% recall, and a 94% F1-score, while weak barriers obtained 96% accuracy, 96% recall, and a 96% F1-score. These results demonstrate the model’s high capacity to accurately identify and differentiate the specific characteristics of each type of sellar barrier, even in complex clinical scenarios [
DISCUSSION
The sellar barrier is an important parameter to consider in the assessment of the risk of intraoperative CSF leakage. A prospective cohort study that included 155 cases of patients with supradiaphragmatic pituitary adenomas operated through an endoscopic approach demonstrated the correlation between the type of sellar barrier in MRI, and its in vivo characteristics observed during surgery, also detecting the differences between patients who developed an intraoperative CSF leak and those who did not.[
In this context, which validates the concept of the sellar barrier as a predictor of intraoperative CSF leakage, it motivates the creation of predictive models with the help of radiomics, which converts images into exploitable data to be subsequently analyzed for decision-making based on AI that includes algorithms capable of modeling themselves and improving accuracy through the analysis of data sets.[
The results obtained demonstrate the exceptional capability of the CNN model in accurately classifying sellar barriers in MRI images, even with a limited dataset. This success is primarily due to the transfer learning approach, which makes it possible to harness the knowledge embedded in pretrained models to address challenges associated with small datasets. In addition, techniques like dropout significantly reduced the risk of overfitting, enhancing the model’s ability to generalize effectively to unseen data.
The findings of this study underscore the potential of CNNs and transfer learning in addressing critical challenges in medical imaging and surgical planning. By achieving a classification accuracy of 97% for sellar barriers, the proposed model demonstrates its capacity to enhance preoperative diagnostics and mitigate complications, such as CSF leaks, in endoscopic endonasal approaches for pituitary adenomas.
This precise classification of sellar barriers carries critical clinical implications. By offering surgeons detailed predictions about the likelihood of CSF leaks, the model supports more informed surgical planning and facilitates the selection of appropriate reconstruction strategies. Such predictive power has the potential to improve surgical outcomes and reduce postoperative complications.
However, a notable concern among clinicians remains the “black box” nature of AI systems. This model, for example, operates using over 4 million parameters, making it nearly impossible to articulate its decision-making process transparently fully. While this complexity is fundamental to achieving high precision, it also creates challenges in building trust among medical professionals. To mitigate these concerns, the integration of an explainability module could be invaluable. Such a system would highlight the primary factors influencing the model’s decisions, offering clinicians a clearer understanding of its rationale and thereby fostering greater confidence in its outputs.
Expanding the dataset used for training is another avenue for improvement. By incorporating a broader and more diverse range of anatomical variations, the model’s robustness and generalizability could be further enhanced. This would not only increase its reliability across different clinical scenarios but also solidify its utility in real-world applications.
It is essential to emphasize that these AI systems are not intended to replace medical professionals. Instead, it will be the physicians who adopt and leverage these technologies who will replace those who do not. Furthermore, technology is an unstoppable force; resisting its adoption not only delays progress but often works against those who oppose it, highlighting the importance of embracing innovation for collective advancement.
Despite the challenges related to system opacity, the demonstrated ability of this model to deliver high-precision classification with minimal data and computational demands underscores its transformative potential in modern neurosurgery. Continued efforts to validate and refine such models in practical clinical environments will be crucial to ensuring their widespread adoption and their role in addressing complex surgical challenges.
Limitations
However, the study also highlights challenges, such as the “black box” nature of AI models, which can hinder clinician trust. Future work should focus on incorporating explainability modules to elucidate the decision-making processes of these models. Expanding the dataset to include diverse anatomical variations will further enhance the model’s generalizability and clinical applicability.
CONCLUSION
The use of the EfficientNetB0 architecture, coupled with transfer learning, proved effective in overcoming the constraints of small datasets. This approach maximizes performance while maintaining computational efficiency. Moreover, the high accuracy and precision across all barrier categories highlight the model’s robustness and its ability to generalize well to new data.
The clinical implications of this work are significant. Providing surgeons with reliable predictions about the integrity of the sellar barrier equips them to make informed decisions, optimizing surgical outcomes and patient safety. In addition, this methodology offers a pathway to integrate AI-powered tools seamlessly into clinical workflows, improving diagnostic accuracy and reducing variability.
The proposed AI model represents a promising step toward the integration of advanced machine learning techniques in neurosurgical practice, paving the way for more precise, reliable, and patient-centered approaches to managing pituitary lesions.
Ethical approval:
The research/study was approved by the Institutional Review Board at Hospital Padilla, Tucumán, Argentina. Approval Number: 2025-107 dated January 3, 2025.
Declaration of patient consent:
The authors certify that they have obtained all appropriate patient consent.
Financial support and sponsorship:
Nil.
Conflicts of interest:
There are no conflicts of interest.
Use of artificial intelligence (AI)-assisted technology for manuscript preparation:
The authors confirm that there was no use of artificial intelligence (AI)-assisted technology for assisting in the writing or editing of the manuscript and no images were manipulated using AI.
Disclaimer
The views and opinions expressed in this article are those of the authors and do not necessarily reflect the official policy or position of the Journal or its management. The information contained in this article should not be considered to be medical advice; patients should consult their own physicians for advice as to their specific medical needs.
References
1. Avanzo M, Stancanello J, Pirrone G, Drigo A, Retico A. The evolution of artificial intelligence in medical imaging: From computer science to machine and deep learning. Cancers (Basel). 2024. 16: 3702
2. Bayar B, Stamm MC. Constrained convolutional neural networks: A new approach towards general purpose image manipulation detection. IEEE Trans Inform Forens Secur. 2018. 13: 2691-706
3. Campero A, Villalonga JF, Basso A. Anatomical risk factors for intraoperative cerebrospinal fluid leaks in transsphenoidal surgery for pituitary adenomas. World Neurosurg. 2019. 50: 1878-87
4. Chen L, Li R, Zhang H, Tian L, Chen N. Intelligent fall detection method based on accelerometer data from a wrist-worn smart watch. Measurement. 2019. 140: 215-26
5. Dhillon A, Verma GK. Convolutional neural network: A review of models, methodologies and applications to object detection. Prog Artif Intell. 2020. 9: 85-112
6. Esposito F, Dusick JR, Fatemi N, Kelly DF. Graded repair of cranial base defects and cerebrospinal fluid leaks in transsphenoidal surgery. Oper Neurosurg (Hagerstown). 2007. 60: 295-9
7. Guo W, Dong Y, Hao GF. Transfer learning empowers accurate pharmacokinetics prediction of small samples. Drug Discov Today. 2024. 29: 103946
8. Guo Z, Shen Y, Aloqaily M, Jararweh Y, Yu K. Probabilistic inference-based modeling for sustainable environmental systems under hybrid cloud infrastructure. Simul Model Pract Theory. 2021. 107: 102215
9. Guo Z, Yu K, Jolfaei A, Ding F, Zhang N, editors. Fuz-spam: Label smoothing-based fuzzy detection of spammers in internet of things. IEEE transactions on fuzzy systems. United States: IEEE computational intelligence society; 2021. p. 1
10. Hosny A, Parmar C, Quackenbush J, Schwartz LH, Aerts HJ. Artificial intelligence in radiology. Nat Rev Cancer. 2018. 18: 500-10
11. Kang X, Zhuo B, Duan P. Dual-path network-based hyperspectral image classification. IEEE Geosci Remote Sens Lett. 2019. 16: 447-51
12. Khan A, Sohail A, Zahoora U, Qureshi AS. A survey of the recent architectures of deep convolutional neural networks. Artif Intell Rev. 2020. 53: 5455-516
13. Khan S, Rahmani H, Shah SA, Bennamoun MA. Guide to convolutional neural networks for computer vision. Synth Lect Comput Vision. 2018. 8: 1-207
14. Kourounis G, Elmahmudi AA, Thomson B, Hunter J, Ugail H, Wilson C. Computer image analysis with artificial intelligence: A practical introduction to convolutional neural networks for medical professionals. Postgrad Med J. 2023. 99: 1287-94
15. Lotan E, Jain R, Razavian N, Fatterpekar GM, Lui YW. State of the art: Machine learning applications in glioma imaging. AJR Am J Roentgenol. 2019. 212: 26-37
16. Makimoto H, Höckmann M, Lin T, Glöckner D, Gerguri S, Clasen L. Performance of a convolutional neural network derived from an ECG database in recognizing myocardial infarction. Sci Rep. 2020. 10: 8445
17. Manoharan S. Image detection, classification and recognition for leak detection in automobiles. J Innov Image Process. 2019. 1: 61-70
18. Maray N, Ngu AH, Ni J, Debnath M, Wang L. Transfer learning on small datasets for improved fall detection. Sensors (Basel). 2023. 23: 1105
19. Meng D, Xiao Y, Guo Z, Jolfaei A, Qin L, Lu X. A data-driven intelligent planning model for UAVs routing networks in mobile internet of things. Comput Commun. 2021. 179: 231-41
20. Peng Y, Jolfaei A, Yu K. A novel real-time deterministic scheduling mechanism in industrial cyber-physical systems for energy internet. IEEE Trans Ind Inform. 2022. 18: 5670-80
21. Renuka B, Sivaranjani B, Lakshmi AM, Muthukumaran N. Automatic enemy detecting defense robot by using face detection technique. Asian J Appl Sci Technol. 2018. 2: 495-501
22. Rivera-Serrano CM, Snyderman CH, Gardner P, Prevedello D, Wheless S, Kassam AB. Nasoseptal “rescue” flap: A novel modification of the nasoseptal flap technique for pituitary surgery. Laryngoscope. 2011. 121: 990-3
23. Senders JT, Zaki MM, Karhade AV, Chang B, Gormley WB, Broekman ML. An introduction and overview of machine learning in neurosurgical care. Acta Neurochir (Wien). 2018. 160: 29-38
24. Sharma P, Berwal YP, Ghai W. Performance analysis of deep learning CNN models for disease detection in plants using image segmentation. Inform Process Agric. 2020. 7: 566-74
25. Soffer S, Ben-Cohen A, Shimon O, Amitai MM, Greenspan H, Klang E. Convolutional neural networks for radiologic images: A radiologist’s guide. Radiology. 2019. 290: 590-606
26. Solari D, Pivonello R, Caggiano C, Guadagno E, Chiaramonte C, Miccoli G. Pituitary adenomas: What are the key features?. What are the current treatments? Where is the future taking US? World Neurosurg. 2019. 127: 695-709
27. Tacconi C, Mellone S, Chiari L, editors. Smartphone-based applications for investigating falls and mobility. Proceedings of the 2011 5th international conference on pervasive computing technologies for healthcare (pervasive health) and workshops. Dublin, Ireland: 2011. p. 258-61
28. Tan L, Yu K, Lin L, Cheng X, Srivastava G, Lin JC. Speech emotion recognition enhanced traffic efficiency solution for autonomous vehicles in a 5G-enabled space-air-ground integrated intelligent transportation system. IEEE Trans Intell Transportation Syst. 2022. 23: 2830-42
29. Topol EJ. As artificial intelligence goes multimodal, medical applications multiply. Science. 2023. 381: adk6139
30. Versaci M, Morabito FC. Image edge detection: A new approach based on fuzzy entropy and fuzzy divergence. Int J Fuzzy Syst. 2021. 23: 918-36
31. Villalonga JF, Fuchssteiner C, Solari D, Campero A, Cavallo LM, Cappabianca P. Endoscopic anatomy of the sellar barrier: From the anatomical model to the operating room. Clin Anat. 2020. 33: 468-74
32. Villalonga JF, Ries-Centeno T, Sáenz A, Solari D, Cervio A, Campero A. The mixed sellar barrier: A new subtype of this novel concept. World Neurosurg. 2019. 132: e5-13
33. Villalonga JF, Solari D, Cavallo LM, Cappabianca P, Prevedello DM, Carrau R. The sellar barrier on preoperative imaging predicts intraoperative cerebrospinal fluid leak: A prospective multicenter cohort study. Pituitary. 2021. 24: 27-37
34. Villalonga JF, Solari D, Cuocolo R, De Lucia V, Ugga L, Gragnaniello C. Clinical application of the “sellar barrier’s concept” for predicting intraoperative CSF leak in endoscopic endonasal surgery for pituitary adenomas with a machine learning analysis. Front Surg. 2022. 9: 934721
35. Wang P, Qiao J, Liu N. An improved convolutional neural network-based scene image recognition method. Comput Intell Neurosci. 2022. 2022: 3464984
36. Yadav SS, Jadhav SM. Deep convolutional neural network based medical image classification for disease diagnosis. J Big Data. 2019. 6: 113
37. Yang L, Yu K, Yang SX, Chakraborty C, Lu Y, Guo T, editors. An intelligent trust cloud management method for secure clustering in 5G enabled internet of medical things. United States: IEEE Transactions on Industrial Informatics; 2021. p. 1
38. Yang Z, Huang Y, Jiang Y, Sun Y, Zhang YJ, Luo P. Clinical assistant diagnosis for electronic medical record based on convolutional neural network. Sci Rep. 2018. 8: 6329
39. Yao P, Wu H, Gao B, Tang J, Zhang Q, Zhang W. Fully hardware-implemented memristor convolutional neural network. Nature. 2020. 577: 641-6
40. Yasin M, Sarıgül M, Avci M. Logarithmic learning differential convolutional neural network. Neural Netw. 2024. 172: 106114
41. Yu K, Tan L, Mumtaz S, Alrubaye S, Al-Dulaimi A, Bashir AK. Securing critical infrastructures: Deep-learning-based threat detection in IIoT. IEEE Commun Mag. 2021. 59: 76-82
42. Zhao L, Chai H, Han Y, Yu K, Mumtaz S. A collaborative V2X data correction method for road safety. IEEE Trans Reliab. 2022. 71: 951-62
43. Zhou DX. Universality of deep convolutional neural networks. Appl Comput Harmon Anal. 2020. 48: 787-94