Next Article in Journal
Channel Characterization and Modeling for 6G UAV-Assisted Emergency Communications in Complicated Mountainous Scenarios
Next Article in Special Issue
Differentiation between Phyllodes Tumors and Fibroadenomas through Breast Ultrasound: Deep-Learning Model Outperforms Ultrasound Physicians
Previous Article in Journal
Pankun: A New Generation of Broadband Ocean Bottom Seismograph
Previous Article in Special Issue
Deep Learning Strategy for Sliding ECG Analysis during Cardiopulmonary Resuscitation: Influence of the Hands-Off Time on Accuracy
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Deep Learning Aided Neuroimaging and Brain Regulation

1
Center for Cognition and Neuroergonomics, State Key Laboratory of Cognitive Neuroscience and Learning, Beijing Normal University, Zhuhai 519087, China
2
Centre for Cognitive and Brain Sciences, Institute of Collaborative Innovation, University of Macau, Macau SAR 999078, China
3
Nanomicro Sino-Europe Technology Company Limited, Zhuhai 519031, China
4
Jiangfeng China-Portugal Technology Co., Ltd., Macau SAR 999078, China
*
Authors to whom correspondence should be addressed.
These authors contributed equally to this work.
Sensors 2023, 23(11), 4993; https://doi.org/10.3390/s23114993
Submission received: 24 April 2023 / Revised: 15 May 2023 / Accepted: 22 May 2023 / Published: 23 May 2023

Abstract

:
Currently, deep learning aided medical imaging is becoming the hot spot of AI frontier application and the future development trend of precision neuroscience. This review aimed to render comprehensive and informative insights into the recent progress of deep learning and its applications in medical imaging for brain monitoring and regulation. The article starts by providing an overview of the current methods for brain imaging, highlighting their limitations and introducing the potential benefits of using deep learning techniques to overcome these limitations. Then, we further delve into the details of deep learning, explaining the basic concepts and providing examples of how it can be used in medical imaging. One of the key strengths is its thorough discussion of the different types of deep learning models that can be used in medical imaging including convolutional neural networks (CNNs), recurrent neural networks (RNNs), and generative adversarial network (GAN) assisted magnetic resonance imaging (MRI), positron emission tomography (PET)/computed tomography (CT), electroencephalography (EEG)/magnetoencephalography (MEG), optical imaging, and other imaging modalities. Overall, our review on deep learning aided medical imaging for brain monitoring and regulation provides a referrable glance for the intersection of deep learning aided neuroimaging and brain regulation.

1. Introduction

In recent years, the medical field has become one of the most important research and application fields in the artificial intelligence (AI) industry. Machine learning and its subset, deep learning, are branches of AI, and have shown promising findings in the medical field, especially when applied to imaging data, which have been used in radiological diagnosis, bioinformatics, genome sequencing, drug development, and histopathological image analysis [1,2]. Segmentation of brain disease lesions can provide imaging biomarkers of disease burden that can help monitor disease progression and the imaging response to treatment [3]. Particularly for histopathological diagnosis, AI has exhibited a potential ability that matches that of medical experts. The AI aided neuroimaging and brain regulation generally employed distinguished deep learning and conventional machine learning (include rule-based learning). Classical machine learning methods such as support vector machine (SVM) or random forest require a well-prepared feature engineering procedure, in other words, need to manually segment morphological features and select import features, which is extremely time consuming and tends to show a large performance difference between different operators [4]. As AI techniques continue to be refined and improved, deep learning has been proposed to dramatically change the health care monitoring and regulation of the brain [5], which can not only improve the reconstruction accuracy of neuroimaging and achieve fast imaging, but also mine a large amount of pathological and genetic data by processing and cross-referencing health and medical big data such as images, pathology, and genes, and help pathologists to evaluate pathological sections faster to improve the efficiency and prognosis of disease diagnosis [6]. Deep learning is a special type of machine learning. The advantage of deep learning is to utilize a neural network-like engineering architecture that can detect and extract import features automatically, whose predicting label and assessment algorithm has developed several varieties including prognosis, immune-score, microsatellite instability, histological subtyping, microenvironment analysis and segmentation, etc. Extensive studies have revealed high accuracies and provided excellent examples of deep learning’s potential in brain health care [7,8,9]. Thanks to advancements in neuroimaging technology, numerous neuroimaging datasets of brains have been gathered from multiple imaging facilities utilizing various scanners and imaging protocols. These datasets have been collected to study both typical and atypical brain development [10]. Recent advancements in medical imaging of the brain have greatly solved challenges from biopsy—an invasive, unrepeatable technique that usually ignore heterogeneity within parenchyma. It uses data characterization algorithms to convert conventional imaging information into data matrices by modern linear algebra and statistics, which can be further extracted into information revealing a certain malady.
Currently, deep learning-aided medical imaging is becoming the hot spot of AI frontier application and the future development trend of precision neuroscience. Clinical decision-making, an essential element of medicine, involves judgement with the integration of comprehensive data [11]. Especially in brain or central nervous system diseases, there are unique challenges in medical decision-making due to their diverse forms and progression as well as the need to consider individual patient factors such as their ability to receive treatment and response to it. The early detection of cancer is crucial in saving thousands of lives. Targeted therapy for cancer heavily relies on its grading [12]. Due to the invasive, time-consuming, and expensive nature of cancer diagnosis, there is an urgent need to develop non-invasive, cost-effective, and efficient tools to characterize and estimate the grade of brain cancer. MRI, PET/CT, EEG/MEG, optical imaging, and other imaging modalities offer quick and safer options for tumor detection during brain scans. The utilization of deep learning in molecular diagnosis, prognosis, and treatment monitoring has resulted in the creation of a structured resource for radiogenomic analysis of brain or central nervous system diseases. Besides greatly reducing the scan time of neuroimaging methods like MRI and PET/CT [13], the deep learning aided medical images acquired better signal to noise ratio, higher contrast-to-noise ratio, and stronger brain or central nervous system disease lesion detection ability. Therefore, radiomics was thought to be the bridge between medical imaging and personalized medicine [14], which is a quantitative approach to medical imaging that involves the extraction and analysis of large amounts of quantitative data from medical images such as MRI and PET/CT. These data are then used to create predictive models that can help doctors make more accurate diagnoses, predict treatment outcomes, and develop personalized treatment plans. Deep learning is a subset of machine learning that uses artificial neural networks to analyze and learn from data. In radiomics, deep learning algorithms can be trained on large datasets of medical images to identify patterns and features that are not visible to the human eye. This can help radiologists and other medical professionals to more accurately diagnose and treat a wide range of medical conditions including cancer, neurological disorders, and cardiovascular disease. By applying deep learning to medical imaging, radiomics has the potential to revolutionize the way we diagnose and treat disease. It can help to identify early-stage disease, predict which patients will respond best to which treatments, and develop new, personalized treatment plans based on a patient’s unique genetic and environmental factors.
Deep learning algorithms have been used to accurately segment brain tumors in MRI scans, diagnosis of Alzheimer’s disease, brain–computer interfaces (BCI), brain stimulation, and so on. For example, a study used a deep learning algorithm to segment brain tumors in MRI scans with high accuracy. The algorithm was trained on a large dataset of MRI scans and was able to accurately segment tumors, even in cases where the tumor was irregularly shaped or located close to the brain’s surface [15]. In another study, Xiaojun Bi and Haibo Wang proposed a discriminative version of a contractive slab and spike convolutional deep Boltzmann machine model (DCssCDBM) with a multi-task learning framework via EEG spectral images based on identification and verification tasks for overfitting reduction for the first time, and the method accurately predicted whether a patient had Alzheimer’s disease or not [16]. Deep learning algorithms have also been used to improve the accuracy and speed of BCI systems, which allow people to control computers or other devices with their thoughts. For example, a study used a novel deep learning algorithm to improve the accuracy of a BCI system for hand movement detection and the algorithm was able to detect hand movements with high accuracy [17]. Deep learning algorithms have also been applied to optimize brain stimulation techniques such as transcranial magnetic stimulation (TMS), which is used to treat depression and other mental health conditions. For example, a study used connectivity measures and an ensemble of pre-trained deep learning models to predict the treatment outcome of repetitive TMS in major depressive disorder to improve the treatment efficacy and reduce health care costs. The methodology possesses effective connectivity, used for transforming EEG signals to images, and provides an informative feature map [18].
Deep learning holds the potential to greatly improve the brain medical image quality, metastasis detection, radiogenomics, and treatment response monitoring, which can assist with volumetric delineation of brain lesions over time, extrapolation of the genotype and biological course from radiographic phenotype, prediction of clinical outcomes, and assessment of the impact of disease and treatment on the surrounding encephalic region [19,20]. By automating the initial image interpretation, deep learning may revolutionize the clinical workflow of radiographic detection, management decisions, interventions, and follow-up care in ways yet to be envisioned [21]. With more and more deep learning aided medical imaging and therapy decision-making strategies entered into the national authoritative professional society and cancer diagnosis and treatment guidelines that are followed by across the country, deep learning assisted theranostic methods are widely recognized by clinical doctors and lay a solid foundation for its large-scale clinical application. Therefore, in vivo image acquisition and data signal analysis, data grouping storage, separate and detailed recording of each sample’s experimental data, and through the professional analysis module for high-precision quantitative analysis of imaging data, can analyze multiple groups of data at the same time to ensure the consistency of the experimental data [22]. Although there are challenges to overcome such as inter-scanner variability, the need for benchmark datasets, and prospective validations for clinical applicability, there is a significant opportunity for the development of optimal solutions for brain or central nervous system disease stratification. These solutions can provide immediate recommendations for further diagnostic decisions, the guidance of deep brain stimulation target identification and personalized treatment plan optimization [23,24,25]. In a word, using deep learning to assist medical imaging for brain theranostics has the characteristics of objectivity, high-accuracy, and high efficiency beyond the abilities of human judgement from qualitative to quantitative imaging. This review is schematically illustrated in Figure 1.

2. Evolution and Classification of Deep Learning Assisted Medical Imaging

2.1. Evolution of Artificial Intelligence in Medical Imaging

The development of AI and deep learning can be traced back to the 1950s and 1960s [26,27]. Early AI research included rule-based systems that relied on human-written rules to solve specific problems. However, the capabilities of these systems were very limited because they needed to manually write all the rules, and they struggled to cope with complex and uncertain situations [28]. In the 1980s and 1990s, machine learning became popular as improved computer power and large amounts of data became available. The idea is for computers to learn patterns from data so they can better handle new data. One of the important machine learning techniques is neural networks, which are algorithms inspired by the human nervous system. However, early neural networks were very shallow, with only a few layers, limiting their capabilities. With the development of deep learning algorithms, neural networks have become deeper and more complex, which can better handle large amounts of data and have achieved many breakthrough results. The most representative examples are the application of deep learning in image and speech recognition as well as its success in fields such as natural language processing and machine translation. In recent years, with the improvement in computer performance and algorithms, the application scope of AI and deep learning has been expanding, gradually penetrating into various fields and achieving more and more success [29], showing its advantage in real scenarios including lung nodules in chest CT, neuroimaging, mammography, and so on.
AI is transforming medical imaging and driving it forward toward the future at a rapid pace. The evolution of AI in medical imaging has been a game-changer in the field of health care. Medical imaging is a critical component of medical diagnosis and treatment, and AI has significantly improved the accuracy, efficiency, and speed of medical imaging processes [30,31]. The early application of AI in medical imaging focused on computer-aided diagnosis, which involved using algorithms to detect and classify lesions or abnormalities in medical images [32,33,34]. However, with the rise of deep learning, the application of AI in medical imaging has become more sophisticated, enabling the development of predictive models, automatic image segmentation, and even image synthesis [35]. One of the significant advantages of AI in medical imaging is its ability to analyze vast amounts of data quickly and accurately, especially in image segmentation, registration, detection, and recognition. This has led to the development of systems that can detect and diagnose diseases with high accuracy such as lung cancer, head and neck cancer, breast cancer, and diabetic retinopathy [36]. Moreover, AI has facilitated the automation of repetitive tasks, freeing up time for medical professionals to focus on more complex cases [37,38]. AI has also enabled the development of personalized medicine, where treatments can be tailored to individual patients based on their genetic makeup, medical history, and imaging data. This has led to better patient outcomes, reduced costs, and improved overall health care efficiency. However, the adoption of AI in medical imaging is not without its challenges. One of the main concerns is the potential for bias in AI algorithms, which can lead to inaccurate diagnoses or treatment recommendations [1]. Moreover, there is a need for transparent and ethical AI practices including the development of regulatory frameworks, to ensure that AI is used safely and effectively. Generally, the evolution of AI in medical imaging has revolutionized the field of health care, with the potential for improving patient outcomes, reducing costs, and increasing overall efficiency. The continued development and adoption of AI in medical imaging will undoubtedly lead to further advancements in health care and personalized medicine, which has the potential to not only revolutionize traditional medical imaging, but also enhance clinical workflows and transform various aspects of the health care industry.

2.2. Convolutional Neural Networks (CNNs)

The application of convolutional neural networks (CNNs) in medical imaging, especially in brain monitoring and modulation, has made remarkable progress [39,40,41]. CNNs are deep learning algorithms, which has been widely used in image recognition and classification. It can automatically extract features and patterns from a large amount of data. In terms of brain monitoring, CNNs can be used to analyze neuroimaging data such as MRI and CT scan results as well as physiological data such as EEG and MEG. CNNs can automatically identify brain structures and activity patterns to help doctors diagnose and treat them [42,43,44,45]. In terms of brain regulation, CNNs can be used in brain–computer interface technology, which is a technology that converts electrical brain signals into machine commands [46]. Using CNNs to analyze EEG signals, EEG patterns such as event-related potentials and bands can be identified to help people realize brain-controlled devices and applications such as wheelchairs and games. In addition, CNNs can also be used to predict and monitor the progression of brain diseases and response to treatment. By analyzing brain images and physiological data, CNNs can automatically identify pathological patterns and features to predict disease progression and treatment effects [47,48,49]. In general, the application of CNNs in medical imaging, especially in brain monitoring and regulation, provides doctors and patients with more accurate, faster, and more effective treatment methods, which is expected to bring more progress and innovation to the medical health field in the future.

2.3. Recurrent Neural Networks (RNNs)

The application of recurrent neural networks (RNNs) in medical imaging, especially in brain monitoring and regulation, has also made some progress [50,51,52]. Unlike CNN, cyclic neural networks can process data with a sequential structure such as time series and speech signals [53,54,55]. In terms of brain monitoring, RNNS can be used to analyze time series data such as EEG and MEG as well as functional magnetic resonance imaging (fMRI) data. RNNs can recognize brain wave patterns and trends, and make predictions and classifications based on the dynamic nature of the data. In terms of brain regulation, RNNS can be used in brain–computer interface technology (BCIs), which is a technique that converts electrical signals in the brain into machine instructions. By analyzing EEG signals with a RNN, patterns and dynamic changes of EEG can be identified to help people realize more refined and flexible brain control devices and applications. In addition, RNNS can be used to predict brain diseases and monitor therapeutic responses [56,57,58]. By analyzing time series and dynamic data, RNNs can automatically identify brain pathological patterns and features to predict disease progression and treatment effects [59]. In general, the application of RNN in medical imaging, especially in brain monitoring and regulation, provides doctors and patients with a more comprehensive, precise, and dynamic treatment style, which is expected to bring more progress and innovation to the medical health field in the future.

2.4. Generative Adversarial Networks (GANs)

The application of a generative adversarial network (GAN) in medical imaging, especially in brain monitoring and regulation, has also received more and more attention and research. GAN is a deep learning algorithm that can be used to generate images and data with specific characteristics and attributes [60,61,62,63,64]. In terms of brain monitoring, GAN can be used to generate virtual brain images with specific brain structures and activity patterns. This can help doctors understand the relationship between different brain structures and activity patterns, and the impact on different brain diseases [65,66,67,68,69,70]. In addition, GAN can also be used to synthesize brain images with different pathological features and trends to help doctors diagnose and treat diseases. In terms of brain regulation, GAN can be used in brain–computer interface technology (BCI) to help train and optimize brain-control devices and applications by generating virtual images and data with specific patterns of brain activity. In addition, GAN can be used to generate virtual images and data with specific brain pathological patterns to help researchers explore mechanisms and treatments for brain diseases [71,72,73,74]. Broadly speaking, the application of GAN in medical imaging, especially in brain monitoring and regulation, provides a more comprehensive, targeted, and innovative treatment for doctors and patients, and is expected to bring more progress and innovation to the medical health field in the future. The applications of different neuron network-assisted brain health care monitoring are illustrated in Figure 2.

3. Deep Learning Aided Neuroimaging for Brain Monitoring and Regulation

Deep learning has shown tremendous potential in the field of neuroimaging and brain regulation. Neuroimaging techniques such as MRI, CT, PET/CT, EEG/MEG, optical imaging, and other imaging modalities generate large amounts of comprehensive and complex data, which can be challenging to analyze and interpret. Deep learning techniques such as CNNs, RNNs, and GANs have been proven to be effective in extracting meaningful information from these data and transforming the neuroimaging from qualitative to quantitative imaging modality. The aforementioned information is merged with additional patient data and processed using advanced bioinformatics software to create models that could potentially enhance the accuracy in the diagnosis, prognosis, and prediction for brain monitoring and regulation.

3.1. Deep Learning Assisted MRI

MRI scans of the brain are considered the most effective approach for identifying chronic neurological disorders such as brain tumors, dementia, stroke, and multiple sclerosis. They are also the preferred method for detecting conditions affecting the pituitary gland, brain vessels, inner ear organs, and eyes due to their high sensitivity. In recent years, several deep learning-based medical image analysis methods have been introduced to facilitate health monitoring and diagnosis using brain MRI scans [75,76,77]. One of the primary applications of deep learning in neuroimaging by MRI is the identification and classification of neurological disorders. For example, CNNs have been used to accurately diagnose Alzheimer’s disease, Parkinson’s disease, and multiple sclerosis by analyzing MRI scans. Deep learning has also shown potential in identifying different stages of brain development, detecting early signs of neurological disorders, and predicting the progression of these disorders.
Recent advancements in the classification of gliomas based on biological genotypes, coupled with the utilization of computational deep learning models based on multi-modal MRI biomarkers, offer a promising avenue for customized and effective treatment plans. In this regard, deep learning-based assessment of gliomas using hand-crafted or auto-extracted features derived from MRI has emerged as a critical tool, as genomic alterations can be correlated with MRI-based phenotypes [78]. Deep learning algorithms have been extensively explored for the purpose of classifying neurodegenerative diseases using medical imaging techniques such as magnetic resonance imaging. Utilizing CNNs on MRI data has emerged as a promising technique for achieving exceptional levels of precision in predicting the progression of neurological conditions such as brain tumors, Alzheimer’s disease, multiple sclerosis, and stroke by capturing image features that are not detectable using traditional methods. However, little attention has been given to utilizing post-mortem immunofluorescence imaging studies of patients’ brains for this purpose. These studies have the potential to be a valuable tool in detecting abnormal chemical changes or pathological post-translational modifications of the Tau polypeptide. Therefore, L. Diaz-Gomez et al. proposed a CNN pipeline that utilized transfer learning to analyze post-mortem immunofluorescence images with different Tau biomarkers for the classification of Tau pathology in Alzheimer’s disease and progressive supranuclear palsy. The ResNet-IFT architecture was used to generate models, and interpretability algorithms such as Guided Grad-CAM and occlusion analysis were employed to interpret the outputs of these models. They tested four different architectures to determine the best classifier, and the results showed that their design was able to classify diseases with an average accuracy of 98.41%. Additionally, they were able to provide an interpretation of the classification, which included different structural patterns in the immunoreactivity of the Tau protein in NFTs present in the brains of patients with progressive supranuclear palsy and Alzheimer’s disease [79]. O. Ozkaraca et al. created a new modular deep learning model to enhance the classification accuracy of MRI images while simultaneously addressing the drawbacks of prevalent transfer learning approaches like DenseNet, VGG16, and basic CNN architectures. They employed brain tumor images from the Kaggle database to train and test their model using two distinct data splitting methods: 80% for training and 20% for testing, and 10-fold cross-validation. Although the proposed deep learning model demonstrated better classification performance compared to other transfer learning methods, it required more processing time [75]. In another study, T. Chattopadhyay et al. utilized 3D CNN to forecast Abeta+ based on 3D brain MRI data derived from 762 elderly participants (mean age: 75.1 years ± 7.6SD; 394F/368M; 459 healthy controls, 67 with MCI, and 236 with dementia) who were scanned as part of the Alzheimer’s Disease Neuroimaging Initiative. The 3D CNN accurately projected Abeta+ with a 76% balanced accuracy from T1w scans [80]. Exploring CNN-generated attention maps, which identify the most significant anatomical features used for CNN-driven decisions, holds the potential to unveil crucial disease mechanisms that contribute to the accumulation of disability. L. Coll et al. predicted a class of multiple sclerosis disability using whole-brain MRI scans as input by a 3D-CNN model, which achieved a mean accuracy of 79% and proved to be superior to the equivalent logistic regression model (77%). The model was also successfully validated in the independent external cohort without any re-training (accuracy = 71%) [81]. Perinatal arterial ischemic stroke has been linked to unfavorable neurological outcomes. However, the evaluation of ischemic lesions and the subsequent development of the brain in newborns requires time-consuming manual inspection of brain tissues and ischemic lesions. Therefore, R. Zoetmulder et al. proposed an automatic method that used CNNs to segment brain tissues and ischemic lesions in the MRI scans of infants suffering from perinatal arterial ischemic stroke. This method eliminates the need for the labor-intensive manual assessment of brain tissues and ischemic lesions [82]. This study indicates that the automatic segmentation of brain tissue and ischemic lesions in the MRI scans of patients is feasible and may allow for the evaluation of the brain development and efficacy of treatment in large datasets.

3.2. Deep Learning Assisted PET/CT

PET/CT provides powerful diagnosis methods for neurodegenerative disorder by identifying disease-specific pathologies. Deep learning techniques have shown great promise in enhancing PET and CT imaging for neuroimaging and brain monitoring/regulation. These techniques can help improve the accuracy, speed, and efficiency of image processing, enabling more effective analysis and interpretation of neuroimaging data. Three-dimensional CNN can be trained to denoise the PET images for each disease cohort of neurodegenerative disorders [83] and predict the diagnosis of dementia with Lewy bodies, Alzheimer’s disease, and mild cognitive impairment [84] as well as amyloid standardized uptake value ratio through PET for Alzheimer’s prognosis [85]. One example of deep learning-assisted neuroimaging is the use of convolutional neural networks (CNNs) to improve the accuracy of PET image segmentation. In one study, the researchers developed a CNN-based segmentation method that achieved higher accuracy (96%), sensitivity (96%), and specificity (94%) than the traditional methods in the evaluation of neuro images for the diagnosis of Alzheimer’s disease, which was evaluated using the 18FDG-PET images of 855 patients including 635 normal control and 220 Alzheimer’s disease patients from the ADNI database, thus capable of discriminating the normal control from the Alzheimer’s disease patients [86].
Another example is the use of deep learning techniques to enhance the quality and resolution of CT imaging in neuroimaging. In a study published in the journal Radiology, the researchers constructed and trained a deep learning-based stenosis and plaque classification algorithm for head and neck CT angiography that achieved 85.6% consistency between radiologists and the DL-assisted algorithm, and reduced the time needed for diagnosis and report writing by the radiologists from 28.8 min ± 5.6 to 12.4 min ± 2.0 (p < 0.001) [87]. In addition to image processing, deep learning techniques have also been used to analyze neuroimaging data such as semantic segmentation and quantification of intracerebral hemorrhage (ICH), perihematomal edema, and intraventricular hemorrhage on non-contrast CT scans of patients with spontaneous ICH [88]. Overall, deep learning techniques hold great promise for improving the accuracy, speed, and efficiency of PET and CT imaging for neuroimaging and brain monitoring/regulation. With further development and refinement, these techniques could revolutionize the field of neuroimaging and contribute to a better understanding of brain function and dysfunction.

3.3. Deep Learning Assisted EEG/MEG

Deep learning has become an increasingly popular tool in EEG/MEG neuroimaging and brain monitoring/regulation. With the ability to analyze large datasets and detect subtle patterns in neural activity, deep learning has shown great potential in enhancing our understanding of brain function and informing clinical applications [89,90]. Deep learning has also been used to improve brain regulation techniques such as EEG neurofeedback. EEG neurofeedback is a non-invasive technique that aims to regulate brain activity by providing real-time feedback to the patient. Deep learning algorithms can analyze EEG data in real-time, detect patterns, and provide targeted feedback to patients to help them regulate their brain activity. One of the main applications of deep learning in EEG/MEG is in the classification of brain states or disorders. For example, Chambon et al. used a CNN to classify EEG data into three different cognitive states, achieving an accuracy of up to 90% [91]. In another study, Lawhern et al. used a deep learning model to classify EEG data into different types of epileptic seizures with high accuracy, demonstrating the potential of deep learning in aiding clinical diagnosis [92].
Moreover, deep learning techniques have been used in the development of brain–computer interfaces (BCIs) that allow patients to control external devices such as prosthetic limbs or computer interfaces using their brain activity. Deep learning algorithms can extract meaningful information from EEG or fMRI data and translate it into commands for external devices. Deep learning has also been used for brain activity prediction and regulation. For instance, M. Dinov et al. used deep reinforcement learning in closed-loop behavioral-and neuro-feedback to track and optimize human performance [93], in which a deep learning model was established to predict individualized EEG signals and applied to a closed-loop system for real-time neurofeedback, achieving the successful regulation of brain activity. Recent technological advances such as wireless recording, deep learning analysis, and real-time temporal resolution have increased interest in EEG-based brain–computer interface approaches [94,95]. A deep learning model was developed for real-time decoding of MEG signals, which was applied to a brain–computer interface system for regulating motor imagery tasks [96]. Epilepsy is a chronic brain disorder in which functional changes may precede structural ones and which may be detectable using existing modalities [97]. Functional connectivity analysis using EEG and resting state-functional magnetic resonance imaging (rs-fMRI) can localize epilepsy [98,99]. Finally, deep learning has been used for feature extraction and representation learning in EEG/MEG data. For example, R. Hussein et al. used a deep learning model to learn features from raw EEG data, which were then used for the classification of epileptic seizures [100]. Similarly, Roy et al. used a deep learning smart health monitoring model with the spectral analysis of scalp EEG to automatically predict epileptic seizures [101].
In summary, deep learning has shown great potential in enhancing EEG/MEG neuroimaging and brain monitoring/regulation. By enabling accurate classification of brain states and disorders, predicting and regulating brain activity, and learning meaningful representations from EEG/MEG data, deep learning has the potential to revolutionize our understanding of brain function and inform clinical applications.

3.4. Deep Learning Assisted Optical Neuroimaging and Others

Optical imaging is a non-invasive technique that uses light to visualize tissue structure and function. Optical neuroimaging holds great promise for imaging guided brain regulation. For example, a study published in Nature Neuroscience in 2020 used deep learning to predict behavior from functional imaging data in mice, demonstrating the potential for using deep learning in real-time behavioral prediction and manipulation [102]. The non-invasive guidance of therapeutic strategies would enable the removal of cancerous tissue while avoiding side effects and systemic toxicity, preventing damage to healthy tissues and decreasing the risk of postoperative problems such as bioluminescence imaging (BLI), fluorescence imaging (FI), Cerenkov luminescence imaging (CLI), and photoacoustic imaging (PAI). BLI is always used in small animal imaging for the in vivo tracking of therapeutic gene expression and cell-based therapy. In contrast, FI is highly promising for clinical translation. The applications of FI include image-guided surgery, radiotherapy, gene therapy, drug delivery, and sentinel lymph node fluorescence mapping. CLI is a novel radioactive optical hybrid imaging strategy for animal and clinical translation. Deep learning has shown significant promise in optical imaging modalities for neuroimaging and brain regulation including photoacoustic imaging and photoacoustic tomography. Photoacoustic imaging is a hybrid imaging technique that combines the advantages of optical and ultrasound imaging. It uses laser light to generate acoustic signals, which are then used to create high-resolution images of tissue structure and function. Deep learning has been applied to photoacoustic imaging for neuroimaging, and there have been several relevant studies in this area. Using deep learning to reconstruct high-resolution images of the cerebral vasculature from photoacoustic tomography data can significantly improve image quality and reduce imaging artifacts in photoacoustic tomography [103]. In 2020, a new deep learning method called Pixel-DL was proposed by S. Guan et al., which involved pixel-wise interpolation based on the physics of photoacoustic wave propagation followed by the use of CNN to reconstruct images. Synthetic data and data from phantoms of the mouse-brain, lung, and fundus vasculature were used to train and test the model. The results showed that Pixel-DL performed similarly or better than the iterative methods and consistently outperformed other CNN-based approaches in correcting artifacts. Furthermore, Pixel-DL is a computationally efficient approach that enables real-time photoacoustic tomography rendering and improves the quality of image reconstruction for limited-view and sparse data [104].
Recent advancements in deep learning assisted optical neuroimaging have greatly solved challenges from biopsy—an invasive, unrepeatable technique that usually ignores heterogeneity within the brain. From a resolution perspective, current optical images for brain disease diagnosis include two branches of cytological and histopathological. The former cytological examination is generally inexpensive, minimally invasive, and easily repeatable compared to histopathological examination. However cytological examinations are also labor intensive and insensitive compared to histopathological examination. Image interpretation is a highly subjective task and deep learning has revealed its ability for a more objective and straightforward diagnosis. Specifically, most deep learning research has focused on optical images on a histopathological scale [105]. Deep learning began to gain more attention in the health care sector due to its promising results in recent years, which use data characterization algorithms to convert conventional imaging information into data matrices by modern linear algebra and statistics that can further be extracted into information revealing certain patterns. These deep learning approaches have been used in radiological diagnosis, bioinformatics, genome sequencing, drug development, and histopathological image analysis. Particularly for histopathological diagnosis, deep learning has surpassed that of clinical experts. The core advantage of deep learning is that it utilizes a complex neural network-like engineering architecture that can detect and extract import features automatically. The predicting label and assessment of the deep learning algorithm has developed several varieties including prognosis, PD-L1 status, microsatellite instability, histological subtyping, microenvironment analysis, and segmentation. Moreover, deep learning can solve the problem that some neuroimaging is difficult to quantify in three dimensions. However, the use of deep learning in neuroimaging and brain regulation also presents challenges. The interpretation of deep learning models is often opaque, making it difficult to understand the reasoning behind the model’s decisions. Moreover, deep learning algorithms require large amounts of data to be trained effectively, which can be challenging to acquire in the field of neuroimaging. In conclusion, deep learning has shown significant promise in various imaging modalities for neuroimaging and brain regulation including fluorescence imaging, photoacoustic imaging, and photoacoustic tomography. These studies demonstrate the potential for using deep learning to improve the image quality, reduce imaging artifacts, and develop predictive models for the diagnosis and treatment of neurological disorders.

4. Conclusions

In recent years, deep learning and sensing technologies have made impressive advances in medical health care monitoring. The following are several specific directions: ① Deep learning-based medical imaging diagnosis: Deep learning technology can make an intelligent diagnosis of medical images to improve the accuracy of the diagnosis. For example, deep learning algorithms can identify image features of diseases such as brain tumors and neurodegenerative disorders and provide accurate diagnosis results. ② Medical monitoring based on sensor technology: Sensor technology enables real-time monitoring of biological signals such as electroencephalography (EEG), electrocardiogram (ECG), etc. By analyzing these biological signals through deep learning technology, accurate physiological parameter measurement and abnormal detection can be achieved, providing important reference information for doctors. ③ Health management based on deep learning: Deep learning technology can analyze a large amount of health data such as biological signals, movement tracks, eating habits, etc. By analyzing these data, more accurate health management suggestions and personalized health intervention programs can be provided to the users. ④ Health risk prediction based on sensor technology and deep learning: Sensor technology can collect a large amount of physiological data such as blood pressure, blood sugar, oxyhemoglobin saturation, blood flow velocity, etc. The analysis of these data through deep learning technology can build health risk prediction models and provide users with personalized prevention and intervention recommendations. In general, the application of deep learning and sensing technology in the field of medical monitoring provides doctors and patients with more intelligent and precise services and treatments, which is expected to bring more progress and innovation in the field of medical health in the future. However, the use of deep learning in neuroimaging and brain regulation also presents challenges. The interpretation of deep learning models is often opaque, making it difficult to understand the reasoning behind the model’s decisions. Moreover, deep learning algorithms require large amounts of data to be trained effectively, which can be challenging to acquire in the field of neuroimaging. In conclusion, deep learning has shown great promise in the field of neuroimaging and brain regulation, with the potential to improve the accuracy and speed of diagnosis and the treatment of neurological disorders as well as enable new forms of brain–computer interfaces. However, the challenges associated with deep learning must be addressed to ensure that these techniques can be used safely and effectively in clinical settings. Overall, this article reviewed the recent progress of how deep learning is being applied in the medical field of neuroimaging and brain regulation. As research in this field continues to grow, we can expect to witness, and even participate in, more innovative applications of deep learning that will improve our understanding of the brain and advance our ability to treat neurological disorders.

Author Contributions

Conceptualization, M.X.; Writing—original draft preparation, M.X. and Y.O.; Writing—review and editing, M.X. and Y.O.; Visualization, Y.O.; Supervision, M.X. and Z.Y.; Funding acquisition, M.X. and Z.Y. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the Start-Up Fund for Introduced Talents and Scientific Research at Beijing Normal University, grant number 28709-312200502501; the National Key R&D Program of China, grant number 2019YFE0110400; the Young Scientists Fund of the National Natural Science Foundation of China, Macao Science and Technology Development Fund, grant number FDCT 0020/2019/AMJ and FDCT 0011/2018/A1; University of Macau, grant number MYRG 2020-00067-FHS, MYRG2019-00082-FHS, and MYRG2018-00081-FHS; the APC was funded by the Start-Up Fund for Introduced Talents and Scientific Research at Beijing Normal University.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Pesapane, F.; Codari, M.; Sardanelli, F. Artificial intelligence in medical imaging: Threat or opportunity? Radiologists again at the forefront of innovation in medicine. Eur. Radiol. Exp. 2018, 2, 35. [Google Scholar] [CrossRef] [PubMed]
  2. Bonacchi, R.; Filippi, M.; Rocca, M.A. Role of artificial intelligence in MS clinical practice. Neuroimage Clin. 2022, 35, 103065. [Google Scholar] [CrossRef] [PubMed]
  3. Diaz-Hurtado, M.; Martinez-Heras, E.; Solana, E.; Casas-Roma, J.; Llufriu, S.; Kanber, B.; Prados, F. Recent advances in the longitudinal segmentation of multiple sclerosis lesions on magnetic resonance imaging: A review. Neuroradiology 2022, 64, 2103–2117. [Google Scholar] [CrossRef] [PubMed]
  4. Guan, X.; Yang, G.; Ye, J.; Yang, W.; Xu, X.; Jiang, W.; Lai, X. 3D AGSE-VNet: An automatic brain tumor MRI data segmentation framework. BMC Med. Imaging 2022, 22, 6. [Google Scholar] [CrossRef]
  5. Saba, L.; Biswas, M.; Kuppili, V.; Cuadrado Godia, E.; Suri, H.S.; Edla, D.R.; Omerzu, T.; Laird, J.R.; Khanna, N.N.; Mavrogeni, S.; et al. The present and future of deep learning in radiology. Eur. J. Radiol. 2019, 114, 14–24. [Google Scholar] [CrossRef]
  6. Chen, R.; Huang, J.; Song, Y.; Li, B.; Wang, J.; Wang, H. Deep learning algorithms for brain disease detection with magnetic induction tomography. Med. Phys. 2021, 48, 745–759. [Google Scholar] [CrossRef]
  7. Graham, S.; Depp, C.; Lee, E.E.; Nebeker, C.; Tu, X.; Kim, H.C.; Jeste, D.V. Artificial Intelligence for Mental Health and Mental Illnesses: An Overview. Curr. Psychiatry Rep. 2019, 21, 116. [Google Scholar] [CrossRef]
  8. Scotton, W.J.; Bocchetta, M.; Todd, E.; Cash, D.M.; Oxtoby, N.; VandeVrede, L.; Heuer, H.; Prospect Consortium, R.C.; Alexander, D.C.; Rowe, J.B.; et al. A data-driven model of brain volume changes in progressive supranuclear palsy. Brain Commun. 2022, 4, fcac098. [Google Scholar] [CrossRef]
  9. Peeken, J.C.; Wiestler, B.; Combs, S.E. Image-Guided Radiooncology: The Potential of Radiomics in Clinical Application. Recent Results Cancer Res. 2020, 216, 773–794. [Google Scholar] [CrossRef]
  10. Wang, L.; Wu, Z.; Chen, L.; Sun, Y.; Lin, W.; Li, G. iBEAT V2.0: A multisite-applicable, deep learning-based pipeline for infant cerebral cortical surface reconstruction. Nat. Protoc. 2023, 18, 1488–1509. [Google Scholar] [CrossRef]
  11. Gillies, R.J.; Kinahan, P.E.; Hricak, H. Radiomics: Images Are More than Pictures, They Are Data1. Radiology 2016, 278, 563–577. [Google Scholar] [CrossRef]
  12. Tandel, S.G.; Biswas, M.; Kakde, O.G.; Tiwari, A.; Suri, H.S.; Turk, M.; Laird, J.R.; Asare, C.K.; Ankrah, A.A.; Khanna, N.N.; et al. A Review on a Deep Learning Perspective in Brain Cancer Classification. Cancers 2019, 11, 111. [Google Scholar] [CrossRef]
  13. Hou, Y.; Liu, Q.; Chen, J.; Wu, B.; Zeng, F.; Yang, Z.; Song, H.; Liu, Y. Application value of T2 fluid-attenuated inversion recovery sequence based on deep learning in static lacunar infarction. Acta Radiol. 2022, 64, 1650–1658. [Google Scholar] [CrossRef]
  14. Lambin, P.; Leijenaar, R.T.H.; Deist, T.M.; Peerlings, J.; de Jong, E.E.C.; van Timmeren, J.; Sanduleanu, S.; Larue, R.T.H.M.; Even, A.J.G.; Jochems, A.; et al. Radiomics: The bridge between medical imaging and personalized medicine. Nat. Rev. Clin. Oncol. 2017, 14, 749–762. [Google Scholar] [CrossRef]
  15. Wang, G.; Li, W.; Ourselin, S.; Vercauteren, T. Automatic brain tumor segmentation using cascaded anisotropic convolutional neural networks. In Proceedings of the Brainlesion: Glioma, Multiple Sclerosis, Stroke and Traumatic Brain Injuries: Third International Workshop, BrainLes 2017, Held in Conjunction with MICCAI 2017, Quebec City, QC, Canada, 14 September 2017; pp. 178–190. [Google Scholar]
  16. Bi, X.; Wang, H. Early Alzheimer’s disease diagnosis based on EEG spectral images using deep learning. Neural Netw. 2019, 114, 119–135. [Google Scholar] [CrossRef]
  17. Tabar, Y.R.; Halici, U. A novel deep learning approach for classification of EEG motor imagery signals. J. Neural Eng. 2017, 14, 016003. [Google Scholar] [CrossRef]
  18. Sadat Shahabi, M.; Nobakhsh, B.; Shalbaf, A.; Rostami, R.; Kazemi, R. Prediction of treatment outcome for repetitive transcranial magnetic stimulation in major depressive disorder using connectivity measures and ensemble of pre-trained deep learning models. Biomed. Signal Process. Control 2023, 85, 104822. [Google Scholar] [CrossRef]
  19. Muhammad, K.; Khan, S.; Ser, J.D.; Albuquerque, V.H.C. Deep Learning for Multigrade Brain Tumor Classification in Smart Healthcare Systems: A Prospective Survey. IEEE Trans. Neural Netw. Learn. Syst. 2021, 32, 507–522. [Google Scholar] [CrossRef] [PubMed]
  20. Park, J.E. Artificial Intelligence in Neuro-Oncologic Imaging: A Brief Review for Clinical Use Cases and Future Perspectives. Brain Tumor Res. Treat. 2022, 10, 69–75. [Google Scholar] [CrossRef]
  21. Bi, W.L.; Hosny, A.; Schabath, M.B.; Giger, M.L.; Birkbak, N.J.; Mehrtash, A.; Allison, T.; Arnaout, O.; Abbosh, C.; Dunn, I.F.; et al. Artificial intelligence in cancer imaging: Clinical challenges and applications. CA Cancer J. Clin. 2019, 69, 127–157. [Google Scholar] [CrossRef]
  22. Alsubai, S.; Khan, H.U.; Alqahtani, A.; Sha, M.; Abbas, S.; Mohammad, U.G. Ensemble deep learning for brain tumor detection. Front. Comput. Neurosci. 2022, 16, 1005617. [Google Scholar] [CrossRef] [PubMed]
  23. Liang, B.; Wei, R.; Zhang, J.; Li, Y.; Yang, T.; Xu, S.; Zhang, K.; Xia, W.; Guo, B.; Liu, B.; et al. Applying pytorch toolkit to plan optimization for circular cone based robotic radiotherapy. Radiat. Oncol. 2022, 17, 82. [Google Scholar] [CrossRef] [PubMed]
  24. Lopes, E.M.; Rego, R.; Rito, M.; Chamadoira, C.; Dias, D.; Cunha, J.P.S. Estimation of ANT-DBS Electrodes on Target Positioning Based on a New Percept (TM) PC LFP Signal Analysis. Sensors 2022, 22, 6601. [Google Scholar] [CrossRef]
  25. Rogers, W.; Thulasi Seetha, S.; Refaee, T.A.G.; Lieverse, R.I.Y.; Granzier, R.W.Y.; Ibrahim, A.; Keek, S.A.; Sanduleanu, S.; Primakov, S.P.; Beuque, M.P.L.; et al. Radiomics: From qualitative to quantitative imaging. Br. J. Radiol. 2020, 93, 20190948. [Google Scholar] [CrossRef]
  26. Buchanan, B.G. A (Very) Brief History of Artificial Intelligence. AI Mag. 2006, 26, 53. [Google Scholar]
  27. Meyers, P.H.; Nice, C.M., Jr. Automated Computer Analysis of Radiographic Images. Arch. Environ. Health 1964, 8, 774–775. [Google Scholar] [CrossRef]
  28. Aloysius, N.; Geetha, M. A review on deep convolutional neural networks. In Proceedings of the 2017 International Conference on Communication and Signal Processing (ICCSP), Chennai, India, 6–8 April 2017; IEEE: Piscataway, NJ, USA, 2017; pp. 588–592. [Google Scholar]
  29. Tang, X. The role of artificial intelligence in medical imaging research. BJR Open 2019, 2, 20190031. [Google Scholar] [CrossRef]
  30. Park, S.H.; Han, K. Methodologic Guide for Evaluating Clinical Performance and Effect of Artificial Intelligence Technology for Medical Diagnosis and Prediction. Radiology 2018, 286, 800–809. [Google Scholar] [CrossRef]
  31. Rajpurkar, P.; Chen, E.; Banerjee, O.; Topol, E.J. AI in health and medicine. Nat. Med. 2022, 28, 31–38. [Google Scholar] [CrossRef]
  32. Bera, K.; Schalper, K.A.; Rimm, D.L.; Velcheti, V.; Madabhushi, A. Artificial intelligence in digital pathology—New tools for diagnosis and precision oncology. Nat. Rev. Clin. Oncol. 2019, 16, 703–715. [Google Scholar] [CrossRef]
  33. Driver, C.N.; Bowles, B.S.; Bartholmai, B.J.; Greenberg-Worisek, A.J. Artificial Intelligence in Radiology: A Call for Thoughtful Application. Clin. Transl. Sci. 2020, 13, 216–218. [Google Scholar] [CrossRef] [PubMed]
  34. Gore, J.C. Artificial intelligence in medical imaging. Magn. Reson. Imaging 2020, 68, A1–A4. [Google Scholar] [CrossRef] [PubMed]
  35. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. ImageNet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  36. Mahmood, H.; Shaban, M.; Rajpoot, N.; Khurram, S.A. Artificial Intelligence-based methods in head and neck cancer diagnosis: An overview. Br. J. Cancer 2021, 124, 1934–1940. [Google Scholar] [CrossRef]
  37. Kim, J.S.; Kim, B.G.; Hwang, S.H. Efficacy of Artificial Intelligence-Assisted Discrimination of Oral Cancerous Lesions from Normal Mucosa Based on the Oral Mucosal Image: A Systematic Review and Meta-Analysis. Cancers 2022, 14, 3499. [Google Scholar] [CrossRef]
  38. Yao, K.; Unni, R.; Zheng, Y. Intelligent nanophotonics: Merging photonics and artificial intelligence at the nanoscale. Nanophotonics 2019, 8, 339–366. [Google Scholar] [CrossRef]
  39. Gruber, N.; Galijasevic, M.; Regodic, M.; Grams, A.E.; Siedentopf, C.; Steiger, R.; Hammerl, M.; Haltmeier, M.; Gizewski, E.R.; Janjic, T. A deep learning pipeline for the automated segmentation of posterior limb of internal capsule in preterm neonates. Artif. Intell. Med. 2022, 132, 102384. [Google Scholar] [CrossRef]
  40. Kim, D.; Lee, J.; Moon, J.; Moon, T. Interpretable deep learning-based hippocampal sclerosis classification. Epilepsia Open 2022, 7, 747–757. [Google Scholar] [CrossRef]
  41. Shabanpour, M.; Kaboodvand, N.; Iravani, B. Parkinson’s disease is characterized by sub-second resting-state spatio-oscillatory patterns: A contribution from deep convolutional neural network. Neuroimage Clin. 2022, 36, 103266. [Google Scholar] [CrossRef]
  42. Thakur, S.P.; Pati, S.; Panchumarthy, R.; Karkada, D.; Wu, J.; Kurtaev, D.; Sako, C.; Shah, P.; Bakas, S. Optimization of Deep Learning Based Brain Extraction in MRI for Low Resource Environments. Brainlesion 2022, 12962, 151–167. [Google Scholar] [CrossRef]
  43. Zou, A.; Ji, J.; Lei, M.; Liu, J.; Song, Y. Exploring Brain Effective Connectivity Networks Through Spatiotemporal Graph Convolutional Models. IEEE Trans. Neural Netw. Learn. Syst. 2022, 1–13. [Google Scholar] [CrossRef]
  44. Jiang, X.; Yan, J.; Zhao, Y.; Jiang, M.; Chen, Y.; Zhou, J.; Xiao, Z.; Wang, Z.; Zhang, R.; Becker, B.; et al. Characterizing functional brain networks via Spatio-Temporal Attention 4D Convolutional Neural Networks (STA-4DCNNs). Neural Netw. 2023, 158, 99–110. [Google Scholar] [CrossRef]
  45. Shen, X.; Lin, L.; Xu, X.; Wu, S. Effects of Patchwise Sampling Strategy to Three-Dimensional Convolutional Neural Network-Based Alzheimer’s Disease Classification. Brain Sci. 2023, 13, 254. [Google Scholar] [CrossRef]
  46. van Dyck, L.E.; Denzler, S.J.; Gruber, W.R. Guiding visual attention in deep convolutional neural networks based on human eye movements. Front. Neurosci. 2022, 16, 975639. [Google Scholar] [CrossRef]
  47. Xie, Y.; Zaccagna, F.; Rundo, L.; Testa, C.; Agati, R.; Lodi, R.; Manners, D.N.; Tonon, C. Convolutional Neural Network Techniques for Brain Tumor Classification (from 2015 to 2022): Review, Challenges, and Future Perspectives. Diagnostics 2022, 12, 1850. [Google Scholar] [CrossRef]
  48. Marini, N.; Otalora, S.; Wodzinski, M.; Tomassini, S.; Dragoni, A.F.; Marchand-Maillet, S.; Morales, J.P.D.; Duran-Lopez, L.; Vatrano, S.; Muller, H.; et al. Data-driven color augmentation for H&E stained images in computational pathology. J. Pathol. Inf. 2023, 14, 100183. [Google Scholar] [CrossRef]
  49. Xu, X.; Lin, L.; Sun, S.; Wu, S. A review of the application of three-dimensional convolutional neural networks for the diagnosis of Alzheimer’s disease using neuroimaging. Rev. Neurosci. 2023. [Google Scholar] [CrossRef]
  50. Kar, K.; Kubilius, J.; Schmidt, K.; Issa, E.B.; DiCarlo, J.J. Evidence that recurrent circuits are critical to the ventral stream’s execution of core object recognition behavior. Nat. Neurosci. 2019, 22, 974–983. [Google Scholar] [CrossRef]
  51. Zhang, X.Y.; Yin, F.; Zhang, Y.M.; Liu, C.L.; Bengio, Y. Drawing and Recognizing Chinese Characters with Recurrent Neural Network. IEEE Trans. Pattern Anal. Mach. Intell. 2018, 40, 849–862. [Google Scholar] [CrossRef]
  52. Mahmud, M.; Kaiser, M.S.; Hussain, A.; Vassanelli, S. Applications of Deep Learning and Reinforcement Learning to Biological Data. IEEE Trans. Neural Netw. Learn. Syst. 2018, 29, 2063–2079. [Google Scholar] [CrossRef]
  53. Cai, L.Y.; Lee, H.H.; Newlin, N.R.; Kerley, C.I.; Kanakaraj, P.; Yang, Q.; Johnson, G.W.; Moyer, D.; Schilling, K.G.; Rheault, F.; et al. Convolutional-recurrent neural networks approximate diffusion tractography from T1-weighted MRI and associated anatomical context. bioRxiv 2023. [Google Scholar] [CrossRef]
  54. Wan, Z.B.; Dong, Y.Q.; Yu, Z.C.; Lv, H.B.; Lv, Z.H. Semi-Supervised Support Vector Machine for Digital Twins Based Brain Image Fusion. Front. Neurosci. 2021, 15, 705323. [Google Scholar] [CrossRef] [PubMed]
  55. Hu, M.D.; Zhong, Y.; Xie, S.X.; Lv, H.B.; Lv, Z.H. Fuzzy System Based Medical Image Processing for Brain Disease Prediction. Front. Neurosci. 2021, 15, 714318. [Google Scholar] [CrossRef] [PubMed]
  56. Liu, S.J.; Ye, M.; Pao, G.M.; Song, S.M.; Jhang, J.; Jiang, H.B.; Kim, J.H.; Kang, S.J.; Kim, D.I.; Han, S. Divergent brainstem opioidergic pathways that coordinate breathing with pain and emotions. Neuron 2022, 110, 857–873. [Google Scholar] [CrossRef]
  57. Munir, K.; Elahi, H.; Ayub, A.; Frezza, F.; Rizzi, A. Cancer Diagnosis Using Deep Learning: A Bibliographic Review. Cancers 2019, 11, 1235. [Google Scholar] [CrossRef]
  58. Zhao, X.M.; Wu, Y.H.; Song, G.D.; Li, Z.Y.; Zhang, Y.Z.; Fan, Y. A deep learning model integrating FCNNs and CRFs for brain tumor segmentation. Med. Image Anal. 2018, 43, 98–111. [Google Scholar] [CrossRef]
  59. Evans, D.A.; Stempel, A.V.; Vale, R.; Ruehle, S.; Lefler, Y.; Branco, T. A synaptic threshold mechanism for computing escape decisions. Nature 2018, 558, 590–594. [Google Scholar] [CrossRef]
  60. Gibson, E.; Li, W.Q.; Sudre, C.; Fidon, L.; Shakir, D.I.; Wang, G.T.; Eaton-Rosen, Z.; Gray, R.; Doel, T.; Hu, Y.P.; et al. NiftyNet: A deep-learning platform for medical imaging. Comput. Methods Programs Biomed. 2018, 158, 113–122. [Google Scholar] [CrossRef]
  61. Nie, D.; Trullo, R.; Lian, J.; Wang, L.; Petitjean, C.; Ruan, S.; Wang, Q.; Shen, D. Medical Image Synthesis with Deep Convolutional Adversarial Networks. IEEE Trans. Biomed. Eng. 2018, 65, 2720–2730. [Google Scholar] [CrossRef]
  62. Wang, Y.; Yu, B.T.; Wang, L.; Zu, C.; Lalush, D.S.; Lin, W.L.; Wu, X.; Zhou, J.L.; Shen, D.G.; Zhou, L.P. 3D conditional generative adversarial networks for high-quality PET image estimation at low dose. Neuroimage 2018, 174, 550–562. [Google Scholar] [CrossRef]
  63. Xue, Y.; Xu, T.; Zhang, H.; Long, L.R.; Huang, X.L. SegAN: Adversarial Network with Multi-scale L (1) Loss for Medical Image Segmentation. Neuroinformatics 2018, 16, 383–392. [Google Scholar] [CrossRef]
  64. Dar, S.U.H.; Yurt, M.; Karacan, L.; Erdem, A.; Erdem, E.; Cukur, T. Image Synthesis in Multi-Contrast MRI With Conditional Generative Adversarial Networks. IEEE Trans. Med. Imaging 2019, 38, 2375–2388. [Google Scholar] [CrossRef]
  65. Lundervold, A.S.; Lundervold, A. An overview of deep learning in medical imaging focusing on MRI. Z. Fur Med. Phys. 2019, 29, 102–127. [Google Scholar] [CrossRef]
  66. Ghassemi, N.; Shoeibi, A.; Rouhani, M. Deep neural network with generative adversarial networks pre-training for brain tumor classification based on MR images. Biomed. Signal Process. Control 2020, 57, 101678. [Google Scholar] [CrossRef]
  67. Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative Adversarial Networks. Commun. Acm 2020, 63, 139–144. [Google Scholar] [CrossRef]
  68. Ravishankar, S.; Ye, J.C.; Fessler, J.A. Image Reconstruction: From Sparsity to Data-Adaptive Methods and Machine Learning. Proc. IEEE 2020, 108, 86–109. [Google Scholar] [CrossRef]
  69. Chlap, P.; Min, H.; Vandenberg, N.; Dowling, J.; Holloway, L.; Haworth, A. A review of medical image data augmentation techniques for deep learning applications. J. Med. Imaging Radiat. Oncol. 2021, 65, 545–563. [Google Scholar] [CrossRef]
  70. Dalmaz, O.; Yurt, M.; Cukur, T. ResViT: Residual Vision Transformers for Multimodal Medical Image Synthesis. IEEE Trans. Med. Imaging 2022, 41, 2598–2614. [Google Scholar] [CrossRef]
  71. Geng, M.F.; Meng, X.X.; Yu, J.Y.; Zhu, L.; Jin, L.J.; Jiang, Z.; Qiu, B.; Li, H.; Kong, H.J.; Yuan, J.M.; et al. Content-Noise Complementary Learning for Medical Image Denoising. IEEE Trans. Med. Imaging 2022, 41, 407–419. [Google Scholar] [CrossRef]
  72. Hu, S.Y.; Lei, B.Y.; Wang, S.Q.; Wang, Y.; Feng, Z.G.; Shen, Y.Y. Bidirectional Mapping Generative Adversarial Networks for Brain MR to PET Synthesis. IEEE Trans. Med. Imaging 2022, 41, 145–157. [Google Scholar] [CrossRef]
  73. Korkmaz, Y.; Dar, S.U.H.; Yurt, M.; Ozbey, M.; Cukur, T. Unsupervised MRI Reconstruction via Zero-Shot Learned Adversarial Transformers. IEEE Trans. Med. Imaging 2022, 41, 1747–1763. [Google Scholar] [CrossRef]
  74. You, S.; Lei, B.; Wang, S.; Chui, C.K.; Cheung, A.C.; Liu, Y.; Gan, M.; Wu, G.; Shen, Y. Fine Perceptive GANs for Brain MR Image Super-Resolution in Wavelet Domain. IEEE Trans. Neural Netw. Learn. Syst. 2022, 1–13. [Google Scholar] [CrossRef] [PubMed]
  75. Ozkaraca, O.; Bagriacik, O.I.; Guruler, H.; Khan, F.; Hussain, J.; Khan, J.; Laila, U.E. Multiple Brain Tumor Classification with Dense CNN Architecture Using Brain MRI Images. Life 2023, 13, 349. [Google Scholar] [CrossRef] [PubMed]
  76. Li, Z.; Fan, Q.; Bilgic, B.; Wang, G.; Wu, W.; Polimeni, J.R.; Miller, K.L.; Huang, S.Y.; Tian, Q. Diffusion MRI data analysis assisted by deep learning synthesized anatomical images (DeepAnat). Med. Image Anal. 2023, 86, 102744. [Google Scholar] [CrossRef] [PubMed]
  77. Chen, Z.; Pawar, K.; Ekanayake, M.; Pain, C.; Zhong, S.; Egan, G.F. Deep Learning for Image Enhancement and Correction in Magnetic Resonance Imaging-State-of-the-Art and Challenges. J. Digit Imaging 2023, 36, 204–230. [Google Scholar] [CrossRef]
  78. Gore, S.; Chougule, T.; Jagtap, J.; Saini, J.; Ingalhalikar, M. A Review of Radiomics and Deep Predictive Modeling in Glioma Characterization. Acad. Radiol. 2021, 28, 1599–1621. [Google Scholar] [CrossRef]
  79. Diaz-Gomez, L.; Gutierrez-Rodriguez, A.E.; Martinez-Maldonado, A.; Luna-Munoz, J.; Cantoral-Ceballos, J.A.; Ontiveros-Torres, M.A. Interpretable Classification of Tauopathies with a Convolutional Neural Network Pipeline Using Transfer Learning and Validation against Post-Mortem Clinical Cases of Alzheimer’s Disease and Progressive Supranuclear Palsy. Curr. Issues Mol. Biol. 2022, 44, 5963–5985. [Google Scholar] [CrossRef]
  80. Chattopadhyay, T.; Ozarkar, S.S.; Buwa, K.; Thomopoulos, S.I.; Thompson, P.M.; Alzheimer’s Disease Neuroimaging, I. Predicting Brain Amyloid Positivity from T1 weighted brain MRI and MRI-derived Gray Matter, White Matter and CSF maps using Transfer Learning on 3D CNNs. bioRxiv 2023. [Google Scholar] [CrossRef]
  81. Coll, L.; Pareto, D.; Carbonell-Mirabent, P.; Cobo-Calvo, A.; Arrambide, G.; Vidal-Jordana, A.; Comabella, M.; Castillo, J.; Rodriguez-Acevedo, B.; Zabalza, A.; et al. Deciphering multiple sclerosis disability with deep learning attention maps on clinical MRI. Neuroimage Clin. 2023, 38, 103376. [Google Scholar] [CrossRef]
  82. Zoetmulder, R.; Baak, L.; Khalili, N.; Marquering, H.A.; Wagenaar, N.; Benders, M.; van der Aa, N.E.; Isgum, I. Brain segmentation in patients with perinatal arterial ischemic stroke. Neuroimage Clin. 2023, 38, 103381. [Google Scholar] [CrossRef]
  83. Daveau, R.S.; Law, I.; Henriksen, O.M.; Hasselbalch, S.G.; Andersen, U.B.; Anderberg, L.; Hojgaard, L.; Andersen, F.L.; Ladefoged, C.N. Deep learning based low-activity PET reconstruction of [(11)C]PiB and [(18)F]FE-PE2I in neurodegenerative disorders. Neuroimage 2022, 259, 119412. [Google Scholar] [CrossRef]
  84. Etminani, K.; Soliman, A.; Davidsson, A.; Chang, J.R.; Martinez-Sanchis, B.; Byttner, S.; Camacho, V.; Bauckneht, M.; Stegeran, R.; Ressner, M.; et al. A 3D deep learning model to predict the diagnosis of dementia with Lewy bodies, Alzheimer’s disease, and mild cognitive impairment using brain 18F-FDG PET. Eur. J. Nucl. Med. Mol. Imaging 2022, 49, 563–584. [Google Scholar] [CrossRef]
  85. Maddury, S.; Desai, K. DeepAD: A deep learning application for predicting amyloid standardized uptake value ratio through PET for Alzheimer’s prognosis. Front. Artif. Intell. 2023, 6, 1091506. [Google Scholar] [CrossRef]
  86. Hamdi, M.; Bourouis, S.; Rastislav, K.; Mohmed, F. Evaluation of Neuro Images for the Diagnosis of Alzheimer’s Disease Using Deep Learning Neural Network. Front. Public Health 2022, 10, 834032. [Google Scholar] [CrossRef]
  87. Fu, F.; Shan, Y.; Yang, G.; Zheng, C.; Zhang, M.; Rong, D.; Wang, X.; Lu, J. Deep Learning for Head and Neck CT Angiography: Stenosis and Plaque Classification. Radiology 2023, 307, 220996. [Google Scholar] [CrossRef]
  88. Kok, Y.E.; Pszczolkowski, S.; Law, Z.K.; Ali, A.; Krishnan, K.; Bath, P.M.; Sprigg, N.; Dineen, R.A.; French, A.P. Semantic Segmentation of Spontaneous Intracerebral Hemorrhage, Intraventricular Hemorrhage, and Associated Edema on CT Images Using Deep Learning. Radiol. Artif. Intell. 2022, 4, e220096. [Google Scholar] [CrossRef]
  89. Moghadam, S.M.; Airaksinen, M.; Nevalainen, P.; Marchi, V.; Hellstrom-Westas, L.; Stevenson, N.J.; Vanhatalo, S. An automated bedside measure for monitoring neonatal cortical activity: A supervised deep learning-based electroencephalogram classifier with external cohort validation. Lancet Digit. Health 2022, 4, e884–e892. [Google Scholar] [CrossRef]
  90. Mughal, N.E.; Khan, M.J.; Khalil, K.; Javed, K.; Sajid, H.; Naseer, N.; Ghafoor, U.; Hong, K.S. EEG-fNIRS-based hybrid image construction and classification using CNN-LSTM. Front. Neurorobot 2022, 16, 873239. [Google Scholar] [CrossRef]
  91. Chambon, S.; Galtier, M.N.; Arnal, P.J.; Wainrib, G.; Gramfort, A. A Deep Learning Architecture for Temporal Sleep Stage Classification Using Multivariate and Multimodal Time Series. IEEE Trans. Neural Syst. Rehabil. Eng. 2018, 26, 758–769. [Google Scholar] [CrossRef]
  92. Lawhern, V.J.; Solon, A.J.; Waytowich, N.R.; Gordon, S.M.; Hung, C.P.; Lance, B.J. EEGNet: A compact convolutional neural network for EEG-based brain-computer interfaces. J. Neural Eng. 2018, 15, 056013. [Google Scholar] [CrossRef]
  93. Dinov, M.; Leech, R. Tracking and optimizing human performance using deep reinforcement learning in closed-loop behavioral-and neuro-feedback: A proof of concept. bioRxiv 2017. [Google Scholar] [CrossRef]
  94. Abiri, R.; Borhani, S.; Sellers, E.W.; Jiang, Y.; Zhao, X. A comprehensive review of EEG-based brain-computer interface paradigms. J. Neural Eng. 2019, 16, 011001. [Google Scholar] [CrossRef] [PubMed]
  95. Petrosyan, A.; Sinkin, M.; Lebedev, M.; Ossadtchi, A. Decoding and interpreting cortical signals with a compact convolutional neural network. J. Neural Eng. 2021, 18, 026019. [Google Scholar] [CrossRef] [PubMed]
  96. Zubarev, I.; Vranou, G.; Parkkonen, L. MNEflow: Neural networks for EEG/MEG decoding and interpretation. SoftwareX 2022, 17, 100951. [Google Scholar] [CrossRef]
  97. Shen, D.; Deng, Y.; Lin, C.; Li, J.; Lin, X.; Zou, C. Clinical Characteristics and Gene Mutation Analysis of Poststroke Epilepsy. Contrast Media Mol. Imaging 2022, 2022, 4801037. [Google Scholar] [CrossRef]
  98. Hosseini, M.P.; Tran, T.X.; Pompili, D.; Elisevich, K.; Soltanian-Zadeh, H. Multimodal data analysis of epileptic EEG and rs-fMRI via deep learning and edge computing. Artif. Intell. Med. 2020, 104, 101813. [Google Scholar] [CrossRef]
  99. Hosseini, M.P.; Tran, T.X.; Pompili, D.; Elisevich, K.; Soltanian-Zadeh, H. Deep Learning with Edge Computing for Localization of Epileptogenicity using Multimodal rs-fMRI and EEG Big Data. In Proceedings of the 2017 IEEE International Conference on Automatic Computing (ICAC), Columbus, OH, USA, 17–21 July 2017; pp. 83–92. [Google Scholar]
  100. Hussein, R.; Palangi, H.; Ward, R.; Wang, Z.J. Epileptic seizure detection: A deep learning approach. arXiv 2018, arXiv:1803.09848. [Google Scholar]
  101. Singh, K.; Malhotra, J. Deep learning based smart health monitoring for automated prediction of epileptic seizures using spectral analysis of scalp EEG. Phys. Eng. Sci. Med. 2021, 44, 1161–1173. [Google Scholar] [CrossRef]
  102. Markowitz, J.E.; Gillis, W.F.; Jay, M.; Wood, J.; Harris, R.W.; Cieszkowski, R.; Scott, R.; Brann, D.; Koveal, D.; Kula, T.; et al. Spontaneous behaviour is structured by reinforcement without explicit reward. Nature 2023, 614, 108–117. [Google Scholar] [CrossRef]
  103. Gröhl, J.; Schellenberg, M.; Dreher, K.; Maier-Hein, L. Deep learning for biomedical photoacoustic imaging: A review. Photoacoustics 2021, 22, 100241. [Google Scholar] [CrossRef]
  104. Guan, S.; Khan, A.A.; Sikdar, S.; Chitnis, P.V. Limited-View and Sparse Photoacoustic Tomography for Neuroimaging with Deep Learning. Sci. Rep. 2020, 10, 8510. [Google Scholar] [CrossRef]
  105. Cao, R.; Nelson, S.D.; Davis, S.; Liang, Y.; Luo, Y.; Zhang, Y.; Crawford, B.; Wang, L.V. Label-free intraoperative histology of bone tissue via deep-learning-assisted ultraviolet photoacoustic microscopy. Nat. Biomed. Eng. 2023, 7, 124–134. [Google Scholar] [CrossRef]
Figure 1. Schematic illustration of the review on deep learning aided neuroimaging and brain regulation.
Figure 1. Schematic illustration of the review on deep learning aided neuroimaging and brain regulation.
Sensors 23 04993 g001
Figure 2. Applications of different neuron network-assisted brain health care monitoring.
Figure 2. Applications of different neuron network-assisted brain health care monitoring.
Sensors 23 04993 g002
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Xu, M.; Ouyang, Y.; Yuan, Z. Deep Learning Aided Neuroimaging and Brain Regulation. Sensors 2023, 23, 4993. https://doi.org/10.3390/s23114993

AMA Style

Xu M, Ouyang Y, Yuan Z. Deep Learning Aided Neuroimaging and Brain Regulation. Sensors. 2023; 23(11):4993. https://doi.org/10.3390/s23114993

Chicago/Turabian Style

Xu, Mengze, Yuanyuan Ouyang, and Zhen Yuan. 2023. "Deep Learning Aided Neuroimaging and Brain Regulation" Sensors 23, no. 11: 4993. https://doi.org/10.3390/s23114993

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop