You are currently viewing a new version of our website. To view the old version click .
Healthcare
  • Article
  • Open Access

18 November 2022

AI-Assisted Diagnosis and Decision-Making Method in Developing Countries for Osteosarcoma

,
,
,
,
and
1
School of Computer Science and Engineering, Central South University, Changsha 410083, China
2
The First People’s Hospital of Huaihua, Huaihua 418000, China
3
The Second People’s Hospital of Huaihua, Huaihua 418000, China
4
Collaborative Innovation Center for Medical Artificial Intelligence and Big Data Decision Making Assistance, Hunan University of Medicine, Huaihua 418000, China
This article belongs to the Special Issue Advances of Decision-Making Medical System in Healthcare

Abstract

Osteosarcoma is a malignant tumor derived from primitive osteogenic mesenchymal cells, which is extremely harmful to the human body and has a high mortality rate. Early diagnosis and treatment of this disease is necessary to improve the survival rate of patients, and MRI is an effective tool for detecting osteosarcoma. However, due to the complex structure and variable location of osteosarcoma, cancer cells are highly heterogeneous and prone to aggregation and overlap, making it easy for doctors to inaccurately predict the area of the lesion. In addition, in developing countries lacking professional medical systems, doctors need to examine mass of osteosarcoma MRI images of patients, which is time-consuming and inefficient, and may result in misjudgment and omission. For the sake of reducing labor cost and improve detection efficiency, this paper proposes an Attention Condenser-based MRI image segmentation system for osteosarcoma (OMSAS), which can help physicians quickly locate the lesion area and achieve accurate segmentation of the osteosarcoma tumor region. Using the idea of AttendSeg, we constructed an Attention Condenser-based residual structure network (ACRNet), which greatly reduces the complexity of the structure and enables smaller hardware requirements while ensuring the accuracy of image segmentation. The model was tested on more than 4000 samples from two hospitals in China. The experimental results demonstrate that our model has higher efficiency, higher accuracy and lighter structure for osteosarcoma MRI image segmentation compared to other existing models.

1. Introduction

Osteosarcoma is one of the most common bone malignancies, which develops from mesenchymal cell line [1]. Although it accounts for 0.2% of human malignant solid tumors, the mortality and disability rates are very high. Osteosarcoma is the most common type of bone malignancy in the elderly and children, and the second peak of its morbidity is over 60 years old [2,3,4]. At present, the five-year survival rate of osteosarcoma patients in China is about 60%, and the ten-year survival rate decreases to only 30%. Moreover, during the five-year or ten-year survival process, most patients still need to experience metastasis, recurrence, multiple operations, etc. [5,6,7,8]. The size, position, architecture, and shape of osteosarcomas will be diverse depending on the patient’s physiology, physical condition, and degree of disease. Additionally, the distribution density of osteosarcoma is not uniform. It is often difficult to distinguish the tumor tissue from the surrounding normal tissue [9,10,11,12,13].
The diversity of osteosarcoma leads to scattered information about it in the medical literature, particularly in imaging results. The images of diverse osteosarcoma generated in the identical environment and process are various as well, and it is hard to tell the difference between healthy tissue and lesion areas manually in some cases [14,15]. Generally, the examination of osteosarcoma is nothing more than X-ray examination, CT scan, and MRI examination. Among the three methods, magnetic resonance owns great soft tissue identification as well as an exceptionally high Contrast Ratio, and its ability to slice in multiple parameters and planes allows for clear visualization of the position and degree of the lesion, in the meantime, the harm to the patient’s body during the detection is also the smallest [16,17].
The mortality of osteosarcoma is very high, but early detection and timely treatment can greatly improve the survival rate [18]. However, the imaging diagnosis and treatment of osteosarcoma has faced barriers due to the poorly developed medical infrastructure in most of developing countries. The need for early diagnosis of osteosarcoma is made difficult by the high price of MRI machines and the lack of top-level talent. To make matters worse, the diagnosis of a single patient will generate 600-700 MRI images. Of the vast quantity of data, frequently less than 20 of them are considered usable [19]. The rest of the redundant images not only slow down the progress of judgment, but also sometimes mislead doctors to a certain extent [20]. In addition, considering that the diagnosis of MRI images is highly subjective and mostly depends on the experience and professional knowledge of doctors, the lack of talents in this field is also a major reason for the low efficiency of diagnoses [21]. The low recognition rate of early diagnosis leads to more difficult later-stage treatment, forming a vicious circle [22,23]. In China, 80% of medical resources are concentrated in large cities with only 10% of the population. In remote areas and some underdeveloped small towns, the lack of medical resources has led to a serious imbalance in the doctor-patient ratio, leaving people with only simple medical care. This series of problems is heavy grief for families and societies [24,25,26]. It is obvious that if we want to promote the medical system in this area more quickly and comprehensively, we need to realize low-cost, efficient, and accurate image segmentation technology to replace the original diagnosis methods, and use machines for initial screening to provide diagnostic assistance for doctors [27,28].
In recent years, with the increasing attention to AI, some medical image detection technologies have also been applied to the diagnosis of osteosarcoma [29,30]. Although these techniques can to some extent detect the position and margins of the tumor, the accuracy of detection is not high due to the variability of osteosarcoma [31,32,33,34]. The effect of existing medical image diagnosis technology in osteosarcoma image segmentation is not as expected [35]. The methods of machine learning are to manually calibrate the original data and then establish the mapping between the image and the calibrated area through a function. Then, the parameters of the function will be trained by using a large amount of data so that the function can extract some information between them that we cannot extract [36,37,38]. Previous models often used elaborate architectures as well as deeper hierarchies to improve their fitting ability and thus give them higher accuracy. While this approach can increase the model’s accuracy to some extent, it tends to reduce the generalization ability of the model, and thus the effect of the improvement is on the weak side often [39,40]. Further, excessively sophisticated architecture will lead to slow and less efficient model training, and the hardware demands will likely be high.
According to the above contents, we propose an MRI image segmentation and AI-assisted diagnosis system based on Attention Condenser for osteosarcoma (OMSAS) to assist doctors. This intelligent medical system is designed to assist doctors in identifying MRI images of patients with osteosarcoma and to automatically segment osteosarcoma, thus providing doctors with more powerful and intuitive tips and aids. The analysis results provided by the system can be used as an auxiliary basis for doctors to diagnose patients, which can effectively improve the efficiency and accuracy of diagnosis and reduce the cost of diagnosis and treatment for patients. Firstly, we carry out data augmentation in the dataset and use a variety of methods to expand, standardize and classify the original dataset. Data preprocessing reduces the degree of over-fitting and makes the model more generalized. In terms of model design, we improve AttendSeg [41], replace the ordinary convolution layer with two blocks we designed and delete several unnecessary layers. We designed a residual network based on Attention Condenser (ACRNet), which uses a residual network structure and reconstructs the most important component Attention Condenser [42] in AttendSeg. Compared with other models combining the residual structure, ACRNet retains the self-attention mechanism in Attention Condenser that learns and generates cohesive embeddings characterizing joint local and cross-channel activation relationships, further enhancing the model’s attention to osteosarcoma regions. The self-attention mechanism enables the model to migrate more attention to refine the details between regions in the osteosarcoma MRI image, better extract various features of the image, and ultimately improve the accuracy of segmentation. In addition, the ACRNet after component reconstruction facilitates for sparser use of larger stand-alone convolution modules to reduce overall network complexity and has a lower number of parameters. Finally, in terms of output results, we choose to change a single output into a composite decision, calculate its average results and generate a binary map to facilitate the physician’s diagnosis.
The specific contributions of the essay are divided into the following points:
(1)
In this article, the region in the image can be focused more accurately through the Attention Condenser. The multilayer condenser structure can further locate the boundary of the tumor, and reduce a large number of unnecessary calculations in the later stage, to improve the efficiency and accuracy of training.
(2)
When using the model to predict, OMSAS uses a compound decision module that turns a single input into multiple copies of the input. Then, the model makes predictions by multiple copies simultaneously, and then combines the multiple outputs to get the final result. Multiple decisions can reduce the wrong prediction caused by some unknown situations and make the output more accurate. Meanwhile, it can improve the stability of the model.
(3)
More than 4000 images provided by the First People’s Hospital of Huaihua and the Second People’s Hospital of Huaihua were used for testing. The result indicates that our ACRNet in OMSAS is better than other existing segmentation models. The model has high training efficiency and prediction accuracy as well as small resource consumption, which is important in assisting doctors to diagnose osteosarcoma in patients.
The paper is arranged as follows: Section 2 introduces some research related to our work. In Section 3, we depict the main structure and design of the Attention Condenser-based MRI image segmentation system for osteosarcoma (OMSAS). Section 4 presents the experimental dataset and shows the performance of the model through evaluation metrics to demonstrate the validity of the method. At the end of the paper, we summarize our work and look forward to future work.

3. System Model Design

This section may be divided by subheadings. It should provide a concise and precise description of the experimental results, their interpretation, as well as the experimental conclusions that can be drawn.
At this stage, in most developing countries, the distribution of medical resources is very unbalanced, and many advanced medical resources are concentrated in the region where very few people are located. Additionally, generally speaking, a particularly good analytical instrument can greatly reduce the working pressure of doctors. However, the high manufacturing and maintenance costs make it very difficult to promote the medical equipment of image-assisted diagnosis.
Considering the osteosarcoma patients living in developing countries, if the early diagnosis is not timely, the lack of technology or the very high cost of later medical treatment will often make them give up their last chance of survival. Therefore, what we need to do is to improve the early diagnosis rate. For hospitals, the prolonged manual diagnosis and the shortage of talents result in the difficulty of diagnosis. It is necessary to introduce an osteosarcoma MRI image segmentation medical system to assist doctor. Moreover, to promote this system more widely, we expect that the built-in process should be simple and efficient enough, the model should have high accuracy, and the requirements for equipment should be as low as possible. Thus, based on these cognitions, this paper proposes a segmentation method for osteosarcoma MRI image based on Attention Condenser (OMSAS), which can accurately depict the tumor area in the image, relieving the burden of reading films for doctors. The overall design of the system is shown in Figure 1.
Figure 1. Architecture diagram of OMSAS.
According to Figure 1, we can see that the system we proposed in this paper is roughly divided into three parts, including data preprocessing, model training, and model application. This chapter will be divided into four sections. In Section 3.1, we introduce the work of data preprocessing. In Section 3.2, we briefly introduce the attention mechanism. Section 3.3 and Section 3.4 elaborate the ACRNet segmentation model proposed in the paper and the loss function, respectively.
Some symbols involved in the chapter are explained, as shown in Table 1.
Table 1. Description of relevant symbols in the chapter.

3.1. Data Preprocessing

We find that the data in the original dataset is not suitable for model training directly, and there are the following problems:
(1)
The osteosarcoma MRI images in the dataset have different brightness and darkness due to different instruments, contrast agent dose, and other external environmental reasons;
(2)
The amount of data is still insufficient for training a model with high accuracy, and the training is prone to instability.
To address the above two problems, we have carried out a series of data preprocessing to optimize the effect of the model by improving the quality of data and eliminating interference factors. The specific operation is shown in Figure 2.
Figure 2. Process of data preprocess.
First, we select the region of all the osteosarcoma MRI images. After the image is binarized by (1), it will be filled to obtain a picture with only pure black and pure white areas. Then, we select the smallest rectangular area in the picture that can just contain the pure white area, which is the selected effective area. Selecting the effective region can simplify the training and recognition of the model, and make the model pay more attention to the feature extraction of the effective region.
p b i n a r y = 1 2 + 1 2 × p μ p i x e l p μ p i x e l
In addition, we regularize all the osteosarcoma MRI images to avoid the impact of some excessively bright and dim images on the training. The brightness gap and different density distribution between images make it easy for the model to take these unnecessary factors into account, which not only reduces the speed of model training, but also may affect the effect of the model. Therefore, we use a unified regularization process to transform all MRI images with (2).
p n o r m = p p m i n p m i n m a x
The regularized image addresses the first problem. At this time, the image has met the minimum standard of training. However, to enhance the robustness of our model and the generalization of it, we also need data augmentation. For data augmentation, we rotate each picture by 90°, 180°, and 270° at first, and then flip each picture horizontally and vertically. After that, the amount of data is 8 times that of the original dataset. We then randomly select some images and add noise to them to reduce the overfitting phenomenon of the model when learning high-frequency features.
After data preprocessing, the quality of the MRI images of osteosarcoma in the dataset was improved, and the high-quality dataset can be used as a reference basis for clinical diagnosis, which is beneficial for doctors to make more effective film reading. At the same time, the enhanced dataset can provide a good fit for the model training.3.2. Brief Introduction of Attention.

3.2. Brief Introduction of Attention

The model of this paper involves the mechanism of attention, so attention will be briefly introduced in this part.
We know that the inspiration of the attention mechanism comes from our physiological perception of the environment. Our visual system will actively select the information we need to pay attention to in the impression and ignore the irrelevant information in the field of vision. Similarly, in the neural network, when we segment an osteosarcoma MRI image or recognize the boundary, we also hope the network to pay more attention to the region we expect, and ignore the irrelevant information. The attention mechanism can be divided into three steps: compression, activation, and attention. For example, we now have an osteosarcoma MRI image input M. and the size of M is (w, h, c). Moreover, w means width, h represents its height, and c is its number of channels. Then, the compression process can be expressed as (3).
z c = 1 × w i = 1 w j = 1 M c , i , j
In order to evaluate the importance of each channel, we need to activate the previous operation results as a whole. The weight calculation formula on each channel is shown in (4), where W ρ and W σ are two weight matrices. After their respective matrix operations, they will be input into σ. The ρ representing the sigmoid function and the ReLU function.
e c = ρ W ρ σ W σ z c
After the activation, we assign M to the weight matrix e c corresponding to the input M. After e c is multiplied by M, we get M after adjusting the channel weight. It is written as (5).
M = M e c
In our model design, attention can be understood as the process of Figure 3. Through this mechanism, the algorithm can pay more attention to useful information, ignore many useless parts, and reduce the amount and complexity of calculation.
Figure 3. Schematic diagram of the structure of the attention mechanism involved in this paper.

3.3. Osteosarcoma MRI Image Segmentation Model

On the osteosarcoma image segmentation model, we designed a residual network based on Attention Condenser (ACRNet), and its overall structure is shown in Figure 4. The whole network is split into three main parts as follows:
(1)
The osteosarcoma MRI images will be rotated and flipped to obtain a group of pictures before each prediction of it.
(2)
The set of pictures from the previous step will be used as input for the residual network with multiple attention condensers, and a set of results will be output.
(3)
This group of results is inversely changed according to the original change direction to obtain a group of results consistent with the original image direction. This group of results is calculated and processed to obtain the final prediction tumor region.
Figure 4. The architecture of osteosarcoma segmentation network ACRNet.
In recent years, a large number of scholars have emerged to conduct research on attention networks. The attention condenser proposed by Alexander Wong et al. [42] uses the self-attention mechanism to reduce the size of the model and does not affect the performance. Attention Condenser is designed to capture spatial channel activation relationships by learning the aggregation embedding of activation relationships to achieve efficient selective attention. ACRNet incorporates a residual network with multiple Attention Condenser, highlighting the superiority of the structure. On the one hand, Attention Condenser, as an independent stand-alone module, jointly models both local and cross-channel activation relations within a unified condensed embedding using a self-concentration mechanism, which facilitates for sparser use of larger stand-alone convolution modules to reduce overall computational complexity of such joint modeling and unnecessary computations in the ACRNet. On the other hand, the multilayer condenser-based ACRNet can localize the boundaries of tumors and focus more accurately on the osteosarcoma region in MRI images, which further improves the accuracy of image segmentation.
Attention Condenser consists of four parts: a condensed layer that reduces the dimension of spatial channels, an embedded structure that represents the activation relationship of joint spatial channels, an extended layer that increases the dimension, and an attention mechanism that applies selective attention. In ACRNet, we use two linear structures and the middle embedding layer to perform the above scaling operation. In addition, we introduce a variable called ratio, which can manually adjust the expansion and contraction times of the condenser, to meet the needs in different situations better.
The residual structure is used many times in the whole network. The residual unit is realized in the form of layer skip connection. Some layers are regarded as a unit. The initial input of the unit and the final output will be added before activation. The residual structure addresses the degenerate problem of the deep neural network. Under the same number of layers, the training speed of the residual network is faster, which can make the operation efficiency of the model higher.
In order to better extract various features of the osteosarcoma MRI image, we use the idea of the block to combine and transform the original general convolution layer into the reservation block and the retracting block. The reservation block includes a 1 × 1 convolution layer, a Layer Normalization, and a ReLU activation function, which only changes the number of channels of input without affecting the size of the image. Adding LayerNorm can normalize the data before entering the activation function to avoid the problem of vanishing gradient in the saturated region of the activation function. Meanwhile, the use of LayerNorm accelerates the convergence speed of the model again.
The retracting block consists of a 3 × 3 and a 1 × 1 convolution layer. Its purpose is to simply extract image features. A retracting block instead of single 3 × 3 convolution layer makes a convolution kernel responsible for only one channel. A channel will be convoluted by only one convolution kernel, and the change of channel is handed over to 1×1 convolution layer. This can significantly reduce the parameters number of the ordinary convolution operation.
In the output, we use compound decision-making to integrate multiple outputs from different angles of the same image, since that for neural networks, every rotated or flipped image is a new image, and the prediction results are not necessarily the same. Unified processing of homologous output can enhance the accuracy and robustness of results. The specific formula for output calculation is as follows.
r j , k = 1 2 1 2 × σ j , k n 2 σ j , k n 2 , σ j , k = i = 0 n p i , j , k ( j = 0 , 1 , 2 , , h ; k = 0 , 1 , 2 , , w )

3.4. Loss Function

Medical image segmentation models often encounter the phenomenon of data imbalance. Training unbalanced data is easy to cause the problem of high precision but low recall. Moreover, in the diagnosis of osteosarcoma, missed diagnosis of tumor area is a more intolerable serious error than the wrong diagnosis of the normal area. Therefore, when designing the model, we choose Tversky Loss as the loss function. Tversky loss is a special dice loss, which is a combination of dice loss and Jaccard coefficient. Generally, the weight of dice loss for FP and FN is equal, so the model will not focus on the improvement of recall Sufficiently during training. Therefore, we use Tversky Loss with parameters β as the loss function in model training. The specific definition is as shown in (7).
T ( α , β ) = i = 0 N i i i = 0 N i i + α i = 1 N i ( 1 i ) + β i = 1 N i ( 1 i ) + ε
α and β control the penalty strength of the loss function for FP and FN, respectively. The higher the β, the greater the penalty of the loss function on FN, and the easier it is to improve the recall. After testing, we adopt α = 0.25, β = 0.75.
By using our model, we can easily perform region detection on osteosarcoma MRI images and the annotation regions are comparable to the “Gold Standard” of doctor annotation. Our model ensures accuracy while greatly reducing the burden on doctors when reading MRI images. For hospitals, the time and money spent on processing related cases are greatly reduced.

4. Performance Evaluation

4.1. Dataset Introduction

The data in the article are collected from the First People’s Hospital of Huaihua and the Second People’s Hospital of Huaihua. During the experiment, we collected more than 4000 MRI images and other related index data of 210 patients with different degrees of osteosarcoma. During the experiments, all segmentation models are built on the same dataset, so the performance tests of the models are comparable. To ensure that the sample is closer to the standard of developing countries, we selected data mostly from patients with not very high income in the lower social class. Additionally, the data are mostly from early imaging and a small percentage from mid to late. We selected three different cross-sectional slices to allow the model to better fit multiple input cases and to achieve stable segmentation results. All MRI images are selected and manually labeled by doctors. We divided the dataset into the training set and the test set in the ratio of 7:3. Of the 210 real cases of patients we owned, 147 are put into training set and 63 are in test set. We used a 10-fold cross-validation to ensure the validity of experimental results.
The training, test environment, and some important parameters settings during the training of the model in this paper are shown in Table 2.
Table 2. Operating environment and some parameter settings.

4.2. Detection Index

In the process of training and testing, better, we will use some common comparative index and single epoch training time (SETT) as metrics to evaluate the performance of our ACRNet model and other segmentation models. Moreover, we will also make a comparison of the number of parameters of the different models.
We used the confusion matrix to analyze and explain the model. TP indicates that both the predicted results and the actual results are osteosarcoma areas, and TN indicates that both the predicted results and the actual results are normal areas. FP represents the normal area misjudged as the disease area, and FN represents the lesion area misjudged as the normal area. The specific definition and calculation formula of the above indicators are as follows.
Definition 1.
Accuracy. It represents the proportion of all samples with correct prediction. Its definition is as shown below.
A c c u r a c y = T P + T N T P + T N + F P + F N
Definition 2.
Precision. It represents the proportion of all samples judged as tumor areas that are actual tumor areas. The formula for Precision is:
P r e c i s i o n = T P T P + F P
Definition 3.
Recall. It represents the percentage of all samples judged as normal areas that are really normal areas. Compared with precision, we expect the emergence of high recall in medical experiments, and false negative is more intolerable than false positive. Its definition is as follows.
R e c a l l = T N T N + F N
Definition 4.
F1-Score. It is an index based on precision and recall, which can represent the robustness of a model. The higher the F1-Score, the greater the robustness. Its definition is as shown below.
F 1 = 2 × P r e c i s i o n × R e c a l l P r e c i s i o n + R e c a l l
Definition 5.
IOU. It represents the intersection and union ratio between the tumor region in the prediction result and the tumor region in the real result, reflecting the coincidence between the prediction result and the real region. Its definition formula is as follows, where I p r e d represents the tumor region in the predicted result and I r e a l represents the tumor region in the real result.
I O U = I p r e d I r e a l I p r e d I r e a l
Definition 6.
DSC. It describes the similarity between the prediction result and the truth and is an index with a value of 0-1. The closer DSC is to 1, the closer the result is to the real situation. Its definition formula is as follows.
D S C = 2 × I p r e d I r e a l I p r e d + I r e a l
Definition 7.
SETT. It represents the average time required for an epoch of model training under the same amount of data, and the unit is seconds (s).

4.3. Comparison Model

For the purpose of further evaluating the segmentation effectiveness as well as the structural complexity of our ACRNet, we used U-Net [51], MSFCN [55], MSRN [56], FPN [57] and FCN-8s [58] to compare various metrics with the ACRNet model, which have achieved some achievements in image segmentation. In the related work, we have made certain presentations.

4.4. Operation Results and Analysis

Firstly, we trained the model without preprocessing. The training and test results are shown in Figure 5. The tumor areas marked in red are manually marked by professional doctors. In the two pictures marked in white, the one on the left is the result of model prediction obtained by training without preprocessing, and the one on the right is the result of model prediction obtained by using the preprocessed data as input. We can obviously find that the segmentation results with the preprocess are at closer proximity to the gold standard, while the output of the model without data preprocessing is not satisfactory. In addition, it can be seen that ACRNet with data preprocessing achieves better segmentation results for images with obscure osteosarcoma boundaries. Therefore, ACRNet makes the model more stable by using the strategy of compound decision to integrate multiple outputs from different angles of the same image, which effectively improves the segmentation effect of the model for the osteosarcoma region.
Figure 5. Comparison of segmentation results of the model with and without data preprocessing.
We trained the model and finally got a trained ACRNet model. Then, we put the prepared test data into it for testing and test it synchronously with other prepared image segmentation models. The tumor region prediction images are shown in Figure 6. The first column represents the original MRI image, which contains three different sections from different patients. The second column is the tumor area manually marked by doctors, which is called “Gold Standard”. Columns 3 to 7 are the predicted results of previous models we choose to compare. The last column is the tumor area map predicted by ACRNet. The DSC metrics of the predicted results of the MRI image under different models are recorded at the bottom of each line, and the highest index is marked in bold form.
Figure 6. Comparison of segmentation results of each model on different images.
According to Figure 6, we can find that in the five selected pictures, ACRNet shows good performance and the DSC metrics are higher than those of the other models. Especially in the image segmentation from the perspective of transverse, the result of ACRNet has obvious advantages, which is about 4 percentage points higher than that of the U-Net model. Based on these five examples of osteosarcoma segmentation, we can draw the conclusion that ACRNet can segment the osteosarcoma images with better accuracy and can be consistent with the manual annotation called Golden Standard.
To compare each model more accurately and digitally, we counted the segmentation results of each model and calculated and sorted out its indicators. We used several indicators selected and defined in section C to compare our model with other models. Table 3 shows different indicators of prediction results of different models under the same data set and the same operating environment. As shown in Table 3, we found that ACRNet showed a good effect on MRI image segmentation of osteosarcoma, and achieved higher values than others in most index comparisons, like IOU, recall, etc. These indicators are relatively high, which indicates that the model has better generalization and robustness, and can identify segmentation more accurately. To compare the computational complexity of the models, on the one hand, we compared the SETT metrics of ACRNet with those of UNet, MSFCN, MSRN, FCN and FCN-8s models. The SETT of ACRNet is only 174s, which is significantly lower than other segmentation models, indicating that it has higher computational efficiency than other models. On the other hand, the number of parameters of ACRNet is also the lowest among all models, only 6.91 M. This shows that ACRNet has lower computational complexity than other models. The low computational complexity can better accommodate medical devices with low configuration (e.g., CPU, GPU, etc.), which undoubtedly saves more human and financial resources for developing countries and can improve the efficiency of diagnosis of osteosarcoma in hospitals.
Table 3. Index comparison of different models in different image segmentation results.
Figure 7 more intuitively and specifically shows the comparison of parameters between different models and the changes of DSC. In Figure 7, the horizontal coordinates represent the different models, and the vertical coordinates are divided into two indicators on both the left and right. The left vertical coordinate indicates the quantity of parameters of these models, and the right vertical coordinate indicates the DSC metric of the model. The chart shows that ACRNet has lower number of parameters than UNet, MSFCN, MSRN, FCN, and FCN-8s models, with only 6.91 M parameters, which is nearly 7.4 M parameters lower than the MSRN model. Meanwhile, ACRNet is not inferior in osteosarcoma image segmentation because of the small number of parameters, with a DSC of 91.76%. Compared with the MSFCN model, which ranked second with a DSC of 89.2%, the DSC of ACRNet was nearly 2.5% higher. This result further shows that our model will take up less memory, and the cost of each training can be much lower than other models, which can save more medical expenses for some underdeveloped areas.
Figure 7. Comparison of Params and DSC between different models.
To achieve a better prediction effect, we often need to train more epochs to make the model fit the dataset better. We often expect a good model to converge to a higher accuracy as soon as possible, and the higher the final convergence value, the better. Based on this, we tested different models as shown in Figure 8. We selected some of the training records of the first 50 epochs from the log file for drawing a line diagram. According to Figure 8, we find that the initial accuracy of most models can reach more than 80%, and ACRNet reaches 88%. With the increase of training epochs, ACRNet began to converge rapidly and reached more than 95% accuracy in the early stage. Compared with the rapid rise of MSFCN and the stable convergence of U-Net, ACRNet not only remains stable after convergence but also the accuracy maintains at a high level. Generally speaking, ACRNet has good convergence efficiency and effects, which means that the model can obtain a good prediction result with fewer training epochs and less training time than other models. It can improve diagnostic efficiency for doctors.
Figure 8. Variation of the accuracy of each model with the number of training epoch.
In addition to comparing accuracy, we also made statistics on recall and F1-Score. Similarly, we took the first 50 epochs for research. Figure 9 shows the recall changes of different models with the change of epoch. MSRN and MSFCN improved rapidly in the early stage, but soon reached a relatively stable value and began to converge. Although ACRNet lagged behind several models at the beginning, there was a rapid improvement since about 20 epochs, and it has been maintained at a higher value than other models ever since. On the recall metrics, although ACRNet starts slightly slower than other models, it can get a more considerable improvement. Figure 10 shows the relation between F1-Score and the number of training epochs, which represents the robustness and stability of the model. ACRNet, FPN, U-Net, and FCN-8s have achieved good performance. ACRNet has less obvious advantages in the early stage of training, which is slightly higher than the other models. However, with the increase of the number of training epochs, according to the previous statistical results, we found that ACRNet can be 1% higher than the second-best model in later F1-Score comparison. From Figure 8, Figure 9 and Figure 10, we can observe that ACRNet model converges quickly and has good stability for each evaluation index such as accuracy, recall, and F1-Score. Therefore, it can be seen that the ACRNet combined with the composite decision-making approach makes the model have good stability.
Figure 9. Variation of the recall of each model with the number of training epoch.
Figure 10. Variation of the F1-Score of each model with the number of training epoch.
During the experiment, we also conducted additional research on the attention part. We visualized the attention mechanism and superimposed the binary mask output by the middle layer with the input picture to obtain a more intuitive information range of attention focus, namely the attention map. The brighter the area, the more focused the attention. We selected 18 maps of different categories for comparison and display, and the results are shown in Figure 11.
Figure 11. Some examples of attention maps from ACRNet.

5. Conclusions

In the article, the ACRNet was trained and tested by more than 4000 osteosarcoma MRI images. Through comparison and statistical analysis, we find that OMSAS has good prediction results in many indicators such as accuracy and recall. OMSAS not only improves the efficiency and accuracy of the model but also saves the cost with fewer parameters and shorter training time. The experimental results show that the segmentation model ACRNet in OMSAS has a better effect and more obvious advantages in osteosarcoma MRI image segmentation, and is suitable for application in developing countries with unbalanced medical resources distribution and poor medical care.
In the future, with further expansion and improvement of the dataset, we can continuously improve the prediction ability of the model, so as to detect the region of more complex osteosarcoma MRI images or even achieve similar good results on MRI of other cancers. At the same time, with the improvement of computing power and the need for later medical treatment, we can expand more modules in the model, such as outputting the size of the tumor area and predicting the future diffusion of tumor area, to provide more professional assistance for medical diagnosis.

Author Contributions

Conceptualization, J.W. and J.L.; Data curation, H.T. and H.H.; Formal analysis, F.G. and J.Z. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Data Availability Statement

Data used to support the findings of this study are currently under embargo while the research findings are commercialized. Requests for data, 12 months after publication of this article, will be considered by the corresponding author. All data analyzed during the current study are included in the submission.

Conflicts of Interest

The authors declare that they have no conflicts of interest.

References

  1. Rathore, R.; Van Tine, B. Pathogenesis and Current Treatment of Osteosarcoma: Perspectives for Future Therapies. J. Clin. Med. 2021, 10, 1182. [Google Scholar] [CrossRef] [PubMed]
  2. Harper, K.; Sathiadoss, P.; Saifuddin, A.; Sheikh, A. A review of imaging of surface sarcomas of bone. Skelet. Radiol. 2020, 50, 9–28. [Google Scholar] [CrossRef] [PubMed]
  3. Yang, C.; Tian, Y.; Zhao, F.; Chen, Z.; Su, P.; Li, Y.; Qian, A. Bone Microenvironment and Osteosarcoma Metastasis. Int. J. Mol. Sci. 2020, 21, 6985. [Google Scholar] [CrossRef] [PubMed]
  4. Zhao, Z.; Shi, J.; Zhao, G.; Gao, Y.; Jiang, Z.; Yuan, F. Large Scale Identification of Osteosarcoma Pathogenic Genes by Multiple Extreme Learning Machine. Front. Cell Dev. Biol. 2021, 9, 2654. [Google Scholar] [CrossRef] [PubMed]
  5. Jafari, F.; Javdansirat, S.; Sanaie, S.; Naseri, A.; Shamekh, A.; Rostamzadeh, D.; Dolati, S. Osteosarcoma: A comprehensive review of management and treatment strategies. Ann. Diagn. Pathol. 2020, 49, 151654. [Google Scholar] [CrossRef]
  6. Li, X.; Qi, H.; Wu, J. Efficient path-sense transmission based on IoT system in opportunistic social networks. Peer-to-Peer Netw. Appl. 2022, 15, 811–826. [Google Scholar] [CrossRef]
  7. Wu, J.; Gou, F.; Tan, Y. A Staging Auxiliary Diagnosis Model for Nonsmall Cell Lung Cancer Based on the Intelligent Medical System. Comput. Math. Methods Med. 2021, 2021, 6654946. [Google Scholar] [CrossRef]
  8. Wu, J.; Gou, F.; Tian, X. Disease Control and Prevention in Rare Plants Based on the Dominant Population Selection Method in Opportunistic Social Networks. Comput. Intell. Neurosci. 2022, 2022, 1489988. [Google Scholar] [CrossRef]
  9. Gou, F.; Wu, J. Message Transmission Strategy Based on Recurrent Neural Network and Attention Mechanism in Iot System. J. Circuits Syst. Comput. 2022, 31, 2250126. [Google Scholar] [CrossRef]
  10. Li, X.; Qi, H.; Wu, J. Node Social Nature Detection OSN Routing Scheme Based on IoT System. IEEE Internet Things J. 2022, 9, 14048–14059. [Google Scholar] [CrossRef]
  11. Gou, F.; Wu, J. Triad link prediction method based on the evolutionary analysis with IoT in opportunistic social networks. Comput. Commun. 2021, 181, 143–155. [Google Scholar] [CrossRef]
  12. Wu, J.; Xia, J.; Gou, F. Information transmission mode and IoT community reconstruction based on user influence in opportunistic social networks. Peer-to-Peer Netw. Appl. 2022, 15, 1398–1416. [Google Scholar] [CrossRef]
  13. Shen, Y.; Gou, F.; Dai, Z. Osteosarcoma MRI Image-Assisted Segmentation System Base on Guided Aggregated Bilateral Network. Mathematics 2022, 10, 1090. [Google Scholar] [CrossRef]
  14. Jiao, Y.; Qi, H.; Wu, J. Capsule network assisted electrocardiogram classification model for smart healthcare. Biocybern. Biomed. Eng. 2022, 42, 543–555. [Google Scholar] [CrossRef]
  15. Zhuang, Q.; Dai, Z.; Wu, J. Deep Active Learning Framework for Lymph Node Metastasis Prediction in Medical Support System. Comput. Intell. Neurosci. 2022, 2022, 4601696. [Google Scholar] [CrossRef] [PubMed]
  16. Li, L.; Gou, F.; Wu, J. Modified Data Delivery Strategy Based on Stochastic Block Model and Community Detection in Opportunistic Social Networks. Wirel. Commun. Mob. Comput. 2022, 2022, 5067849. [Google Scholar] [CrossRef]
  17. Lv, B.; Liu, F.; Gou, F.; Wu, J. Multi-Scale Tumor Localization Based on Priori Guidance-Based Segmentation Method for Osteosarcoma MRI Images. Mathematics 2022, 10, 2099. [Google Scholar] [CrossRef]
  18. Liu, F.; Gou, F.; Wu, J. An Attention-Preserving Network-Based Method for Assisted Segmentation of Osteosarcoma MRI Images. Mathematics 2022, 10, 1665. [Google Scholar] [CrossRef]
  19. Shen, Y.; Gou, F.; Wu, J. Node Screening Method Based on Federated Learning with IoT in Opportunistic Social Networks. Mathematics 2022, 10, 1669. [Google Scholar] [CrossRef]
  20. Wang, Z.; Jiao, Y.; Wu, J. User-optimized data transmission scheduling based on edge community service in opportunistic social network. IET Commun. 2022, 16, 1838–1852. [Google Scholar] [CrossRef]
  21. Kamel, Z.S.A.S.A.; El-Shafey, M.H.R.; Hassanien, O.A.; Nagy, H.A. Can dynamic magnetic resonance imaging replace static magnetic resonance sequences in evaluation of temporomandibular joint dysfunction? Egypt. J. Radiol. Nucl. Med. 2021, 52, 19. [Google Scholar] [CrossRef]
  22. Ponti, M.; Kittler, J.; Riva, M.; de Campos, T.; Zor, C. A decision cognizant Kullback–Leibler divergence. Pattern Recognit. 2017, 61, 470–478. [Google Scholar] [CrossRef]
  23. Binsar, F.; Eryanto, E.; Wahyudi, I.; Sugandi, Y.; Suroso, J.S. Risk of Invalidation of Data in Hospital Information Systems in In-donesia. In Proceedings of the 2020 International Conference on Information Management and Technology (ICIMTech), Bandung, Indonesia, 13–14 August 2020. [Google Scholar]
  24. Cui, R.; Chen, Z.; Wu, J.; Tan, Y.; Yu, G. A Multiprocessing Scheme for PET Image Pre-Screening, Noise Reduction, Segmentation and Lesion Partitioning. IEEE J. Biomed. Health Inform. 2020, 25, 1699–1711. [Google Scholar] [CrossRef]
  25. Yu, G.; Chen, Z.; Wu, J.; Tan, Y. A diagnostic prediction framework on auxiliary medical system for breast cancer in developing countries. Knowl.-Based Syst. 2021, 232, 107459. [Google Scholar] [CrossRef]
  26. Tian, X.; Yan, L.; Jiang, L.; Xiang, G.; Li, G.; Zhu, L.; Wu, J. Comparative transcriptome analysis of leaf, stem, and root tissues of Semiliquidambar cathayensis reveals candidate genes involved in terpenoid biosynthesis. Mol. Biol. Rep. 2022, 49, 5585–5593. [Google Scholar] [CrossRef] [PubMed]
  27. Gou, F.; Wu, J. Data Transmission Strategy Based on Node Motion Prediction IoT System in Opportunistic Social Networks. Wirel. Pers. Commun. 2022, 126, 1751–1768. [Google Scholar] [CrossRef]
  28. Li, L.; Gou, F.; Long, H.; He, K.; Wu, J. Effective Data Optimization and Evaluation Based on Social Communication with AI-Assisted in Opportunistic Social Networks. Wirel. Commun. Mob. Comput. 2022, 2022, 4879557. [Google Scholar] [CrossRef]
  29. Wu, J.; Xiao, P.; Huang, H.; Gou, F.; Zhou, Z.; Dai, Z. An artificial intelligence multiprocessing scheme for the diagnosis of osteosarcoma MRI images. IEEE J. Biomed. Health Inform. 2022, 26, 4656–4667. [Google Scholar] [CrossRef]
  30. Qin, Y.; Li, X.; Wu, J.; Yu, K. A management method of chronic diseases in the elderly based on IoT security environment. Comput. Electr. Eng. 2022, 102, 108188. [Google Scholar] [CrossRef]
  31. Wu, J.; Guo, Y.; Gou, F.; Dai, Z. A medical assistant segmentation method for MRI images of osteosarcoma based on DecoupleSegNet. Int. J. Intell. Syst. 2022, 37, 8436–8461. [Google Scholar] [CrossRef]
  32. Wu, J.; Liu, Z.; Gou, F.; Zhu, J.; Tang, H.; Zhou, X.; Xiong, W. BA-GCA Net: Boundary-Aware Grid Contextual Attention Net in Osteosarcoma MRI Image Segmentation. Comput. Intell. Neurosci. 2022, 2022, 3881833. [Google Scholar] [CrossRef] [PubMed]
  33. Wu, J.; Zhou, L.; Gou, F.; Tan, Y. A Residual Fusion Network for Osteosarcoma MRI Image Segmentation in Developing Countries. Comput. Intell. Neurosci. 2022, 2022, 7285600. [Google Scholar] [CrossRef] [PubMed]
  34. Xiong, W.; Chen, H.; Jiao, Y.; Yang, M.; Zhou, X. A User Cache Management and Cooperative Transmission Mechanism Based on Edge Community Computing in Opportunistic Social Networks. IET Commun. 2022, 16, 2045–2058. [Google Scholar] [CrossRef]
  35. Yu, G.; Chen, Z.; Wu, J.; Tan, Y. Medical decision support system for cancer treatment in precision medicine in developing countries. Expert Syst. Appl. 2021, 186, 115725. [Google Scholar] [CrossRef]
  36. Liu, F.; Zhu, J.; Lv, B.; Yang, L.; Sun, W.; Dai, Z.; Gou, F.; Wu, J. Auxiliary Segmentation Method of Osteosarcoma MRI Image Based on Transformer and U-Net. Comput. Intell. Neurosci. 2022, 2022, 9990092. [Google Scholar] [CrossRef]
  37. Wu, J.; Yu, L.; Gou, F. Data transmission scheme based on node model training and time division multiple access with IoT in opportunistic social networks. Peer-to-Peer Netw. Appl. 2022, 15, 2719–2743. [Google Scholar] [CrossRef]
  38. Wu, J.; Yang, S.; Gou, F.; Zhou, Z.; Xie, P.; Xu, N.; Dai, Z. Intelligent Segmentation Medical Assistance System for MRI Images of Osteosarcoma in Developing Countries. Comput. Math. Methods Med. 2022, 2022, 7703583. [Google Scholar] [CrossRef]
  39. Chen, H.; Liu, J.; Cheng, Z.; Lu, X.; Wang, X.; Lu, M.; Li, S.; Xiang, Z.; Zhou, Q.; Liu, Z.; et al. Development and external validation of an MRI-based radiomics nomogram for pretreatment prediction for early relapse in osteosarcoma: A retrospective multicenter study. Eur. J. Radiol. 2020, 129, 109066. [Google Scholar] [CrossRef]
  40. Georgeanu, V.; Mamuleanu, M.L.; Selişteanu, D. Convolutional neural networks for automated detection and classification of bone tumors in magnetic resonance imaging. In Proceedings of the 2021 IEEE International Conference on Artificial Intelligence, Robotics, and Communication (ICAIRC), Fuzhou, China, 25–27 June 2021; pp. 5–7. [Google Scholar]
  41. Wen, X.; Famouri, M.; Hryniowski, A.; Wong, A. AttendSeg: A Tiny Attention Condenser Neural Network for Semantic Segmentation on the Edge. arXiv 2021, arXiv:2104.14623. [Google Scholar]
  42. Wong, A.; Famouri, M.; Pavlova, M.; Surana, S. Tinyspeech: Attention condensers for deep speech recognition neural networks on edge devices. arXiv 2020, arXiv:2008.04245. [Google Scholar]
  43. Nasor, M.; Obaid, W. Segmentation of osteosarcoma in mri images by k-means clustering, chan-vese segmentation, and iterative gaussian filtering. IET Image Process. 2021, 15, 1310–1318. [Google Scholar] [CrossRef]
  44. Kayal, E.B.; Kandasamy, D.; Sharma, R.; Bakhshi, S.; Mehndiratta, A. Segmentation of osteosarcoma tumor using diffusion weighted MRI: A comparative study using nine segmentation algorithms. Signal Image Video Process. 2019, 14, 727–735. [Google Scholar] [CrossRef]
  45. Nabid, R.A.; Rahman, M.L.; Hossain, M.F. Classification of Osteo-sarcoma Tumor from Histological Image Using Sequential RCNN. In Proceedings of the 2020 11th International Conference on Electrical and Computer Engineering (ICECE), Dhaka, Bangladesh, 17–19 December 2020; pp. 363–366. [Google Scholar] [CrossRef]
  46. Arunachalam, H.B.; Mishra, R.; Daescu, O.; Cederberg, K.; Rakheja, D.; Sengupta, A.; Leonard, D.; Hallac, R.; Leavey, P. Viable and necrotic tumor assessment from whole slide images of osteosarcoma using machine-learning and deep-learning models. PLoS ONE 2019, 14, e0210706. [Google Scholar] [CrossRef] [PubMed]
  47. Dionísio, F.; Oliveira, L.; Hernandes, M.; Engel, E.; Rangayyan, R.; Azevedo-Marques, P.; Nogueira-Barbosa, M. Manual and semiautomatic segmentation of bone sarcomas on MRI have high similarity. Braz. J. Med. Biol. Res. 2020, 53, e8962. [Google Scholar] [CrossRef]
  48. Badashah, S.J.; Basha, S.S.; Ahamed, S.R.; Rao, S.P.V.S. Fractional-Harris hawks optimization-based generative adversarial network for osteosarcoma detection using Renyi entropy-hybrid fusion. Int. J. Intell. Syst. 2021, 36, 6007–6031. [Google Scholar] [CrossRef]
  49. Anisuzzaman, D.; Barzekar, H.; Tong, L.; Luo, J.; Yu, Z. A deep learning study on osteosarcoma detection from histological images. Biomed. Signal Process. Control 2021, 69, 102931. [Google Scholar] [CrossRef]
  50. Shirly, S.; Ramesh, K. Review on 2D and 3D MRI Image Segmentation Techniques. Curr. Med. Imaging Former. Curr. Med. Imaging Rev. 2019, 15, 150–160. [Google Scholar] [CrossRef]
  51. Ronneberger, O.; Fischer, P.; Brox, T. U-net: Convolutional networks for biomedical image segmentation. In Proceedings of the International Conference on Medical Image Computing and Computer-Assisted Intervention, Munich, Germany, 5–9 October 2015; Springer: Berlin/Heidelberg, Germany, 2015; pp. 234–241. [Google Scholar]
  52. Song, S.; Gao, T. Research on image segmentation algorithm based on threshold. In Proceedings of the 2021 13th International Conference on Measuring Technology and Mechatronics Automation (ICMTMA), Beihai, China, 16–17 January 2021; pp. 306–308. [Google Scholar] [CrossRef]
  53. Gao, J.; Wang, B.; Wang, Z.; Wang, Y.; Kong, F. A wavelet transform-based image segmentation method. Opt.-Int. J. Light Electron. Opt. 2019, 208, 164123. [Google Scholar] [CrossRef]
  54. Liu, F. Susan: Segment unannotated image structure using adversarial network. Magn. Reson. Med. 2018, 81, 3330–3345. [Google Scholar] [CrossRef]
  55. Huang, L.; Xia, W.; Zhang, B.; Qiu, B.; Gao, X. MSFCN-multiple supervised fully convolutional networks for the osteosarcoma segmentation of CT images. Comput. Methods Programs Biomed. 2017, 143, 67–74. [Google Scholar] [CrossRef]
  56. Rui, Z.; Lin, H.; Wei, X.; Bo, Z.; Qiub, B. Multiple supervised residual network for osteosarcoma segmentation in ct images. Comput. Med. Imaging Graph. 2018, 63, 1–8. [Google Scholar]
  57. Lin, T.Y.; Dollár, P.; Girshick, R.; He, K.; Hariharan, B.; Belongie, S. Feature pyramid networks for object detection. In Proceedings of the 30th IEEE Conference on Computer Vision and Pattern Recognition, CVPR 2017, Honolulu, HI, USA, 21–26 July 2017; pp. 936–944. [Google Scholar] [CrossRef]
  58. Shelhamer, E.; Long, J.; Darrell, T. Fully convolutional networks for semantic segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 640–651. [Google Scholar] [CrossRef]
  59. Gou, F.; Liu, J.; Zhu, J.; Wu, J. A Multimodal Auxiliary Classification System for Osteosarcoma Histopathological Images Based on Deep Active Learning. Healthcare 2022, 10, 2189. [Google Scholar] [CrossRef] [PubMed]
  60. Wang, L.; Yu, L.; Zhu, J.; Tang, H.; Gou, F.; Wu, J. Auxiliary Segmentation Method of Osteosarcoma in MRI Images Based on Denoising and Local Enhancement. Healthcare 2022, 10, 1468. [Google Scholar] [CrossRef] [PubMed]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.