MRI Based Brain Tumor Segmentation Using FPGA Accelerated Neural Network
MRI Based Brain Tumor Segmentation Using FPGA Accelerated Neural Network
*Correspondence:
caow@[Link]; Abstract
shizhifeng@[Link]
†
Background: Brain tumor segmentation is a challenging problem in medical image
Siyu Xiong, Guoqing Wu
have contributed equally to processing and analysis. It is a very time-consuming and error-prone task. In order
this work to reduce the burden on physicians and improve the segmentation accuracy, the
1
State Key Laboratory computer-aided detection (CAD) systems need to be developed. Due to the powerful
of ASIC an System, Fudan
University, Shanghai, China feature learning ability of the deep learning technology, many deep learning-based
3
Huashan Hospital Affiliated methods have been applied to the brain tumor segmentation CAD systems and
to Fudan University, achieved satisfactory accuracy. However, deep learning neural networks have high
Shanghai, China
Full list of author information computational complexity, and the brain tumor segmentation process consumes
is available at the end of the significant time. Therefore, in order to achieve the high segmentation accuracy of brain
article tumors and obtain the segmentation results efficiently, it is very demanding to speed
up the segmentation process of brain tumors.
Results: Compared with traditional computing platforms, the proposed FPGA accel-
erator has greatly improved the speed and the power consumption. Based on the
BraTS19 and BraTS20 dataset, our FPGA-based brain tumor segmentation accelerator
is 5.21 and 44.47 times faster than the TITAN V GPU and the Xeon CPU. In addition, by
comparing energy efficiency, our design can achieve 11.22 and 82.33 times energy
efficiency than GPU and CPU, respectively.
Conclusion: We quantize and retrain the neural network for brain tumor segmenta-
tion and merge batch normalization layers to reduce the parameter size and computa-
tional complexity. The FPGA-based brain tumor segmentation accelerator is designed
to map the quantized neural network model. The accelerator can increase the seg-
mentation speed and reduce the power consumption on the basis of ensuring high
accuracy which provides a new direction for the automatic segmentation and remote
diagnosis of brain tumors.
Keywords: Brain tumor segmatation, FPGA acceleration, Neural network
Background
Brain glioma is the most common malignant tumor caused by the cancerization of glial
cells in the brain and spinal cord. It has the characteristics of high incidence, high recur-
rence, high mortality and low cure rate. The purpose of brain tumor segmentation is to
separate the tumor tissue from the healthy brain tissue such as white matter, gray matter
© The Author(s) 2021. Open Access This article is licensed under a Creative Commons Attribution 4.0 International License, which permits
use, sharing, adaptation, distribution and reproduction in any medium or format, as long as you give appropriate credit to the original
author(s) and the source, provide a link to the Creative Commons licence, and indicate if changes were made. The images or other third
party material in this article are included in the article’s Creative Commons licence, unless indicated otherwise in a credit line to the mate-
rial. If material is not included in the article’s Creative Commons licence and your intended use is not permitted by statutory regulation or
exceeds the permitted use, you will need to obtain permission directly from the copyright holder. To view a copy of this licence, visit http://
[Link]/licenses/by/4.0/. The Creative Commons Public Domain Dedication waiver ([Link]
main/zero/1.0/) applies to the data made available in this article, unless otherwise stated in a credit line to the data.
Xiong et al. BMC Bioinformatics (2021) 22:421 Page 2 of 15
and cerebrospinal fluid [1]. It plays an important role in the diagnosis and treatment of
the brain glioma.
The glioma image segmentation is helpful for surgical planning and can improve the
survival rate. Currently, most of the segmentation of brain tumor images is performed
by physicians. In clinical practice, due to the increasing number of brain tumor images,
manual segmentation of different areas of brain tumors becomes an error-prone and
time-consuming task for physicians. In addition, physicians’ cognitions may have dif-
ferent influences on the formulation of subsequent treatment plans and operations.
Therefore, automated methods are needed for high accuracy brain tumor location and
segmentation.
In the past, CPU was used to complete computations in CAD systems. Due to the
unsatisfactory speed of processing data by the CPU, many GPU acceleration methods
have been gradually studied [2, 3]. For some machine learning algorithms in bioinfor-
matics, FPGA acceleration has also become a new direction [4, 5].
Related work
Automated brain tumor segmentation has attracted widespread attention in the research
community and has been continuously studied. Before 2010, most researchers used
standard image processing methods, such as threshold-based method [6] and region-
based method [7]. Suzuki et al. used an iterative thresholding algorithm for segmenta-
tion [6], but when the image contrast is low, it becomes difficult to select the threshold.
In 2005, it was proved that region growth is an effective brain tumor segmentation
method. Compared with other non-region-based methods, the amount of computation
is less, especially for homogeneous tissues and regions [7]. Although they are simple to
implement and have small amount of computation, the segmentation accuracy does not
meet the practical expectation. Hence it is mostly used for two-dimensional segmenta-
tion only [8]. Subsequently, machine learning [9] has been gradually applied to medi-
cal image analysis. Many researchers have proposed brain tumor segmentation based on
classification or clustering methods [10–12]. Fletcher-Heath et al. used an unsupervised
fuzzy clustering algorithm, which combines domain knowledge and image processing
technology to achieve tumor segmentation [10]. Zhou et al. proposed a method based
on one-class support vector machine(SVM) to extract brain tumors from Magnetic
Resonance Imaging(MRI) [11]. Subbanna et al. presented a fully automated hierarchi-
cal probabilistic framework for segmenting brain tumor based on multiwindow Gabor
filters and an adapted Markov Random Field (MRF) framework [12]. Compared with
conventional segmentation methods, these methods can improve accuracy. However,
methods with higher accuracy are still needed in clinical practice.
In the past ten years, with the huge increase in computing power, deep learning meth-
ods have continued to advance. Deep neural networks can thoroughly learn hierarchi-
cal features from input images instead of pre-defined manual features. There are many
well-known deep learning networks, such as Convolutional Neural Networks(CNNs)
and Recurrent Neural Networks(RNNs), which are gradually applied in various tasks of
medical image analysis, such as breast image analysis [13] and chest X-ray image analysis
[14]. At the same time, the segmentation of brain tumors based on deep learning net-
works have also aroused the interest of researchers.
Xiong et al. BMC Bioinformatics (2021) 22:421 Page 3 of 15
In 2014, D. Zikic et al. studied the possibility of directly applying CNNs to brain
tumor segmentation, which achieved higher segmentation accuracy than traditional
machine learning methods [15]. In 2016, Brosch et al. proposed a segmentation
method based on a deep 3D convolutional encoder network, which composes of two
interrelated paths, namely a convolution path and a deconvolution path. Each image
contains a repetitive structure with corresponding changes. Therefore, only a few
images are needed to train a network [16]. In 2017, a multi-path CNN network for
brain tumor segmentation was proposed as an extension of single-path feedforward
CNN [17]. Multi-path CNN can extract different features from different modalities.
In 2019 and 2020, Muhammad Sharif and Javaria Amin et al. proposed several brain
tumor segmentation algorithms [18–21] to further improve the segmentation accu-
racy and reduce the processing time. Our design is based on the 3D U-Net network
proposed in [22], which extends the U-Net network [23] and replaces all 2D opera-
tions with 3D operations. It is a multi-path CNN network which can achieve great
accuracy in the segmentation task.
Challenge
Although significant progress has been made in the brain tumor segmentation, there
are still problems and challenges to be solved. Firstly, brain gliomas are mutations
of glial cells. Due to the wide spatial distribution of glial cells, gliomas can appear
anywhere in the brain. Moreover, the shapes and sizes of brain tumors in different
patients have great uncertainty, which means before segmentation process, almost
no prior information can be provided to describe the shape and size of a tumor.
Location uncertainty and morphological uncertainty have brought great challenges
to accurately locate brain tumors. Secondly, MRI which provides the tissue details
can be imaged in multiple directions. The 3D imaging method is more conducive to
the detection of brain tumors, so we mainly focus on MRI to segment brain tumors.
However, the MRI computation in the automated process is complicated, which usu-
ally requires more time for image analysis. There is demanding performance require-
ment for the segmentation processing platform.
It can be concluded that deep learning methods can achieve high accuracy in the
brain tumor segmentation task. However, the brain tumor segmentation process may
consume significant time and computing resources. Therefore, speeding up the CNN-
based 3D brain tumor segmentation is the key to high accuracy brain tumor detec-
tion and obtaining detection results efficiently. In order to speed up the segmentation
process, GPU is currently used for brain tumor segmentation, but it can be further
improved in terms of speed and power consumption.
CNN realizes brain tumor segmentation in two processes: training and inference.
Training is an iterative process to train the parameters. During the training process,
the output of the model is compared with the expected result to update the param-
eters to minimize the difference. This process is repeated until the output results
converge to a value that reduces the gap to an acceptable range. Training is a typical
offline operation which can be done in advance. The inference process is real-time, so
our work is focused on accelerating the inference process.
Xiong et al. BMC Bioinformatics (2021) 22:421 Page 4 of 15
Method
We have designed an FPGA-based brain tumor segmentation inference accelerator
which speeds up the segmentation process. It can be shown that our FPGA-based design
outperforms traditional computing technologies such as CPU and GPU implementa-
tions. In this section, we firstly describe the hardware and software platforms, and the
dataset. Then the quantization process for brain tumor segmentation neural network
and the hardware acceleration architecture are presented.
Fig. 2 Hardware platform for brain tumor segmentation where the red is a Xilinx Alveo U280 accelerator card
Fig. 3 Input images with four MRI modalities and the corresponding segmentation output
Software Xilinx Vivado integrated environment and Synopsys VCS simulator are
used for hardware design and simulation. GPU reference designs based on single-
precision floating point are compiled by CUDA 10.1 with PyTorch of python 3.8.3
and cuDNN 7.6.3 library.
Dataset
The BraTS19 and BraTS20 dataset [24–26] are used to test the performance of brain
tumor segmentation with four modalities for each case. A single modality of brain
tumor may lead to inaccurate segmentation because it does not provide detailed
information. Multi-modality images can complement each other effectively, which
can improve the segmentation accuracy. Figure 3 shows one exemplar with four MRI
modalities, flair, t1, t2 and t1ce. Each represents a unique MRI modality. The image
on the right is the segmentation result in Fig. 3.
Xiong et al. BMC Bioinformatics (2021) 22:421 Page 6 of 15
Y =W · X + B (1)
x̂i =(xi − µ)/ σ 2 + ǫ,
(2)
yi =γ x̂i + β
By combining Eqs. 1 and 2 , the parameters after merging the BN layer can be obtained
as shown in Eqs. 3, 4. Equation 5 describes the output of the merged layer.
γ
Wmerged =W · √ (3)
σ2 + ǫ
γ
Bmerged =(B − µ) · √ (4)
σ2 + ǫ
During the merging process, the 3D U-Net structure is traversed to find adjacent con-
volutional layers and BN layers. The merged weights and bias are obtained according
to Eqs. 3 and 4 . Then the original convolutional layer and BN layer are replaced by the
merged convolutional layer. The core computation equation of the deconvolutional layer
is the same as that of the convolutional layer, so the above merging method is also effec-
tive for BN layers and deconvolutional layers.
Adam 30 2 0.01
of an 8-bit signed fixed-point number, that is − 128 to 127. In order to facilitate the
FPGA to perform scaling operations, the scaling factors are constrained to be power of
2, so that the scaling operations can be achieved through the shift operations. The calcu-
lation of the bit number of the shift operations bs is shown in Eq. 6 where the bit width
bw is 8, |W| is the absolute value of W, ceil and max are the round-up and the maximum
functions respectively.
In the quantization process of 3D U-Net, the weight and bias scale factors of each 3D
convolutional or deconvolutional layer are calculated separately.
as input to run the network forward, and calculate the pixel accuracy score (ACC) and
dice similarity coefficient score (DSC). Compared with the network before the quantiza-
tion, the loss is almost negligible. After quantization, the activity data and parameters
are 8-bit signed fixed-point numbers, which can reduce storage resources and imple-
ment in the PFGA hardware efficiently.
Control mechanism
The execution of each instruction of the MIPS CPU is divided into 5 stages: instruction
fetch (IF), instruction decode (ID), execute (EX), memory access (MEM), and register
write back (WB). Usually these 5 stages are made into one 5-stage pipeline, and the exe-
cution of each stage of the pipeline is fixed at 1 clock cycle, so that the pipeline improves
the processing speed. Our accelerator learns from the CPU’s approach and divides the
execution of each instruction into 4 stages: configuration (CF), load data (LD), execution
(EX), and write back (WB). Because the numbers of cycles consumed by these 4 stages
are different and not fixed, it cannot form a pipeline like the MIPS CPU. The double-
buffering strategy is adopt, which can make the stages with the longest cycles to cover
up the time of other stages. Figure 5a, b are diagrams of the execution stages without
and under the double-buffering strategy. This is the case where the execution time of the
EX stage is longer than the other 3 stages. Under the double-buffering strategy, it can
be seen that when multiple such instructions are executed, the EX stage will mask most
of the running time of CF, LD, and WB, reducing the overall computation time signifi-
cantly. Similarly, if LD becomes the longest time for each stage, the LD stage can cover
most of the running time of CF, EX and WB.
Fig. 5 a The execution of each stage without the double-buffering strategy, b the execution of each stage
under the double-buffering strategy
Xiong et al. BMC Bioinformatics (2021) 22:421 Page 10 of 15
one layer of 3D U-Net must be devided into several blocks. Different partition results
will have different effects on the execution efficiency of the CNN model on FPGA. Con-
straints are set according to chip resources and other conditions to obtain partition
result. According to the partition result, address allocation is performed on the input
feature map of the first layer and the static data such as the weights and bias data of each
layer.
After obtaining the partition result and the data addresses respectively, we start
to schedule the execution stages. The goal of scheduling is to try to conceal the load
data time or the computation time of the computation block to improve the efficiency
of the execution pipeline. Firstly, we use the layer as a unit to generate a sequence of
computation blocks according to the partition result. In this step, the computation pro-
cess of each layer is divided into several computation blocks in a certain order. Secondly,
according to the instruction set configuration rules of the chip and the label informa-
tion of each computation block, we generate the instructions of the binary computation
block sequence. Finally, the binary instructions are stored in the external memory for the
instruction acquisition module of the design to obtain the relevant instructions.
Result
Performance The BraTS19 and BraTS20 datasets are used to train the network and
test. In BraTS19, we use 240(HGG/LGG) cases for training and 100 cases for testing.
In BraTS20, 260(HGG/LGG) cases are for training and 109 cases are for testing. The
DSC, execution time, and energy consumption of our method are shown in Table 2. In
BraTS19 and BraTS20, our FPGA method achieves 0.871 DSC and 0.882 DSC. The aver-
age execution time is 0.15s and the FPGA energy consumptions is 45W.
Table 3 is comparison provided with the recent method which also test on the BraTS
dataset. It demonstrates that the suggested method provided accurate and efficient seg-
mentation results. Moreover, our FPGA approach also reduces power consumption by
more than half compared to CPU or GPU solutions.
The performance and energy efficiency of FPGA, CPU and GPU designs are also com-
pared. The execution time of testing a single image is accidental because it is too short
and often not accurate enough. Therefore, the number of test images per second are
measured by testing 200 MRIs segmentation tasks in BraTS19 and BraTS20 datasets to
obtain the average. Performance is measured by the number of frames per second (FPS)
of images processed, and energy efficiency refers to the ratio of performance and power
(FPS/MW) in the computation process. Figure 7 shows the comparison of performance
and energy efficiency among FPGA, CPU and GPU implementations. It can be seen that
the performance of FPGA is 5.21 times higher than CPU and 44.47 times faster than
GPU, and the energy efficiency ratio is 11.22 times of GPU and 82.33 times of CPU. Both
processing speed and energy efficiency have been greatly improved.
Resource usage Our design is tested on the U280 card. The FPGA resource utilization
is listed in Table 4. It can be seen that our FPGA design uses storage and computing
resources reasonably. Figure 8 is the floorplan of the FPGA-based design after place-
ment and routing. There are 6 computation cores inside the FPGA chip. Each of them is
marked in a different color.
Xiong et al. BMC Bioinformatics (2021) 22:421 Page 12 of 15
Fig. 7 Performance and energy efficiency comparison among CPU, GPU, and FPGA
Conclusion
Brain tumor is one of the most common cancers which has the characteristics of high
morbidity, high recurrence, and high mortality. Brain tumor segmentation is a very
effective method to identify potentially cancerous tissue. However, this increases the
burden on physicians, and the physicians’ status and experiences greatly affect the
analysis results. Therefore, many CAD systems have been developed. In these systems,
Xiong et al. BMC Bioinformatics (2021) 22:421 Page 14 of 15
the first step of high-precision brain tumor segmentation is crucial for the subsequent
treatment process. We propose and implement an FPGA-based brain tumor segmen-
tation inference accelerator, which can speed up segmentation and reduce power con-
sumption. Based on BraTS19 and BraTS20, the performance and power consumption
of our FPGA accelerator are better than traditional computing technology. The aver-
age speed is 5.21 times and 44.47 times higher than that of CPU and GPU. In addi-
tion, the energy efficiency is 11.22 times and 82.33 times higher than that of CPU
and GPU. The design of FPGA acceleration hardware provides a new direction for the
improvement of automated brain tumor segmentation.
Abbreviations
CPU: Central processing unit; GPU: Graphics processing unit; FPGA: Field programmable gate array; CAD: Computer-
aided detection; SVM: Support vector machine; MRI: Magnetic resonance imaging; MRF: Markov random field; CNN:
Convolutional neural network; RNN: Recurrent neural network.
Acknowledgements
Not applicable.
Authors’ contributions
ZS, WC conceived the project. WC, XF, SX designed and implemented the FPGA-based architecture. GW, XZ completed
the quantification work. ZH completed the partition and scheduling process. SX performed the analysis and wrote the
manuscript. LW, JY, SD, ZS critically revised the manuscript. All authors have read and approved of the final manuscript.
Funding
This paper is supported by the National Natural Science Foundation of China (61971143), Shanghai Municipal Science
and Technology Major Project (2018SHZDZX01), Shanghai Muncipal Health Commission Project (2018ZHYL0107).
Declarations
Ethics approval and consent to participate
Not applicable.
Competing interests
The authors declare that they have no competing interests.
Author details
1
State Key Laboratory of ASIC an System, Fudan University, Shanghai, China. 2 School of Information Science and Tech-
nology, Fudan University, Shanghai, China. 3 Huashan Hospital Affiliated to Fudan University, Shanghai, China. 4 School
of Computer Science, Fudan University, Shanghai, China.
References
1. Gordillo N, Montseny E, Sobrevilla P. State of the art survey on MRI brain tumor segmentation. Magn Reson Imaging.
2013;31:1426–38.
2. Ahmed N, Lévy J, Ren S, Mushtaq H, Bertels K, Al-Ars Z. GASAL2: a GPU accelerated sequence alignment library for
high-throughput NGS data. BMC Bioinform. 2019;20:520.
3. Awan MG, Deslippe J, Buluc A, Selvitopi O, Hofmeyr S, Oliker L, Yelick K. ADEPT: a domain independent sequence
alignment strategy for GPU architectures. BMC Bioinform. 2020;21:1–29.
4. Jin Z, Bakos JD. Extending the BEAGLE library to a multi-FPGA platform. BMC Bioinform. 2013;14:25.
5. Sanaullah A, Yang C, Alexeev Y, Yoshii K, Herbordt MC. Real-time data analysis for medical diagnosis using FPGA-
accelerated neural networks. BMC Bioinform. 2018;19:490.
6. Suzuki H, Toriwaki J. Automatic segmentation of head MRI images by knowledge guided thresholding. Comput
Med Imaging Graph. 1991;15:233–40.
7. Salman YM, Assal MA, Badawi AM, Alian SM, El-Bayome ME-. Validation techniques for quantitative brain tumors
measurements. In: 2005 IEEE Engineering in Medicine and Biology 27th Annual Conference, 2005;pp. 7048–7051.
Xiong et al. BMC Bioinformatics (2021) 22:421 Page 15 of 15
8. Vijayakumar C, Gharpure DC. Development of image-processing software for automatic segmentation of brain
tumors in MR images. J Med Phys. 2011;36:147–58.
9. Alpaydin E. Introduction to machine learning. London: MIT Press; 2020.
10. Fletcher-Heath LM, Hall LO, Goldgof DB, Murtagh FR. Automatic segmentation of non-enhancing brain tumors in
magnetic resonance images. Artif Intell Med. 2001;21:43–63.
11. Zhou J, Chan K, Chong V, Krishnan SM. Extraction of brain tumor from MR images using one-class support vector
machine. In: 2005 IEEE engineering in medicine and biology 27th annual conference; 2006. p. 6411–14.
12. Subbanna NK, Precup D, Collins DL, Arbel T. Hierarchical probabilistic GABOR and MRF segmentation of brain
tumours in MRI volumes. In: International conference on medical image computing and computer-assisted inter-
vention; 2013. p. 751–8.
13. Kooi T, van Ginneken B, Karssemeijer N, den Heeten A. Discriminating solitary cysts from soft tissue lesions in mam-
mography using a pretrained deep convolutional neural network. Med Phys. 2017;44:1017–27.
14. Cicero M, Bilbily A, Colak E, Dowdell T, Gray B, Perampaladas K, Barfett J. Training and validating a deep convolutional
neural network for computer-aided detection and classification of abnormalities on frontal chest radiographs. Inves-
tig Radiol. 2017;52:281–7.
15. Zikic D, Ioannou Y, Brown M, Criminisi A. Segmentation of brain tumor tissues with convolutional neural networks.
In: MICCAI workshop on multimodal brain tumor segmentation challenge (BRATS); 2014. p. 36–39.
16. Brosch T, Tang LY, Yoo Y, Li DK, Traboulsee A, Tam R. Deep 3D convolutional encoder networks with shortcuts
for multiscale feature integration applied to multiple sclerosis lesion segmentation. IEEE Trans Med Imaging.
2016;35:1229–39.
17. Sedlar S. Brain tumor segmentation using a multi-path CNN based metho. In: International MICCAI Brainlesion
workshop; 2017. p. 403–22.
18. Amin J, Sharif M, Raza M, Saba T, Anjum MA. Brain tumor detection using statistical and machine learning method.
Comput Methods Programs Biomed. 2019;177:69–79.
19. Sharif M, Amin J, Nisar MW, Anjum MA, Muhammad N, Shad SA. A unified patch based method for brain tumor
detection using features fusion. Cogn Syst Res. 2020;59:273–86.
20. Amin J, Sharif M, Anjum MA, Raza M, Bukhari SAC. Convolutional neural network with batch normalization for
glioma and stroke lesion detection using MRI. Cogn Syst Res. 2020;59:304–11.
21. Sharif M, Amin J, Raza M, Anjum MA, Afzal H, Shad SA. Brain tumor detection based on extreme learning. Neural
Comput Appl. 2020;56:1–13.
22. Çiçek Ö, Abdulkadir A, Lienkamp SS, Brox T, Ronneberger O. 3D U-Net: learning dense volumetric segmentation
from sparse annotation. In: International conference on medical image computing and computer-assisted interven-
tion; 2016. p. 424–32.
23. Ronneberger O, Fischer P, Brox T. U-net: convolutional networks for biomedical image segmentation. In: Interna-
tional conference on medical image computing and computer-assisted intervention; 2015. p. 234–41.
24. Menze BH, Jakab A, Bauer S, Kalpathy-Cramer J, Farahani K, Kirby J, Burren Y, Porz N, Slotboom J, Wiest R, et al. The
multimodal brain tumor image segmentation benchmark (brats). IEEE Trans Med Imaging. 2014;34(10):1993–2024.
25. Bakas S, Akbari H, Sotiras A, Bilello M, Rozycki M, Kirby JS, Freymann JB, Farahani K, Davatzikos C. Advancing the
cancer genome atlas glioma MRI collections with expert segmentation labels and radiomic features. Sci Data.
2017;4(1):1–13.
26. Bakas S, Reyes M, Jakab A, Bauer S, Rempfler M, Crimi A, Shinohara RT, Berger C, Ha SM, Rozycki M, et al. Identifying
the best machine learning algorithms for brain tumor segmentation, progression assessment, and overall survival
prediction in the brats challenge; 2018. arXiv preprint arXiv:1811.02629.
27. Luo C, Cao W, Wang L, Leong PH. RNA: an accurate residual network accelerator for quantized and reconstructed
deep neural networks. IEICE Trans Inf Syst. 2019;102:1037–45.
28. Zhao X, Wu Y, Song G, Li Z, Zhang Y, Fan Y. A deep learning model integrating FCNNs and CRFs for brain tumor
segmentation. Med Image Anal. 2018;43:98–111.
29. Amin J, Sharif M, Yasmin M, Fernandes SL. Big data analysis for brain tumor detection: deep convolutional neural
networks. Futur Gener Comput Syst. 2018;87:290–7.
30. Chen S, Ding C, Liu M. Dual-force convolutional neural networks for accurate brain tumor segmentation. Pattern
Recognit. 2019;88:90–100.
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.