Summary
English

Automatically Generated

Swin-PSAxialNet: An Efficient Multi-Organ Segmentation Technique

Published: July 05, 2024
doi:

Summary

The present protocol describes an efficient multi-organ segmentation method called Swin-PSAxialNet, which has achieved excellent accuracy compared to previous segmentation methods. The key steps of this procedure include dataset collection, environment configuration, data preprocessing, model training and comparison, and ablation experiments.

Abstract

Abdominal multi-organ segmentation is one of the most important topics in the field of medical image analysis, and it plays an important role in supporting clinical workflows such as disease diagnosis and treatment planning. In this study, an efficient multi-organ segmentation method called Swin-PSAxialNet based on the nnU-Net architecture is proposed. It was designed specifically for the precise segmentation of 11 abdominal organs in CT images. The proposed network has made the following improvements compared to nnU-Net. Firstly, Space-to-depth (SPD) modules and parameter-shared axial attention (PSAA) feature extraction blocks were introduced, enhancing the capability of 3D image feature extraction. Secondly, a multi-scale image fusion approach was employed to capture detailed information and spatial features, improving the capability of extracting subtle features and edge features. Lastly, a parameter-sharing method was introduced to reduce the model’s computational cost and training speed. The proposed network achieves an average Dice coefficient of 0.93342 for the segmentation task involving 11 organs. Experimental results indicate the notable superiority of Swin-PSAxialNet over previous mainstream segmentation methods. The method shows excellent accuracy and low computational costs in segmenting major abdominal organs.

Introduction

Contemporary clinical intervention, including the diagnosis of diseases, the formulation of treatment plans, and the tracking of treatment outcomes, relies on the accurate segmentation of medical images1. However, the complex structural relationships among abdominal organs2make it a challenging task to achieve accurate segmentation of multiple abdominal organs3. Over the past few decades, the flourishing developments in medical imaging and computer vision have presented both new opportunities and challenges in the field of abdominal multi-organ segmentation. Advanced Magnetic Resonance Imaging (MRI)4 and Computed Tomography (CT) technology5 enable us to acquire high-resolution abdominal images. The precise segmentation of multiple organs from CT images holds significant clinical value for the assessment and treatment of vital organs such as the liver, kidneys, spleen, pancreas, etc.6,7,8,9,10 However, manual annotation of these anatomical structures, especially those requiring intervention from radiologists or radiation oncologists, is both time-consuming and susceptible to subjective influences11. Therefore, there is an urgent need to develop automated and accurate methods for abdominal multi-organ segmentation.

Previous research on image segmentation predominantly relied on Convolutional Neural Networks (CNNs), which improve segmentation efficiency by stacking layers and introducing ResNet12. In 2020, the Google research team introduced the Vision Transformer (VIT) model13, marking a pioneering instance of incorporating Transformer architecture into the traditional visual domain for a range of visual tasks14. While convolutional operations can only contemplate local feature information, the attention mechanism in Transformers enables the comprehensive consideration of global feature information.

Considering the superiority of Transformer-based architectures over traditional convolutional networks15, numerous research teams have undertaken extensive exploration in optimizing the synergy between the strengths of Transformers and convolutional networks16,17,18,19. Chen et al. introduced the TransUNet for medical image segmentation tasks16, which leverages Transformers to extract global features from images. Due to the high cost of network training and the failure to utilize the concept of feature extraction hierarchy, the advantages of Transformer have not been fully realized.

To address these issues, many researchers have started experimenting with incorporating Transformers as the backbone for training segmentation networks. Liu et al.17 introduced the Swin Transformer, which employed a hierarchical construction method for layered feature extraction. The concept of Windows Multi-Head Self-Attention (W-MSA) was proposed, significantly reducing computational cost, particularly in the presence of larger shallow-level feature maps. While this approach reduced computational requirements, it also isolated information transmission between different windows. To address this problem, the authors further introduced the concept of Shifted Windows Multi-Head Self-Attention (SW-MSA), enabling information propagation among adjacent windows. Building upon this methodology, Cao et al. formulated the Swin-UNet18, replacing the 2D convolutions in U-Net with Swin modules and incorporating W-MSA and SW-MSA into the encoding and decoding processes, achieving commendable segmentation outcomes.

Conversely, Zhou et al. highlighted that the advantage of conv operation could not be ignored when processing high-resolution images19. Their proposed nnFormer employs a self-attention computation method based on local three-dimensional image blocks, constituting a Transformer model characterized by a cross-shaped structure. The utilization of attention based on local three-dimensional blocks significantly reduced the training load on the network.

Given the problems with the above study, an efficient hybrid hierarchical structure for 3D medical image segmentation, termed Swin-PSAxialNet, is proposed. This method incorporates a downsampling block, Space-to-depth (SPD)20 block, capable of extracting global information21. Additionally, it adds a parameter shared axial attention (PSAA) module, which reduces the learning parameter count from quadratic to linear and will have a good effect on the accuracy of network training and the complexity of training models22.

Swin-PSAxialNet network
The overall architecture of the network adopts the U-shaped structure of nnU-Net23, consisting of encoder and decoder structures. These structures engage in local feature extraction and the concatenation of features from large and small-scale images, as illustrated in Figure 1.

Figure 1
Figure 1: Swin-PSAxialNet schematic diagram of network architecture. Please click here to view a larger version of this figure.

In the encoder structure, the traditional Conv block is combined with the SPD block20 to form a downsampling volume. The first layer of the encoder incorporates Patch Embedding, a module that partitions the 3D data into 3D patches Equation 1, (P1, P2, P3) represents non-overlapping patches in this context, Equation 2 signifies the sequence length of 3D patches. Following the embedding layer, the next step involves a non-overlapping convolutional downsampling unit comprising both a convolutional block and an SPD block. In this setup, the convolutional block has a stride set to 1, and the SPD block is employed for image scaling, leading to a fourfold reduction in resolution and a twofold increase in channels.

In the decoder structure, each upsample block after the Bottleneck Feature layer consists of a combination of an upsampling block and a PSAA block. The resolution of the feature map is increased twofold, and the channel count is halved between each pair of decoder stages. To restore spatial information and enhance feature representation, feature fusion between large and small-scale images is performed between the upsampling blocks. Ultimately, the upsampling results are fed into the Head layer to restore the original image size, with an output size of (H × W × D × C, C = 3).

SPD block architecture
In traditional methods, the downsampling section employs a single stride with a step size of 2. This involves convolutional pooling at local positions in the image, limiting the receptive field, and confining the model to the extraction of features from small image patches. This method utilizes the SPD block, which finely divides the original image into three dimensions. The original 3D image is evenly segmented along the x, y, and z axes, resulting in four sub-volume bodies. (Figure 2) Subsequently, the four volumes are concatenated through "cat" operation, and the resulting image undergoes a 1 × 1 × 1 convolution to obtain the down-sampled image20.

Figure 2
Figure 2: SPD block diagram. Please click here to view a larger version of this figure.

PSAA block architecture
In contrast to traditional CNN networks, the proposed PSAA block is more effective in conducting global information focus and more efficient in network learning and training. This enables the capture of richer images and spatial features. The PSAA block includes axial attention learning based on params sharing in three dimensions: Height, Width, and Depth. In comparison to the conventional attention mechanism that performs attention learning for each pixel in the image, this method independently conducts attention learning for each of the three dimensions, reducing the complexity of self-attention from quadratic to linear. Moreover, a learnable keys-queries parameter sharing mechanism is employed, enabling the network to perform attention mechanism operations in parallel across the three dimensions, resulting in faster, superior, and more effective feature representation.

Protocol

The present protocol was approved by the Ethics Committee of Nantong University. It involves the intelligent assessment and research of acquired non-invasive or minimally invasive multimodal data, including human medical images, limb movements, and vascular imaging, utilizing artificial intelligence technology. Figure 3 depicts the overall flowchart of multi-organ segmentation. All the necessary weblinks are provided in the Table of Materials.

Figure 3
Figure 3: Overall flowchart of the multi-organ segmentation. Please click here to view a larger version of this figure.

1. Dataset collection

  1. Download the AMOS2022 dataset24 (see Table of Materials).
  2. Select 160 images as the training dataset and 40 images as the testing dataset.
  3. Terminally execute mkdir ~/PaddleSeg/contrib/MedicalSeg/data/raw_data.
  4. Open the path of raw_data.
  5. Place the dataset in the raw_data directory, and ensure correct correspondence with the file names in the "imagesTr" and "labelsTr" directories.
  6. Create dataset.json file.
  7. Add dataset information such as numTraining, numTest, and labels name into the "dataset.json" file.
    NOTE: The dataset for this study consists of 200 high-quality CT datasets selected from the AMOS2022 official 500 CT dataset for training and testing purposes. The dataset encompasses 11 organs, including the spleen, right kidney, left kidney, gallbladder, esophagus, liver, stomach, aorta, inferior vena cava, pancreas, and bladder.

2. Environment configuration

  1. Install the Paddlepaddle environment (see Table of Materials) and the corresponding CUDA version25.
    NOTE: Utilizing Paddlepaddle version 2.5.0 and CUDA version 11.7 is recommended.
  2. Git clone the PaddleSeg repository25.
  3. Terminally execute cd ~/PaddleSeg/contrib/MedicalSeg/.
  4. Execute pip install -r requirements.txt.
  5. Execute pip install medpy.

3. Data preprocessing

  1. Modify the "yml" file, including parameters such as data_root, batch_size, model, train_dataset, val_dataset, optimizer, lr_scheduler, and loss.
  2. Specify the required parameters for execution, including "-config, –log_iters, –precision, –nnunet, –save_dir, –save_interval, etc.
    NOTE: The original "raw_data" is transformed into three types of data files: "Cropped," "Decathlon," and "Preprocessed," catering to different training and evaluation requirements to enhance model performance and training effectiveness.

4. Model training and comparison

NOTE: As a widely used baseline in the field of image segmentation, nnU-Net23 serves as a baseline model in the study. The specific model comparison process is as follows.

  1. Train nnU-Net
    1. Configure with a batch size of 2 and 80,000 iterations, log every 20 iterations, and utilize fp16 precision.
    2. Set the save_dir as the "output" folder within the working directory, with a save interval set at 1,000 iterations.
    3. Add parameter –use_vdl for VisualDL parameter visualization.
    4. Run train.py.
    5. Create four yml files, specify folds 1-4, and train the model with five-fold cross-validation.
  2. Train Swin-UNet
    1. Maintain consistency with the previously established experimental conditions.
    2. Remove parameter –nnunet and use –swinunet.
    3. Run the train.py file.
    4. Train folds 1-4 individually and then subject to ensemble processing.
  3. Train TransUNet
    1. Follow the procedures outlined in step 4.1.
    2. Open yml file.
    3. Modify type: nnunet to type: transunet.
    4. Modify parameter –nnunet to –transunet.
    5. Run the train.py file.
    6. Train folds 1-4 individually and process ensemble processing.
  4. Train UNETR
    1. Maintain consistent experimental variables.
    2. Repeat the procedures outlined previously like Swin-UNet and TransUNet (step 4.2 and step 4.3).
    3. Run train.py.
    4. Train folds 1-4 individually and process ensemble processing.
  5. Train nnFormer
    1. Modify the parameters like Swin-UNet and TransUNet.
    2. Run single fold train.py.
    3. Run ensemble processing.

5. Ablation experiment

  1. PSAA block for the entire network
    1. Keep variables consistent with step 4.1.
    2. Set volume_shape = (64,192,192) and Axial_attention = True.
    3. Run single fold train.py.
    4. Process ensemble processing across five folds.
  2. SPD block for the entire network
    1. Keep variables consistent with step 4.1.
    2. Set SPD_enable = True and Axial_attention = False.
    3. Run single fold train.py.
    4. Process ensemble processing across five folds.
  3. PSAA block for upsampling
    NOTE: Do not use SPD block for downsampling.
    1. Use the PSAA block for upsampling.
    2. Keep variables consistent with step 4.1.
    3. Set Axial_attention = False in the downsampling part and SPD_enable = False.
    4. Run single fold train.py.
    5. Process five folds ensemble processing.
  4. PSAA block for upsampling
    NOTE: Use SPD block for downsampling.
    1. Use the SPD block for downsampling and the PSAA block for upsampling.
    2. Keep variables consistent with step 4.1.
    3. Set Axial_attention = False in the downsampling part and SPD_enable = True.
    4. Run single fold train.py.
    5. Process five folds ensemble processing.
    6. Set the number of iterations to 2,00,000 because of the superior results.

Representative Results

This protocol employs two metrics to evaluate the model: Dice Similarity Score (DSC) and 95% Hausdorff Distance (HD95). DSC measures the overlap between voxel segmentation predictions and ground truth, while 95% HD assesses the overlap between voxel segmentation prediction boundaries and ground truth, filtering out 5% of outliers. The definition of DSC26 is as follows:

Equation 3      (1)

Where T and P represent the ground truth values and predicted values for all voxels, respectively. The definition of 95% HD26 is as follows:

Equation 4      (2)

Where dT,P, is the 95th percentile distance of the maximum HD95 between the ground truth and predicted voxels, and dT,P, represents the 95th percentile distance of the maximum HD95 between the predicted voxels and ground truth.

Swin-PSAxialNet is compared with various mainstream multi-organ segmentation algorithms in Table 1. The results demonstrate improved segmentation performance for abdominal organs in the proposed algorithm. The Dice accuracy of the segmentation method surpasses the average Dice value of other methods by 2-3 percentage points and even outperforms nnFormer by 1 percentage point. Swin-PSAxialNet also exhibits high accuracy in boundary segmentation, with an average HD95 coefficient of 5.63, the lowest among all compared algorithms.

Methods Lkn Spl Liv Pan Aro Bla Sto Gbd Pos Eso Rkn Avg
DSC HD95
nnU-Net 96.3 96.8 97.21 85.2 95.22 86.57 91.47 84.7 91.79 82.5 95.44 91.2 7.13
Swin-Unet 96.03 96.68 97.22 85.74 95.3 87.82 91.77 84.64 91.67 83.26 96.14 91.48 9.68
TransUNet 96.42 96.74 97.05 85.09 94.86 85.2 91.81 83.33 91.52 82.22 96.12 90.94 13.77
UNETR 95.94 96.48 96.85 84.14 94.62 85.01 90.18 81.16 90.79 80.34 95.15 90.06 12.92
nnFormer 96.59 97.22 97.33 86.79 95.31 89.74 92.82 84.88 92.43 84.24 96.32 92.15 6.59
Swin-PSAxialNet 96.85 97.67 97.64 88.39 95.97 92.03 94.4 87.78 93.15 86.24 96.59 93.34 5.63

Table 1: Comparison of mainstream segmentation algorithms. The abbreviations stand for the following organs: Lkn for left kidney, Spl for spleen, Liv for liver, Pan for pancreas, Aro for aorta, Bla for bladder, Sto for stomach, Gbd for gall bladder, Pos for postcava, Eso for esophagus, Rkn for right kidney, and DSC for the average Dice coefficient, HD95 for the average 95% Hausdorff Distance.

Moreover, the introduction of a parameter-sharing mechanism and improvements to the transformer structure provide significant advantages in both model complexity and inference speed (Table 2, Figure 4, and Figure 5).

Models Params (M) FLOPs (G) Inference Time (s)
nn-Unet 17.96 398.54 9.07
UNETR 89.58 39.67 12.51
nnFormer 134.78 152.43 11.24
Swin-PSAxialNet 37.64 43.15 10.31

Table 2: Comparison table of network model indicators.

Figure 4
Figure 4: Dice variation curve of model training network. (A) The current network. (B) nnFormer. (C) nnU-Net. (D) UNETR. Please click here to view a larger version of this figure.

Figure 5
Figure 5: Model training results comparison. The figure highlights the results of the current network, the results of nnFormer, the results of nnU-Net, and the results of UNETR. Please click here to view a larger version of this figure.

This study conducted comprehensive ablation experiments on the algorithm network, using the PSAA block and SPD block as variables, and comparing single-fold training with five-fold training. The Dice coefficients from the experimental results are presented in Table 3. These results demonstrate that utilizing the SPD block for downsampling and the PSAA block for upsampling effectively enhances segmentation performance compared to the unmodified nnU-Net (Figure 6).

Models PSAA Block for the entire network SPD Block for the entire network PSAA Block only for upsampling PSAA Block for upsampling, use SPD
Single-fold training 92.17 90.51 92.24 92.39
Five-fold training 92.82 91.26 92.79 93.34

Table 3: The results of the ablation experiment of Swin-PSAxialNet. The results of the ablation experiment of Swin PSAxialNet are presented. In the ablation experiment, the use of PSAA Block and SPD Block were compared, and the training results of DSC are shown in the table.

Figure 6
Figure 6: The segmentation results. Please click here to view a larger version of this figure.

Discussion

The segmentation of abdominal organs is a complicated work. Compared to other internal structures of the human body, such as the brain or heart, segmenting abdominal organs seems more challenging because of the low contrast and large shape changes in CT images27,28. Swin-PSAxialNet is proposed here to solve this difficult problem.

In the data collection step, this study downloaded 200 images from the AMOS2022 official website24. In the environment configuration step, the experimental setup included the Paddlepaddle environment along with its corresponding CUDA version. During the data preprocessing step, random cropping and random flipping were applied to enhance the data robustness. In the model training and comparison steps, based on the baseline model nnU-Net, the experiment added the SPD Block for downsampling and the PSAA feature extraction block for upsampling, achieving excellent segmentation results. Swin-PSAxialNet was compared with various mainstream segmentation models and demonstrated promising results on the public dataset AMOS2022. It achieved higher segmentation accuracy than the other segmentation methods mentioned in this study16,17,18,19. In the ablation experiment section, the usage of the PSAA block and SPD block was varied to find the optimal segmentation network. The results of the ablation experiment indicated that using the PSAA module only for upsampling achieves one percentage point higher than using it for the entire network.

The innovations of this study are as follows: Firstly, the SPD module is incorporated into the network to constitute the downsampling unit alongside the Conv block, thereby enhancing the network's feature extraction capabilities. Secondly, the introduction of the PSAA feature extraction block effectively reduces network complexity while improving training accuracy. Lastly, the introduction of a parameter-sharing mechanism significantly reduces the complexity of the training model.

This study has limitations in the following aspects: It has only been applied to the AMOS2022 dataset, and results have not been validated on other datasets. Again, although nnU-Net has achieved high performance in abdominal organ segmentation tasks after a series of optimizations as an automated segmentation framework, some parameters still have a certain degree of subjectivity, such as initial learning rate, number of iterations, network architecture, etc. Future research will explore methods for the automatic calculation of these parameters29.

Swin-PSAxialNet has achieved excellent segmentation results in abdominal organ segmentation tasks, demonstrating strong generalization capabilities and stability. In future research, additional datasets will be employed to investigate the algorithm's generalization capabilities further and make adjustments to the model. Moreover, the algorithm's structure can undergo further optimization to enhance the overall accuracy and robustness of the network. The aim is to integrate the network with multimodal capabilities to transform it into a medical multimodal model.

Disclosures

The authors have nothing to disclose.

Acknowledgements

This study was supported by the '333' Engineering Project of Jiangsu Province ([2022]21-003), the Wuxi Health Commission General Program (M202205), and the Wuxi Science and Technology Development Fund (Y20212002-1), whose contributions have been invaluable to the success of this work." The authors thank all the research assistants and study participants for their support.

Materials

AMOS2022 dataset None None Datasets for network training and testing. The weblink is: https://pan.baidu.com/s/1x2ZW5FiZtVap0er55Wk4VQ?pwd=xhpb
ASUS mainframe ASUS https://www.asusparts.eu/en/asus-13020-01910200
CUDA version 11.7 NVIDIA https://developer.nvidia.com/cuda-11-7-0-download-archive
NVIDIA GeForce RTX 3090 NVIDIA https://www.nvidia.com/en-in/geforce/graphics-cards/30-series/rtx-3090-3090ti/
Paddlepaddle environment  Baidu None Environmental preparation for network training. The weblink is: https://www.paddlepaddle.org.cn/
PaddleSeg Baidu None The baseline we use: https://github.com/PaddlePaddle/PaddleSeg

References

  1. Liu, X., Song, L., Liu, S., Zhang, Y. A review of deep-learning-based medical image segmentation methods. Sustain. 13 (3), 1224 (2021).
  2. Popilock, R., Sandrasagaren, K., Harris, L., Kaser, K. A. CT artifact recognition for the nuclear technologist. J Nucl Med Technol. 36 (2), 79-81 (2008).
  3. Rehman, A., Khan, F. G. A deep learning based review on abdominal images. Multimed Tools Appl. 80, 30321-30352 (2021).
  4. Schenck, J. F. The role of magnetic susceptibility in magnetic resonance imaging: Mri magnetic compatibility of the first and second kinds. Med Phys. 23 (6), 815-850 (1996).
  5. Palmisano, A., et al. Myocardial late contrast enhancement ct in troponin-positive acute chest pain syndrome. Radiol. 302 (3), 545-553 (2022).
  6. Chen, E. -. L., Chung, P. -. C., Chen, C. -. L., Tsai, H. -. M., Chang, C. -. I. An automatic diagnostic system for CT liver image classification. IEEE Trans Biomed Eng. 45 (6), 783-794 (1998).
  7. Sagel, S. S., Stanley, R. J., Levitt, R. G., Geisse, G. J. R. Computed tomography of the kidney. Radiol. 124 (2), 359-370 (1977).
  8. Freeman, J. L., Jafri, S., Roberts, J. L., Mezwa, D. G., Shirkhoda, A. CT of congenital and acquired abnormalities of the spleen. Radiographics. 13 (3), 597-610 (1993).
  9. Almeida, R. R., et al. Advances in pancreatic CT imaging. AJR Am J Roentgenol. 211 (1), 52-66 (2018).
  10. Eisen, G. M., et al. Guidelines for credentialing and granting privileges for endoscopic ultrasound. Gastrointest Endosc. 54 (6), 811-814 (2001).
  11. Sykes, J. Reflections on the current status of commercial automated segmentation systems in clinical practice. J Med Radiat Sci. 61 (3), 131-134 (2014).
  12. He, K., Zhang, X., Ren, S., Sun, J. Deep residual learning for image recognition. 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). , 770-778 (2016).
  13. Dosovitskiy, A., et al. An image is worth 16×16 words: Transformers for image recognition at scale. arXiv. arXiv. , 11929 (2020).
  14. Ramachandran, P., et al. Stand-alone self-attention in vision models. Adv Neural Inf Process Syst. 32, (2019).
  15. Vaswani, A., et al. Attention is all you need. Adv Neural Inf Process Syst. 30, (2017).
  16. Chen, J., et al. TransUNET: Transformers make strong encoders for medical image segmentation. arXiv. , 04396 (2021).
  17. Liu, Z., et al. Swin transformer: Hierarchical vision transformer using shifted windows. Proc IEEE Int Conf Comput Vis. , 10012-10022 (2021).
  18. Cao, H., et al. Swin-UNET: Unet-like pure transformer for medical image segmentation. Comput Vis ECCV. , 205-218 (2022).
  19. Zhou, H. -. Y., et al. nnformer: Interleaved transformer for volumetric segmentation. arXiv. arXiv. , 03201 (2021).
  20. Sunkara, R., Luo, T. No more strided convolutions or pooling: A new CNN building block for low-resolution images and small objects. Mach Learn Knowl Discov Databases. , 443-459 (2023).
  21. Çiçek, &. #. 2. 1. 4. ;., Abdulkadir, A., Lienkamp, S. S., Brox, T., Ronneberger, O. 3D U-Net: learning dense volumetric segmentation from sparse annotation. Med Comput Vis Bayesian Graph Models Biomed Imaging. 2016, 424-432 (2016).
  22. Kim, H., et al. Abdominal multi-organ auto-segmentation using 3D-patch-based deep convolutional neural network. Scientific reports. 10 (1), 6204 (2020).
  23. Isensee, F., et al. nnu-net: Self-adapting framework for u-net-based medical image segmentation. arXiv. , 10486 (2018).
  24. Ji, Y., et al. A large-scale abdominal multi-organ benchmark for versatile medical image segmentation. Adv Neural Inf Process Syst. 35, 36722-36732 (2022).
  25. Liu, Y., et al. Paddleseg: A high-efficient development toolkit for image segmentation. arXiv. , 06175 (2021).
  26. Hatamizadeh, A., et al. UNETR: Transformers for 3D medical image segmentation. Proc IEEE Int Conf Comput Vis. , 574-584 (2022).
  27. Zhou, H., Zeng, D., Bian, Z., Ma, J. A semi-supervised network-based tissue-aware contrast enhancement method for ct images. Nan Fang Yi Ke Da Xue Xue Bao. 43 (6), 985-993 (2023).
  28. Ma, J., et al. Abdomenct-1k: Is abdominal organ segmentation a solved problem. IEEE Trans Pattern Anal Mach Intell. 44 (10), 6695-6714 (2021).
  29. Galván, E., Mooney, P. Neuroevolution in deep neural networks: Current trends and future challenges. IEEE Trans Artif Intell. 2 (6), 476-493 (2021).
This article has been published
Video Coming Soon
Keep me updated:

.

Cite This Article
Ma, L., Zhang, D., Wang, Z., Han, J., Wang, X., Zhou, X., Yang, W., Lu, P. Swin-PSAxialNet: An Efficient Multi-Organ Segmentation Technique. J. Vis. Exp. (209), e66459, doi:10.3791/66459 (2024).

View Video