An End-to-End Data-Adaptive Pancreas Segmentation System with an Image Quality Control Toolbox

,


Introduction
Pancreatic cancer is a malignant tumor and is recognized with a low survival rate, and pancreatic diseases are characterized by rapid occurrence and continuous progression [1,2].Even among all types of cancer, it is one of the most dangerous and deadly.It is estimated that, in 2021, statistically, about 60,430 new cases of pancreatic cancer would be diagnosed in the US, and 48,220 people would die from this disease [3].Because pancreatic cancer is difcult to diagnose in its early stage, the rate of diagnosis is almost as high as the mortality rate.If not treated timely, a signifcant portion of patients with pancreas disease would be diagnosed with metastatic symptoms [4].For pancreatic cancer, the common treatment options are surgical resection, neoadjuvant radiotherapy, radiotherapy, and chemotherapy [5,6], of which surgery is acknowledged to achieve a better prognosis but with somewhat invasiveness.Regardless of the chosen treatment option, precise localization and segmentation of the pancreas are crucial for physicians to diagnose and assess the patient's condition early in the treatment phase [7].
Computed tomography (CT) and magnetic resonance imaging (MRI) are particularly important examination procedures and tools in diagnosing pancreatic diseases [8].CT, especially contrast-enhanced CT, is the frst choice for pancreatic examinations in hospitals and has the advantages of more rapid imaging and clearer strips than MRI.However, the irregularity and variability of the pancreas in morphology and low contrast in its surrounding tissues lead to a high demand for experience and prior knowledge of radiologists in the early diagnosis of pancreatic diseases [9][10][11][12].Nowadays, precision medicine requires the clinical process upgrading from qualitative observation to quantitative analysis and diagnosis [13,14].Terefore, accurate automatic segmentation of pancreatic tissues in CT images can greatly accelerate the early diagnosis process for physicians and assist in the appropriate treatment process.
In recent years, with technological breakthroughs in deep learning, computer-aided pancreas segmentation has yielded a series of promising methodological studies [15][16][17][18][19][20][21][22][23][24][25][26].Limited by the small object characteristics of the pancreas, several studies have employed the two-stage cascade approach for semantic segmentation [7,26].Tese methods frst locate the pancreas in the entire CT sequence with bounding boxes and then perform pixel-level segmentation of the pancreas on the basis of box localization.nnU-Net adopts a two-stage strategy to perform segmentation on pancreatic healthy tissues and lesion tissues autonomously by confguring itself and achieves the best accuracy in the Medical Segmentation Decathlon challenge [27].Zhu et al. [15], Fu et al. [17], and Oktay and Chen [18] introduced attention modules and spatial information to a three-dimensional (3D) segmentation structure to capture the consistency information of pancreatic tissue in CT images.More recently, a series of AI-based semantic segmentation or object detection methods has been applied to clinical practice for quantitative data measurement and morphometric analysis [28][29][30].For example, pancreatic fstula prediction after pancreaticoduodenal surgery is based on quantitative pancreatic volume measurements in CT images, and gallbladder resection is guided by fat pairs from quantitative measurements of the pancreas in CT images.
However, the more challenging issue in the clinical application of computer-aided pancreas segmentation is that the heterogeneity of across domain images leads to poor generalization of models [31].Tis is manifested by the fact that supervised models trained on a single dataset, even if trained with accurate expert manual annotations, are subject to signifcant model inference errors once they are deployed to other medical centers [32][33][34][35][36][37][38].Normalized images are essential for good performance of deep learning models.Te variation in medical images in terms of populations, scanning devices, scanning parameters, or imaging protocols will lead to varying quality [33,39,40].Tis heterogeneity exists in both CT and MRI images [33,34,36].Besides, pancreatic disease and pancreatic cancer dramatically change the morphology of the pancreatic tissue, specifcally demonstrated by difuse enlargement, inhomogeneous density, and ambiguous boundaries, which make the data quality inconsistent among medical centers [41].Liu et al. [33] and Wang et al. [36] illustrated the heterogeneity across medical sites in terms of the patient cohort and image quality.Several research studies on medical image segmentation have previously demonstrated signifcant performance degradation of single-site models when deployed to other sites.Lerousseau and Xiao [42] proposed a new weakly supervised multi-instance learning method as a tool for pancreas tumor segmentation which achieved promising performance by taking full advantage of less annotated data at the pixel level.However, this method yielded about 15-26% performance degradation when it was tested for other publicly available datasets.Obviously, interdomain generalization greatly limits the clinical application of automatic pancreas segmentation models.In the feld of natural images, several studies have addressed this issue by using transfer learning or federation learning algorithms [43][44][45][46][47].However, the efectiveness of approaches based on natural images is unsatisfactory or even worse for CT images due to the gap between the two types of images [48].Tere is one research study on semisupervised segmentation pancreas tasks on the NIH-TCIA dataset and achieved a dice similarity coefcient score of 78.27% by using a portion of annotated images [49].Moreover, to date, few research studies related to unsupervised segmentation or domain adaptation of the pancreas have been conducted.
Terefore, to tackle this serious real-clinical problem, we propose an end-to-end data-adaptive pancreas segmentation system with an image quality control toolbox in this paper.Te system focuses on the pancreas segmentation model construction with heterogeneous cross-domain CT images in the presence of insufcient annotations in medical centers.Te main contributions of this paper are summarized as follows: (1) Te system utilizes adversarial learning to construct data-adaptive segmentation models with the assistance of domain discriminators.Besides, we employ a collaboration center to perform feature-level transfer learning without data sharing across domains.(2) A multifunctional image quality control toolbox is designed to standardize the quality of images from various medical centers in terms of the intensity range, feld of view, region of interest, etc. (3) Te system works in an end-to-end mode, which only requires physicians to select images, set up personalized parameters, and then wait for automatic model construction and inferences on unlabeled data.(4) Te system ofers a variety of pancreas-related features including textual information and imaging results, which can assist physicians in quantitative and precise clinical diagnosis of the pancreas.
We experimentally demonstrate the efectiveness of the system on a public dataset and an in-hospital dataset and validate the robustness of the system on a small dataset.Te data-adaptive pancreas segmentation system we developed is able to diagnose a larger number of people quickly and efectively in the clinical practice and to obtain meaningful pancreas segmentation results.

Methods
Accurate segmentation of pancreatic tissue is an essential stage in clinical diagnosis of pancreatic diseases.Te variability of CT images from diferent medical centers afects the generalizability of automatic pancreatic segmentation tools.Subtle diferences in image features result in sudden decreases in segmentation accuracy for deep learning models.To address such a problem, improvements can be carried out in terms of both data alignment and segmentation model construction methods, respectively.On the one hand, some image quality control approaches for pancreas CT images are used to process diferent source images.On the other hand, the automatic pancreatic segmentation model should have the ability to adapt to variations in the data domain in terms of the methodology and system design.To address these issues, a novel end-to-end data-adaptive segmentation system for the pancreas with an image control toolbox is proposed for pancreatic data quality normalization and assisting in pancreas segmentation model generalization.In this chapter, the overall framework of the system and the construction and functions of each module are shown.

2.1.
Framework.Te overall framework of the proposed data-adaptive pancreas segmentation system embedded with a data quality control toolbox is shown in Figure 1.Te system consists of two parts: the local clients of medical centers and a collaborative center on the cloud server.Te primary tasks of the local client are integrating and processing data, constructing segmentation models, and visualizing segmentation results.Te collaborative center is mainly responsible for transfer learning of image features among multiple medical centers.
Te medical center client consists of four modules, namely, data organization module, image quality control module, segmentation module, and visualization module.Te data organization module is mainly operated by imaging physicians to establish the patients' cohort to study and extract the pancreas CT images from the in-hospital database.Te image quality control module performs standardized preprocessing on the previously selected pancreas CT images in order to normalize the data and reduce the variation of images from diferent sources.Te segmentation module is mainly equipped with the predefned semantic segmentation model in the system to train annotated data and cooperates with the collaborative center to assist in the construction of the data-adaptive segmentation model for unannotated data.Besides, the segmentation module uses the well-trained model to predict the mask of pancreas tissues in input images.Te visualization module presents multidimensional results of the segmented pancreas tissues in terms of visualized images and structured text.
Te collaborative center mainly contains feature discriminators.Te feature transmission and learning process between the medical center and the collaborative center is encapsulated as the transfer learning module.Te transfer learning module mainly accomplishes adversarial learning between image features from multiple centers.Te feature discriminator optimizes the segmentation network of unlabeled images by balancing the Nash equilibrium of the domain classifcation loss and the segmentation loss.Tus, the segmentation network is adaptive to the new images without the need of annotation.

Image Quality Control
Toolbox.Te image quality control toolset provides a variety of image processing means to standardize various qualities of images from multiple sources, mainly including the intensity value cutof, rotation augmentation, and superresolution reconstruction.Te CT intensity of abdominal organs is in the range of (−160, 240) HU, and the range for the pancreas is kept in (−100, 240) HU.Tis scale preserves pancreatic tissue features and removes background information.Rotational augmentation refers to amplify CT images by rotating them at the axial plane with degrees in the range of ( ± 5 °, ± 10 °).Tis operation is not performed in the ordinary sense of augmenting data to improve model performance but rather to attenuate the angle bias introduced by the feld of view or body position during scanning so as to eliminate heterogeneity.Superresolution reconstruction could efectively improve the image quality, thus reducing the quality inconsistency caused by scanning devices, imaging protocols, slice thickness, and so on.In this study, it is concerned that the efective abdominal region in CT images fuctuates because of diferent scanning felds of view, so the frst step in superpixel reconstruction is framing out the region of interest.Pancreatic CT images were binarized to measure image region properties.Ten, the maximum connected region containing the region of interest was found by the region-growing algorithm.Te rectangular area bounded by the diagonal vertices of the maximum connected region in the image is considered the valid abdominal area.Ten, the truncated 3D volume is interpolated in 3D cubic interpolation (system default settings) and reconstructed to the same resolution to feed into the network.Te default size is 512 × 512 resolution in an axial plane and 1 mm thickness in a sagittal direction.Furthermore, the system provides various image interpolation methods to system users.Te reconstruction algorithm library includes nearest neighbor interpolation, bilinear interpolation, Lanczos interpolation, and bicubic interpolation to support multiple requirements for medical studies.

Segmentation Module.
Te segmentation module is embedded with a deep learning semantic segmentation network applicable to the pancreas segmentation task.Tis module is mainly responsible for the training of labeled data models, the construction of adaptive unlabeled data models in cooperation with discriminators, and the prediction of pancreas masks for input samples.
In this paper, a 3D ResUnet structure integrated with an attention mechanism is designed as the backbone model of the segmentation module.U-Net [50] is a widely used semantic segmentation network, which has the advantages of the small amount of data needed, high data utilization, and Journal of Healthcare Engineering short training periods.On the basis of U-Net, the ResUnet model designed in this research introduces a deep residual structure which enables weighted interactions of image features at diferent scales, thus improving the segmentation performance of small targets like pancreatic tissue.Te residual connection simplifes training and eliminates the degradation of the transmission of information among low and high scales, resulting in fewer parameters in networks.Besides, an attention mechanism is introduced into 3D spatial channels to enlarge the model capacity and enhance the network feature representation capability.Te attention structure is inspired by the design paradigm of the squeeze and excitation (SE) network [50] and extended to 3D to handle 3D image representations.Te specifc structure is represented as follows: the input 3D feature map is squeezed based on pooling operations to obtain the feature vector in channels and feature activation is performed by the multilayer perception (MLP) operation of fully connected layers.After obtaining the weighting coefcients of the important channels in the feature map, the obtained coefcients are then linearly weighted into the input 3D feature map in a dot-product manner.
Te feature maps at four scales (32 × 32, 64 × 64, 128 × 128, and 256 × 256) of the decoder are transmitted to the collaborative center for adversarial learning.Te feature maps at various scales contain not only the shallow boundary information but also the precise pancreas target information so as to ensure the efectiveness of domain adaptation.Moreover, to more strictly constrain the segmentation task on small object tasks, the model employs a linear combination of the dice loss and cross-entropy loss function as optimization criteria.Te loss function is formulated as where α and λ are the linear coefcients that range from 0 to 1. Tese parameters can be customized by system users according to the needs of practical research purposes.In the experiments of this paper, α � 0.8 and λ � 0.5.

Feature Transfer Module.
Te feature transfer module is mainly responsible for adversarial learning of image features among centers.Adversarial learning is performed by the discriminator in the collaborative center.Te multiscale image features generated by the segmentation module of the medical center are transferred to the collaborative center as the four inputs of the discriminator.Each of the input feature maps sequentially undergoes a 3D convolutional layer and an activation function, where the step size of the 3D convolutional layer is set to 2. Terefore, the spatial scale of image features is decreased by half and thus concatenated with the next-scale feature map and fed to the next layer.Upon weighted feature fusion of multiple scales, the features are fed into the average pooling layer and the fully connected layer in turn to obtain the fnal domain classifcation results.Te workfow of feature transfer is shown in Figure 2. Te labeled source-domain data are denoted as x, and the labeled dataset is defned as S(x); the unlabeled target-domain data are denoted as z, and the unlabeled dataset is defned as T(z).At frst, the labeled dataset is trained with annotated images by the segmentation module to obtain an initial pancreas segmentation model.Ten, x and z are fed into the segmentation model in pairs to generate multiscale feature maps F S (x) and F T (z), respectively, which are then transmitted to the collaborative center.In the collaborative center, the feature maps from the two branches are trained by the discriminator H(•) for domain identifcation.Given feature maps from the source domain with label 1 and target-domain feature maps with label 0, then the optimization condition of the discriminator is to seek the weights that maximize the diference of domain features, and the loss function is defned as where E x∼S(x) and E z∼T(z) denote the mathematical expectations for the two parts.Meanwhile, the feature maps from the target-domain images are labeled with a change to 1 and are fed into the discriminator in a single branch.Te discriminator backpropagates the source-domain features with respect to the labels, thus amplifying such common image features and updating the pancreas segmentation model for the target domain.In the subsequent feature transfer process, the discriminator and the target domain segmentation model are continuously updated and frozen alternately as in the above step, thus searching for the Nash equilibrium of the two optimization functions in the adversarial process.Te fnal optimization purpose of adversarial learning is to discover general image features between domains and utilize them to guide the segmentation of the pancreas.

Visualization Module.
When the pancreas segmentation model for unannotated images is constructed by the feature transfer module, doctors could select CT images to study from the hospital local database and perform standardized preprocessing for images with custom parameters in the image quality control toolbox.Subsequently, the visualization module performs postprocessing on the pancreas mask output by the segmentation module and displays CT images and corresponding segmentation results.Te presentation includes visualized images and structured text information.
Segmentation is essentially the prediction of whether each pixel in the image is a target foreground or background, so there are usually a number of isolated and noisy points in the segmentation mask.In this study, a conditional random feld (CRF) model and a hole-flling algorithm are used as postprocessing operations in visualization modules to further optimize the segmentation mask to eliminate the anomalous structure and smooth the boundaries.In addition, the module ofers a parallel visualization comparison between segmentation results and physician annotations in each slice-of-interest so that the physician can check annotations.Te visualized image results include original CT images and segmented pancreas masks in the form of 2D slices and 3D volumes.Besides, the 3D reconstruction of surface distances between masks and annotations are also displayed.Te module with visualized images also provides support for window dragging, rotating, zooming, and other operations to display images more comprehensively.Structured text information covers volume, size, and occupancy depth of pancreas tissues, surface distances between masks and annotations, etc.

Experimental Results
2.6.1.Datasets.Te NIH-TCIA dataset [23] is employed as the labeled source-domain dataset in this study.Te NIH-TCIA dataset is collected by the National Institutes of Health Clinical Center, which is currently the authoritative and commonly adopted public dataset for pancreas segmentation.We employed 70 cases of in-hospital CT images, collected from the First Afliated Hospital of the Zhejiang University School of Medicine, as the unlabeled target domain dataset, noted as the Zheyi dataset.Te annotations of Zheyi images were all manually outlined and cross-validated   by professional physicians.Notably, the annotations of the Zheyi dataset are not only used for the training process of the system but are also used for evaluation.Te NIH-TCIA dataset contains 82 enhanced CT sequences, and the Zheyi dataset includes 70 instances.Te axial resolution of CT images in the two datasets is 512 × 512.Slice thickness of CT images in the NIH-TCIA dataset ranges from 0.5 mm to 1 mm, and the number of slices is in the range of 181 to 466, which are relatively high-resolution CT images.In contrast, the CT images of the Zheyi dataset range from 2.5 mm to 3 mm in layer thickness, and the slice number varies from 76 to 107.

Experimental Details.
Te system previously presented is expected to be able to cope with new sources of unlabeled CT images to construct an efective segmentation network.Taking NIH-TCIA as the labeled data center and Zheyi as the unlabeled data center, we validated the data adaptability and pancreas segmentation performance of the designed system for Zheyi images.We utilized PyTorch [51] in the Python environment to implement models and algorithms.Te experiments were carried out with an NVI-DIA TITAN V GPU with 12 GB memory and 2 Intel Xeon E5-2630 v4 CPUs.To guarantee the stability and reliability of the system, all trials are performed with a 5-fold crossvalidation approach.Te execution time for necessary steps in the dataadaptive chain is listed in Table 1.Tese time data are statistically derived from the mean time of all sequences of the NIH-TCIA dataset.As can be seen in the table, the time for the data quality control module is mainly distributed over superresolution reconstruction.Te duration of the complete data processing is about 16 seconds.Te inference time of the segmentation model is only 2.37 seconds, and the postprocessing time consumes an average of 6.77 seconds.While 3D reconstruction takes up the majority of the time cost of the visualization module, text analysis is relatively less time consuming.
Te experiments are conducted in the following steps: (1) Te predefned model in the segmentation module performs supervised learning on the NIH-TCIA center to obtain the original baseline model.( 2) Te optimized baseline models are derived as in the frst step from NIH-TCIA images with various image quality control measures.(3) Te original and optimized baseline models are tested on Zheyi images, which undergo the same quality control measures as NIH-TCIA images corresponding to the model, respectively.In this way, the segmentation performance without the proposed dataadaptive system can be observed.( 4) Data adaptation training is carried out on NIH-TCIA and Zheyi images to get the segmentation model applicable to unannotated Zheyi data, so as to investigate the segmentation efectiveness after data adaptation.Te same data adaptation trainings are also performed on images on which various processing means in the quality control toolbox have been taken.
Te results are mainly evaluated by the dice similarity coefcient (DSC) and mean intersection over union (mIoU), which indicate the similarity between the pancreas mask generated by the segmentation model and ground truth.Te Hausdorf distance measures the deviation between the predicted mask and ground truth mask and is calculated as the distance of points in the two masks to each other's surfaces.Te DSC and mIoU are defned as follows: where X represents the pancreas mask generated by the segmentation model and Y is the ground truth.Te Hausdorf distance is computed as where S(X) is the point set of the predicted pancreas mask and x represents the points in it.Similarly, S(Y) is the point set of the ground truth mask and y represents the points in it.In the formula, d(x, S(Y)) indicates the distance from the point x to the surface formed by the point set S(Y).We display the average DSC, mIoU, and the Hausdorf distance on the test samples to demonstrate the average performance of the proposed pancreas segmentation system.In addition, considering the comprehensive presentation of segmentation masks, the visualization module will present the textual information and multidimensional images of the pancreas segmentation results.Te textual information contains the volume diference, Hausdorf distance, centerof-mass distance, and average symmetric surface distance (ASSD) to evaluate the segmentation results from multiple perspectives.
2.6.3.Baseline Performance.Figure 3 displays CT images from NIH and Zheyi datasets processed with multiple quality control methods.As can be observed, the raw data from the two datasets exhibit diferences in various aspects  2 that CT images of the two datasets difer signifcantly in terms of layer thickness.Te range of intensity preserves most of information of abdominal organs and decreases the noise interference.Te rotation augmentation with minor angles increases the diversity of FOV.Te superresolution reconstruction algorithm frames out valid abdominal regions and eliminates redundant background information, so as to diminish heterogeneity caused by the clinical imaging process.It is observed from Figure 3 that, with the data quality control means, the visualized CT sequences narrowed the variation within and between datasets.We frst train the supervised baseline models on the NIH-TCIA dataset using raw and processed images and compare segmentation performance with existing pancreas segmentation methods.It is observed that the models trained with quality-controlled images outperform those trained on raw images, demonstrating that the image quality control toolbox designed in this paper is efective in decreasing intradomain heterogeneity.As summarized in Table 3, our model achieves a DSC of 85.45% after corresponding quality control means, which is superior to the performance of existing methods.Te results show the mean value and standard deviation of the 5-fold cross-validation, which more reliably refect the model's performance on the whole dataset.Te results indicate that the designed network in segmentation modules is of sufcient confdence to serve as the baseline models for subsequent adaptation experiments.
2.6.4.Data-Adaptive Performance.Table 4 lists the results of the unlabeled Zheyi dataset of directly tested and with data adaptation by the proposed system.When the NIH-TCIA baseline model is tested directly on the Zheyi dataset, only a DSC of 58.79% is obtained, indicating that almost half of the tissues are segmented incorrectly.Figure 4 presents the segmentation masks for three CT slices selected from the Zheyi dataset.As can be clearly observed, the model trained on the NIH-TCIA dataset exhibits signifcant degradation in pancreas segmentation on unseen images with heterogeneity.Te masks in such cases do not efectively capture the pancreatic tissue information without data adaptation.
After data-adaptive training, the DSC score increases to 72.73% (a gain of 13.94%).Notably, the models trained with quality-controlled data demonstrate better performance when oriented to images from new sources.With rotation augmentation and superpixel reconstruction, the performance increases from 61.95% to 75.43% (a gain of 16.64%).To ensure the reliability of the experimental results, paired ttests were performed on segmentation results of both the models with and without data adaptation.As listed in Table 4, the segmentation performance is signifcantly improved as observed, by the p value less than 0.01.Moreover, Table 3: Baselines segmentation performance on the NIH-TCIA dataset, where "sup-pixel" refers to superpixel reconstruction.

Model Image options
Mean DSC (%) Li et al. [19] Raw images 83.06 ± 5.57 Cai et al. [20] Raw images 82.40 ± 6.70 Yu et al. [16] Raw images 84.Journal of Healthcare Engineering after the image quality control, the diference in segmentation performance is more between before and after data adaptation.As can be observed in Figure 4, the pancreatic tissue can be correctly segmented out after the designed system.As expected, the model with data adaptation constructed by the system is capable of efcient segmentation for pancreas tissues and is of great signifcance in clinical decision-making.
2.6.5.Visualization.Te visualization module provides structured textual information about the segmentation results and visualized images to assist physicians.We selected a CT sequence from the Zheyi dataset and allowed the physician to make corrections to annotations.In the designed system, the visualization results of this sample are shown in Figure 5.For the segmentation mask, statistics such as pancreas volume and size are and displayed, and pancreas images in multiple dimensions are reconstructed.For physician manual corrections, the system then indicates the deviation between manual and system corrections in metrics such as the volume diference, mass distance, average symmetric surface distance (ASSD), Hausdorf distance, and dice similarity coefcient.In addition, we calculate the 3D surface distances between two masks and reconstruct them as images.

Discussion
3.1.System Functionalities.In this research, a multifunctional image quality control toolbox was developed to standardize CT images from various aspects.Te general rotation augmentation is employed to enhance the abundance of samples.However, the pancreas as a segmentation target is relatively small in volume in comparison with the whole abdomen.Terefore, the ordinary rotation augmentation operation is of little signifcance for the task.In this paper, an augmentation approach with minor rotation angles is designed to solve this problem.First, the statistical analysis of images is performed frst to obtain the angle bias range of the abdomen, and then, we set a reasonable range of rotation angles with regard to data distribution characteristics.Tis operation reduces the discrepancy in images caused by various scanning felds of view and patient body positions with precise angle settings and enriches the samples in quantity.Furthermore, superresolution  As can be seen in Tables 3 and 4, the images with the quality control show less heterogeneity not only in intradomains but also in interdomains.Te normalization of the images in terms of the angle, intensity scale, and region of interest lays the foundation for the subsequent transfer of the segmentation model.We performed paired t-tests for the ablation study in the image quality control module, and the statistical results are presented in Table 5. Te results are statistically signifcantly improved by rotation and rotation plus superpixel reconstruction with a p value less than 0.01.Te superpixel reconstruction only yields obviously improved results with a p value less than 0.05.
Moreover, in the visualization module, the reconstructed pancreas tissue displayed in Figure 5 of the visualization module demonstrates better spatial continuity, which facilitates the radiologist's diagnosis.

System Robustness.
With the development of technology, the standardization and industrialization of deep learning are becoming more and more mature.As stated in the introduction, deep learning models have excellent performance in real-world applications but still face many challenges.Tis research improves the generalization of the segmentation model in the presence of new datasets, which addresses the potential problems of automatic pancreas segmentation systems in practical deployment and applications.However, the robustness of the deep learning model is also an issue that needs to be focused.
Robustness typically denotes the property of a system to maintain its primary performance in the presence of fuctuations in some parameters [52,53].Normally, robustness is used to evaluate how stable a system is against uncertain utilization environments.It is widely known that deep learning systems are driven by big data, and thus, more data lead to richer feature extraction and higher quality model construction [54][55][56].Terefore, when attention is paid to the small amount of data, we investigate whether the system is afected by performance degradation.
We adopted an external dataset for the robustness experiments.Te dataset was collected by Vanderbilt University and contained 30 sequences, which was marked as the BTCV dataset [57,58].Te resolution of CT images in this dataset is 512 × 512 pixels, the slice number ranges in [85,198], and the layer thickness is in the interval of [2.5, 5] mm.Te experiments were performed with the NIH dataset as a labeled center and the BTCV dataset as an unlabeled center.As can be seen in Table 6, after the proposed dataadaptive pancreas segmentation system with the image quality control, the performance on the BTCV dataset reached a DSC of 74.97%, which is 14.83% improvement  result is consistent with that when the Zheyi dataset served as an unlabeled dataset, which means that the system still achieves a superior data adaptation performance.Tis result indicates that the designed system maintains a robust performance with respect to small datasets.

System Efectiveness.
In clinical diagnosis, there is a demand for modern data analysis technology to mine CT image information and assist clinicians in improving diagnosis efciency, thus refning the medical treatment process.Te data-adaptive pancreas segmentation system proposed in this study utilizes the domain-invariant features from source-domain images with annotations for transfer learning to adapt the model to CT data features from different domains, thus implementing data-adaptive pancreatic segmentation.We developed a comprehensive image quality control toolbox to rectify data quality diferences among diferent data domains.Te image quality is controlled in several dimensions, including the image intensity range, scanning feld of view, CT layer thickness, and valid region of interest.In addition, the image quality control toolbox provides physicians with more discriminative CT images.In terms of system architecture, we employ an adversarial learning scheme to implement feature distribution acquisition and feature adaptation among domains.In addition, this system fully takes into consideration the diference in the contribution of semantic information at diferent scales for domain adaptation and adopts weighted connections to realize the stitching of multiscale information to achieve a more stable and smooth adversarial learning structure.Te efectiveness of the system has been validated with public datasets and real in-hospital data, and the robustness of the system has been demonstrated on a small dataset.In summary, the system enables the establishment of the data-adaptive segmentation model by transfer learning, both interhospital and intrahospital across time lengths.Te system eliminates the need for time-consuming and tedious annotation work by radiologists, which is of signifcant relevance to the automation of hospital treatment processes in real-time medical scenarios.In addition, it is a meaningful research area to combine image semantic segmentation techniques with other text analysis tasks to design an automatic pancreatic disease diagnosis system applicable to richer medical scenarios.In the subsequent research, we will combine natural language processing and image interpretability to further improve the system, optimize the pancreatic disease diagnosis process, and promote the efciency of physicians.

Conclusions
In this paper, we designed an end-to-end data-adaptive pancreas segmentation system with an image quality control toolbox.Te system aims to address the problem of poor generalization capability exhibited by existing pancreas segmentation networks when oriented to data from diferent medical centers.For the visual task of label-free semantic segmentation, this research utilizes an adversarial learning method to obtain domain-invariant supervised information and construct the data-adaptive pancreas segmentation model.In addition, a functional image quality control toolbox was designed to provide multiple image preprocessing methods.Te system works in an end-to-end manner and is easy to operate by physicians.Te experimental results of public datasets and in-hospital datasets demonstrated that the end-to-end data-adaptive pancreas segmentation tool proposed in this paper can efectively assist in pancreas segmentation, and the generalization of segmentation networks was enhanced when facing images from diferent sources.Tis system is of considerable relevance in medical diagnosis and treatment and greatly promotes the development of precision and automated medical processes.

Figure 1 :
Figure 1: Overall framework of the data-adaptive pancreas segmentation system.Te system consists of two parts: the local clients of medical centers and the collaborative center on the cloud server.

F
Ti H Unsupervised Training for segmentation Model M Ti D S & Y S M S ~ G S = F S ( x s , θ s ) Feature encoding vectors extracted from deep layers are detached for adversarial training.

Figure 2 :
Figure 2: Workfow for the construction of a data-adaptive pancreas segmentation model for unlabeled data.

Figure 3 :
Figure 3: CT slice presentation with various image quality control means from two domains.

Figure 4 :
Figure 4: Qualitative segmentation results with diferent methods on several axial slices from the Zheyi dataset: manual label (red), prediction label (yellow), and overlap (green).(a) Ground truth.(b) Without adaptation.(c) With adaptation without the image quality control.(d) With adaptation with the image quality control.

Figure 5 :
Figure 5: System interface display: visualization of multidimensional segmentation results in multiple dimensions and comparison with manual annotations.

Table 1 :
Execution time for processing steps in the system.
like the intensity distribution and scanning feld of view (FOV) upon observation by a human.Moreover, it is also evident from the statistics shown in Table

Table 2 :
Datasets in the experiment.

Table 4 :
Adaptation performance of the system on the Zheyi dataset, where "sup-pixel" refers to superpixel reconstruction and "w/o" indicates without.Te maximum value is presented in bold.

Table 5 :
Paired t-test results for the ablation study of raw images and images with the quality control, where "sup-pixel" refers to superpixel reconstruction.