上海大良网站建设,铁岭网站开发公司,seo网站诊断流程,天水模板型网站建设Title
题目 SAMUS: Adapting Segment Anything Model for Clinically-Friendly and Generalizable Ultrasound Image Segmentation
SAMUS#xff1a;适应临床友好型和泛化的超声图像分割的Segment Anything模型 01
文献速递介绍
医学图像分割是一项关键技术#xff0c;用…
Title
题目 SAMUS: Adapting Segment Anything Model for Clinically-Friendly and Generalizable Ultrasound Image Segmentation
SAMUS适应临床友好型和泛化的超声图像分割的Segment Anything模型 01
文献速递介绍
医学图像分割是一项关键技术用于辨识和突出显示医学图像中的特定器官、组织和病变是计算机辅助诊断系统的一个组成部分刘等2021。为自动医学图像分割提出了众多深度学习模型展示了巨大的潜力Ronneberger, Fischer, 和 Brox2015吴等2022。然而这些模型是为特定对象量身定做的并且在应用于其他对象时需要重新训练给临床使用带来了极大的不便。
段落任何模型SAM作为一个多用途的视觉分割基础模型由于其在多样对象上的显著分割能力和强大的零次学习泛化能力获得了相当的赞誉Kirillov等2023。根据用户提示包括点、边界框和粗略遮罩SAM能够对应分割对象。因此通过简单的提示SAM可以轻松适应各种分割应用。这一范式使得将多个单一医学图像分割任务整合进一个统一框架即一个通用模型成为可能大大促进了临床部署黄等2023。
尽管构建了迄今为止最大的数据集即SA-1BSAM在医学领域遭遇快速性能下降原因是可靠的临床注释稀缺黄等2023。一些基础模型已经被提出通过在医学数据集上调整SAM来适应医学图像分割马和王2023吴等2023。然而与SAM相同它们在特征建模之前对输入图像进行16×的无重叠分割这破坏了识别小目标和边界所需的局部信息使它们难以分割具有复杂/线状形状、弱边界、小尺寸或低对比度的临床对象。此外它们中的大多数要求输入尺寸为1024×1024由于生成的长输入序列给GPU消耗带来了巨大负担。
在本文中我们介绍SAMUS将SAM的卓越分割性能和强大的泛化能力转移到医学图像分割领域同时降低了计算复杂度。SAMUS继承了SAM的ViT图像编码器、提示编码器和掩码解码器对图像编码器进行了量身定制的设计。首先我们通过减小输入尺寸来缩短ViT分支的序列长度以降低计算复杂度。然后开发了特征适配器和位置适配器以从自然域微调ViT图像编码器到医学域。为了补充ViT图像编码器中的局部即低级信息我们引入了一个并行的CNN分支图像编码器与ViT分支并行运行并提出了一个跨分支注意力模块使ViT分支中的每个补丁都能从CNN分支吸收局部信息。 Abstract-Background
摘要 Segment anything model (SAM), an eminent universal im age segmentation model, has recently gathered considerable attention within the domain of medical image segmenta tion. Despite the remarkable performance of SAM on natu ral images, it grapples with significant performance degrada tion and limited generalization when confronted with med ical images, particularly with those involving objects of low contrast, faint boundaries, intricate shapes, and diminu tive sizes. In this paper, we propose SAMUS, a universal model tailored for ultrasound image segmentation. In con trast to previous SAM-based universal models, SAMUS pur sues not only better generalization but also lower deploy ment cost, rendering it more suitable for clinical applications. Specifically, based on SAM, a parallel CNN branch is intro duced to inject local features into the ViT encoder through cross-branch attention for better medical image segmenta tion. Then, a position adapter and a feature adapter are de veloped to adapt SAM from natural to medical domains and from requiring large-size inputs (1024×1024) to small-size inputs (256×256) for more clinical-friendly deployment. A comprehensive ultrasound dataset, comprising about 30k im ages and 69k masks and covering six object categories, is collected for verification. Extensive comparison experiments demonstrate SAMUS’s superiority against the state-of-the-art task-specific models and universal foundation models under both task-specific evaluation and generalization evaluation. Moreover, SAMUS is deployable on entry-level GPUs, as it has been liberated from the constraints of long sequence encoding. The code, data, and models will be released at https://github.com/xianlin7/SAMUS.
段落任何模型SAM作为一个杰出的通用图像分割模型最近在医疗图像分割领域内获得了相当多的关注。尽管SAM在自然图像上的性能表现显著但面对医学图像尤其是那些涉及低对比度、界限模糊、形状复杂和尺寸微小的对象时它却面临着显著的性能下降和有限的泛化能力。在本文中我们提出了SAMUS一个专为超声图像分割定制的通用模型。与之前基于SAM的通用模型相比SAMUS不仅追求更好的泛化能力也追求更低的部署成本使其更适合临床应用。
具体而言基于SAM通过引入一个并行的CNN分支通过跨分支注意力将局部特征注入ViT编码器以实现更好的医学图像分割。然后开发了位置适配器和特征适配器以将SAM从自然领域适配到医学领域并从需要大尺寸输入1024×1024到小尺寸输入256×256进行适配以实现更加临床友好的部署。收集了一个全面的超声数据集包括大约30k图像和69k掩模覆盖六个对象类别用于验证。广泛的比较实验表明SAMUS在任务特定评估和泛化评估下都优于最先进的任务特定模型和通用基础模型。
此外SAMUS可以部署在入门级GPU上因为它已经摆脱了长序列编码的限制。代码、数据和模型将在https://github.com/xianlin7/SAMUS发布 Conclusions
结论 In this paper, we propose SAMUS, a universal founda tion model derived from SAM, for clinically-friendly and generalizable ultrasound image segmentation. Specifically, we present a parallel CNN branch image encoder, a fea ture adapter, a position adapter, and a cross-branch atten tion module to enrich the features for small-size objects and boundary areas while reducing GPU consumption. Further more, we construct a large ultrasound image dataset US30K, consisting of 30,106 images and 68,570 masks for eval uation and potential clinical usage. Experiments on both seeable and unseen domains demonstrate the outstanding segmentation ability and strong generalization ability of SAMUS. Moreover, the GPU memory cost of SAMUS is merely 28% of that required to train the entire SAM, and SAMUS is about 3× faster than SAM for inference.
在本文中我们提出了SAMUS一个基于SAM衍生的通用基础模型用于临床友好和可泛化的超声图像分割。具体而言我们展示了一个并行的CNN分支图像编码器、一个特征适配器、一个位置适配器和一个跨分支注意力模块以丰富小尺寸对象和边界区域的特征同时减少GPU消耗。此外我们构建了一个大型超声图像数据集US30K包含30,106张图像和68,570个掩模用于评估和潜在的临床使用。在可见和未见领域的实验展示了SAMUS的卓越分割能力和强大的泛化能力。此外SAMUS的GPU内存成本仅为训练整个SAM所需的28%且SAMUS的推理速度大约是SAM的3倍。 Method
方法 As depicted in Fig. 8, the overall architecture of SAMUS is inherited from SAM, retaining the structure and param eters of the prompt encoder and the mask decoder without any adjustment. Comparatively, the image encoder is care fully modified to address the challenges of inadequate local features and excessive computational memory consumption, making it more suitable for clinically-friendly segmentation. Major modifications include reducing the input size, over
lapping the patch embedding, introducing adapters to the ViT branch, adding a CNN branch, and introducing cross branch attention (CBA). Specifically, the input spatial reso ution is scaled down from 1024 × 1024 pixels to 256 × 256 pixels, resulting in a substantial reduction in GPU memory cost due to the shorter input sequence in transformers. The overlapped patch embedding uses the same parameters as the patch embedding in SAM while its patch stride is half to the original stride, well keeping the information from patch boundaries. Adapters in the ViT branch include a position adapter and five feature adapters. The position adapter is to accommodate the global position embedding in shorter sequences due to the smaller input size. The first feature adapter follows the overlapped patch embedding to align in put features with the required feature distribution of the pre trained ViT image encoder. The remaining feature adapters are attached to the residual connections of the feed-forward network in the global transformer to fine-tune the pre-trained image encoder. In terms of the CNN branch, it is parallel to the ViT branch, providing complementary local information to the latter through the CBA module, which takes the ViT branch features as the query and builds global dependency with features from the CNN branch. It should be noted that CBA is only integrated into each global transformer. Finallythe outputs of both the two branches are combined as the fi nal image feature embedding of SAMUS.
如图8所示SAMUS的总体架构继承自SAM保留了提示编码器和掩码解码器的结构和参数没有任何调整。相比之下图像编码器经过仔细修改以解决不足的局部特征和过多的计算内存消耗问题使其更适合临床友好的分割。主要修改包括减小输入尺寸、重叠的补丁嵌入、向ViT分支引入适配器、增加CNN分支并引入跨分支注意力CBA。具体而言输入的空间分辨率从1024×1024像素降低到256×256像素由于变换器中输入序列的缩短显著降低了GPU内存成本。重叠的补丁嵌入使用与SAM中的补丁嵌入相同的参数而其补丁步长是原始步长的一半很好地保留了补丁边界的信息。ViT分支中的适配器包括一个位置适配器和五个特征适配器。位置适配器是为了适应更小输入尺寸下的短序列中的全局位置嵌入。第一个特征适配器跟随重叠的补丁嵌入以将输入特征与预训练ViT图像编码器所需的特征分布对齐。其余特征适配器附加在全局变换器的前馈网络的残差连接上以微调预训练的图像编码器。就CNN分支而言它与ViT分支平行通过CBA模块提供补充的局部信息给后者CBA模块将ViT分支特征作为查询并与CNN分支的特征建立全局依赖关系。应该注意的是CBA仅集成在每个全局变换器中。最后两个分支的输出合并为SAMUS的最终图像特征嵌入。 Figure
图 Figure 1: Structure comparison of different SAM-based foundation models for medical image segmentation.
图1基于SAM的不同医学图像分割基础模型的结构比较。 Figure 2: Overview of the proposed SAMUS.
图2提出的SAMUS的概述。 Figure 3: Comparison between SAMUS and task-specific methods evaluated on seeable (marked in blue) and unseen datasets(marked in orange).
图3在可见用蓝色标记和未见用橙色标记数据集上评估的SAMUS与任务特定方法的比较。 Figure 4: Qualitative comparisons between SAMUS and task-specific methods. From top to bottom are examples of segmenting thyroid nodule, breast cancer, and myocardium.
图4SAMUS与任务特定方法之间的定性比较。从上到下分别是分割甲状腺结节、乳腺癌和心肌的示例。 Figure 5: Qualitative comparisons between SAMUS and foundation models. From top to bottom are examples of seg menting thyroid nodule, breast cancer, and myocardium.
图5SAMUS与基础模型之间的定性比较。从上到下分别是分割甲状腺结节、乳腺癌和心肌的示例。 Figure 6: Segmentation and generalization ability compari son of our SAMUS and other foundation models on seeable (in light color) and unseen (in dark color) US30K data.
图6我们的SAMUS与其他基础模型在可见的浅色和未见的深色US30K数据上的分割和泛化能力比较。 Figure 7: Comparison of SAMUS and foundation models on GPU memory cost, model parameters, computational com plexity, inference speed, performance, and generalization.
图7SAMUS与基础模型在GPU内存成本、模型参数、计算复杂度、推理速度、性能和泛化能力上的比较。 Table
表 Table 1: Summary of the datasets in US30K. LV, MYO, and LA are short for the left ventricle, myocardium, and left atrium.
表1US30K数据集摘要。LV、MYO和LA分别代表左心室、心肌和左心房。 Table 2: Quantitative comparison of our SAMUS and SOTA task- specific methods on segmenting thyroid nodule (TN3K), breast cancer (BUSI), left ventricle (CAMUS-LV), myocardium (CAMUS-MYO), and left atrium (CAMUS- LA). The perfor mance is evaluated by the Dice score (%) and Hausdorff distance (HD). The best results are marked in bold.
表2我们的SAMUS与SOTA任务特定方法在分割甲状腺结节TN3K、乳腺癌BUSI、左心室CAMUS-LV、心肌CAMUS-MYO和左心房CAMUS-LA上的定量比较。性能通过Dice得分%和Hausdorff距离HD评估。最佳结果以粗体标记。 Table 3: Quantitative comparison of our SAMUS and other foundation models on seeable US30K data. The performance is evaluated by the Dice score (%) and Hausdorff distance (HD).
表3我们的SAMUS与其他基础模型在可见的US30K数据上的定量比较。性能通过Dice得分%和Hausdorff距离HD评估。 Table 4: Ablation study on different component combinations of SAMUS on the thyroid nodule and breast cancer segmentation. F-Adapter and P-Adapter represent the feature adapter and the position adapter respectively
表4SAMUS在甲状腺结节和乳腺癌分割上不同组件组合的消融研究。F-Adapter和P-Adapter分别代表特征适配器和位置适配器。 Table 5: Ablation study of different prompts. Pt1, pt2, and pt3 represent the single-point prompt in different (randomly deter mined) foreground positions. Multipoint prompts are generated by random sampling on the foreground areas.
表5不同提示的消融研究。Pt1、pt2和pt3代表在不同随机确定的前景位置的单点提示。多点提示通过在前景区域的随机抽样生成。