Preprint
Article

This version is not peer-reviewed.

PRONOBIS:A Robotic System for Automated Ultrasound-Based Prostate Reconstruction and Biopsy Planning

A peer-reviewed article of this preprint also exists.

Submitted:

15 April 2025

Posted:

16 April 2025

You are already at the latest version

Abstract
This paper presents the PRONOBIS project, an ultrasound-only, robotically navigated prostate scanning and biopsy treatment planning. The proposed system addresses the challenges of precise prostate reconstruction and inter-operator variability by performing fully automated prostate scanning, real-time ultrasound image processing, 3D prostate reconstruction, and biopsy needle position planning. Fully automated prostate scanning is achieved by using a robotic arm with an ultrasound system. Real-time ultrasound image processing utilizes deep learning algorithms for precise prostate segmentation. To create a high-quality prostate segmentation dataset, this paper proposes a deep-learning based medical annotation platform - MedAP. For precise segmentation of the entire prostate sweep, DAF3D and MicroSegNet models are evaluated, and additional computer vision postprocessing methods are proposed. The 3D visualization and prostate reconstruction are performed based on the segmentation results and robotic positional data, enabling robust and user-friendly biopsy treatment planning. The real-time sweep scanning and segmentation operates at 30Hz, which enables full sweep scanning in 15 to 20 seconds, depending on the prostate size. The system is evaluated on prostate phantoms by reconstructing the sweep and by performing dimensional analysis, which indicates 92\% and 98\% volumetric accuracy on used phantoms. 3D prostate reconstruction takes approximately 3 seconds and enables a fast and detailed insight for precise biopsy needle position planning.
Keywords: 
;  ;  ;  ;  ;  

1. Introduction

Prostate cancer is the second most common malignant neoplasm in men worldwide and the fourth leading cause of cancer-related death [1]. Diagnosis relies on histopathological evaluation of prostate biopsies, typically performed following an elevated prostate-specific antigen (PSA) level and/or an abnormal digital rectal examination (DRE), even though most patients are asymptomatic. The prostate is anatomically divided into four main zones—peripheral, transitional, central, and anterior—with approximately 70% of prostate cancers arising in the peripheral zone [2].
Multiparametric MRI (mpMRI) is now recommended before biopsy to detect suspicious lesions, which are subsequently targeted using either cognitive or fusion approaches; current evidence suggests no significant difference in accuracy between these methods [3]. However, a key challenge remains: the precise targeting of lesions, as targeted prostate biopsy is highly operator-dependent and subject to significant inter-operator variability [4]. This variability raises concerns regarding the consistent application of the mpMRI diagnostic pathway and the risk of improper diagnosis.
To address these challenges, current initiatives aim to standardize the entire diagnostic workflow—including MRI acquisition, interpretation, biopsy planning, and execution—through robust quality assurance and control measures [4]. One proposed solution is the integration of robotic prostate platforms and artificial intelligence to minimize inter-operator bias and reduce the learning curve [5], although studies on this approach have yielded conflicting results [6].
Robotic biopsy techniques that combine transrectal ultrasound (TRUS) with MR imaging have been introduced. Although designs vary, these systems typically fuse mpMRI-identified lesions with TRUS images to enable a robotic arm to target the areas of interest accurately. The system determines the needle’s penetration angle and depth by positioning a needle guide equipped with a stop bar; the surgeon then manually inserts and fires the needle gun at the preset location and depth [7,8]. The targeted biopsy results have higher success rates than systematic biopsy [9,10]. Therefore, the novel approach that enables clinicians to pick the exact positions for biopsy needle impact exclusively via ultrasound, is crucial. Incorporating targeted biopsy into the workflow requires an end-to-end software solution that enables biopsy procedure planning.
One of the challenges for robotic biopsy are safety measures and robustness of the robots that operate in close proximity to the human body. By using a redundant 7 DOF robot arm it is possible to optimize each trajectory avoiding joint limits and singularities [11], crucial to perform complex movements in space restricted areas. Zhang et al. presented a design of a 7-DOF manipulator for TRUS probe [12]. The manipulator was evaluated in an environment that indicated the mechanism can reach any position in working space of the biopsy procedure. The robotic arm structure of continuous body type for prostate biopsy procedures in confined spaces has been introduced [13]. Two sets of drive lines run through the robot and are fixed to the end nodes. The end position is controlled by adjusting the angle of the servo motor. Experiments indicate the introduced robot design achieves 2.5 mm positioning error which is a significant improvement for robotic biopsy procedures.
The robotic platform for prostate biopsy, the iSR’obot Mona Lisa™ (Biobot Surgical, Singapore) [14], allows clinicians to create biopsy procedure planning by fusing MRI and ultrasound images. Additionally, a novel Vector MRI/ultrasound fusion transperineal biopsy technique, utilizing electromagnetic needle tracking, has been developed [15]. Ipsen et al. [16] demonstrated 4D (volumetric) ultrasound image acquisition with motion compensation over longer periods of time. The study presented the approach evaluated on five volunteers for periods of 30 minutes with dynamic pressure adjustments with target force of 10N, to determine the feasibility and safety of the proposed approach.
A transperineal prostate biopsy approach that proposes a robotically guided fusion of MRI and ultrasound images [17] has been introduced. After scanning the phantom used in the experimental phase, it creates a 3D model, enabling biopsy planning with the average error of 1.44 mm, which is below the clinical threshold required to distinguish cancerous from healthy tissue. Furthermore, the approach is similar to traditional biopsy approaches and can be adopted in a clinical setting. Another high precision robot introduced by Stoianovici et al. is designed for transperineal prostate biopsy with an accuracy of 2.55 mm [18]. The promising accuracy results of targeting the exact positions for the biopsy needle using robots provide the foundation for our research.
The robotically assisted biopsy needle positioning study resulted in a robot, guided by an MRI image, that achieved average precision of 2.39 mm, compared to 3.71 mm with the manual approach [19]. Although a robot can eliminate human error, it introduces challenges such as the robot positioning and the pressure applied to the patient, that causes prostate deformation. A recent study addressed the challenges by introducing robot that minimizes prostate deformation [20] with 4 DOF probe manipulator used for TRUS; needle targeting was achieved with a precision of 1 mm. To minimize prostate deformation, the approach involved three essential aspects; optimization of the biopsy trajectory, optimization of the order of biopsy points, and a prostate coordinate system that the team developed.
The ultrasound image segmentation plays a crucial role to ensure adequate precision in prostate scanning, 3D reconstruction, and robot-assisted biopsy procedure planning, The ultrasound image intensity inhomogeneities and weak object boundaries make segmentation difficult for conventional image processing methods [21], such as simple standard or adaptive thresholding, Otsu’s binarization [22] or contour detection that do not completely fulfill the requirements of high accuracy, robustness and real-time performance [23]. Deep learning-based approaches demonstrate superior accuracy and robustness due to ability to learn the local and global image features. The utilization of Convolutional Neural Networks (CNN) for medical image segmentation presents a relevant solution. One of the most significant CNN architectures for medical image segmentation is U-Net [24], which is the encoder-decoder network, capable of capturing local image features and producing accurate high-resolution segmentations. The latest research utilizes the Transformer architecture for image processing. The Vision Transformer (ViT) architecture for classification [25] demonstrates remarkable performance that surpasses state-of-the-art CNN models. One of the most relevant algorithms that combines CNNs and ViT models for medical image segmentation is TransUNet [26]. The TransUNet hybrid architecture combines U-Net and ViT to eliminate the limitations of separate solutions, precisely, it enables segmentation of images with significant inter-patient variations in terms of specific object size, shape and tissue texture. One of the current state-of-the-art segmentation models for micro-ultrasound prostate images is MicroSegNet model [27], based on the TransUNet architecture with a modified annotation-guided binary cross-entropy (AG-BCE) loss function and multi-scale deep supervision (MSDS) that modifies the segmentation performance at different decoder layers, assuring precise segmentation globally and locally.
The proposed article presents robotically navigated prostate scanning and biopsy treatment planning. Our method integrates robotics, ultrasound imaging, and deep learning algorithms for image analysis, with the objective of generating visualizations for biopsy needle placement. A key challenge addressed is achieving an ultrasound-only reconstruction process, comparable to hybrid MR-ultrasound methods. The ultrasound-only prostate reconstruction and biopsy treatment is fast, cost-effective, and less physically demanding for the patient. Fully automated, robotic prostate sweep scanning ensures positional tracking of sweep slices for precise prostate reconstruction. The utilization of deep learning algorithms with image post-processing enables real-time, precise segmentation across the whole prostate sweep. For precise segmentation performance, the model is trained on a hybrid dataset consisting of real prostate and phantom prostate images, annotated using our Medical Annotation Platform (MedAP), that utilizes zero-shot CNN architectures. To select the optimal deep learning algorithm for the ultrasound medical prostate images, DAF3D and MicroSegNet architectures are trained on a hybrid dataset and compared, with the selection of MicroSegNet. 3D visualization and prostate reconstruction, based on the segmentation results and robotic positional information, are evaluated on two prostate phantoms (CIRS Prostate Training Phantom Model 070L and CIRS Tissue Equivalent Ultrasound Prostate Phantom Model 053L) achieving accurate dimensional and volumetric reconstructions. The 3D prostate reconstruction takes approximately 3 seconds and enables detailed insight for biopsy needle position planning.. The proposed system is planned to be integrated as part of the RONNA robotic system for neuronavigation which was already clinically validated in brain biopsies and EVD placement [28,29]

2. Materials and Methods

The proposed robotic approach integrates multiple components to create a comprehensive solution. The goal is to seamlessly incorporate the robotic system into the existing prostate biopsy procedure, maintaining simplicity for clinicians while enhancing patient comfort. The complete workflow of the procedure is illustrated in Figure 1.

2.1. Hardware and Software

KUKA’s LBR iiwa collaborative industrial robot, is utilized for this research. Although custom-built robotic systems, such as the Mona Lisa [14], provide significant design flexibility and a tailored approach to robotic architecture and procedural flow, they require extensive development. In contrast, using a medical robot arm KUKA LBR greatly reduces the time needed for development, offering a streamlined, cost-effective, and easily deployable solution with proven performance in industrial and medical environments. In this stage of development, the clinical ultrasound system P25 and BCL10-5 TRUS probe, developed by SonoScape are utilized. The TRUS probe transmits a 7.5 MHz signal and receives between 6 and 16 MHz signals. The acquired ultrasound image is 60 mm wide, with depth between 3 and 90 mm. The robot with an attached TRUS probe is shown in Figure 2 and Figure 3.
Robot operating system 2 (ROS2) [30] is used for robot control and communication within the system components; robot, ultrasound, segmentation and visualization. ROS2 provides integration of system components without compromising the integrity or functionality of existing elements. Integration between the robot and ROS2 Humble framework is achieved by using LBR-Stack [31]. LBR-Stack takes KUKA’s network communications protocols and integrates them into the ROS2 framework to enable seamless usage of these robots.
The proposed end-to-end software was developed using Python programming language version 3.10 for calibration, robot control and segmentation, while for the process of reconstruction and visualization version 3.12 was used because it offers a speed improvement. The deep learning algorithms used for segmentation purposes require CUDA enabled PyTorch 2.5.1. To implement reconstruction and visualization, we used the following libraries: OpenCV 4.10. and PyQt6. The complete solution uses the HP Z4 workstation with Intel Xeon W-2245 processor, 64 GB RAM memory, Quadro RTX 5000 GPU with 16 GB VRAM memory.

2.1.1. Robot Setup

After the TRUS probe is mounted on the robot, calibration of the TCP must be performed, for which KUKA’s four-point algorithm was used. Since the TRUS image needs to be calibrated with the TCP of the robot, a digital cross-hair was created to position a 3D printed cone in the center of the image, as shown in Figure 2(b).
Upon successful calibration, the TRUS probe is placed inside the specified prostate phantom using the robot’s compliance mode. The segmentation starts the process and the robot rotates the TRUS probe around its axis to acquire the US images. Once the segmentation process finds no prostate in the image, the algorithm sends the end signal and the robot changes its rotational direction to scan the other half. Prostate images, fused with the synchronized robot arm data are constantly being published via ROS2 for visualization purposes. Once the whole prostate sweep is complete, the program stops. The detailed image of the prostate scanning equipment and the prostate sweep scanning procedure is shown in Figure 3.

2.2. Prostate Segmentation

Prostate segmentation processes a raw ultrasound image and outputs a binary segmentation mask. Our segmentation algorithm utilizes a hybrid CNN-ViT model to achieve robust image segmentation by combining the strengths of both approaches. It enables automated analysis of ultrasound prostate images by combining local feature extraction via CNN with global context extraction via ViT. Trained on a large, diverse and well-balanced dataset, this combination delivers state-of-the-art segmentation performance.

2.2.1. MedAP

As part of the segmentation training dataset development, we have developed the Medical Annotation Platform (https://github.com/CRTA-Lab/MedAP), an advanced annotation framework designed for semi-automatic segmentation of medical images, shown in Figure 4. MedAP leverages state-of-the-art, zero-shot segmentation models to achieve rapid and accurate extraction of objects from medical images. The platform combines deep learning models with human annotation expertise in case the deep learning models struggle to segment complex structures, such as granular-appearing tissues in medical ultrasound images.
The zero-shot segmentation model used for MedAP is Meta AI’s Segment Anything Model (SAM) [32], a state-of-the-art transformer-based model designed to segment objects in images and videos with minimal user input. The model represents an innovative approach to the segmentation process by utilizing various prompt-driven annotations. SAM is selected for use in the MedAP due to its versatility in prompting options such as positive and negative Point prompting, Bounding Box prompts and Free-form text.
The MedAP annotation pipeline is presented in Figure 5. The process begins with importing ultrasound prostate sweeps, followed by MedAP-assisted annotation using prompts or manual tools to ensure precise segmentation and high dataset quality. The annotation platform stores the original image and segmentation mask pair, alongside the image metadata, including the image parameters and probe angle information required for 3D reconstruction. The created dataset is used to evaluate two deep learning segmentation models, DAF3D and MicroSegNet, on prostate phantom images in order to determine the most suitable solution for robust segmentation. DAF3D and MicroSegNet are state-of-the-art ultrasound segmentation models with specialized architectures, presented in this article and evaluated on the prostate phantom segmentation task, crucial for performing our prostate phantom reconstruction.

2.2.2. Deep Attentive Features for Prostate Segmentation (DAF3D)

DAF3D is a 3D CNN with built-in attention modules that leverage information encoded in the different layers of the CNN called multilevel features [33]. The feature extraction portion of the network is based on the 3D ResNeXt architecture [34]. The network also employs 3D atrous spatial pyramid pooling for resampling the attentive features at different rates in order to achieve a more accurate prostate representation. The model uses a hybrid loss function that combines binary cross-entropy loss and Dice loss to preserve the boundary details while producing a compact segmentation result. The model achieves values of 0.90 and 0.82 for the Dice score and the Jaccard score in the test set, respectively. After the segmentation step, the predicted masks are saved and their contours are plotted on the original image.

2.2.3. MicroSegNet

The MicroSegNet model [27], designed for micro ultrasound prostate segmentation, has introduced the annotation-guided binary cross entropy loss AG-BCE, which enables training attention on hard to segment regions. The model also introduces multi-scale deep supervision that enables robust performance, invariant to object appearance on the image. Although MicroSegNet achieved a Dice score of 0.939 and a Hausdorff distance of 2.02 m m , this approach struggles with the challenge of annotating the edges of the prostate and producing empty masks if the prostate is not present in the image. To address these challenges, MicroSegNet combined with the image post-processing algorithm eliminates false positive segmentation results. The image post-processing portion includes a filtering algorithm that calculates the area and the center of the segmentation mask at the prostate edges.
As part of our research, the Graphical User Interface (GUI), that enables intuitive prostate procedure planning was developed in collaboration with clinicians (T.H, T.K, T.Z.), who indicated the base requirements and useful tools for our software. The application shows the reconstructed prostate and allows the clinician to pick the biopsy target locations. The clinician can pick an ultrasound slice with suspected lesions and choose the exact target location to perform the biopsy. Upon choosing the target(s), they will appear on the reconstructed 3D visualization model.

3. Results

3.1. Model Performance Evaluation

To achieve real-time ultrasound prostate scanning and segmentation, crucial for 3D visualization and biopsy treatment planning, the first step is to compare existing deep learning solutions based on their segmentation performance. Following a thorough literature review based on the Dice score coefficient and Jaccard index, two state-of-the-art models, DAF3D and MicroSegNet, are further evaluated on the dataset that combines the open-source MicroSegNet dataset (https://zenodo.org/records/10475293) consisting of 75 prostate sweeps ranging from 40 to 60 slices per sweep, and our six sweep dataset consisting of CIRS 053L and CIRS 070L prostate phantoms with up to 300 slices per sweep (https://www.kaggle.com/datasets/lukaiktar/crta-pronobis-prostate-phantom-ultrasound-dataset). Each model was trained on the five uniform hybrid dataset splits to analyze the performance on non-biased sets of images with different augmentations. The best performing model is then trained on complete dataset and used for real-time segmentation.
The Dice score, commonly known as the Dice coefficient is defined as (1) and Jaccard score or Jaccard index is defined as (2). For binary image segmentation, y represents the ground truth segmentation and y ˜ represents the model segmentation prediction:
D ( y , y ˜ ) = 2 | y y ˜ | | y | + | y ˜ |
J ( y , y ˜ ) = | y y ˜ | | y y ˜ |
DAF3D and MicroSegNet models are evaluated on the hybrid dataset splits, consisting of 55 training and 20 testing real prostate sweeps from the open-source MicroSegNet dataset, along with 3 sweeps from the 053L phantom and 3 sweeps from the 070L phantom. As phantom sweeps consist 300 slices, they are randomly divided into the smaller sets and allocated to the train and test datasets. The dataset is split into five equal folds with data augmentation, to perform uniform and non-biased training evaluation. Each fold consists of 18 054 original and augmented images divided as follows: 80 % for training, 10 % for validation and 10% test images. The MicroSegNet model is trained on images of size 128x128 pixels, for 30 epochs with batch size 8, and adaptive learning rate that starts with 0.001. The DAF3D model is trained on the same folds with the same image size, for 50 epochs, adaptive batch size and mixed precision training in order to reduce computational costs. A learning rate scheduler is also used so that the learning rate is reduced once the Dice score plateaus to achieve optimal model performance. The different training parameters are derived from the optimal hyperparameter values from the respective research papers [27,33]. The training results across the folds are presented in Table 1:
MicroSegNet performed better on the hybrid dataset, thus the MicroSegNet is chosen for further training process and exploitation. The chosen model is fine-tuned on a complete hybrid dataset consisting of 2 152 real prostate images and 3 042 phantom prostate images, further enhanced with 5 194 augmented images. The augmentation procedure includes random vertical and horizontal flipping, random rotations in the range of ± 10° and random Gaussian noise addition. The model is trained on images size 224x224 pixels for 30 epochs with batch size 8 and adaptive learning rate 0.001, patch size 16 pixels and weight for the hard regions equals 4. The model achieved a Dice score of 0.943988 and a Jaccard score of 0.886911. The examples of segmentation results on prostate phantoms 035L and 070L are shown in Image 6. The model performance has been enhanced with a postprocessing algorithm that filters the segmentation of the prostate edges by applying the curvature factor threshold and area threshold, as described in Algorithm 1. A postprocessing step is needed for the complete prostate sweep segmentation because of frequent false positive model predictions on the edges, caused by unusual prostate shapes and indistinctive outer contours. The overall system processing speed is 30 Hz, which is considered real-time for transformer-based segmentation performance.
Algorithm 1 Postprocessing procedure for segmentation images
  • Require: List of segmentation masks M, probe angles A, angle threshold θ , area threshold T a , curvature threshold T c
  • Ensure: Filtered segmentation masks M
  • for each ( m i , a i ) in ( M , A )  do
  •      if  | a i | > θ  then
  •           a r e a computeArea ( m i )
  •          if  a r e a > T a  then
  •              m i emptyMask ( )
  •          end if
  •      end if
  •       c u r v a t u r e computeCurvature ( m i )
  •      if  c u r v a t u r e > T c  then
  •           m i emptyMask ( )
  •      end if
  •      Append m i to M
  • end for
  • return  M
The state-of-the-art model results on prostate sweep with included edges and background images are primarily achieved by area and curvature filtering on the edges. Segmentation results fused with the robot arm positional information are stored in the DICOM format for further analysis, such as mask contour extraction and 3D visualization. An intuitive and detailed 3D prostate visualization, enriched with the list of recorded 2D slices, is required for biopsy treatment planning.

3.2. Prostate Reconstruction

After the segmentation of each ultrasound slice, the prostate gland is reconstructed using the contours of segmented masks. As discussed in Section 2, the robot arm, along with the ultrasound image, sends information about the angle of the probe, synchronized in time with image acquisition. The obtained slices, used for reconstruction and prostate visualization are not parallel because the probe rotates around its axis. The 3D reconstruction is achieved by transforming each slice of the sweep from its coordinate frame (the slice is assumed to be positioned in the xy plane) to the visualization frame x y z by using equations (3 - 5). The visualization frame is aligned with the frame of the initial slice, which will be in the middle of the prostate due to such probe placement in standard clinical practice.
x = x
y = c o s ( ϕ ) y
z = s i n ( ϕ ) ( y + r )
Using this method, we calculate the point-cloud representation of the prostate gland’s outer surface. Ultrasound sweep and the corresponding prostate representation enable the clinician to analyze suspect lesions and mark biopsy targets, as shown in Figure 7. This approach is tested and evaluated on prostate phantoms, as the 3D reconstructed phantom’s dimensions are compared to the specification measurements of the CIRS 053L and CIRS 070L phantoms. CIRS 070L phantom specifications do not provide the dimensions, so it is evaluated only on the specified volume. The reconstruction was tested on 10 phantom sweeps performed with the same pressure force applied perpendicular to the surface plane, along the image y direction, of 6.5 N for CIRS 053L and 9 N for CIRS 070L. The reconstructed point clouds are used to generate a surface approximation in MeshLab, then measured and compared with the datasheet specifications to assess reconstruction accuracy. For the dimensional measurements we use the outer dimensions of the bounding box aligning with x y z coordinate frame, and the volume is calculated by SolidWorks software. After 10 measurements, the average dimensions of the CIRS 053L prostate phantom, given in Table 3, are 55.9x42.9x37.3 mm and the volume of 54 058.2 m m 3 , for the given dimensions of the fantom of 50x45x40 mm and the volume of 53 c m 3 . The volume difference of the reconstructed models is within 2% of the specified volume. There are larger differences in dimensions, which stem mostly from the deformations caused by the presence of the probe but are also affected by the inaccuracies in probe calibration, inaccuracies in segmentation, and differences in measuring methodology. The average volume of CIRS 070L is 53 217.6 m m 3 , with a given volume of 49 c m 3 . This phantom has a larger difference in volume than the first one, most probably due to the unclear prostate dimensions used for volume calculation. The presented robotically navigated ultrasound-only prostate scanning and 3D reconstruction evaluated on our phantoms demonstrated results comparable to the volumetric and dimensional results of MR-ultrasound solutions [17].
Table 2. Reconstruction performance over 10 sweeps - phantom CIRS 053L.
Table 2. Reconstruction performance over 10 sweeps - phantom CIRS 053L.
Length / [mm] Width / [mm] Height / [mm] Volume / [ m m 3 ]
Mesured 55.9 ± 0.47 42.9 ± 0.42 37.3 ± 0.61 54058.2 ± 652.4
Ground truth 50 45 40 53000
Table 3. Reconstruction performance over 10 sweeps - phantom CIRS 070L. Dimensional ground truth data is not specified.
Table 3. Reconstruction performance over 10 sweeps - phantom CIRS 070L. Dimensional ground truth data is not specified.
Length / [mm] Width / [mm] Height / [mm] Volume / [ m m 3 ]
Mesured 58.0 ± 0.16 43.9 ± 0.43 37.4 ± 0.19 53217.6 ± 546.6
Ground truth - - - 49000
Figure 7. Graphical User Interface. The program is divided into three different interactive visualizations. In the top left corner, a reconstructed 3D model of the prostate gland is shown. The green cylinders represent biopsy targets picked by the clinician. The clinician can zoom in/out and rotate the model (using mouse and keyboard) to examine it. Next to the model is a small representation of the male human body, to check the prostate orientation. The bottom part contains all the US images recorded during the sweep, while the right top corner shows the selected US image on which the clinician can select an area from which a sample will be taken via biopsy.
Figure 7. Graphical User Interface. The program is divided into three different interactive visualizations. In the top left corner, a reconstructed 3D model of the prostate gland is shown. The green cylinders represent biopsy targets picked by the clinician. The clinician can zoom in/out and rotate the model (using mouse and keyboard) to examine it. Next to the model is a small representation of the male human body, to check the prostate orientation. The bottom part contains all the US images recorded during the sweep, while the right top corner shows the selected US image on which the clinician can select an area from which a sample will be taken via biopsy.
Preprints 156054 g007

4. Discussion

This paper presents a robotically navigated proof-of-concept system for prostate scanning and biopsy treatment planning. It utilizes a robot arm with attached TRUS probe for real-time sweep scanning, followed by deep learning-based segmentation, necessary for 3D prostate reconstruction. The accurate prostate reconstruction enables thorough analysis and biopsy needle path planning. On contrary to hybrid methods that combine MR and ultrasound imaging, our ultrasound-only, scanning and prostate reconstruction system, evaluated on CIRS Prostate Training Phantom Model 070L and CIRS Tissue Equivalent Ultrasound Prostate Phantom Model 053L, results in 92% and 98% volumetric accuracy for 070L and 053L, respectively. This paper utilizes a hybrid dataset of open-source prostate sweeps and our prostate phantom sweeps to enable fully automated real-time prostate scanning and segmentation at 30Hz. To create a balanced dataset, we developed MedAP, a Medical Annotation Platform utilizing zero-shot segmentation models. State-of-the-art ultrasound segmentation models were evaluated, and the best-performing algorithm was enhanced with postprocessing to discard outliers at prostate edges, achieving a Dice score of 0.944 and a Jaccard score of 0.887.
To facilitate the selection of target points for prostate needle path planning, we have developed a GUI that helps the clinician examine the prostate and select the exact points to target during the biopsy procedure. The presented end-to-end software system is conceptualized, designed and validated with valuable insights from urological clinicians who the system is aimed for. Our work so far presents foundation for a more sophisticated approach that aims to address the challenges of patient movement, like presented in [35], as well as prostate deformation, especially tissue inflammation caused by biopsy needle penetration. Furthermore, an important consideration for future work is the validation of biopsy precision and accuracy by measuring the distance between the target and location where the sample is taken from.
Also, a vital consideration for the future is to automate the process of targeting locations i.e. lesions and other areas of interest. Furthermore, human robot interaction in surgical scenarios will be considered in the next step to facilitate intuitive interaction with the system [36].

Author Contributions

“Conceptualization, M.Š., F.Š., B.Š., T.H., T.K. and B.J.; methodology, M.M, L.M, J.J and L.Š.; software, M.M, L.M, J.J and L.Š.; validation, M.M., L.M, J.J, L.Š., F.Š., B.Š and M.Š.; formal analysis, M.M., L.M, J.J. and L.Š.; investigation, M.Š., T.Z., M.M, L.M, J.J and L.Š.; resources, M.Š, F.Š., B.Š., T.H., T.K. and T.Z.; data curation, M.M, L.M, J.J and L.Š.; writing—original draft preparation, M.M, L.M, J.J., L.Š. and T.Z.; writing—review and editing, M.M, L.M, J.J., L.Š., T.Z., F.Š, B.Š. and M.Š.; visualization, M.M, L.M, J.J. and L.Š.; supervision, M.Š.; project administration, M.Š. All authors have read and agreed to the published version of the manuscript.“

Funding

NPOO and the Croatian Science Foundation through the project PRONOBIS – Robotically navigated prostate biopsy (project code: NPOO.C3.2.R3-I1.04.0181)

Informed Consent Statement

Not applicable

Acknowledgments

The authors would like to acknowledge the NPOO and the Croatian Science Foundation through the project PRONOBIS – Robotically navigated prostate biopsy (project code: NPOO.C3.2.R3-I1.04.0181). Also, we would like to acknowledge the academician Željko Kaštelan together with academician Bojan Jerbić who initiated the PRONOBIS project idea.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
PSA Prostate-Specific Antigen
DRE Digital Rectal Examination
MRI Magnetic Resonance Imaging
mpMRI Multiparametric MRI
DOF Degree of Freedom
ROS Robot Operatinf System
TRUS Transrectal Ultrasound
CNN Convolutional Neural Network
ViT Vision Transformer
AG-BCE Annotation-Guided Binary Cross-Entropy
MSDS Multi-Scale Deep Supervision
ROS Robot Operating System
TCP Tool Center Point
MedAP Medial Annotation Platform
SAM Segment Anything
DICOM Digital Imaging and Communications in Medicine
NIfTI (Neuroimaging Informatics Technology Initiative
DAF3D Deep Attentive Features for Prostate Segmentation
GUI Graphical User Interface

References

  1. The GLOBOCAN 2022 cancer estimates: Data sources, methods, and a snapshot of the cancer burden worldwide - PubMed.
  2. McNeal, J.E.; Redwine, E.A.; Freiha, F.S.; Stamey, T.A. Zonal distribution of prostatic adenocarcinoma. Correlation with histologic pattern and direction of spread. The American Journal of Surgical Pathology 1988, 12, 897–906. [Google Scholar] [CrossRef] [PubMed]
  3. Wegelin, O.; van Melick, H.H.E.; Hooft, L.; Bosch, J.L.H.R.; Reitsma, H.B.; Barentsz, J.O.; Somford, D.M. Comparing Three Different Techniques for Magnetic Resonance Imaging-targeted Prostate Biopsies: A Systematic Review of In-bore versus Magnetic Resonance Imaging-transrectal Ultrasound fusion versus Cognitive Registration. Is There a Preferred Technique? European Urology 2017, 71, 517–531. [Google Scholar] [CrossRef]
  4. Barrett, T.; de Rooij, M.; Giganti, F.; Allen, C.; Barentsz, J.O.; Padhani, A.R. Quality checkpoints in the MRI-directed prostate cancer diagnostic pathway. Nature Reviews. Urology 2023, 20, 9–22. [Google Scholar] [CrossRef] [PubMed]
  5. Patel, M.I.; Muter, S.; Vladica, P.; Gillatt, D. Robotic-assisted magnetic resonance imaging ultrasound fusion results in higher significant cancer detection compared to cognitive prostate targeting in biopsy naive men. Translational Andrology and Urology 2020, 9, 601–608. [Google Scholar] [CrossRef]
  6. Rouvière, O.; Jaouen, T.; Baseilhac, P.; Benomar, M.L.; Escande, R.; Crouzet, S.; Souchon, R. Artificial intelligence algorithms aimed at characterizing or detecting prostate cancer on MRI: How accurate are they when tested on independent cohorts? - A systematic review. Diagnostic and Interventional Imaging 2023, 104, 221–234. [Google Scholar] [CrossRef]
  7. Maris, B.; Tenga, C.; Vicario, R.; Palladino, L.; Murr, N.; De Piccoli, M.; Calanca, A.; Puliatti, S.; Micali, S.; Tafuri, A.; et al. Toward autonomous robotic prostate biopsy: a pilot study. International Journal of Computer Assisted Radiology and Surgery 2021, 16, 1393–1401. [Google Scholar] [CrossRef] [PubMed]
  8. Wetterauer, C.; Trotsenko, P.; Matthias, M.O.; Breit, C.; Keller, N.; Meyer, A.; Brantner, P.; Vlajnic, T.; Bubendorf, L.; Winkel, D.J.; et al. Diagnostic accuracy and clinical implications of robotic assisted MRI-US fusion guided target saturation biopsy of the prostate. Scientific Reports 2021, 11, 20250. [Google Scholar] [CrossRef]
  9. Lee, A.Y.; Yang, X.Y.; Lee, H.J.; Law, Y.M.; Huang, H.H.; Lau, W.K.; Lee, L.S.; Ho, H.S.; Tay, K.J.; Cheng, C.W.; et al. Multiparametric MRI-ultrasonography software fusion prostate biopsy: initial results using a stereotactic robotic-assisted transperineal prostate biopsy platform comparing systematic vs targeted biopsy. BJU International 2020, 126, 568–576. [Google Scholar] [CrossRef]
  10. Porpiglia, F.; De Luca, S.; Passera, R.; Manfredi, M.; Mele, F.; Bollito, E.; De Pascale, A.; Cossu, M.; Aimar, R.; Veltri, A. Multiparametric-Magnetic Resonance/Ultrasound Fusion Targeted Prostate Biopsy Improves Agreement Between Biopsy and Radical Prostatectomy Gleason Score. Anticancer Research 2016, 36, 4833–4840. [Google Scholar] [CrossRef]
  11. Chou, W.; Liu, Y. An Analytical Inverse Kinematics Solution with the Avoidance of Joint Limits, Singularity and the Simulation of 7-DOF Anthropomorphic Manipulators. 48, 117–132. [CrossRef]
  12. Zhang, Y.; Liang, D.; Sun, L.; Guo, X.; Jiang, J.; Zuo, S.; Zhang, Y. Design and experimental study of a novel 7-DOF manipulator for transrectal ultrasound probe. Science Progress 2020, 103, 0036850420970366. [Google Scholar] [CrossRef] [PubMed]
  13. Duan, H.; Zhang, Y.; Liu, H. Continuous Body Type Prostate Biopsy Robot for Confined Space Operation. IEEE Access 2023, 11, 113667–113677. [Google Scholar] [CrossRef]
  14. Ho, H.; Yuen, J.S.P.; Mohan, P.; Lim, E.W.; Cheng, C.W.S. Robotic transperineal prostate biopsy: pilot clinical study. Urology 2011, 78, 1203–1208. [Google Scholar] [CrossRef]
  15. Fletcher, P.; De Santis, M.; Ippoliti, S.; Orecchia, L.; Charlesworth, P.; Barrett, T.; Kastner, C. Vector Prostate Biopsy: A Novel Magnetic Resonance Imaging/Ultrasound Image Fusion Transperineal Biopsy Technique Using Electromagnetic Needle Tracking Under Local Anaesthesia. European Urology 2023, 83, 249–256. [Google Scholar] [CrossRef] [PubMed]
  16. Ipsen, S.; Wulff, D.; Kuhlemann, I.; Schweikard, A.; Ernst, F. Towards automated ultrasound imaging—robotic image acquisition in liver and prostate for long-term motion monitoring. Physics in Medicine & Biology 2021, 66, 094002. [Google Scholar] [CrossRef]
  17. Wang, W.; Pan, B.; Fu, Y.; Liu, Y. Development of a transperineal prostate biopsy robot guided by MRI-TRUS image. The International Journal of Medical Robotics and Computer Assisted Surgery 2021, 17, e2266. [Google Scholar] [CrossRef]
  18. Stoianovici, D.; Kim, C.; Petrisor, D.; Jun, C.; Lim, S.; Ball, M.W.; Ross, A.; Macura, K.J.; Allaf, M. MR Safe Robot, FDA Clearance, Safety and Feasibility Prostate Biopsy Clinical Trial. IEEE/ASME transactions on mechatronics: a joint publication of the IEEE Industrial Electronics Society and the ASME Dynamic Systems and Control Division 2017, 22, 115–126. [Google Scholar] [CrossRef]
  19. Tilak, G.; Tuncali, K.; Song, S.E.; Tokuda, J.; Olubiyi, O.; Fennessy, F.; Fedorov, A.; Penzkofer, T.; Tempany, C.; Hata, N. 3T MR-guided in-bore transperineal prostate biopsy: A comparison of robotic and manual needle-guidance templates: Robotic Template for MRI-Guided Biopsy. Journal of Magnetic Resonance Imaging 2015, 42, 63–71. [Google Scholar] [CrossRef]
  20. Lim, S.; Jun, C.; Chang, D.; Petrisor, D.; Han, M.; Stoianovici, D. Robotic Transrectal Ultrasound Guided Prostate Biopsy. IEEE TRANSACTIONS ON BIOMEDICAL ENGINEERING 2019, 66, 2527–2537. [Google Scholar] [CrossRef]
  21. Li, X.; Li, C.; Fedorov, A.; Kapur, T.; Yang, X. Segmentation of prostate from ultrasound images using level sets on active band and intensity variation across edges. Medical Physics 2016, 43, 3090–3103. [Google Scholar] [CrossRef]
  22. Otsu, N. A Threshold Selection Method from Gray-Level Histograms. IEEE Transactions on Systems, Man, and Cybernetics 1979, 9, 62–66. [Google Scholar] [CrossRef]
  23. Minaee, S.; Boykov, Y.Y.; Porikli, F.; Plaza, A.J.; Kehtarnavaz, N.; Terzopoulos, D. Image Segmentation Using Deep Learning: A Survey. pp. 1–1. [CrossRef]
  24. Ronneberger, O.; Fischer, P.; Brox, T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In Proceedings of the Medical Image Computing and Computer-Assisted Intervention – MICCAI 2015; Navab, N.; Hornegger, J.; Wells, W.M.; Frangi, A.F., Eds., Cham, 2015; pp. 234–241.
  25. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An Image is Worth 16x16 Words: Transformers for Image Recognition at Scale, 2021, [arXiv:cs.CV/2010.11929].
  26. Chen, J.; Lu, Y.; Yu, Q.; Luo, X.; Adeli, E.; Wang, Y.; Lu, L.; Yuille, A.L.; Zhou, Y. TransUNet: Transformers Make Strong Encoders for Medical Image Segmentation.
  27. Jiang, H.; Imran, M.; Muralidharan, P.; Patel, A.; Pensa, J.; Liang, M.; Benidir, T.; Grajo, J.R.; Joseph, J.P.; Terry, R.; et al. MicroSegNet: A deep learning approach for prostate segmentation on micro-ultrasound images. Computerized Medical Imaging and Graphics 2024, 112, 102326. [Google Scholar] [CrossRef]
  28. Dlaka, D.; Švaco, M.; Chudy, D.; Jerbić, B.; Šekoranja, B.; Šuligoj, F.; Vidaković, J.; Romić, D.; Raguž, M. Frameless stereotactic brain biopsy: A prospective study on robot-assisted brain biopsies performed on 32 patients by using the RONNA G4 system. The international journal of medical robotics + computer assisted surgery: MRCAS 2021, 17, e2245. [Google Scholar] [CrossRef]
  29. Raguž, M.; Dlaka, D.; Orešković, D.; Kaštelančić, A.; Chudy, D.; Jerbić, B.; Šekoranja, B.; Šuligoj, F.; Švaco, M. Frameless stereotactic brain biopsy and external ventricular drainage placement using the RONNA G4 system. Journal of Surgical Case Reports 2022, 2022, rjac151. [Google Scholar] [CrossRef]
  30. Macenski, S.; Foote, T.; Gerkey, B.; Lalancette, C.; Woodall, W. Robot Operating System 2: Design, architecture, and uses in the wild. Science Robotics 2022, 7, eabm6074. [Google Scholar] [CrossRef] [PubMed]
  31. Huber, M.; Mower, C.E.; Ourselin, S.; Vercauteren, T.; Bergeles, C. LBR-Stack: ROS 2 and Python Integration of KUKA FRI for Med and IIWA Robots. Journal of Open Source Software 2024, 9, 6138. [Google Scholar] [CrossRef]
  32. Kirillov, A.; Mintun, E.; Ravi, N.; Mao, H.; Rolland, C.; Gustafson, L.; Xiao, T.; Whitehead, S.; Berg, A.C.; Lo, W.Y.; et al. Segment Anything. In Proceedings of the 2023 IEEE/CVF International Conference on Computer Vision (ICCV). IEEE, pp. 3992–4003. [CrossRef]
  33. Wang, Y.; Dou, H.; Hu, X.; Zhu, L.; Yang, X.; Xu, M.; Qin, J.; Heng, P.A.; Wang, T.; Ni, D. Deep Attentive Features for Prostate Segmentation in 3D Transrectal Ultrasound. IEEE Transactions on Medical Imaging 2019, 38, 2768–2778. [Google Scholar] [CrossRef] [PubMed]
  34. Xie, S.; Girshick, R.; Dollar, P.; Tu, Z.; He, K. Aggregated Residual Transformations for Deep Neural Networks. In Proceedings of the 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). IEEE, pp. 5987–5995. [CrossRef]
  35. Suligoj, F.; Heunis, C.M.; Sikorski, J.; Misra, S. RobUSt–An Autonomous Robotic Ultrasound System for Medical Imaging. IEEE Access 2021, 9, 67456–67465. [Google Scholar] [CrossRef]
  36. Zhang, J.; Li, Y.; Hu, F.; Chen, P.; Zhang, H.; Song, L.; Yu, Y. Human-Robot Interaction of a Craniotomy Robot Based on Fuzzy Model Reference Learning Control. Transactions of FAMENA 2024, 48, 155–171. [Google Scholar] [CrossRef]
Figure 1. Proposed workflow of robotically guided prostate biopsy. The robot performs prostate sweep scanning by rotating the ultrasound probe around its axis. The ultrasound images are captured on the PC, and analyzed in real-time. Using the segmentation masks, a 3D visualization is created and displayed on the GUI. Using GUI commands, the user can choose target coordinates for the biopsy needle placement on the individual slice images. The target coordinates are sent to the robot so it can place the aiming reticle attached to the probe in position for the clinician to take a biopsy sample at the selected location.
Figure 1. Proposed workflow of robotically guided prostate biopsy. The robot performs prostate sweep scanning by rotating the ultrasound probe around its axis. The ultrasound images are captured on the PC, and analyzed in real-time. Using the segmentation masks, a 3D visualization is created and displayed on the GUI. Using GUI commands, the user can choose target coordinates for the biopsy needle placement on the individual slice images. The target coordinates are sent to the robot so it can place the aiming reticle attached to the probe in position for the clinician to take a biopsy sample at the selected location.
Preprints 156054 g001
Figure 2. Calibration of the US probe on the robot. (a) US probe mounted on the robot submerged in water for the calibration. (b) US image of the calibration cone with the aiming reticle for the calibration of the TCP.
Figure 2. Calibration of the US probe on the robot. (a) US probe mounted on the robot submerged in water for the calibration. (b) US image of the calibration cone with the aiming reticle for the calibration of the TCP.
Preprints 156054 g002
Figure 3. Ultrasound scanning procedure. A robot mounted US probe is inserted into the prostate phantom. By rotating the probe around its axis, the robot can perform a scan of the whole prostate.
Figure 3. Ultrasound scanning procedure. A robot mounted US probe is inserted into the prostate phantom. By rotating the probe around its axis, the robot can perform a scan of the whole prostate.
Preprints 156054 g003
Figure 4. MedAP - Medical Annotation Platform, examples. (a) CIRS 053L - easy visible edges that can be segmented with SAM. (b) CIRS 070L - partially indistinctive edges on the right that are manually annotated using polygons.
Figure 4. MedAP - Medical Annotation Platform, examples. (a) CIRS 053L - easy visible edges that can be segmented with SAM. (b) CIRS 070L - partially indistinctive edges on the right that are manually annotated using polygons.
Preprints 156054 g004
Figure 5. MedAP Workflow for training and testing dataset creation. The image shows complete data flow with prompting options and segmentation results.
Figure 5. MedAP Workflow for training and testing dataset creation. The image shows complete data flow with prompting options and segmentation results.
Preprints 156054 g005
Figure 6. Segmentation results. (a) - (c) are phantom 1 images and (d) - (f) are phantom 2 images. (a) and (d) show original images, (b) and (e) show segmentation results while (c) and (f) show segmentation masks.
Figure 6. Segmentation results. (a) - (c) are phantom 1 images and (d) - (f) are phantom 2 images. (a) and (d) show original images, (b) and (e) show segmentation results while (c) and (f) show segmentation masks.
Preprints 156054 g006
Table 1. Comparison of segmentation performance for the two models; DAF3D and MicroSegNet.
Table 1. Comparison of segmentation performance for the two models; DAF3D and MicroSegNet.
Fold no. DAF3D MicroSegNet
Dice Score Jaccard Score Dice Score Jaccard Score
1 0.905432 0.829091 0.936580 0.885725
2 0.897736 0.819591 0.931808 0.881644
3 0.899132 0.820070 0.932482 0.884236
4 0.906388 0.831164 0.930665 0.877613
5 0.903041 0.825669 0.926850 0.873178
Average 0.902346 0.825117 0.931677 0.880479
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.
Copyright: This open access article is published under a Creative Commons CC BY 4.0 license, which permit the free download, distribution, and reuse, provided that the author and preprint are cited in any reuse.
Prerpints.org logo

Preprints.org is a free preprint server supported by MDPI in Basel, Switzerland.

Subscribe

Disclaimer

Terms of Use

Privacy Policy

Privacy Settings

© 2025 MDPI (Basel, Switzerland) unless otherwise stated