KEYWORDS: Medical image reconstruction, Cone beam computed tomography, Education and training, Deep learning, Brain, Computed tomography, Error analysis, Data modeling, 3D modeling, Monte Carlo methods
Deep Learning (DL) image synthesis has gained increasing popularity for the reconstruction of CT and cone-beam CT (CBCT) images, especially in combination with physically-principled reconstruction algorithms. However, DL synthesis is challenged by the generalizability of training data and noise in the trained model. Epistemic uncertainty has proven as an efficient way of quantifying erroneous synthesis in the presence of out-of-domain features, but its estimation with Monte Carlo (MC) dropout requires a large number of inference runs, variable as a function of the particular uncertain feature. We propose a single-pass method–the Moment Propagation Model–which approximates the MC dropout by analytically propagating the statistical moments through the network layers, removing the need for multiple inferences and removing errors in estimations from insufficient dropout realizations. The proposed approach jointly computes the change of the expectation and the variance of the input (first two statistical moments) through each network layer, where each moment undergoes a different numerical transformation. The expectation is initialized as the network input; the variance is solely introduced at dropout layers, modeled as a Bernoulli process. The method was evaluated using a 3D Bayesian conditional generative adversarial network (GAN) for synthesis of high-quality head MDCT from low-quality intraoperative CBCT reconstructions. 20 pairs of measured MDCT volumes (120kV, 400 to 550mAs) depicting normal head anatomy, and simulated CBCT volumes (100 to120kV, 32 to 200mAs) were used for training. Scatter, beam-hardening, detector lag and glare were added to the simulated CBCT and were corrected (assuming unknown) prior to reconstruction. Epistemic uncertainty was estimated for 30 heads (outside of the training set) containing simulated brain lesions using the proposed single-pass propagation model, and results were compared to the standard 200-pass dropout approach. Image quality and quantitative accuracy of the estimated uncertainty of lesions and other anatomical sites were further evaluated. The proposed propagation model captured >2HU increase in epistemic uncertainty caused by various hyper- and hypo-density lesions, with <0.31HU error over the brain compared to the reference MC dropout result at 200 inferences and <0.1HU difference to a converged MC dropout estimate at 100 inference passes. These findings indicate a potential 100-fold increase in computational efficiency of neural network uncertainty estimation. The proposed moment propagation model is able to achieve accurate quantification of epistemic uncertainty in a single network pass and is an efficient alternative to conventional MC dropout.
KEYWORDS: Video, Education and training, Deformation, 3D acquisition, Voxels, 3D image processing, Imaging systems, Endoscopy, 3D image reconstruction, Visualization
Purpose: Navigating deep-brain structures in neurosurgery, especially under deformation from CSF egress, remains challenging due to the limitations of current robotic systems relying on rigid registration. This study presents the initial steps towards vision-based navigation leveraging Neural Radiance Fields (NeRF) to enable 3D neuroendoscopic reconstruction on the Robot-Assisted Ventriculoscopy (RAV) platform. Methods: An end-to-end 3D reconstruction and registration method using posed images was developed and integrated with the RAV platform. The hyperparameters for training the dual-branch network were first identified. Further experiments were conducted to evaluate reconstruction accuracy using projected error (PE) while varying the volume density threshold parameter. Results: A 3D volume was reconstructed using a simple linear trajectory for data acquisition with 300 frames and corresponding camera poses. The density volume threshold was varied to obtain an optimal value of 96.55 percentile, with a corresponding PE of 0.65 mm. Conclusions: Initial methods for end-to-end neuroendoscopic video reconstruction were developed in phantom studies. Experiments identified the optimal parameters, yielding a geometrically accurate reconstruction along with fast network convergence runtime of < 30 s. The method is highly promising for future clinical translation in realistic neuroendoscopic scenes. Future work will also develop a direct surface-to-volume registration method for improving reconstruction accuracy and runtime.
Unlike traditional ultrasound (US) transducers with rigid casing, flexible array transducers can be deformed to patientspecific geometries, thus potentially removing user dependence during real-time monitoring in radiotherapy. Proper transducer geometry estimation is required for the transducer's delay-and-sum (DAS) beamforming algorithm to reconstruct B-mode US images. The main contribution of this work is to track each element's position of the transducer to improve the quality of reconstructed images. An NDI Polaris Spectra infrared tracker was used to localize the custom design optical markers and interfaced using the Plus toolkit to estimate the transducer geometry in real-time. Each marker was localized with respect to a reference marker. Each element's coordinate position and azimuth angle were estimated using a polygon fitting algorithm. Finally, DAS was used to reconstruct the US image from radio-frequency channel data. Various transducer curvatures were emulated using gel padding placed on a CIRS phantom. The geometric accuracy of localizing the optical markers attached to the transducer surface was evaluated using 3D Cone-Beam Computed Tomography (CBCT). The tracked element positions' deviations compared to the CBCT images were measured to be 0.50±0.29 mm. The Dice score for the segmented target structure from reconstructed US images was 95.1±3.3% for above mentioned error in element position. We have obtained a high accuracy (<1mm error) when tracking the element positions with different random curvatures. The proposed method can be used for reconstructing US images to assist in real-time monitoring of radiotherapy, with minimal user dependence.
Neurosurgical techniques often require accurate targeting of deep-brain structures even in the presence of deformation due intervention and egress of Cerebrospinal Fluid (CSF) during surgical access. Prior work reported Simultaneous Localization and Mapping (SLAM) methods for endoscopic guidance using 3D reconstruction. In this work, methods for correcting the geometric distortion of a neuroendoscope are reported in a form that have been translated intraoperative use in first clinical studies. Furthermore, SLAM methods are evaluated in first clinical studies for real-time 3D endoscopic navigation with near real-time registration in the presence of deep-brain tissue deformation. A custom calibration jig with swivel mounts was designed and manufactured for neuroendoscope calibration in the operating room. The process is potentially suitable to intraoperative use while maintaining sterility of the endoscope, although the current calibration system was used in the Operating Room (OR) immediately following the case for offline analysis. A six by seven checkerboard pattern was used to obtain corner locations for calibration, and the method was evaluated in terms of Reprojection Error (RPE). Neuroendoscopic video was acquired under an IRB-approved clinical study, demonstrating rich vascular features and other structures on the interior walls of the lateral ventricles for 3D point-cloud reconstruction. Geometric accuracy was evaluated in terms of Projected Error (PE) on a ground truth surface defined from MR or cone-beam CT (CBCT) images. Intraoperative neuroendoscope calibration was achieved with sub-pixel [0.61 ± 0.20 px] error. The calibration yielded a focal length of 816.42 px and 822.71 px in X and Y directions respectively, along with radial distortion coefficients of -0.432 (first order term [𝑘1]) and 0.158 (second order term [𝑘2]). The 3D reconstruction was performed successfully with a PE of 0.23 ± 0.15 mm compared to the ground truth surface. The system for neuroendoscopic guidance based on SLAM 3D point-cloud reconstruction provided a promising platform for the development of 3D neuroendoscopy. The studies reported in this work presented an important means of neuroendoscope calibration in the OR and provided preliminary evidence for accurate 3D video reconstruction in first clinical studies. Future work aims to further extend the clinical evaluation and improve reconstruction accuracy using ventricular shape priors.
High-precision image-guided neurosurgery – especially in the presence of brain shift – would benefit from intraoperative image quality beyond the conventional contrast-resolution limits of cone-beam CT (CBCT) for visualization of the brain parenchyma, ventricles, and intracranial hemorrhage. Deep neural networks for 3D image reconstruction offer a promising basis for noise and artifact reduction, but generalizability can be challenged in scenarios involving features previously unseen in training data. We propose a 3D deep learning reconstruction framework (termed “DL-Recon”) that integrates learning-based image synthesis with physics-based reconstruction to leverage strengths of each. A 3D conditional GAN was developed to generate synthesized CT from CBCT images. Uncertainty in the synthesis image was estimated in a spatially varying, voxel-wise manner via Monte-Carlo dropout and was shown to correlate with abnormalities or pathology not present in training data. The DL-Recon approach improves the fidelity of the resulting image by combining the synthesized image (“DL-Synthesis”) with physics-based reconstruction (filtered back-projection (FBP) or other approaches) in a manner weighted by uncertainty – i.e., drawing more from the physics-based method in regions where model uncertainty is high. The performance of image synthesis, uncertainty estimation, and DL-Recon was investigated for the first time in real CBCT images of the brain. Variable input to the synthesis network was tested – including uncorrected FBP and precorrection with a simple (constant) scatter estimate – hypothesizing the latter to improve synthesis performance. The resulting uncertainty estimation was evaluated for the first time in real anatomical features not included in training (abnormalities and brain shift). The performance of DL-Recon was evaluated in terms of image uniformity, noise, and soft-tissue contrast-to-noise ratio in comparison to DL-Synthesis and FBP with a comprehensive artifact correction framework. DL-Recon was found to leverage the strengths of the learning-based and physics-based reconstruction approaches, providing a high degree of image uniformity similar to DL-Synthesis while accurately preserving soft-tissue contrast as in artifact-corrected FBP.
Purpose: Internal fixation of pelvic fractures is a challenging task requiring the placement of instrumentation within complex three-dimensional bone corridors, typically guided by fluoroscopy. We report a system for two- and three-dimensional guidance using a drill-mounted video camera and fiducial markers with evaluation in first preclinical studies.Approach: The system uses a camera affixed to a surgical drill and multimodality (optical and radio-opaque) markers for real-time trajectory visualization in fluoroscopy and/or CT. Improvements to a previously reported prototype include hardware components (mount, camera, and fiducials) and software (including a system for detecting marker perturbation) to address practical requirements necessary for translation to clinical studies. Phantom and cadaver experiments were performed to quantify the accuracy of video-fluoroscopy and video-CT registration, the ability to detect marker perturbation, and the conformance in placing guidewires along realistic pelvic trajectories. The performance was evaluated in terms of geometric accuracy and conformance within bone corridors.Results: The studies demonstrated successful guidewire delivery in a cadaver, with a median entry point error of 1.00 mm (1.56 mm IQR) and median angular error of 1.94 deg (1.23 deg IQR). Such accuracy was sufficient to guide K-wire placement through five of the six trajectories investigated with a strong level of conformance within bone corridors. The sixth case demonstrated a cortical breach due to extrema in the registration error. The system was able to detect marker perturbations and alert the user to potential registration issues. Feasible workflows were identified for orthopedic-trauma scenarios involving emergent cases (with no preoperative imaging) or cases with preoperative CT.Conclusions: A prototype system for guidewire placement was developed providing guidance that is potentially compatible with orthopedic-trauma workflow. First preclinical (cadaver) studies demonstrated accurate guidance of K-wire placement in pelvic bone corridors and the ability to automatically detect perturbations that degrade registration accuracy. The preclinical prototype demonstrated performance and utility supporting translation to clinical studies.
Purpose: Neuro-endoscopic surgery requires accurate targeting of deep-brain structures in the presence of deep-brain deformations (up to 10 mm). We report a deep learning-based method to solve deformable MR-to-CBCT registration using a joint synthesis and registration (JSR) network. Method: The JSR network first encodes the MR and CBCT images into latent variables via MR and CBCT encoders, which are then decoded by two branches: image synthesis branches for MR-CT and CBCT-CT synthesis; and a registration branch for intra-modality registration in an intermediate (synthetic) CT domain. The two branches are jointly optimized, encouraging the encoders to extract features pertinent to both synthesis and registration. The algorithm was trained and tested on a dataset of 420 paired volumes presenting a wide range of simulated deformations. The JSR method was trained in a semi-supervised manner and evaluated in comparison to an alternative, state-of-the-art, inter-modality registration method (VoxelMorph). Results: The JSR method achieved Dice of 0.67 ± 0.11, surface distance error (SD) of 0.47 ± 0.26 mm, and target registration error (TRE) of 2.23 ± 0.80 mm in a simulation study – each superior to the alternative methods considered in this work. Moreover, JSR maintained diffeomorphism and exhibited a fast runtime of 2.55 ± 0.03 s. Conclusion: The JSR algorithm demonstrates accurate, near real-time deformable registration of preoperative MRI to intraoperative CBCT and is potentially suitable to intraoperative guidance of intracranial neurosurgery.
Purpose: Recent neurosurgical techniques require accurate targeting of deep-brain structures even in the presence of deformation due to egress of cerebrospinal fluid (CSF) during surgical access. Prior work reported Structure-from-Motion (SfM) based methods for endoscopic guidance using 3D reconstruction. We are developing feature detection and description methods for a real-time 3D endoscopic navigation system using simultaneous localization and mapping (SLAM) to for accurate and near real-time registration. Methods: Feature detectors and descriptors were evaluated in SLAM reconstruction in anthropomorphic phantom studies emulating neuroendoscopy. The experimental system utilized a mobile UR3e robot (Universal Robots, Denmark) and ventriculoscope (Karl Storz, Tuttlingen, Germany) affixed to the end effector as a repeatable ventriculoscopy platform. Experiments were conducted to quantify optimal feature detection parameters in scale-space. Neuroendoscopic images acquired in traversal of the lateral and third ventricles provided a rich feature space of vessels and other structures on ventricular walls supporting feature detection and 3D point-cloud reconstruction. Performance was evaluated in terms of the mean number of features detected per frame and the algorithm runtime. Results: Parameter search in scale-space for feature detection demonstrated the dependence on the mean number of features per image and the points of diminishing return in parameter selection (e.g., the number of octaves and scale levels) and tradeoffs in runtime. Nominal parameters were identified as 3 octaves and 9 scale levels, with a mean number of features detected as 492 and 806 respectively. Conclusions: The system for neuroendoscopic guidance based on SLAM 3D point-cloud reconstruction provided a promising platform for the development of robot-assisted endoscopic neurosurgery. The studies reported in this work provided an essential basis for rigorous selection of parameters for feature detection. Future work aims to further develop the SLAM framework, assess the geometric accuracy of reconstruction, and translate methods to clinical studies.
Purpose: A method for fluoroscopic guidance of a robotic assistant is presented for instrument placement in pelvic trauma surgery. The solution uses fluoroscopic images acquired in standard clinical workflow and helps avoid repeat fluoroscopy commonly performed during implant guidance.
Approach: Images acquired from a mobile C-arm are used to perform 3D–2D registration of both the patient (via patient CT) and the robot (via CAD model of a surgical instrument attached to its end effector, e.g; a drill guide), guiding the robot to target trajectories defined in the patient CT. The proposed approach avoids C-arm gantry motion, instead manipulating the robot to acquire disparate views of the instrument. Phantom and cadaver studies were performed to determine operating parameters and assess the accuracy of the proposed approach in aligning a standard drill guide instrument.
Results: The proposed approach achieved average drill guide tip placement accuracy of 1.57 ± 0.47 mm and angular alignment of 0.35 ± 0.32 deg in phantom studies. The errors remained within 2 mm and 1 deg in cadaver experiments, comparable to the margins of errors provided by surgical trackers (but operating without the need for external tracking).
Conclusions: By operating at a fixed fluoroscopic perspective and eliminating the need for encoded C-arm gantry movement, the proposed approach simplifies and expedites the registration of image-guided robotic assistants and can be used with simple, non-calibrated, non-encoded, and non-isocentric C-arm systems to accurately guide a robotic device in a manner that is compatible with the surgical workflow.
Purpose: A method and prototype for a fluoroscopically-guided surgical robot is reported for assisting pelvic fracture fixation. The approach extends the compatibility of existing guidance methods with C-arms that are in mainstream use (without prior geometric calibration) using an online calibration of the C-arm geometry automated via registration to patient anatomy. We report the first preclinical studies of this method in cadaver for evaluation of geometric accuracy. Methods: The robot is placed over the patient within the imaging field-of-view and radiographs are acquired as the robot rotates an attached instrument. The radiographs are then used to perform an online geometric calibration via 3D-2D image registration, which solves for the intrinsic and extrinsic parameters of the C-arm imaging system with respect to the patient. The solved projective geometry is then be used to register the robot to the patient and drive the robot to planned trajectories. This method is applied to a robotic system consisting of a drill guide instrument for guidewire placement and evaluated in experiments using a cadaver specimen. Results: Robotic drill guide alignment to trajectories defined in the cadaver pelvis were accurate within 2 mm and 1° (on average) using the calibration-free approach. Conformance of trajectories within bone corridors was confirmed in cadaver by extrapolating the aligned drill guide trajectory into the cadaver pelvis. Conclusion: This study demonstrates the accuracy of image-guided robotic positioning without prior calibration of the Carm gantry, facilitating the use of surgical robots with simpler imaging devices that cannot establish or maintain an offline calibration. Future work includes testing of the system in a clinical setting with trained orthopaedic surgeons and residents.
Purpose. Deep brain stimulation is a neurosurgical procedure used in treatment of a growing spectrum of movement disorders. Inaccuracies in electrode placement, however, can result in poor symptom control or adverse effects and confound variability in clinical outcomes. A deformable 3D-2D registration method is presented for high-precision 3D guidance of neuroelectrodes. Methods. The approach employs a model-based, deformable algorithm for 3D-2D image registration. Variations in lead design are captured in a parametric 3D model based on a B-spline curve. The registration is solved through iterative optimization of 16 degrees-of-freedom that maximize image similarity between the 2 acquired radiographs and simulated forward projections of the neuroelectrode model. The approach was evaluated in phantom models with respect to pertinent imaging parameters, including view selection and imaging dose. Results. The results demonstrate an accuracy of (0.2 ± 0.2) mm in 3D localization of individual electrodes. The solution was observed to be robust to changes in pertinent imaging parameters, which demonstrate accurate localization with ≥20° view separation and at 1/10th the dose of a standard fluoroscopy frame. Conclusions. The presented approach provides the means for guiding neuroelectrode placement from 2 low-dose radiographic images in a manner that accommodates potential deformations at the target anatomical site. Future work will focus on improving runtime though learning-based initialization, application in reducing reconstruction metal artifacts for 3D verification of placement, and extensive evaluation in clinical data from an IRB study underway.
Purpose: Emerging deep-brain stimulation (DBS) procedures require a high degree of accuracy in placement of neuroelectrodes, even in the presence of deformation due to cerebrospinal fluid (CSF) egress during surgical access. We are developing ventriculoscope and hand-eye calibration methods for a robot-assisted guidance system to augment accurate electrode placement through transventricular approach. Methods: The ventriculoscope camera was modelled and calibrated for lens distortion using three different checkerboards, followed by evaluation on a separate board. The experimental system employed a benchtop UR3e robot (Universal Robots, Denmark) and ventriculoscope (Karl Storz, Tuttlingen, Germany) affixed to the end effector – referred to as the robotassisted ventriculoscopy (RAV) platform. Performance was evaluated in terms of three error metrics (RPE, FCE and PDE). Experiments were conducted to estimate the camera frame of reference using hand-eye calibration methods, and evaluated using a ChAruco board, using five different solvers and residual calibration error as the metric. Results: Camera calibration demonstrated subpixel (0.81 ± 0.11) px reprojection error and projection distance error (PDE) <0.5 mm. The error was observed to converge for any checkerboard used given a sufficient number of calibration images. The hand-eye calibration exhibited sub-mm residual error (0.26 ± 0.18) mm insensitive to the solver used. Conclusions: The RAV system demonstrates sub-mm ventriculoscope camera calibration error and robot-to-camera handeye residual error, providing a valuable platform for the development of advanced 3D guidance systems for emerging DBS approaches. Future work aims to develop structure-from-motion (SfM) methods to reconstruct a 3D optical scene using endoscopic video frames and further testing using rigid and deformable anatomical phantoms as well as cadaver studies.
Purpose: Deep-brain stimulation via neuro-endoscopic surgery is a challenging procedure that requires accurate targeting of deep-brain structures that can undergo deformations (up to 10 mm). Conventional deformable registration methods have the potential to resolve such geometric error between preoperative MR and intraoperative CT but at the expense of long computation time. New advances in deep learning methods offer benefits to inter-modality image registration accuracy and runtime using novel similarity metrics and network architectures. Method: An unsupervised deformable registration network is reported that first generates a synthetic CT from MR using CycleGAN and then registers the synthetic CT to the intraoperative CT using an inverse-consistent registration network. Diffeomorphism of the registration is maintained using deformation exponentiation “squaring and scaling” layers. The method was trained and tested on a dataset of CT and T1-weighted MR images with randomly simulated deformations that mimic deep-brain deformation during surgery. The method was compared to a baseline method using inter-modality deep learning registration, VoxelMorph. Results: The methods were tested on 10 pairs of CT/MR images from 5 subjects. The proposed method achieved a Dice score of 0.84±0.04 for the lateral ventricles, 0.72±0.09 for the 3rd ventricle, and 0.63±0.10 for the 4th ventricle, with target registration error (TRE) of 0.95±0.54 mm. The proposed method showed statistically significant improvement in both Dice score and TRE in comparison to inter-modality VoxelMorph, while maintaining a fast runtime of less than 3 seconds for a typical MR-CT pair of volume images. Conclusion: The proposed unsupervised image synthesis and registration network demonstrates the capability for accurate volumetric deformable MR-CT registration with near real-time performance. The method will be further developed for application in intraoperative CT (or cone-beam CT) guided neurosurgery.
Purpose: Intraoperative cone-beam CT (CBCT) plays an important role in neurosurgical guidance but is conventionally limited to high-contrast bone visualization. This work reports a high-fidelity artifacts correction pipeline to advance image quality beyond conventional limits and achieve soft-tissue contrast resolution even in the presence of multiple metal objects – specifically, a stereotactic head frame. Methods: A new metal artifact reduction (MAR) method was developed based on a convolutional neural network (CNN) that simultaneously estimates metal-induced bias and metal path length in the projection domain. To improve generalizability of the network, a physics-based method was developed to generate highly accurate simulated, metalcontaminated projection training data. The MAR method was integrated with previously proposed artifacts correction methods (lag, glare, scatter, and beam-hardening) to form a high-fidelity artifacts correction pipeline. The proposed methods were tested using an intraoperative CBCT system (O-arm, Medtronic) emulating a realistic setup in stereotactic neurosurgery, including nominal (20 cm) and extended (40 cm) field of view (FOV) protocols. Results: The physics-based data generation method provided accurate simulation of metal in projection data, including scatter, polyenergetic, quantum noise, and electronic noise effects. The artifacts correction pipeline was able to accommodate both 20 cm and 40 cm FOV protocols and demonstrated ~80% improvement in image uniformity and ~20% increase in contrast-to-noise ratio (CNR). Fully corrected images in the smaller FOV mode exhibited ~32% increase in CNR compared to the 40 cm FOV mode, showing the method’s ability to handle truncated metal objects outside the FOV. Conclusion: The image quality of intraoperative CBCT was greatly improved with the proposed artifacts correction pipeline, with clear improvement in soft-tissue contrast resolution (e.g., cerebral ventricles) even in the presence of a complex metal stereotactic frame. Such capability gives clearer visualization of structures of interest for intracranial neurosurgery, and it provides an important basis for future work aiming to deformably register preoperative MRI to intraoperative CBCT. Ongoing work includes clinical studies now underway.
Purpose: Pelvic fracture fixation is a challenging procedure that commonly relies on 2D fluoroscopic guidance to place guidewires within complex bone corridors. Prior work reported on a video-on-drill navigation system for guidewire insertion as a potential solution. Here, we assess performance across a range of hardware components to help guide the design of future system prototypes with respect to clinical requirements. Methods: The video-on-drill system uses a camera rigidly mounted on the drill and multimodality fiducial markers (optical and radio-opaque) to provide real-time trajectory visualization. This work reports on the selection of a new camera+lens configuration. Configurations were assessed across two cameras (referred to as the ArduCam and ELP) and five lens options (A45-A90). Clinical requirements were specified by an orthopaedic-surgeon in terms of the nominal drill operating distance (𝐷) and operating area (𝐴). Performance was evaluated in terms of the accuracy of fiducial marker pose estimation (δ𝐷), and the field of view (FOV). Results: At matched FOV, the accuracy for the ELP camera was significantly better (p < 0.01) with median δ𝐷 of 1.26 mm (1.1 mm IQR) compared to the ArduCam [median δ𝐷 of 1.85 mm (2.0 mm IQR)]. The accuracy of the A45 and A55 lens was found to be suitable (δ𝐷 < 2 mm) while providing sufficient FOV at nominal drill operating conditions. Conclusion: With respect to application requirements, the camera+lens combination (ELP+Ardu55) was identified to provide the best performance, serving as an important precursor to future design iterations of the video-on-drill system.
Purpose: Percutaneous fracture fixation is a challenging procedure that requires accurate interpretation of fluoroscopic images to insert guidewires through narrow bone corridors. We present a guidance system with a video camera mounted onboard the surgical drill to achieve real-time augmentation of the drill trajectory in fluoroscopy and/or CT.
Approach: The camera was mounted on the drill and calibrated with respect to the drill axis. Markers identifiable in both video and fluoroscopy are placed about the surgical field and co-registered by feature correspondences. If available, a preoperative CT can also be co-registered by 3D–2D image registration. Real-time guidance is achieved by virtual overlay of the registered drill axis on fluoroscopy or in CT. Performance was evaluated in terms of target registration error (TRE), conformance within clinically relevant pelvic bone corridors, and runtime.
Results: Registration of the drill axis to fluoroscopy demonstrated median TRE of 0.9 mm and 2.0 deg when solved with two views (e.g., anteroposterior and lateral) and five markers visible in both video and fluoroscopy—more than sufficient to provide Kirschner wire (K-wire) conformance within common pelvic bone corridors. Registration accuracy was reduced when solved with a single fluoroscopic view (TRE = 3.4 mm and 2.7 deg) but was also sufficient for K-wire conformance within pelvic bone corridors. Registration was robust with as few as four markers visible within the field of view. Runtime of the initial implementation allowed fluoroscopy overlay and/or 3D CT navigation with freehand manipulation of the drill up to 10 frames / s.
Conclusions: A drill-mounted video guidance system was developed to assist with K-wire placement. Overall workflow is compatible with fluoroscopically guided orthopaedic trauma surgery and does not require markers to be placed in preoperative CT. The initial prototype demonstrates accuracy and runtime that could improve the accuracy of K-wire placement, motivating future work for translation to clinical studies.
Purpose. We report the initial development of an image-based solution for robotic assistance of pelvic fracture fixation. The approach uses intraoperative radiographs, preoperative CT, and an end effector of known design to align the robot with target trajectories in CT. The method extends previous work to solve the robot-to-patient registration from a single radiographic view (without C-arm rotation) and addresses the workflow challenges associated with integrating robotic assistance in orthopaedic trauma surgery in a form that could be broadly applicable to isocentric or non-isocentric C-arms. Methods. The proposed method uses 3D-2D known-component registration to localize a robot end effector with respect to the patient by: (1) exploiting the extended size and complex features of pelvic anatomy to register the patient; and (2) capturing multiple end effector poses using precise robotic manipulation. These transformations, along with an offline hand-eye calibration of the end effector, are used to calculate target robot poses that align the end effector with planned trajectories in the patient CT. Geometric accuracy of the registrations was independently evaluated for the patient and the robot in phantom studies. Results. The resulting translational difference between the ground truth and patient registrations of a pelvis phantom using a single (AP) view was 1.3 mm, compared to 0.4 mm using dual (AP+Lat) views. Registration of the robot in air (i.e., no background anatomy) with five unique end effector poses achieved mean translational difference ~1.4 mm for K-wire placement in the pelvis, comparable to tracker-based margins of error (commonly ~2 mm). Conclusions. The proposed approach is feasible based on the accuracy of the patient and robot registrations and is a preliminary step in developing an image-guided robotic guidance system that more naturally fits the workflow of fluoroscopically guided orthopaedic trauma surgery. Future work will involve end-to-end development of the proposed guidance system and assessment of the system with delivery of K-wires in cadaver studies.
Purpose: Metal artifacts remain a challenge for CBCT systems in diagnostic imaging and image-guided surgery, obscuring visualization of metal instruments and surrounding anatomy. We present a method to predict C-arm CBCT orbits that will avoid metal artifacts by acquiring projection data that is least affected by polyenergetic bias. Methods: The metal artifact avoidance (MAA) method operates with a minimum of prior information, is compatible with simple mobile C-arms that are increasingly prevalent in routine use, and is consistent with either 3D filtered backprojection (FBP), more advanced (polyenergetic) model-based image reconstruction (MBIR), and/or metal artifact reduction (MAR) post-processing methods. MAA consists of the following steps: (i) coarse localization of metal objects in the field of view (FOV) via two or more low-dose scout views, coarse backprojection, and segmentation (e.g., with a U-Net); (ii) a simple model-based prediction of metal-induced x-ray spectral shift for all source-detector vertices (gantry rotation and tilt angles) accessible by the imaging system; and (iii) definition of a source-detector orbit that minimizes the view-to-view inconsistency in spectral shift. The method was evaluated in anthropomorphic phantom study emulating pedicle screw placement in spine surgery. Results: Phantom studies confirmed that the MAA method could accurately predict tilt angles that minimize metal artifacts. The proposed U-Net segmentation method was able to localize complex distributions of metal instrumentation (over 70% Dice coefficient) with 6 low-dose scout projections acquired during routine pre-scan collision check. CBCT images acquired at MAA-prescribed tilt angles demonstrated ~50% reduction in “blooming” artifacts (measured as FWHM of the screw shaft). Geometric calibration for tilted orbits at prescribed angular increments with interpolation for intermediate values demonstrated accuracy comparable to non-tilted circular trajectories in terms of the modulation transfer function. Conclusion: The preliminary results demonstrate the ability to predict C-arm orbits that provide projection data with minimal spectral bias from metal instrumentation. Such orbits exhibit strongly reduced metal artifacts, and the projection data are compatible with additional post-processing (metal artifact reduction, MAR) methods to further reduce artifacts and/or reduce noise. Ongoing studies aim to improve the robustness of metal object localization from scout views and investigate additional benefits of non-circular C-arm trajectories.
Pelvic trauma surgical procedures rely heavily on guidance with 2D fluoroscopy views for navigation in complex bone corridors. This “fluoro-hunting” paradigm results in extended radiation exposure and possible suboptimal guidewire placement from limited visualization of the fractures site with overlapped anatomy in 2D fluoroscopy. A novel computer visionbased navigation system for freehand guidewire insertion is proposed. The navigation framework is compatible with the rapid workflow in trauma surgery and bridges the gap between intraoperative fluoroscopy and preoperative CT images. The system uses a drill-mounted camera to detect and track poses of simple multimodality (optical/radiographic) markers for registration of the drill axis to fluoroscopy and, in turn, to CT. Surgical navigation is achieved with real-time display of the drill axis position on fluoroscopy views and, optionally, in 3D on the preoperative CT. The camera was corrected for lens distortion effects and calibrated for 3D pose estimation. Custom marker jigs were constructed to calibrate the drill axis and tooltip with respect to the camera frame. A testing platform for evaluation of the navigation system was developed, including a robotic arm for precise, repeatable, placement of the drill. Experiments were conducted for hand-eye calibration between the drill-mounted camera and the robot using the Park and Martin solver. Experiments using checkerboard calibration demonstrated subpixel accuracy [−0.01 ± 0.23 px] for camera distortion correction. The drill axis was calibrated using a cylindrical model and demonstrated sub-mm accuracy [0.14 ± 0.70 mm] and sub-degree angular deviation.
Purpose. Fracture reduction is a challenging part of orthopaedic pelvic trauma procedures, resulting in poor long-term prognosis if reduction does not accurately restore natural morphology. Manual preoperative planning is performed to obtain target transformations of target bones – a process that is challenging and time-consuming even to experts within the rapid workflow of emergent care and fluoroscopically guided surgery. We report a method for fracture reduction planning using a novel image-based registration framework. Method. An objective function is designed to simultaneously register multi-body bone fragments that are preoperatively segmented via a graph-cut method to a pelvic statistical shape model (SSM) with inter-body collision constraints. An alternating optimization strategy switches between fragments alignment and SSM adaptation to solve for the fragment transformations for fracture reduction planning. The method was examined in a leave-one-out study performed over a pelvic atlas with 40 members with two-body and three-body fractures simulated in the left innominate bone with displacements ranging 0–20 mm and 0°–15°. Result. Experiments showed the feasibility of the registration method in both two-body and three-body fracture cases. The segmentations achieved Dice coefficient of median 0.94 (0.01 interquartile range [IQR]) and root mean square error (RMSE) of 2.93 mm (0.56 mm IQR). In two-body fracture cases, fracture reduction planning yielded 3.8 mm (1.6 mm IQR) translational and 2.9° (1.8° IQR) rotational error. Conclusion. The method demonstrated accurate fracture reduction planning within 5 mm and shows promise for future generalization to more complicated fracture cases. The algorithm provides a novel means of planning from preoperative CT images that are already acquired in standard workflow.
Access to the requested content is limited to institutions that have purchased or subscribe to SPIE eBooks.
You are receiving this notice because your organization may not have SPIE eBooks access.*
*Shibboleth/Open Athens users─please
sign in
to access your institution's subscriptions.
To obtain this item, you may purchase the complete book in print or electronic format on
SPIE.org.
INSTITUTIONAL Select your institution to access the SPIE Digital Library.
PERSONAL Sign in with your SPIE account to access your personal subscriptions or to use specific features such as save to my library, sign up for alerts, save searches, etc.