Development and feasibility evaluation of an AR-assisted radiotherapy positioning system

1. Introduction

Radiotherapy is one of the primary treatments for cancers, and more than 50% of patients receive radiation therapy during the course of their illness (1, 2). A critical step of radiotherapy is lying the patient in the correct position on the couch of the linear accelerator for the accuracy of radiation dose delivery. Currently, patient positioning based on treatment room lasers and markers on skin or fixation devices is still routine in most radiotherapy departments. Many techniques were developed to improve the patient positioning accuracy, such as cone beam computed tomography (CBCT) and MRI-Linac. Although these techniques have improved patient positioning, there are some disadvantages: First, it makes radiotherapy more complex and expensive. Second, additional radiation dosages were delivered to the patient, which may cause unexpected consequences. Third, the complicated treatment procedures can increase the therapist’s workload, resulting in fatigue, such as fatigue caused by switching attention between printed treatment plans, screens, lasers, and markers. Last but not least, these techniques can only reflect the positioning errors at the time of scanning but cannot provide real-time and non-rigid positioning guidance (3, 4).

Augmented reality (AR) is a promising visualization technology developed on the basis of virtual reality. It allows people to experience a scenario where virtual and real objects coexist (5). In recent years, AR technology has been increasingly used in medicine, such as education and training (68) and hologram-guided surgery (911). Talbot et al. first utilized the AR technique to guide the positioning of radiotherapy patients (12), which was subsequently explored and improved by Tarutani et al. (13) and Johnson et al. (14). However, there are some limitations for these methods. On the one hand, the assembly based on display devices, cameras, and computing devices increases the complexity of the system, which is not conducive to convenient technical implementation. The user’s AR experience is significantly compromised due to the phantom’s AR contour being displayed on a two-dimensional screen. On the other hand, the virtual-real patient alignment is based on the operator’s human eye judgment, and there is a lack of effective object tracking methods.

Microsoft HoloLens 2 is a portable head-mounted AR device that integrates multiple necessary hardware and multi-functions such as computing, holographic display, and intelligent interaction, which may provide a solution with AR characteristics for patient positioning. In this paper, an AR-assisted patient positioning system based on HoloLens 2 was developed. A three-dimensional (3D) virtual model generated by treatment planning CT was anchored to the treatment position and visualized by the therapist with HoloLens 2. The innovation is that the therapist can adjust the couch under the guidance of this intuitive hologram and virtual coordinate derived from object tracking, until the real model and virtual model were registered. In addition, Vuforia SDK was used for isocenter calibration, virtual and real space coordinate system establishment, and patient tracking (15). The feasibility and accuracy of the system were evaluated in the clinical environment. As far as we know, our system is the first radiotherapy positioning method solely based on a head-mounted AR device, providing 3D object tracking and virtual coordinate indication.

2. Methods and materials2.1 System overview

The system provides assistance for radiotherapy therapists to perform radiotherapy positioning under an intuitive holographic guidance. In our proposed AR-assisted method, HoloLens 2 was the only required hardware. Moreover, a proprietary SDK, Vuforia, was introduced and worked with the front-facing cameras of HoloLens 2 for automatic registration and real-time tracking, which improved the stability of anchored hologram in physical space to a certain extent and achieved good registration accuracy. In addition to system development and data preparation, we also designed a complete experimental process from optimal conditions exploration to phantom testing in the actual radiotherapy positioning clinical environment, as shown in Figure 1.

FIGURE 1

www.frontiersin.org

Figure 1 The overall experimental for the proposed AR-assisted positioning system.

2.2 3D reconstruction and visualization of CT simulation image

One “ISO cube” (5 × 5 × 5 cm), one cube phantom (10 × 10 × 10 cm), and an anthropomorphic phantom (ATOM Dosimetry verification phantom) were used in this experiment. Lead markers were attached to the surface of phantoms according to the clinical routine. Vuforia markers were also attached to the “ISO cube” and two phantoms. CT simulation images were obtained with SIEMENS SOMATOM Confidence and uploaded to Varian Eclipse treatment planning system. Slice thickness was 3 mm. The treatment plan was formulated on the basis of the simulation CT. To simplify the experimental procedures, the reference points determined by three tiny lead markers were artificially defined as the isocenter. The DICOM RT images were exported to Python package and 3D Slicer programs to automatically reconstruct 3D models. Then, the 3D models and pre-designed user interface (UI) were imported into the AR scene and deployed to HoloLens 2 for holographic visualization (Figure 2).

FIGURE 2

www.frontiersin.org

Figure 2 The flow from CT simulation image acquisition to holographic model visualization.

2.3 Registration between real and virtual space

The lasers have been professionally calibrated before the experiment. An “ISO cube” (5 × 5 × 5 cm) used for daily QA (quality assurance) was used as a “registration cube”, and it was placed at an isocentric position so that its geometric center coincided with the mechanical isocenter of the linear accelerator with the aid of room lasers and the mechanical front-pointer (Figure 3). After the therapist put on the HoloLens 2 and initiated the test procedure, the front-facing camera of the HoloLens 2 detected the Vuforia marker attached to the “ISO cube”, the virtual cube would automatically be registered to the “ISO cube” (Figure 3). The therapist could perform the voice command “fixed” or the gesture interaction (Figure 3) to anchor the virtual “ISO cube”, and the correlation between virtual and real space coordinates was established. At the same time, the holographic phantom would be displayed on the HoloLens 2 automatically.

FIGURE 3

www.frontiersin.org

Figure 3 The establishment of the correlation between virtual and real space: precise placement of the “ISO cube” in the isocentric position with the aid of lasers and the mechanical front-pointer (A, B); automatic registration of virtual “ISO cube” and real “ISO cube” based on the HoloLens 2’s detection of Vuforia marker, and the arrow points to the virtual UI interface (C); the gesture interaction to anchor the correlation between virtual and real space (D).

2.4 Cube phantom and ATOM phantom test2.4.1 Cube phantom test

Like the procedure for positioning “ISO cube”, a cube phantom (10 × 10 × 10 cm) was placed at the isocenter of the linear accelerator. After the HoloLens 2 detected the Vuforia marker attached to the cube phantom, the position coordinates of the Vuforia marker in the established virtual space coordinate system were displayed in the virtual UI interface. Meanwhile, the coordinates of the geometric center of the cube phantom were also calculated on the basis of the known space relationships between it and Vuforia marker and were displayed. The “Offset” in the virtual UI interface reflected the deviation between the virtual and real geometric center (Figure 4). In addition, the spatial deviation was calculated with the following formula, and the results were averaged to determine the overall error:

FIGURE 4

www.frontiersin.org

Figure 4 Display of offset between virtual and real geometric centers in virtual UI interface (A) and the units of coordinate are in centimeters; the schematic diagram of partial exploratory conditions (B).

where X, Y, and Z denote the coordinates of the real geometric center in directions of lateral, vertical, and longitudinal, respectively, of the established virtual space coordinates. The factors that can impact the system’s accuracy were tested, including different angles, different distances, different room light intensities, and with or without camera occlusion and head movement (Figure 4). Different angles, including registration angles and tracking angles, refer to the angle between the HoloLens 2’s front-facing camera and the Vuforia marker on the “ISO cube” or the Vuforia marker on the cube phantom, respectively: 0° refers to the direction perpendicular to the Vuforia marker; 30°, 45°, 60°, and 90° are the included angles between the HoloLens 2’s front camera and the vertical direction, respectively, in the horizontal plane. The distances between the front camera of HoloLens 2 and the Vuforia marker are 30, 40, 50, 60, and 70 cm. The light intensity in the treatment room for daily use is divided into normal light intensity (405.0 Lux) and low light intensity (230.0 Lux). For camera occlusion, the therapist artificially covers the front camera of HoloLens 2 for a few seconds when the holographic phantom was displayed on the HoloLens 2. The spatial drift of hologram caused by light changes, camera occlusion, and head movement can be evaluated with the variations of the above formula:

D=(X2−X1)2+(Y2−Y1)2+(Z2−Z1)2(2)

where X1, Y1, and Z1 and X2, Y2, and Z2, denote the coordinates of the real geometric center before and after light intensity, camera occlusion, and head movement changes, respectively.

2.4.2 Anthropomorphic phantom test

An anthropomorphic phantom was used to test the HoloLens 2–based patient positioning system in the clinical environment (Figure 5). After the registration between real and virtual space was established under the optimal exploratory conditions and the holographic anthropomorphic phantom (blue model) in the treatment position was displayed, the therapist moved the treatment couch with the rough guidance of holographic phantom and the fine instructions of virtual coordinates until the virtual coordinates prompting 0,0,0. Then, CBCT was performed to evaluate the positioning errors (Figure 6).

FIGURE 5

www.frontiersin.org

Figure 5 The AR scene of the therapist’s perspective shows a virtual anthropomorphic phantom in the treatment position.

FIGURE 6

www.frontiersin.org

Figure 6 The registration between CBCT (A, D) and simulated positioning CT (B, C).

2.4.3 Statistical analysis

The experimental data were presented as mean ± SD and were analyzed by paired Student’s t-tests. P<0.05 was considered to represent a statistically significant difference.

3. Results3.1 Exploration of optimal conditions

The factors affecting accuracy and stability were confounding, and a method of controlling variates was adopted in the measurement process. The registration angle, tracking angle, and tracking distance affect the recognition and tracking of the Vuforia marker, which has further influence on the coordinate construction and the spatial position accuracy of the holographic model in the AR scenes. For the registration angle and tracking angle, we tested the system performance at 0°, 30°, 45°, and 60°, respectively, with a total of 40 measurements; for the distance, we tested between 30 and 70 cm with a step of 10 cm, and the total number of tests is 50. In addition, the light intensity, camera occlusion, and head movement may affect the accuracy and stability of signal feedback and detection of the depth sensors. The tests for the above three factors were performed 20 times, respectively. The results are summarized in Tables 13. The results of different registration angles were first recorded under the conditions that the tracking angle was consistent with the registration angle, the distance was 50 cm, light intensity was normal, and there was no camera occlusion and head movement. The results showed that there was no significant difference between 0° and 45° (P = 0.84), and significant differences between 0° and other angles (as shown in Table 1). Because 0° was easier to control, it was chosen as one of the optimal conditions. Next, the results of different tracking angles were recorded under the registration angle of 0°, and other conditions were the same as above, demonstrating significant differences between 0° and other angles (as shown in Table 2). Therefore, the tracking angle was consistent with the registration angle, and both were 0°, which can be regarded as one of the best conditions for further exploration. Then, the results of different tracking distances were recorded under the conditions that the registration angle and the tracking angle were 0°, light intensity was normal, the space deviation was minimized to 1.4 ± 0.3 mm at a distance of 40 cm, and there were significant differences between 40 cm and other distances (as shown in Table 3). The results of different light intensity in the treatment room under above optimal conditions were recorded, demonstrating no significant differences between the normal and the low light intensity (P = 0.83). The hologram drift produced by the light changes is 0.7 ± 0.4 mm. However, the program will not work without lights. The hologram drift caused by camera occlusion and head movement were 1.0 ± 0.6 mm and 1.5 ± 0.9 mm, respectively (Table 4). On the basis of the results above, the optimal conditions can be summarized as follows: the registration angle and the tracking angle were consistent at 0°, the distance was 40 cm, light intensity was normal, and there was no camera occlusion and head movement.

TABLE 1

www.frontiersin.org

Table 1 Errors at different registration angles.

TABLE 2

www.frontiersin.org

Table 2 Errors at different tracking angles.

TABLE 3

www.frontiersin.org

Table 3 Errors at different tracking distances.

TABLE 4

www.frontiersin.org

Table 4 The hologram drifts are caused by the light intensity change, camera occlusion, and head movement.

3.2 Accuracy results of the anthropomorphic phantom

The positioning errors in the X, Y, and Z directions were 3.1 ± 1.9 mm, 3.0 ± 2.8 mm, and 4.6 ± 2.8 mm, respectively. In addition, the angle deviation of Rtn was 0.26 ± 0.14°.

4. Discussion

In this study, an innovative AR-assisted radiotherapy positioning system was developed with HoloLens 2 as the only core hardware and provided patient tracking and virtual coordinate indication for radiation therapist, instead of relying on the human eye for virtual-real alignment (1214, 16). Compared with the related works, this improves the convenience (12, 16), accuracy (17), and practicality of AR guidance systems. The feasibility and accuracy of this method were evaluated in the actual clinical environment. Because of the high-precision requirement of radiotherapy positioning, it is necessary to fully explore the positioning method based on the HoloLens 2 and Vuforia SDK, avoiding the factors that increase the offset between the virtual object and real object. Opposite ideas were adopted in the two stages of the experimental process. We adopted fixing the real object and then watching the virtual coordinates to explore optimal conditions based on the cube phantom. On the contrary, in the anthropomorphic phantom positioning stage, the real object was transferred under the guidance of the virtual coordinates to simulate the actual process. The results from the optimal condition exploration stage show that some factors indeed influence the accuracy of the proposed AR-assisted system, such as the distance between HoloLens 2 and Vuforia marker, as well as the registration and tracking angles, which were related to the size and plane attributes of the Vuforia marker used, respectively. In addition, camera occlusion and head movement caused hologram drift, and the randomness of drift may be brought more significant positioning errors to a certain extent.

Furthermore, the change of indoor light intensity for daily use had no significant effect on the results. Our experiments found that the virtual coordinates were prone to fluctuations, so we obtained relatively stable readings. However, the optimal application conditions obtained were not fully and accurately applied to anthropomorphic phantom–based positioning tests, except for the registration step based on the “ISO cube”. For one thing, the acquisition of real-time coordinates depended on the real-time tracking of the Vuforia marker attached to the anthropomorphic phantom. However, it was difficult for the therapist to maintain the optimal distance, tracking angle, and stationary head. For another, unlike the cube phantom, although we used a relatively flat chest, the surface still has a specific curvature, which led to the pattern distortion of the Vuforia marker and may impact the recognition and tracking process. We also found that the heavier anthropomorphic phantom caused the bed to settle, unlike the lighter cube phantom. At the same time, the virtual phantom remained in its original position, resulting in an inherent positional deviation between the virtual and real phantom that ultimately affected positioning accuracy. The good thing was that the bed settlement could be roughly corrected according to the general outline of the virtual phantom.

The results of the anthropomorphic phantom showed that, although the accuracy is in millimeter scale, it is still not up to the requirements of clinical use at present. However, considering the advantages that this method does bring. First, the virtual 3D model is directly presented in the treatment position so that the patient positioning process is more intuitive and straightforward. Second, the decrease of CBCT frequency reduces the cost of treatment and non-treatment dose for patients. Third, the valuable information is presented in real time in virtual form, which realizes paperless and screenless and solves the problem of attention shift, and optimizes human ergonomics with the adjustment of virtual UI interface (18) in the process of positioning. Fourth, unlike CBCT, which can only collect position information at a particular moment, AR-assisted positioning can continuously monitor and correct the patient’s position. Fifth, in combination with other artificial intelligence technology can realize the automatic identification of patients or fixed appliances to avoid human error. Finally, it helps radiologists remotely guide the patient’s radiotherapy positioning and bring high-quality patient education by presenting previously invisible beams, target areas, and organs at risk.

Therefore, it is necessary for us to continue to explore ways to reduce the positioning error in future work. On the one hand, we will consider two or more HoloLens 2 to collaborate, share, and exchange information and realize the complementation of spatial information. On the other hand, considering the limitation caused by the planar Vuforia marker and inspired by the surface-guided radiation therapy technology, we will try to adopt binocular stereo vision or structured light to obtain the overall surface information of the phantom as the basis for 3D holographic image reconstruction. In the future, we look forward to advancing hardware and software to make automatic registration more accurate and the spatial hologram more stable.

5. Conclusion

We developed an AR-assisted positioning system for radiotherapy based on HoloLens 2 and evaluated its feasibility and accuracy. The results showed that the system’s accuracy is in millimeters, which roughly meets clinical requirements and still needs to be further improved. However, considering the advantages, including intuitive visual guidance, radiation-free position verification, and intelligent interaction, the proposed AR method has a promising future.

Data availability statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Author contributions

All authors contributed to the study's conception and design. Material preparation, data collection, and analysis were performed by GZ, XL, and LW. All authors read and approved the final manuscript.

Funding

This work was supported by the National Natural Science Foundation of China (grant numbers 8217102892, 81972863, 81627901, and 82030082) and Natural Science Foundation of Shandong Province (grant numbers ZR2019LZL012).

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

1. Delaney G, Jacob S, Featherstone C, Barton M. The role of radiotherapy in cancer treatment: estimating optimal utilization from a review of evidence-based clinical guidelines. Cancer (2005) 104(6):1129–37. doi: 10.1002/cncr.21324

PubMed Abstract | CrossRef Full Text | Google Scholar

3. Brock KK, Hawkins M, Eccles C, Moseley JL, Moseley DJ, Jaffray DA, et al. Improving image-guided target localization through deformable registration. Acta Oncol (2008) 47(7):1279–85. doi: 10.1080/02841860802256491

PubMed Abstract | CrossRef Full Text | Google Scholar

4. Guckenberger M, Meyer J, Vordermark D, Baier K, Wilbert J, Flentje M. Magnitude and clinical relevance of translational and rotational patient setup errors: A cone beam CT study. Int J Radiat Oncol Biol Phys (2006) 65(3):934–42. doi: 10.1016/j.ijrobp.2006.02.019

PubMed Abstract | CrossRef Full Text | Google Scholar

5. Cipresso P, Giglioli I, Raya MA, Riva G. The past, present, and future of virtual and augmented reality research: A network and cluster analysis of the literature. Front Psychol (2018) 9:2086. doi: 10.3389/fpsyg.2018.02086

PubMed Abstract | CrossRef Full Text | Google Scholar

6. Gnanasegaram JJ, Leung R, Beyea JA. Evaluating the effectiveness of learning ear anatomy using holographic models. J Otolaryn Head Neck Surg (2020) 49(1):63. doi: 10.1186/s40463-020-00458-x

CrossRef Full Text | Google Scholar

7. Condino S, Turini G, Parchi PD, Viglialoro RM, Piolanti N, Gesi M, et al. How to build a patient-specific hybrid simulator for orthopaedic open surgery: Benefits and limits of mixed-reality using the Microsoft HoloLens. J Healthc Eng (2018) 2018:5435097. doi: 10.1155/2018/5435097

PubMed Abstract | CrossRef Full Text | Google Scholar

8. House PM, Pelzl S, Furrer S, Lanz M, Simova O, Voges B, et al. Use of the mixed reality tool “VSI patient education” for more comprehensible and imaginable patient educations before epilepsy surgery and stereotactic implantation of DBS or stereo-EEG electrodes. Epilepsy Res (2020) 159:106247. doi: 10.1016/j.eplepsyres.2019.106247

PubMed Abstract | CrossRef Full Text | Google Scholar

9. Saito Y, Sugimoto M, Imura S, Morine Y, Ikemoto T, Iwahashi S, et al. Intraoperative 3D hologram support with mixed reality techniques in liver surgery. Ann Surg (2020) 271(1):e4–7. doi: 10.1097/SLA.0000000000003552

PubMed Abstract | CrossRef Full Text | Google Scholar

10. Peh S, Chatterjea A, Pfarr J, Schäfer JP, Weuster M, Klüter T, et al. Accuracy of augmented reality surgical navigation for minimally invasive pedicle screw insertion in the thoracic and lumbar spine with a new tracking device. Spine J (2020) 20(4):629–37. doi: 10.1016/j.spinee.2019.12.009

PubMed Abstract | CrossRef Full Text | Google Scholar

11. Li Y, Chen X, Wang N, Zhang W, Li D, Zhang L, et al. A wearable mixed-reality holographic computer for guiding external ventricular drain insertion at the bedside. J Neurosurg (2018), 131(5):1–8. doi: 10.3171/2018.4.JNS18124

CrossRef Full Text | Google Scholar

12. Talbot J, Meyer J, Watts R, Grasset R. A method for patient set-up guidance in radiotherapy using augmented reality. Australas Phys Eng Sci Med (2009) 32(4):203–11. doi: 10.1007/BF03179240

PubMed Abstract | CrossRef Full Text | Google Scholar

13. Tarutani K, Takaki H, Igeta M, Fujiwara M, Okamura A, Horio F, et al. Development and accuracy evaluation of augmented reality-based patient positioning system in radiotherapy: A phantom study. In Vivo (2021) 35(4):2081–7. doi: 10.21873/invivo.12477

PubMed Abstract | CrossRef Full Text | Google Scholar

14. Johnson PB, Jackson A, Saki M, Feldman E, Bradley J. Patient posture correction and alignment using mixed reality visualization and the HoloLens 2. Med Phys (2022) 49(1):15–22. doi: 10.1002/mp.15349

PubMed Abstract | CrossRef Full Text | Google Scholar

15. Frantz T, Jansen B, Duerinck J, Vandemeulebroucke J. Augmenting microsoft's HoloLens with vuforia tracking for neuronavigation. Healthc Technol Lett (2018) 5(5):221–5. doi: 10.1049/htl.2018.5079

PubMed Abstract | CrossRef Full Text | Google Scholar

16. Li C, Lu Z, He M, Sui J, Lin T, Xie K, et al. Augmented reality-guided positioning system for radiotherapy patients. J Appl Clin Med Phys (2022) 23(3):e13516. doi: 10.1002/acm2.13516

PubMed Abstract | CrossRef Full Text | Google Scholar

17. Cosentino F, John NW, Vaarkamp J. RAD-AR: RADiotherapy - augmented reality, in: 2017 International Conference on Cyberworlds (CW), Chester, (2017) (UK: IEEE). pp. 226–8.

Google Scholar

18. Janabi HF, Aydin A, Palaneer S, Macchione N, Al-Jabir A, Khan MS, et al. Effectiveness of the HoloLens mixed-reality headset in minimally invasive surgery: a simulation-based feasibility study. Surg Endosc (2020) 34(3):1143–9. doi: 10.1007/s00464-019-06862-3

PubMed Abstract | CrossRef Full Text | Google Scholar

留言 (0)

沒有登入
gif