Time‑aware deep neural networks for needle tip localization in 2D ultrasound

dc.contributor.authorMwikirize, Cosmas
dc.contributor.authorKimbowa, Alvin B.
dc.contributor.authorImanirakiza, Sylvia
dc.contributor.authorKatumba, Andrew
dc.contributor.authorNosher, John L.
dc.contributor.authorHacihaliloglu, Ilker
dc.date.accessioned2022-12-01T19:10:30Z
dc.date.available2022-12-01T19:10:30Z
dc.date.issued2021
dc.description.abstractAccurate placement of the needle is critical in interventions like biopsies and regional anesthesia, during which incorrect needle insertion can lead to procedure failure and complications. Therefore, ultrasound guidance is widely used to improve needle placement accuracy. However, at steep and deep insertions, the visibility of the needle is lost. Computational methods for automatic needle tip localization could improve the clinical success rate in these scenarios. Methods We propose a novel algorithm for needle tip localization during challenging ultrasound-guided insertions when the shaft may be invisible, and the tip has a low intensity. There are two key steps in our approach. First, we enhance the needle tip features in consecutive ultrasound frames using a detection scheme which recognizes subtle intensity variations caused by needle tip movement. We then employ a hybrid deep neural network comprising a convolutional neural network and long short-term memory recurrent units. The input to the network is a consecutive plurality of fused enhanced frames and the corresponding original B-mode frames, and this spatiotemporal information is used to predict the needle tip location. Results We evaluate our approach on an ex vivo dataset collected with in-plane and out-of-plane insertion of 17G and 22G needles in bovine, porcine, and chicken tissue, acquired using two different ultrasound systems. We train the model with 5000 frames from 42 video sequences. Evaluation on 600 frames from 30 sequences yields a tip localization error of 0.52 ± 0.06 mm and an overall inference time of 0.064 s (15 fps). Comparison against prior art on challenging datasets reveals a 30% improvement in tip localization accuracy. Conclusion The proposed method automatically models temporal dynamics associated with needle tip motion and is more accurate than state-of-the-art methods. Therefore, it has the potential for improving needle tip localization in challenging ultrasound-guided interventions.en_US
dc.identifier.citationMwikirize, C., Kimbowa, A. B., Imanirakiza, S., Katumba, A., Nosher, J. L., & Hacihaliloglu, I. (2021). Time-aware deep neural networks for needle tip localization in 2D ultrasound. International Journal of Computer Assisted Radiology and Surgery, 16(5), 819-827. https://doi.org/10.1007/s11548-021-02361-wen_US
dc.identifier.urihttps://doi.org/10.1007/s11548-021-02361-w
dc.identifier.urihttps://nru.uncst.go.ug/handle/123456789/5636
dc.language.isoenen_US
dc.publisherInternational Journal of Computer Assisted Radiology and Surgeryen_US
dc.subjectNeedle tip localizationen_US
dc.subjectUltrasounden_US
dc.subjectLSTMen_US
dc.subjectMinimally invasive proceduresen_US
dc.titleTime‑aware deep neural networks for needle tip localization in 2D ultrasounden_US
dc.typeArticleen_US
Files