Docket #: S21-073
Deep Learning Enabled Hybrid CT-MRI with Highly Sparse Sensory Data
Among the many medical imaging modalities, CT and MRI scans are utilized most often for imaging bone and soft tissue respectively. As such, physicians often require both images to fully diagnose patients and determine treatment plans. To better harness these advantages, this work has developed a deep-learning model for creating hybrid CT-MRI images from highly sparse sensory data. The key component of this technology utilizes the assumption that each hybrid, CT, and MRI image can be reconstructed from a shared latent feature space. Therefore, the training set only requires a mini-batch of CT and MRI images each which is fed into the cross-modal auto encoder framework. Preliminary testing has shown that analytic reconstruction of just the CT or MRI images show large streak artifacts due to under-sampled data, but the hybrid image has high similarity with the true images. The degree of fusion between CT and MRI can also be tuned using a fusion hyperparameter allowing physicians to create the most useful image.
Stage of Research
Applications
- Treatment planning for radiation therapy
- Hybrid image guided interventions (e.g., radiation therapy or surgery)
- Effectively fuse different modality images: low-dose CT and fast MRI reconstruction
Advantages
- Higher image quality for improved diagnostics and image guidance: hybrid images preserve high bone contrast and clear soft tissue dynamic range
- Creation of MRI image from CT sensory data, and vice vera.
- Reduced human intervention in regularization term design
Publications
- L. Zhu, Y. Chen, L. Liu, L. Xing and L. Yu, "Multi-sensor Learning Enables Information Transfer across Different Sensory Data and Augments Multi-modality Imaging," in IEEE Transactions on Pattern Analysis and Machine Intelligence, doi: 10.1109/TPAMI.2024.3465649.
Related Links
Similar Technologies
-
Radiotransparent audio-visual system to avoid pediatric patient anesthesia during radiation therapy and imaging. S23-306Radiotransparent audio-visual system to avoid pediatric patient anesthesia during radiation therapy and imaging.
-
Deep learning model for 3D computed tomography (CT) image reconstruction with single or a few views S18-464Deep learning model for 3D computed tomography (CT) image reconstruction with single or a few views
-
Automated radiation therapy treatment planning using a context-aware foundation model S24-203Automated radiation therapy treatment planning using a context-aware foundation model