tFUSFormer: Physics-guided super-resolution Transformer for simulation of transcranial focused ultrasound propagation in brain stimulation
This repository contains the implementation of neural networks designed to predict the simulation of transcranial focused ultrasound (tFUS) propagation in brain stimulation. tFUS brain stimulation is a non-invasive medical technique that uses focused sound waves to modulate neuronal activity in specific brain regions, offering potential therapeutic and research applications.
- Fast super-resolution convolutional neural network (FSRCNN), squeeze-and-excitation super-resolution residual network (SE-SRResNet), super-resolution generative adversarial network (SRGAN), and tFUSFormer architecture for accurate prediction of tFUS focal region.
- Custom loss functions that combine MSE, IoU loss, and distance function to enhance prediction accuracy.
- Data pre-processing and loading modules for efficient handling of the tFUS simulation data.
- Evaluation metrics to assess model performance, including IoU score and distance D between maximum pressure points.
- python 3.9
- torch-1.11.0+cu113
- cudatoolkit-11.3.1
To set up the project environment:
- Clone the repository:
git clone https://github.com/iangilan/tFUSFormer.git
The dataset used in this project consists of:
- You can download training, validation, and test datasets from here.
- Due to privacy concerns, CT images are excluded.
You can download the pre-trained models from here.
- Locate your test dataset in your local storage.
- Edit config.py according to the user's need.
- Edit the data loaders
tFUS_dataloader.py
to load and preprocess the data. - Train the model using
python train_model.py
. - Evaluate the model's performance on test data using
python test_model.py
.
- The
FSRCNN_1ch
model is a 3D FSRCNN that consists of encoder and decoder blocks, designed for extracting features and predicting a focal region. - The
SESRResNet_1ch
model is a 3D SESRResNet, designed for extracting features and predicting a focal region. - The
SRGAN_1ch
model is a 3D SRGAN, designed for extracting features and predicting a focal region. - The
tFUSFormer_1ch
model is a 3D tFUSFormer, designed for extracting features and predicting a focal region. - The
tFUSFormer_5ch
model is a 3D tFUSFormer, designed for extracting five physics features and predicting a focal region. - The architectures of all models are defined in
models.py
.
The model utilizes a combined loss function that incorporates MSE loss, IoU loss, and a distance function to mitigate specific challenges in tFUS focal volume prediction.
The model is evaluated based on the intersection over union (IoU) score and the distance between the maximum pressure points of the ground truth full-width at half maximum (FWHM) and the predicted FWHM of the focal volume, offering a comprehensive assessment of its prediction accuracy.
If you use this tool in your research, please cite the following paper:
- M. Shin, M. Seo, S.-S. Yoo, K. Yoon. "tFUSFormer: Physics-guided super-resolution Transformer for simulation of transcranial focused ultrasound propagation in brain stimulation.", TBD
- BibTeX formatted citation
@misc{shin2024tfusformer,
title={{tFUSFormer}: {P}hysics-guided super-resolution {T}ransformer for simulation of transcranial focused ultrasound propagation in brain stimulation},
author={Minwoo Shin and Minjee Seo and Seung-Schik Yoo and and Kyungho Yoon},
year={2024},
eprint={},
archivePrefix={},
primaryClass={},
doi={} }
For the SE-SRResNet, please cite the following paper:
- M. Shin, Z. Peng, H.-J. Kim, S.-S. Yoo, K. Yoon. "Multivariable-incorporating super-resolution residual network for transcranial focused ultrasound simulation", Computer Methods and Programs in Biomedicine, 237, 2023, 107591
- BibTeX formatted citation
@article{shin2023,
author = "Minwoo Shin and Zhuogang Peng and Hyo-Jin Kim and Seung-Schik Yoo and Kyungho Yoon",
title = "{Multivariable-incorporating super-resolution residual network for transcranial focused ultrasound simulation}",
year = {2023},
volume = {237},
pages = {107591},
journal = {Comput. Meth. Programs Biomed.},
doi = {10.1016/j.cmpb.2023.107591} }
For the manifold discovery and analysis (MDA) analysis, please cite the following paper:
- Md Tauhidul Islam, Zixia Zhou, Hongyi Ren, Masoud Badiei Khuzani, Daniel Kapp, James Zou, Lu Tian, Joseph C. Liao and Lei Xing. 2023, "Revealing Hidden Patterns in Deep Neural Network Feature Space Continuum via Manifold Learning", Nature Communications, 14(1), p.8506.
- BibTeX formatted citation
@Article{Islam2023,
author={Islam, Md Tauhidul and Zhou, Zixia and Ren, Hongyi and Khuzani, Masoud Badiei and Kapp, Daniel and Zou, James and Tian, Lu and Liao, Joseph C. and Xing, Lei},
title={Revealing hidden patterns in deep neural network feature space continuum via manifold learning}, journal={Nature Communications},
year={2023},
month={Dec},
day={21},
volume={14},
number={1},
pages={8506},
doi={10.1038/s41467-023-43958-w} }
For any queries, please reach out to Minwoo Shin.