Rishit Dagli1,2, Atsuhiro Hibi2,3,4, Rahul G. Krishnan1,5, Pascal Tyrrell2,4,6
Departments of 1 Computer Science; 2 Medical Imaging, University of Toronto, Canada
3 Division of Neurosurgery, St Michael's Hospital, Unity Health Toronto, Canada
4 Institute of Medical Science; Departments of 5 Laboratory Medicine and Pathobiology; 6 Statistical Sciences, University of Toronto, Canada
This work presents NeRF-US, a method to train NeRFs in-the-wild for sound fields like ultrasound imaging data. Check out our website to view some results of this work.
This codebase is forked from the awesome Ultra-NeRF and Nerfbusters repository.
- First, install the pip package by running:
pip install nerfus
or you could also install the package from source:
git clone https://github.com/Rishit-Dagli/nerf-us
cd nerf-us
pip install -e .
- Now install the dependencies, if you use the
virtualenv
you could run:
pip install -r requirements.txt
If you use conda
you could run:
conda env create -f environment.yml
conda activate nerfus
- Install Nerfstudio and dependencies. Installation guide can be found install nerfstudio
We also use the branch nerfbusters-changes. You may have to run the viewer locally if you want full functionality.
cd path/to/nerfstudio
pip install -e .
pip install torch==1.13.1 torchvision functorch --extra-index-url https://download.pytorch.org/whl/cu117
pip install git+https://github.com/NVlabs/tiny-cuda-nn/#subdirectory=bindings/torch
- Install
binvox
to voxelize cubes
mkdir bins
cd bins
wget -O binvox https://www.patrickmin.com/binvox/linux64/binvox?rnd=16811490753710
cd ../
chmod +x bins/binvox
Check out the Tips section for tips on installing the requirements.
For data preparation, the cubes
directory contains modules for processing 3D data, including dataset handling (datasets3D.py
), rendering (render.py
), and visualization (visualize3D.py
). The data_modules
directory further supports data management with modules for 3D cubes and a general datamodule for the diffusion model.
The diffusion model is primarily implemented in the models
directory, which includes the core model definition (model.py
), U-Net architecture (unet.py
), and related utilities. The lightning
directory contains the training logic for the diffusion model, including loss functions (dsds_loss.py
) and the trainer module (nerfus_trainer.py
). The NeRF component is housed in the nerf
directory, which includes experiment configurations, utility functions, and the main pipeline for NeRF-US (nerfus_pipeline.py
).
.
βββ config (configuration files for the datasets and models)
β βββ shapenet.yaml (configuration file for the shapenet dataset)
β βββ synthetic-knee.yaml (configuration file for the diffusion model)
βββ environment.yml (conda environment file)
βββ nerfus (main codebase)
β βββ bins
β β βββ binvox (binvox executable)
β βββ cubes (making cubes from 3D data)
β β βββ __init__.py
β β βββ binvox_rw.py
β β βββ datasets3D.py
β β βββ render.py
β β βββ utils.py
β β βββ visualize3D.py
β βββ data
β βββ data_modules (data modules for cubes)
β β βββ __init__.py
β β βββ cubes3d.py
β β βββ datamodule.py (data module for diffusion model)
β βββ download_nerfus_dataset.py (script to download the diffusion model dataset)
β βββ lightning (training lightning modules for diffusion model)
β β βββ __init__.py
β β βββ dsds_loss.py (loss for diffusion model)
β β βββ nerfus_trainer.py (training code for diffusion model)
β βββ models (model definition for diffusion model)
β β βββ __init__.py
β β βββ fp16_util.py
β β βββ model.py
β β βββ nn.py
β β βββ unet.py
β βββ nerf (main codebase for the NeRF)
β β βββ experiment_configs (configurations for the Nerfacto experiments)
β β β βββ __init__.py
β β β βββ nerfacto_experiments.py
β β β βββ utils.py
β β βββ nerfbusters_utils.py (utils for nerfbusters)
β β βββ nerfus_config.py (nerfstudio method configurations for the NeRF-US)
β β βββ nerfus_pipeline.py (pipeline for NeRF-US)
β βββ run.py (training script for diffusion model)
β βββ utils (utility functions for the NeRF training)
β βββ __init__.py
β βββ metrics.py
β βββ utils.py
β βββ visualizations.py
βββ requirements.txt (requirements file we use)
First, download either the synthetic knee cubes or the synthetic phantom cubes dataset:
.
βββ config
β βββ shapenet.yaml
β βββ synthetic-knee.yaml
βββ nerfus
β βββ bins
β β βββ binvox
β βββ data
β | βββ syn-knee
| | βββ syn-spi
We can now train the 3D diffusion model using the following command:
python nerfus/run.py --config config/synthetic-knee.yaml --name synthetic-knee-experiment --pt
This also automatically downloads Nerfbusters checkpoint on which we run adaptation.
Contrary to many other NeRF + Diffusion models we do not first train a NeRF and then continue training with the diffusion model as regularizee. Instead, we train we train it with the diffusion model from scratch.
We run the training using our method using Nerfstudio commands:
ns-train nerfus --data path/to/data nerfstudio-data --vis wandb
For our baselines, and experiments we directly use the Nerfstudio commands to train on the 10 individual datasets. For our ablation study, we do 3 ablations:
- for training without the border probability we just set the corresponding
lambda
to 0 (this could easily be made faster) - for training without the scaterring density we just set the corresponding
lambda
to 0 (this could easily be made faster) - for training without ultrasound rendering, we just use standard nerstudio commands
We can use standard Nerfstudio methods to run evaluations.
We share some tips on running the code and reproducing our results.
- Installing Nerfstudio especially on HPC systems can be tricky. We recommend installing
open3d
, andtiny-cuda-nn
before installing Nerfstudio separately and install it from source. We also recommend building these packages on the same GPU you plan to run it on. - When you install PyTorch especially on HPC systems, you will often end up with atleast two version of CUDA: one which is installed when you install PyTorch and is not a full version of CUDA and the other which is in the system. We highly recommend manually installing the same version of CUDA as in the system that PyTorch automatically installs.
- There are some parts of the code that do not run properly with PyTorch 2.x.
- We use virtualenv and use the
requirements.txt
file to install the required packages. While, we provide a condaenvironment.yml
(especially due to some Nerfstudio problems people might face), we have not tested this but expect it to work.
- We have currently optimized the code for and run all of the experiments on a A100 - 80 GB GPU. However, we have also tested the code on a A100 - 40 GB GPU where the inference and evaluation seem to work pretty well.
- In general, we would recommend a GPU with at least above 40 GB vRAM.
- We would recommend having at least 32 GB CPU RAM for the code to work well.
- While training the diffusion model, we recommend doing full-precision adaptation and not use FP-16.
This code base is built on top of, and thanks to them for maintaining the repositories:
If you find NeRF-US helpful, please consider citing:
@misc{dagli2024nerfusremovingultrasoundimaging,
title={NeRF-US: Removing Ultrasound Imaging Artifacts from Neural Radiance Fields in the Wild},
author={Rishit Dagli and Atsuhiro Hibi and Rahul G. Krishnan and Pascal N. Tyrrell},
year={2024},
eprint={2408.10258},
archivePrefix={arXiv},
primaryClass={cs.CV},
url={https://arxiv.org/abs/2408.10258},
}