[Paper notes] AIR-Net: GAN-based registration model

This article is the reading notes of the paper "Adversarial Image Registration with Application for MR and TRUS Image Fusion".

The article proposes a GAN-based supervised network model AIR-Net for multi-modal registration of MRI and transrectal ultrasound (TRUS) images, in which MRI is treated as a fixed image and transrectal ultrasound image is treated as a floating image .

Insert picture description here

Model by generator GGG and discriminatorDDD consists of two parts, in which the generator directly estimates a conversion parameterT est T_{est}from a floating image to a fixed imageTe s t; Then use an image sampler to use the estimated conversion parameters T est T_{est}Te s tAnd ground-truth conversion parameters T gt T_{gt}TgtThe floating image is registered; and the discriminator is used to determine whether the image pair passes T est T_{est}Te s tStill T gt T_{gt}Tgt For alignment.

The article regards three-dimensional images as multi-channel two-dimensional images,

The detailed structure of the generator is as follows:

One hollow convolutional layer (128 channels, dilation rate of 2) is used to expand the receptive field, two convolutional layers (128 channels, step size of 2) are used to reduce the resolution, and one contains three convolutions with residual connections The residual module of the accumulation layer (128 channels), a convolution layer (the size of the convolution kernel is 1 × 1 1\times11×1 , the number of channels is 8) used to reduce the number of parameters, two fully connected layers are used to get the final output, the output of the first fully connected layer is 256 dimensions, and the dimension and conversion of the output of the second fully connected layer The number of parameters is the same (if it is a three-dimensional rigid matching criterion, there are 6 parameters, and if it is a three-dimensional affine matching criterion, there are 12 parameters). If there is no special statement for the above convolution layer, the size of the convolution kernel is3 × 3 3\times33×3 , followed by a ReLU activation function.

The detailed structure of the discriminator is almost the same as that of the generator. The only difference is that the output dimension of the last fully connected layer is 1, and it is followed by a Sigmoid activation function.


判别器的损失为:
L ( D ) = − E T ∼ p g t ( T ) [ D ( I f , I m ) ] + E T ∼ p z ( T ) [ D ( I f , T ( I m ) ] \mathcal{L}(D)=-\mathbb{E}_{T \sim p_{g t}(T)}\left[D\left(I_{f}, I_{m}\right)\right]+\mathbb{E}_{T \sim p_{z}(T)}\left[D\left(I_{f}, T\left(I_{m}\right)\right]\right. L(D)=ETpgt(T)[D(If,Im)]+ETpfrom(T)[D(If,T(Im)]
Where $I_f $ and $I_m$ are fixed image and floating image respectively, ET ∼ pgt (T) [D (I f, I m)] \mathbb{E}_{T \sim p_{gt}(T)} \left[D\left(I_{f}, I_{m}\right)\right]ETpgt(T)[D(If,Im) ] Represents the expected loss of the discriminator of the aligned MR-TRUS image pair, andET ∼ pz (T) [D (I f, T (I m)] \mathbb{E}_{T \sim p_{ z}(T)}[D(I_{f}, T(I_{m})]ETpfrom(T)[D(If,T(Im) ] Represents the expected loss of the discriminator for randomly aligned image pairs.

生成器的损失为:
L ( G ) = E T ∼ p z ( T ) [ 1 − D ( I f , T e s t ( T ( I m ) ) ) + α ∥ T e s t − T − 1 ∥ 2 ] \mathcal{L}(G)=\mathbb{E}_{T \sim p_{z}(T)}\left[1-D\left(I_{f}, T_{e s t}\left(T\left(I_{m}\right)\right)\right)+\alpha\left\|T_{e s t}-T^{-1}\right\|^{2}\right] L(G)=ETpfrom(T)[1D(If,Te s t(T(Im)))+aTe s tT12 ]
where∣ ∣ T eset − T − 1 ∣ ∣ 2 ||T_{eset}-T^{-1}||^2TesetT12 is the Euclidean distance between the estimated transformation and the randomly generated transformation.

In training, a method similar to WGAN (Wasserstein GAN) is used. In order to ensure the stability of training, after each update of the parameters of the discriminator network, the parameters of the discriminator network need to be clipped to a certain range, and the value of the clipping parameter is set to 0.1. And if the generator network is not trained once, the discriminator network will be updated twice.

Guess you like

Origin blog.csdn.net/zuzhiang/article/details/107418421