MY NOTES:
结合之前读过类似的image denoising 或者图像重建的文章,最后都是转化成一个优化问题,在上次讲过的论文中,作者使用稀疏矩阵的方法来解决优化问题,最后的优化问题还是转化成一个迭代问题求解。
在本文中,同样优化问题改编为两部迭代求解最优化的问题,但是在迭代中,迭代的每一步,作者引入CRNN的框架,不再需要提前确定一些超参数来求解,完全依靠模型自己去训练和学习。
其中的CRNN block 可以作为梯度下降求解最优化问题的梯度下降迭代项。
本文最主要的就是图片中的那个结构。
Object:Accelerate the data acquisition of dynamic MRI.
The key issue of the problem is how to exploit the temporal correlations of the MR sequences to resolve aliasing artefacts.
However, the challenges of these optimisation based approaches are the following:
Deep learning based methods are gaining popularity for their accuracy and efficiency.
Unlike traditional approaches, the prior information and regularisation are learnt implicitly from data, without having to specify them in the training objective.
In this work, our contribution:
we propose a novel convolutional recurrent neural network (CRNN) method to reconstruct high quality dynamic MR image sequences from under-sampled data, termed CRNN-MRI.
We then show how this algorithm can be seen as a network architecture.
In particular, the proposed method consists of a CRNN block which acts as the proximal operator and a data consistency layer corresponding to the classical data fidelity term.
In addition, the CRNN block employs recurrent connections across each iteration step, allowing reconstruction information to be shared across the multiple iterations of the process.
As a consequence, the unique CRNN architecture jointly learns representations in a recurrent fashion evolving over both time sequences as well as iterations of the reconstruction process, effectively combining the benefits of traditional iterative methods and deep learning.
x ∈ \in ∈ C D C^D CD(complexed-valued MR images to be reconstructed)
D = D x D y T D_x D_y T DxDyT
y ∈ C M \in C^M ∈CM (M << D)
y is the under-sampled k-space measurements
D x D y D_x D_y DxDy is the width and height of frame;
T is the number of frames.
our problem: is to reconstruct x from y
argmin x _x x R ( x ) + λ ∣ ∣ y − F u x ∣ ∣ 2 2 R(x) + \lambda|| y - F_ux ||_2 ^2 R(x)+λ∣∣y−Fux∣∣22 (1)
F u F_u Fu: an undersampling Fourier encoding matrix
R: regularisation terms on x
λ \lambda λ: adjustment of data fidelity.
Always we can use variable splitting technique to handle the model above.
We introduce another variable z, and (1) can be transformed to (2)
argmin x , z _{x,z} x,z R ( z ) + λ ∣ ∣ y − F u x ∣ ∣ 2 2 R(z) + \lambda|| y - F_ux ||_2^2 R(z)+λ∣∣y−Fux∣∣22 + μ ∣ ∣ x − z ∣ ∣ 2 2 \mu||x-z||_2^2 μ∣∣x−z∣∣22 ~~~~~~~~~~~~~~~~~~~~ (2)
μ \mu μ: penalty parameter
Then we apply alternate minimisation over x and z, and then we get:
z ( i + 1 ) z^{(i+1)} z(i+1)= argmin z _z z R(z) + μ ∣ ∣ x − z ∣ ∣ 2 2 \mu||x-z||_2^2 μ∣∣x−z∣∣22 ~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~~ (3a)
x ( i + 1 ) = a r g m i n x λ ∣ ∣ y − F u x ∣ ∣ 2 2 + μ ∣ ∣ x − z ( i + 1 ) ∣ ∣ 2 2 x^{(i+1)} = argmin_x \lambda|| y - F_ux ||_2^2+\mu||x-z^{(i+1)}||_2^2 x(i+1)=argminxλ∣∣y−Fux∣∣22+μ∣∣x−z(i+1)∣∣22 ~~~~~~~~~~~~~ (3b)
Then during our iteration, x ( 0 ) = x μ = F u H y x^{(0)}=x_{\mu}=F_u^Hy x(0)=xμ=FuHy, which is a zero-filled reconstruction taken as an initialisation.
z can be seen as an intermediate state of optimisation process.
For MRI reconstruction, 3b is often regarded as a data consistency(DC) steps.
In this formula, F is full Fourier encoding matrix.
λ 0 = λ μ \lambda_0 = \frac{\lambda}{\mu} λ0=μλ is ratio of regularisation parameters.
Λ \Lambda Λ is diagonal matrix.
But in Deep-ADMM method , they unroll the traditional optimisation algorithm to.
x ( 0 ) → z ( 1 ) → x ( 1 ) → z ( 2 ) → x ( 2 ) → . . . . . . . . → z ( N ) → x ( N ) x^{(0)}\rightarrow z^{(1)}\rightarrow x^{(1)}\rightarrow z^{(2)}\rightarrow x^{(2)}\rightarrow ........\rightarrow z^{(N)}\rightarrow x^{(N)} x(0)→z(1)→x(1)→z(2)→x(2)→........→z(N)→x(N)
during this iteration, each state transition at stage(i) is an operation such as convolutions independently parametered by θ \theta θ.
But our model use CRNN, in which, each stage(i), as a learnt, recurrent, forward encoding step.
f i f_i fi ( x ( i − 1 ) , z ( i − 1 ) ; θ , y , λ , Ω ) (x^{(i-1)},z^{(i-1)}; \theta, y, \lambda, \Omega ) (x(i−1),z(i−1);θ,y,λ,Ω)
x r e c = f N ( f ( N − 1 ) ( . . . . . f 1 ( x μ ) ) ) x_{rec} = f_N(f_{(N-1)}(.....f_1(x_{\mu}))) xrec=fN(f(N−1)(.....f1(xμ)))
x r e c x_{rec} xrec: prediction of the network
x μ x_{\mu} xμ: sequence of under-sampled image with length T and also is the input of the network
f i ( x μ ; θ , λ , Ω ) f_i(x_{\mu};\theta,\lambda,\Omega) fi(xμ;θ,λ,Ω) is the network function for each iteration of optimization step.
N is the number of iterations.
The model can be written as:
x r n n ( i ) = x r e c ( i − 1 ) + C R N N ( x r e c ( i − 1 ) ) x_{rnn}^{(i)}=x_{rec}^{(i-1)} + CRNN(x_{rec}^{(i-1)}) xrnn(i)=xrec(i−1)+CRNN(xrec(i−1))
x r e c ( i ) x_{rec}^{(i)} xrec(i) = DC( x r n n ( i ) ; y , λ 0 , Ω x_{rnn}^{(i)}; y,\lambda_0,\Omega xrnn(i);y,λ0,Ω)
DC can be treated as a network layer.
CRNN block is to encode the update steps, and can be seen as one step of gradient descent in object minimisation.
this block can be seen as the distance required to move to the next state in the network.
CRNN: is the learnable block,
x r e c ( i ) x_{rec}^{(i)} xrec(i) is progressive reconstruction of the under-sampled image x μ x_{\mu} xμ at iteration i with x r e c ( 0 ) = x μ x_{rec}^{(0)} = x_{\mu} xrec(0)=xμ.
x r n n ( i ) x_{rnn}^{(i)} xrnn(i) is the intermediate reconstruction image before the DC layer.
x r e c ≈ x x_{rec}~\approx~x xrec ≈ x
x r n n ≈ z x_{rnn}~\approx~z xrnn ≈ z in equation 3
y is the acquired k-space sample.
we want to reconstruct x from y.
each state: CRNN update its internal state H given an input.
Furthermore, CRNN block comprised of 5 components:
That is the end of the model.
Appendix:
The example about bidirectional RNN