Optical coherence tomography (OCT) as an interferometric imaging technique, suffers from massive noise. Denoisingmethods are applied essentially to improve image quality in OCT community. The conventional methods rely onpost image processing algorithms such as non-local mean filtering, block-matching and 3D filtering algorithm. However, these conventional noise reduction methods could inevitably cause the destruction of image details, reduce the contrast at the edge of OCT images, and result in a degeneration of image quality. Current deep learning methods often ignorethespecificity of system, therefore haven’t taken advantages of the unique characteristics of different systems. In this work, we present a deep learning noise reduction method using the network architecture trained from synthetic OCTsignalswith random noise that are generated from the noise formation model characterized by our custom-built specificSD-OCT (Spectrum-Domain optical coherence tomography) system. We analyze the signal formation process and the noisegeneration pathway of our system, thereby enabling the construction of a noise formation model. DN-Unet (DenoisingUnity Network) is applied to train the datasets generated by our proposed noise formation model and the multi-to-singlestrategy is developed to enhance the network capability. Preliminary empirical results collectively showthat the networkcan reach an average of 25 dB signal to noise ratio (SNR) improvement while preserving detail structures, whichdemonstrates the effectiveness of our noise reduction method. This method has the potential to be adopted byother systems without the need for large number of golden-standard image generation.
|