Single Image-based Enhancement Techniques for Underwater Optical Imaging

Article information

J. Ocean Eng. Technol. 2020;34(6):442-453
Publication date (electronic) : 2020 December 24
doi : https://doi.org/10.26748/KSOE.2020.030
1Graduate Student, Department of control and instrument engineering, Korea Maritime and Ocean University, Busan, Korea
2Senior Researcher, Maritime ICT R&D Center, Korea Institute of Ocean Science and Technology, Busan, Korea
Corresponding author Soo Mee Kim: +82-51-664-3041, smeekim@kiost.ac.kr
Received 2020 June 5; Revised 2020 November 25; Accepted 2020 November 27.

Abstract

Underwater color images suffer from low visibility and color cast effects caused by light attenuation by water and floating particles. This study applied single image enhancement techniques to enhance the quality of underwater images and compared their performance with real underwater images taken in Korean waters. Dark channel prior (DCP), gradient transform, image fusion, and generative adversarial networks (GAN), such as cycleGAN and underwater GAN (UGAN), were considered for single image enhancement. Their performance was evaluated in terms of underwater image quality measure, underwater color image quality evaluation, gray-world assumption, and blur metric. The DCP saturated the underwater images to a specific greenish or bluish color tone and reduced the brightness of the background signal. The gradient transform method with two transmission maps were sensitive to the light source and highlighted the region exposed to light. Although image fusion enabled reasonable color correction, the object details were lost due to the last fusion step. CycleGAN corrected overall color tone relatively well but generated artifacts in the background. UGAN showed good visual quality and obtained the highest scores against all figures of merit (FOMs) by compensating for the colors and visibility compared to the other single enhancement methods.

1. Introduction

Optical images generated by the reflected light from an object are useful and intuitive for visual monitoring in unknown underwater environments. On the other hand, reflected light is typically scattered and absorbed by water and floating particles. This light attenuation causes low visibility with haziness and color cast in the resulting optical image (Mobley, 1994). The color cast effect produces a greenish or bluish hue in underwater optical images due to different attenuations depending on the light wavelength. Red light is attenuated more than green and blue light because of its longer wavelength. Depending on the transmission distance, water attenuation can cause limited visibility due to the loss of light intensity and contrast.

Considerable efforts have been made to compensate for the limited visibility and color cast effect of underwater optical images. The focus of most studies is on estimating a distance map because light attenuation of a specific wavelength is dependent on the distance from the camera. Stereo imaging has been used in some studies to obtain a distance map or enhance image quality directly (Roser et al., 2014; Zhang et al., 2014). Stereo imaging has been used to recover underwater images through a physical image formation model and an estimated distance map. One other approach has enhanced the underwater image by utilizing multi-directional light sources and fusing these different light images (Treibitz and Schechner, 2012). This study focused solely on single image enhancement as opposed to multi-images. This is because the utilization of multi-image enhancements requires additional hardware devices that are not available as a general imaging platform in the context of an underwater environment.

Single image enhancement improves image quality using the information extracted from a given underwater image. A single image enhancement technique is to adopt prior physical knowledge of light. Dark channel prior (DCP) is popular and was proposed to improve the haze of outdoor images using a light transmission map estimated from the darkest color channel (He et al., 2011). DCP has been applied widely to remove the turbidity of underwater images. On the other hand, it is imperfect for estimating a light transmission map due to nearly zero red channel values in the underwater environment. Underwater DCP estimates light transmission, excluding the red channel (Drews et al., 2013). For single image enhancement, the maximum intensity prior calculates the difference between the maximum values of red, green, and blue channels to estimate the light transmission map (Carlevaris-Bianco et al., 2010).

Domain transform methods were applied to single image enhancement. A homomorphic filter was used to help suppress noise and amplify details in the image frequency domain (Luo et al., 2019). The wavelet transform has been used for denoising (Jian and Wen, 2017) or fusing images (Khan et al., 2016; Wang et al., 2017a). The gradient-domain transform method has been found to recover the original gradient instead of the image intensity itself based on the image formation model (Li et al., 2012; Mi et al., 2016; Park and Sim, 2017; Zetian et al., 2018). Typical image processing techniques have been applied for contrast and turbidity enhancements. Multiple image processing steps have been established to improve the contrast, noise, and color successively (Arnold-Bos et al., 2005; Bazeille et al., 2006; Ghani and Isa, 2014; Ghani and Isa, 2015). The image fusion method was also proposed to combine the characteristics enhanced in multiple image processing steps (Ancuti et al., 2017). The combined technique of domain transform and fusion was proposed to dehaze general color images in air (Cho et al., 2018).

Recent research has found that deep neural networks for underwater optical imaging can enhance underwater images directly (Anwar et al., 2018; Fabbri et al., 2018; Guo et al., 2019; Hou et al., 2018; Li et al., 2019a; Li et al., 2019b; Sun et al., 2018; Uplavikar et al., 2019; Wang et al., 2017b) or estimate inherent information, such as background light intensity and transmission maps (Cao et al., 2018; Li et al., 2018a; Li et al., 2018b). Training convolutional neural networks require huge pairs of underwater images and clean images, but clean images are difficult to obtain in an underwater environment. Therefore, some researchers have used indoor datasets of color images and the corresponding depth information (Anwar et al., 2018; Cao et al., 2018; Hou et al., 2018; Uplavikar et al., 2019) or applied unsupervised networks, such as a generative adversarial network (GAN), to produce the image pairs (Fabbri et al., 2018; Guo et al., 2019; Li et al., 2019b). A detailed review of underwater image enhancement techniques can be found elsewhere (Anwar and Li, 2019; Wang et al., 2019; Yang et al., 2019).

In this study, six single image enhancement methods were considered: original DCP, gradient transform method with Tarel’s and Peng’s transmission maps, image fusion of successive three image processing steps, and two GANs. Although original DCP and Tarel’s methods were designed to enhance the haze outdoor environments, they were applied to underwater images because underwater images suffer from haziness. The gradient-domain transform was selected from the domain transform methods available because it simplifies the image formation model, and two different methods proposed by Tarel and Hautière (2009) and Peng and Cosman (2017) were applied to estimate a transmission map. Peng and Cosman (2017). proposed specific ways to estimate a depth map and background lights considering the underwater environments. The image fusion method should be an effective way to improve by combining appropriate image processing techniques. Two GANs, CycleGAN and underwater GAN, which are designed for underwater image enhancement, were applied because GAN is a new remarkable deep learning algorithm for various applications of image processing. The enhancement performance was assessed quantitatively in terms of underwater image quality measure (UIQM) considering the colorfulness, sharpness, and contrast, underwater color image quality evaluation (UCIQE) considering color saturation, chroma, and contrast and gray world (GW) assumption as color correction metrics, and blur metric. The comparison results would help other researchers better understand the advantages and disadvantages of single image enhancement methods for their studies.

This paper is organized as follows. Section 2 describes the single image enhancement methods considered in this study. Qualitative and quantitative comparisons are reported in Section 3. Sections 4 and 5 discuss the results and outline the conclusions of this study.

2. Single Image Enhancement

The light transmitted through water can be defined by the simple image formation model as follows (Fattal, 2008):

(1) I(x)=J(x)t(x)+A(1t(x))
where I is the measured light intensity; J is the recovered original light intensity; t is the light transmission map, and A is the uniform background light. The intensity is conventionally expressed with three representative color channels: red (R), green (G), and blue (B). The single image enhancement method was considered for the recovery of the original light intensity. The information is derived solely from a single underwater image using this method. This includes methods that derive the transmission map and the background light to solve Eq. (1) with observation-based assumptions and those that combine conventional image processing techniques to enhance the deteriorated underwater color image quality. Recently, deep learning has been the center of research attention with regard to image enhancement. This section describes the DCP, gradient transform, image fusion, and two GAN methods for single image enhancement.

2.1 Dark Channel Prior Enhancement

He et al. (2011) reported that at least one of the R, G, B-color channels in a pixel tends to approach zero in a colorful image. They proposed DCP to estimate the transmission from the darkest color channel in a local patch of given observed light, I. The dark channel image, Jdark, of a clean original image can be defined by the minimum operator (min[ ]) over a local patch and three color channels. This will approach zero according to DCP as follows:

(2) Jdark(x)=minyΩ(x)[minc(r,g,b)[Jc(y)]]0
where c indicates each color channel in an image, and Ω(x) is a local patch centered at x. The patch size is a key parameter determining the performance of DCP enhancement. The transmission can be estimated by substituting the dark channel image in Eq. (2) into Eq. (1), as follows:
(3) t˜(x)=softmat[1minyΩ(x)[minc{r,g,b}[Ic(y)A]]]
The intensity of background light was first determined from the dark channel (Idark) of the underwater image, I, reflecting the amount of haziness in an underwater image. The pixel with the highest intensity in I was chosen as the background light among candidates in the top 10 percent of the brightest pixels in Idark. The initial transmission was estimated using the background light and Idark It was then refined using a soft matting operator to reduce the halo effect around the edges. Finally, the enhanced image was recovered with the refined transmission, (x), as follows:
(4) J˜(x)=I(x)Amax(t˜(x),ε)+A
where ε is the lowest transmission value and was set to 0.1 to avoid the zero denominator. The smaller patch size was used to estimate the initial transmission where the higher color saturation appeared in the final image (). The large patch size produced a halo effect around the edges. In this paper, the same patch size of 15×15 was used for all underwater images.

2.2 Gradient Transform Enhancement

The gradient transform method was derived by adopting a gradient of Eq. (1) (Li et al., 2012). Assuming that the transmission is constant in a local patch, the gradient of I may be represented simply as follows:

(5) I(x)tJ(x)

The gradient transform method estimates the transmission for obtaining the original image gradient, ∇J, from a given ∇I. The enhanced original image, , was then reconstructed from its gradient, ∇J, via the Poisson equation solver assuming Dirichlet boundary conditions. Based on the calculus of variations, the cost function can be defined using Eq. (6) to recover the image intensity from its gradient:

(6) argminJ˜F(J˜,J)dxdy=argminJ˜J˜J2dxdy
where is an estimate of J and can be determined by minimizing the integral of the difference between the derivatives of J and , as follows:
(7) FJ˜ddxFJ˜xddyFJ˜y=0

The Euler-Lagrange equation was adopted to minimize the integral of F( ) in (6) and the Poisson equation in Eq. (8) was then derived from Eq. (7):

(8) 2J˜=divJ

Here, ∇2 is Laplacian operator, and div is the divergence operator. Using the Dirichlet boundary condition, a boundary image, B of was defined as containing all zero pixel values, except for the boundary pixels of . Table 1 lists the estimated final recovered image, , from the Poisson equation (Simchony et al., 1990).

Image reconstruction from a given gradient

This paper considered two transmission estimation methods. Tarel and Hautière (2009) proposed patch-based processing, such as DCP, and adopted the Median operator Along Lines (MAL). This preserves the edges without the halo effect as well as the corners, unlike the classical median operator, as shown in Eq. (9):

(9) m(x)=MALyΩ(x)[minc[I(x)]]n(x)=m(x)MALyΩ(x)[|minc[I(y)]m(y)|]t(x)=1max[min[n(x),minc[I(x)]],0]

Ω(x) was defined as the line segments over a square patch centered at x. Each line segment must pass through the center of the square patch. The MAL performs a classical median filter along each line segment and then calculates the median of the median values of each line segment. The number of line segments and the patch size are key parameters determining the enhancement performance. This study set five line segments and a 61×61 patch size.

Peng and Cosman (2017) estimated the transmission from a depth map based on image blurriness and light absorption by water and floating particles and the attenuation coefficients (μc) corresponding to RGB colors, as follows:

(10) tc(x)=eμcd(x),c{r,g,b}

The depth map, d(x) can be calculated with two relative depth estimates, d0 and dmix . This is based on different light absorption phenomena of RGB colors and the mixed additional distance information relative to the closest distance as follows:

(11) d(x)=dact(d0+dmix(x))
where dact is a scaling constant to convert the sum of two depth estimates to the actual distance in meters and was set to 8 m. d0 represents the normalized closest distance from the underwater camera to the scene, as follows:
(12) d0=1maxx,c[|AcIc(x)|max[Ac,1Ak]]
k indicates the color channel with a maximum absolute difference between the background light and the input intensities.

dmix was determined using three depth-related values accommodating the different attenuations of RGB color signals and blurriness of the underwater image, as follows:

(13) dmix(x)=β{αddiff(x)+(1α)dred(x)}+(1β)dblur(x)

The mixing weights, α and β were determined from the average intensities of background light and the red channel in the input image as follows:

(14) α=11+e32(meanc[Ac]0.5),β=11+e32(meanx[Ir(x)]0.1)
ddiff and dred, expressed by Eqs. (15) and (16), respectively, were estimated assuming that the intensity of the red channel (Ir) easured increased as the scene point to the camera became closer.
(15) ddiff=1fN[maxyΩ(x)Ir(y)maxyΩ(x)[Ig(y),Ib(y)]]
(16) dred(x)=1fN[maxyΩ(x)Ir(y)]

The local patch size was set to 5×5 in max[ ] operator. fN[ ] normalizes the input value in the range of 0 to 1. dblur is estimated by the blurriness in the underwater image as follows:

(17) dblur(x)=1fN[[Fill[Iblur(x)]]

Blurriness was defined by the mean of edge information extracted at various Gaussian kernel sizes, as follows:

(18) Iblur(x)=maxΩ(x)[1Ll=1L|Igray(x)Gaussσ[Igray(x)]|]

Igray is a gray-scaled image of the underwater color image. Gauss is a Gaussian filter with a kernel size, σ of 2lL+1 at different kernel levels, {l = 1, ⋯, L}. L was set to 4. Fill[ ] operator performs a morphological opening to compensate for the sparse regions and holes in Iblur to become denser (Vincent, 1993). Calculation of background light intensity is required to determine the closest distance and the mixing weights. The background light intensity was determined by combining the maximum and minimum of the three background light components, Achigh, Acmean, and Acvar, as expressed in Eq. (19):

(19) Ac=γcmax[Achigh,Acmean,Acvar]+(1γc)min[Achigh,Acmean,Acvar]

Achigh is the mean value of the input image pixels, which correspond to the same locations of the top 0.1% pixels of an image blurriness map, Pblur (= softmat[Fill[Iblur]]). Acmean and Acvar are mean values of the input image pixels belonging to sub-regions that have the largest mean in Pblur and the lowest variance in the input image, respectively. The sub-regions were selected through an iterative process where an image was divided into four quadrants, and the region with the largest mean and the lowest variance among the quadrants were selected. This iteration continues until the quadrant size is 1,024 times smaller than the initial image. γc is the weight calculated by the exponential of the rate of pixels greater than 0.5, as shown in Eq. (20):

(20) γc=11+e32(rate[Ic>0.5]0.2)

First, the transmission of the red channel, tr, was determined using the estimated distance map in Eq. (11), background light in Eq. (19), and the attenuation coefficient, μr, was set to 1/7 for Ocean Type-I water (Zhao et al., 2015). The attenuation coefficient ratios between the red and other colors were used for transmission conversion from red to blue and green colors as follows:

(21) tk(x)=tr(x)μkμr=tr(x)Ar(0.00113λk+1.62517)Ak(0.00113λr+1.62517),k{g,b}

The wavelengths, λc of the red, green and blue colors are 620, 540, and 450 nm, respectively.

2.3 Image Fusion Enhancement

Image fusion enhances features in an image and fuses the enhanced images (Ancuti et al., 2017). This consists of three processing steps: white balancing, feature enhancing, and image fusing. In the image fusion method, white balancing compensates the red and blue signals giving them flatter distributions compared to the green signal, which tends to maintain its intensity through water somewhat.

(22) Ikwb=Ik+(IgmeanIkmean)(1Ik)Ig,k{r,b}
where Iwb is a white-balanced image for the red and blue channels. Icmean is an average value across all image pixels for each color channel.

After white balancing to correct for color casting by water, two feature enhancing techniques were applied independently. First, contrast-limited adaptive histogram equalization (CLAHE) was adopted instead of gamma correction in the original method to enhance image contrast (Icontrast). CLAHE performs an ordinary histogram equalization over the predefined patch with a limited clip level. The parameters of CLAHE are the patch size, Ω, and a histogram clip level, hl. Furthermore, unsharp masking was applied to sharpen features, such as an edge in an image, as follows:

(23) Iedge=Iwb+(IwbGaussσ[Iwb])

Three weight maps of Laplacian contrast, saturation, and saliency were determined to fuse two images enhanced from the previous step. The Laplacian contrast weight map, ωLap, depends on global contrast in an image and was calculated as the absolute value of the Laplacian filtered luminance signal of each enhanced image, as follows:

(24) ωLapE=|Laplacianσ[IYE]|,E{contraast,edge}
where IY is the luminance channel of an image. The kernel size, σ, of the Laplacian filter was set to 3×3. The saturation weight map, ωSat, reflects chromatic information in an image and was defined by the average difference between the luminance and three-color signal, as follows:
(25) ωSatE=13c{r,g,b}(IcEIYE)2

The saliency weight map, ωSal, represents the prominent features of an image in CIELAB color space (Achantay et al. 2009), and can be expressed using Eq. (26):

(26) ωSalE=c{L*,a*,b*}(meanx[IcE]Gaussσ[IcE])2

The Gaussian kernel size was set to 3×3. The final weight map for each enhanced image from the previous step, combined three weight maps with normalization, as follows:

(27) ωE=ωsumEωsumcontrast+ωsumedge

ωsumE=ωLapE+ωSatE+ωSalE+0.1. To fuse two enhanced images, Icontrast and Iedge, with the corresponding weights, ωcontrast and ωedge, multiscale fusion was applied to avoid artifacts in the low-frequency components of the fused image. Multiscale fusion performs Laplacian (Ll) and Gaussian (Gl) pyramids repetitively on the enhanced images and the weight maps, as shown in Eq. (28):

(28) J˜(x)=plELl[IE(x)]Gl[ωE(x)]

The Laplacian pyramid at each lth level quantifies the difference between an input image and its Gaussian filtered image after sampling-down the operation by a factor of 2. The size of the Gaussian kernel in the pyramid computation was set to 5×5. The number of pyramid levels, pl, was set to 3.

2.4 Generative Adversarial Network Enhancement

A convolutional neural network is a widely used deep learning network for many image processing applications. Robust training in a deep learning network requires huge datasets and a reliable ground truth. On the other hand, it is difficult to construct ground truth for underwater color images. Fabbri et al. (2018) proposed underwater GAN (UGAN) to increase the reliability of synthetic training data and enhance underwater color imagery using synthetic data. GAN typically consists of the generator and discriminator networks, and the training process is performed by optimizing the loss function in Eq. (29):

(29) minGmaxD𝔼j[log(D(J))]+𝔼I[log(1D(G(I)))]

The generator network (G) competes against discriminator (D) during training to generate exquisite fake images and deceive the discriminator. The discriminator is trained to distinguish a fake image from a generator. The loss function of UGAN was defined with the Wasserstein GAN loss function and gradient penalties.

(30) UGAN=minGmaxDWGAN(G,D)+η1L1(G)+η2GDL

Wasserstein loss function, WGAN, was proposed to solve the training issue induced by the general GAN loss function, which measures a difference between real and generated data distributions in Eq. (29) (Arjovsky et al., 2017). WGAN is expressed as follows:

(31) WGAN=𝔼[D(J)]𝔼[D(G(I))]+η0𝔼[(z^D(z^)21)2]

The weight, η0 was set to 10. L1 is defined by the L1-norm of the difference between the ground truth and the image predicted, JP, by the generator, as follows:

(32) L1=𝔼[JG(I)1]=𝔼[JJP1]
GDL is a function with respect to the gradient difference between the ground truth and the predicted image (Mathieu et al. 2015), as follows:
(33) GDL=x,y||Jx||JPx|+|Jy||JPy||

The hyper-parameters, η1 and η2 were set to 100 and 1, respectively.

Fig. 1 summarizes the architecture of the discriminator and generator networks of UGAN. The discriminator network adopted the PatchGAN model (Isola et al., 2017) that was comprised of five convolutional layers, and the resulting output feature map was 32×32. Each convolutional layer of the PatchGAN model performs the convolution operation, followed by ReLU activation. The generator has a U-Net model consisting of encoder and decoder sections in a U-shape (Ronneberger et al., 2015). The encoder and decoder sections in U-Net have eight convolutional and seven deconvolutional layers, respectively. Each convolutional layer of the U-Net model performs successive convolutions with a 4×4 kernel and stride of 2, batch normalization, and leaky ReLU activation. The deconvolution layer also includes a 4×4 convolution kernel with a stride of 2 and ReLU activation, without batch normalization. The outputs from the second to seventh convolutional layers were connected to the outputs of deconvolutional layers in U-Net. UGAN was trained with the paired synthetic datasets of underwater and clean ground truth images. To construct the paired datasets, an unsupervised network, cycleGAN, was applied to a mapping function of clean to underwater images: F: JI (Zhu et al., 2018). CycleGAN consists of a Resnet-9 block generator and a PatchGAN discriminator outputting a 70×70 feature map. CycleGAN was trained with two batches and 100 epochs using 6,050 clean images and 5,202 underwater images. This included 1,813 images from ImageNet (Deng et al., 2009) and 3,389 real images taken in Korean waters. The trained cycleGAN generated pairs of clean and underwater images where 5,202 underwater images were used for training. From the paired training images, UGAN was trained with 32 batches and 20 epochs.

Fig. 1

UGAN network architecture: The generator transforms an underwater image into a fake clean image through a u-shape network consisting of an encoder and decoder. The discriminator takes the generated image as input and produces a 32×32 patch image to distinguish between the real clean image and the fake clean image.

3. Results

The performances of the DCP, two gradient transform methods, image fusion, cycleGAN, and UGAN were compared by evaluating the image quality of real underwater images taken in Korean waters. The enhancement was evaluated quantitatively based on UIQM, UCIQE, GW assumption, and blur metric.

UIQM was calculated using the weighted sum of colorfulness (UICM), sharpness (UISM), and contrast (UIConM) measures (Panetta et al., 2016). This study used the same weighting as Panetta et al. (2016); a1= 0.0282, a2= 0.2983, and a3= 0.0339. The computation of UIQM is expressed in Eq. (34):

(34) UIQM=a1UICM+a2UISM+a3UIConM

UICM was measured by the mean and standard deviation in two combined color domains of the image to be evaluated; red-green and yellow-blue colors, as follows:

(35) UICM=0.0268μρ,rg2+μρ,yb2+0.1586σρ,rg2+σρ,yb2

In the combined color domains, Irg = IrIg and Iyb = (IrIg)/2− Ib. μα and σρ are the mean and standard deviation of Irg and Iyb. ρ is a ratio to trim the upper and lower intensity pixels. The UISM sharpness was evaluated based on the weighted sum of the difference between the maximum and minimum values in the edge map for each color channel, as shown in Eq. (36). The edge map was determined by the Sobel operator as follows:

(36) UISM=cδcEME[Sobel[Ic]]
(37) EME[Edgen]=2Nn=1Nlog[max[Edgen]min[Edgen]]
where δc is a weight for each color channel. δr = 0.299; δg= 0.587; δb= 0.114. EME[ ] divides the edge image, Sobel[Ic], into N sub-blocks and calculates the logarithmic difference between the maximum and minimum values for each sub-block. It then outputs the mean of the difference in values over the sub-blocks. UIConM divided a gray-scaled image, Igray into N sub-blocks and then calculates the parameterized logarithmic image processing operators, Θ, ⊕, and ⊗ (Panetta et al., 2011), as follows:
(38) UIConM=1Nn=1Nmax[Igray,n]Θmin[Igray,n]max[Igray,n]min[Igray,n]×logmax[Igray,n]Θmin[Igray,n]max[Igray,n]min[Igray,n]

UCIQE is calculated by a weighted sum of three measures in terms of the chroma (σchr ) and luminance (conLum ) in CIELAB color space and saturation (μsat) (Yang and Sowmya 2015), as expressed in Eq.(39):

(39) UCIQE=b1σchr+b2conLum+b3μsat

In CIELAB color space, [L,a,b] of an image, the standard deviation of chroma (σchr ) was calculated by a2+b2. The contrast in the luminance channel (conLum) was determined by the mean value difference in the top 1% and bottom 1% of pixels in the luminance channel, IL,top_1%IL,bottom_1%, respectively. The saturation measure (μsat) was calculated by a2+b2/L. The weighting for chroma, luminance, and saturation in UCIQE was set to b1=0.4680, b2=0.2745, and b3=0.2576, respectively. Larger UIQM and UCIQE values indicate better enhancement performance.

The GW assumption is that an equal mixture of RGB color channels should be neutral gray under a color-balanced situation. An underwater GW assumption was used to measure the degree of color correction (Berman et al., 2018). Here, the calculation process for the GW value was modified with a standard deviation over three mean color values. The color balance improves as the GW becomes lower, as follows:

(40) GW=12c{r,g,b}(mean[Ic]mean[Ir,Ig,Ib])2

The blur metric is derived based on observing that humans find it difficult to perceive differences between a blurred and re-blurred image (Crété-Roffet et al., 2007). The blur measure compares the horizontal and vertical derivatives of an input and its blurred images by the max operator. Blur metric was normalized from 0 to 1; a larger blur metric value indicated more blurring.

(41) Bmetric=max[IhorDhorIhor,IverDverIver]

Directional sum, difference, and derivative operators were applied to input (I) and its blurred images (Ib = blur[IY]) in the luminance channel according to Eqs. (42) and (43). The blur[ ] operator averaged nine elements horizontally and vertically.

(42) Idir=sumdir[|dir[IY]|],dir{hor,ver}
(43) Ddir=sumdir[max[0,|dir[IY]||dir[Ib]|]]

If the input image already has high blurriness, the difference between the directional derivatives of the input and blurred image would be small, resulting in a larger blur metric value.

Figs. 2 to 4 present images enhanced using single enhancement methods from real underwater images. Three underwater images were captured with a dominant greenish color tone and exposure to a strong light source from standard definition (SD) videos recorded in the late afternoon. Figs. 2 and 3 were taken in May 2015 at Jangmok Port, Geoje-si, Gyeongsangnam-do using a camera equipped with a Pro 4 ROV (VideoRay, USA). Fig. 4 was recorded in November 2017 at Yokjido, Tongyeong-si, Gyeongsangnam-do, using an SD camera equipped with a BlueROV (BlueRobotics, USA). All scenes were illuminated by LED lighting equipped in the ROVs. DCP, (b) in Figs. 2 to 4, tends to emphasize a green color in the overall scene and darkens the background signal. DCP was ineffective in correcting the greenish color cast effect, which is the main artifact of underwater imagery. Gradient transform with Tarel and Peng transmissions were sensitive to the brightness change in an underwater image, as shown in (c) and (d) of Figs. 23. This was also found to be ineffective in enhancing the object details, as illustrated in Figs. 4(c) and (d). Two gradient-Tarel and Peng inadequately corrected for the overall greenish color tone in all underwater images. Image fusion moderately compensated for the overall color tones in (e) of Figs. 2 to 4. On the other hand, it did not enhance the image contrast of objects or reduce the haziness in the background. CycleGAN provided good correction of the overall colors and enhanced the object details, but it generated artifacts in the background, as shown around objects in Fig. 3(f). UGAN provided good visual quality in terms of compensating for the greenish color tone, enhancing the object details, and reducing haziness comparing to the other enhancement methods in (g) of Figs. 2 to 4.

Fig. 2

(a) Real original underwater image of 481×416, showing a dominant greenish color tone and exposure to strong light at the top, and the enhanced images by (b) DCP, (c) gradient-Tarel, (d) gradient-Peng, (e) image fusion, (f) cycleGAN, and (g) UGAN.

Fig. 4

(a) Real original underwater image of 711×406, showing low contrast and visibility, and the enhanced images by (b) DCP, (c) Gradient-Tarel, (d) Gradient-Peng, (e) Image fusion, (f) CycleGAN, and (g) UGAN.

Fig. 3

(a) Real original underwater image of 481×416, showing a dominant greenish color tone and expose to strong light in the middle, and the enhanced images by (b) DCP, (c) gradient-Tarel, (d) gradient-Peng, (e) image fusion, (f) cycleGAN, and (g) UGAN.

Tables 2 to 4 compare the figures of merit (FOM), i.e., UIQM, UCIQE, GW, blur metric, and computational time, evaluated from the original and the enhanced underwater images in Figs. 2 to 4. UIQM and UCIQE were difficult to interpret because they focus on different aspects depending on the example image. Unlike a visual comparison, two gradient transform methods and DCP obtained a high UIQM and UCIQE for Figs. 2 and 3. For all underwater images, DCP obtained the highest UCIQEs for Figs. 2 and 3 because of high color scores with respect to σchr, and μsat. This might be due to DCP’s tendency to increase the saturation of underwater images for a specific color. Gradient transform methods tended to emphasize the bright region exposed to light, which increases the sharpness and contrast score. Thus, two gradient transform methods had the highest UIQM and UCIQE because of the high sharpness (UISM) and contrast (conLum) measures, respectively. Image fusion had the best and the worst scores in terms of the GW color quality and blur metrics, respectively. This shows that image fusion is effective in balancing underwater colors, but makes the images blurry. CycleGAN and UGAN showed high UIQM and UCIQE in all figures. Furthermore, the two GANs had better GW scores than the original underwater images and the same blur scores as the original image for Figs. 2 and 3. Marginal differences were observed between the two GANs in terms of all FOMs for Figs. 2 and 3. In terms of Fig. 4, UGAN had higher UIQM and GW scores than cycleGAN. The computational time associated with each enhancement method was compared on an Intel Core i7-7700HQ CPU. The training time of cycleGAN and UGAN are approximately 46 hours and 12 hours on a Nvidia Quadro P4000 GPU, respectively. The two GANs ran on Python with the Tensorflow framework. The other methods ran on MATLAB. Image fusion and two GANs had the best and moderate computational time performance, respectively, compared to the other single enhancement methods.

Figure of merits and computational time of original underwater image in Fig. 2 and single enhancements

Figure of merits and computational time of original underwater image in Fig. 3 and single enhancements

Figure of merits and computational time of original underwater image in Fig. 4 and single enhancements

4. Discussion

This study applied single image enhancement approaches, DCP, gradient transform method with Tarel and Peng transmission maps, image fusion, and two GANs of cycleGAN and UGAN, to compensate for the color cast effect and low visibility due to the light attenuation underwater. The enhanced performance was evaluated in terms of UIQM, UCIQE, GW, and blur metrics with real underwater images taken in Korean waters.

The original DCP was proposed to improve the hazy outdoor images by saturating the inherent color in a haze-free image. Saturation with a specific color produced strong greenish or bluish hues on the underwater image; both are already dominant color tones in underwater images. In DCP, the user needs to adjust the patch size of the min operator in Eq. (3). The DCP result was evaluated with different patch sizes of 11×11 to 21×21 with intervals of 2; however, there were no different effects on color correction depending on patch sizes.

The gradient transform method recovers the underwater image from its gradient enhanced with an estimated transmission map and the gradient of the given input image. The gradient transform method with Tarel and Peng transmissions produced bright regions that appear as though they are exposed to strong light in Figs. 2 and 3. This method was also ineffective in reducing the color-cast effect and low visibility of underwater images. The adjustable parameters in the Tarel transmission estimation are the number of line segments and a patch size in Eq. (9). Successive numbers of line segments from 3 to 8 and different patch sizes from 41×41 to 101×101 in intervals of 20 were applied. On the other hand, there were no significant differences between the final enhanced images with different adjustable parameters. An estimation of Peng transmission can depend on the patch size for depth-related values in Eqs. (15) and (16) and the number of Gaussian kernel levels in Eq. (18). This study examined the dependency on different patch sizes from 5 to 17 in intervals of 3 and found little dependency on the patch size.

Image fusion had the worst blur scores in Figs. 2 to 4 and resulted in a loss of object detail. It contained a contrast enhancement step, which was replaced with CLAHE instead of gamma correction. The blurry results were attributed to the last fusion step, with multiple levels of Laplacian and Gaussian pyramids. When more than three levels of pyramids were set, the final result increased the blurriness. In contrast, small pyramid level numbers resulted in noisier images.

(44) J˜(x)=EIE(x)ωE(x),E{contraast,edge}

If naïve fusion is performed, as per Eq. (44), there is an increase in the sharpness of underwater images compared to multiscale fusion, but it produces random pattern artifacts in the background.

GAN might be affected by the generality of the training data to express underwater color images. Training data were constructed with ImageNet and real underwater images taken in Korean waters to identify the turbidity and color cast. The inclusion of real underwater imagery in Korean waters improved the color tone and visibility compared to the training data in the ImageNet database. The pre-trained cycleGAN was applied to improve underwater image quality and generate underwater and clean image pairs for training UGAN. CycleGAN and UGAN were compared to confirm the performances of unsupervised and supervised learning for single image enhancement because they do not require additional data, such as a depth map for training. CycleGAN had incomplete enhancement performance producing artifacts in the background. UGAN acted moderately to enhance the color balance and visibility of underwater images compared to the other enhancement methods, but it was not effective in dehazing the underwater images. To enhance visibility, a new network architecture needs to be constructed and trained with the paired underwater images and additional information like depth maps (Li et al., 2018b).

UIQM and UCIQE are popular FOMs used in studies of underwater enhancement. The method to combine three quality measures in UIQM and UCIQE is dependent on the weights in Eqs. (34) and (39). On the other hand, there were no proposed ways to determine the weights properly. In this study, it was difficult to interpret the general tendencies of UIQM and UCIQE because these FOMs fluctuated in each case by case, and the quantitative interpretations based on the scores were different from the visual comparisons. It is necessary to normalize the individual image quality measure of UIQM and UCIQE and to set reasonable weights to evaluate the overall image quality adequately. The GW assumption and blur metric were adopted to evaluate color balancing and blurriness in the enhanced underwater image. Individual FOMs, such as GW and blur metrics, reflected the consistent quantitative image quality measures compared to the combined FOMs of UIQM and UCIQE for underwater images.

5. Conclusions

In this study, six single underwater image enhancement approaches were compared: DCP, two gradient transforms, image fusion, and two GANs. The enhancement performances were evaluated qualitatively and quantitatively. DCP caused saturation of the underwater images to either a greenish or bluish color tone and reduced the brightness of the background signal. The gradient transform methods with two transmission maps were sensitive to the light source and highlighted the region exposed to light. Image fusion provided reasonable color correction, but the object details were lost due to the last fusion step. CycleGAN corrected the overall color tone well, but generated artifacts in the background. Although UGAN was not rated with the best scores for three sample images, it showed fairly good visual quality and quantitative scores for all FOMs.

Acknowledgements

This study was supported in part by research grants from Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education (NRF- 2018R1D1A1B07049296), and by the project titled ‘Development of the support vessel and systems for the offshore field test and evaluation of offshore equipment’, funded by the Ministry of Oceans and Fisheries (MOF), Korea.

References

Achantay R, Hemamiz S, Estraday F, Susstrunk S. 2009. Frequency Tuned Salient Region Detection. In : Proceeding of IEEE Conference on Computer Vision and Pattern Recognition. Miami, FL, USA. p. 1597–1604. https://doi.org/10.1109/CVPR.2009.5206596 .
Ancuti CO, Ancuti C, Vleeschouwer CD, Bekaert P. 2017;Color Balance and Fusion for Underwater Image Enhancement. IEEE Transactions on Image Processing 27(1):379–393. https://doi.org/10.1109/TIP.2017.2759252 .
Anwar S, Li C, Porikli F. 2018. Deep Underwater Image enhancement. arXiv:180703528. Retrieved from https://arxiv.org/abs/1807.03528 .
Anwar S, Li C. 2019. Diving Deeper into Underwater Image Enhancement: A Survey. arXiv:1907.07863. Retrieved 2019 from https://arxiv.org/abs/1907.07863 .
Arjovsky M, Chintala S, Bottou L. 2017. Wasserstein GAN. arXiv:1701.07875. Retrieved from https://arxiv.org/abs/1701.07875 .
Arnold-Bos A, Malkasse J.-P, Kervern G. 2005. A Preprocessing Framework for Automatic Underwater Images Denoising. In : Proceeding of European Conference on Propagation and Systems. Brest, France. p. 1–8. Retrieved from https://hal.archives-ouvertes.fr/hal-00494314 .
Bazeille S, Quidu I, Jaulin L, Malkasse J.-P. 2006;Automatic Underwater Image Pre-processing. Proceeding of Caracterisation Du Milieu Marin :16–19. Retrieved from https://hal.archives-ouvertes.fr/hal-00504893/ .
Berman D, Levy D, Avidan S, Treibitz T. 2018. Underwater single Image Color Restoration Using Haze-lines and a New Quantitative Dataset. arXiv: 1811.01343. Retrieved from https://arxiv.org/abs/1811.01343 .
Cao K, Peng Y, Cosman P. 2018. Underwater Image Restoration Using Deep Networks to Estimate Background Light and Scene Depth. In : Proceeding of IEEE Southwest Symposium on Image Analysis and Interpretation. Las Vegas, NV, USA. p. 1–4. https://doi.org/10.1109/SSIAI.2018.8470347 .
Carlevaris-Bianco N, Mohan A, Eustice RM. 2010. Initial Results in Underwater Single Image Dehazing. In : Proceeding of Oceans 2010 MTS/IEEE. Seattle, WA, USA. p. 1–8. https://doi.org/10.1109/OCEANS.2010.5664428 .
Cho Y, Jeong J, Kim A. 2018;Model-assisted Multiband Fusion for Single Image Enhancement and Applications to Robot Vision. IEEE Robotics and Automation Letters 30(4):2822–2829. https://doi.org/10.1109/LRA.2018.2843127 .
Crété-Roffet F, Dolmière T, Ladret P, Nicolas M. 2007. The Blur Effect: Perception and Estimation with a New No-Reference Perceptual Blur Metric. Proceeding SPIE Electronic Imaging Symposium Conf Human Vision and Electronic Imaging San Jose, USA. p. EI6492–16. Retrieved from https://hal.archives-ouvertes.fr/hal-00232709 .
Deng J, Dong W, Socher R, Li L.-J, Li K, Fei-Fei L. 2009. Imagenet: A Large-scale Hierarchical Image Database. In : Proceeding of IEEE Conference on Computer Vision and Pattern Recognition. Miami, FL, USA. p. 248–255. https://doi.org/10.1109/CVPR.2009.5206848 .
Drews P, Nascimento E, Moraes F, Botelho S, Campos M. 2013. Transmission Estimation in Underwater Single Images. In : Proceeding of IEEE International Conference on Computer Vision Workshops. Sydney, NSW, Australia. p. 825–830. https://doi.org/10.1109/ICCVW.2013.113 .
Fabbri C, Islam MJ, Sattar J. 2018. Enhancing Underwater Imagery Using Generative Adversarial Networks. In : Proceeding of IEEE International Conference on Robotics and Automation. Brisbane, QLD, Australia. p. 7159–7165. https://doi.org/10.1109/ICRA.2018.8460552 .
Fattal R. 2008;Single Image Dehazing. ACM Transactions on Graphics 27(3):1–9. https://doi.org/10.1145/1360612.1360671 .
Ghani ASA, Isa NAM. 2014;Underwater Image Quality Enhancement Trough Composition of Dual-intensity Images and Rayleigh-stretching. SpringerPlus 3:757. https://doi.org/10.1186/2193-1801-3-757 .
Ghani ASA, Isa NAM. 2015;Underwater Image Quality Eenhancement Through Integrated Color Model with Rayleigh Distribution. Applied Soft Computing 27:219–230. https://doi.org/10.1016/j.asoc.2014.11.020 .
Guo Y, Li H, Zhuang P. 2019;Underwater Image Enhancement Using a Multiscale Dense Generative Adversarial Network. IEEE Journal of Oceanic Engineering 45(3):862–870. https://doi.org/10.1109/JOE.2019.2911447 .
He K, Sun J, Tang X. 2011;Single Image Haze Removal Using Dark Channel Prior. IEEE Transactions on Pattern Analysis and Machine Intelligence 33(12):2341–2353. https://doi.org/10.1109/TPAMI.2010.168 .
Hou M, Liu R, Fan X, Luo Z. 2018. Joint Residual Learning for Underwater Image Enhancement. In : Proceeding of IEEE International Conference on Image Processing. Athens, Greece. p. 4043–4047. https://doi.org/10.1109/ICIP.2018.8451209 .
Isola P, Zhu J, Zhou T, Efros AA. 2017. Image-to-Image Translation with Conditional Adversarial Networks. arXiv: 1611.07004. Retrieved from https://arxiv.org/abs/1611.07004 .
Jian S, Wen W. 2017;Study on Underwater Image Denoising Agorithm Based on Wavelet Transform. Journal of Physics: Conference Series 806:012006. Retrieved from https://iopscience.iop.org/article/10.1088/1742-6596/806/1/012006 .
Khan A, Ali SSA, Malik A, Anwer A, Meriaudeau F. 2016. Underwater Image Enhancement by Wavelet Based Fusion. In : Proceeding of IEEE International Conference on Underwater System Technology: Theory and Applications. Penang, Malaysia. p. 83–88. https://10.1109/USYS.2016.7893927 .
Li WJ, Gu B, Huang JT, Wang SY, Wang MH. 2012;Single Image Visibility Enhancement in Gradient Domain. IET Image Processing 6(5):589–595. https://doi.org//10.1049/iet-ipr.2010.0574 .
Li C, Guo C, Guo J. 2018a;Emerging from Water: Underwater Image Color Correction Based on Weakly Supervised Color Transfer. IEEE Signal Processing Letters 25(3):232–327. https://doi.org/10.1109/LSP.2018.2792050 .
Li J, Skinner K, Eustice R, Johnson-Roberson M. 2018b;Watergan: Unsupervised Generative Network to Enable Real-time Color Correction of Monocular Underwater Images. IEEE Robotics and Automation Letters 3(1):387–394. https://doi.org/10.1109/LRA.2017.2730363 .
Li C, Guo C, Ren W, Cong R, Hou J, Kwong S. 2019a. An Underwater Image Enhancement Dataset and Beyond. arXiv: 1901.05495. Retrieved from https://arxiv.org/abs/1901.05495 .
Li H, Li J, Wang W. 2019b. A Fusion Adversarial Underwater Image Enhancement Network with a Public Test Dataset. arXiv: 1906.06819. Retrieved from https://arxiv.org/abs/1906.06819 .
Luo M, Fang Y, Ge Y. 2019;An Effective Underwater Image Enhancement Method Based on CLAHE-HF. Journal of Physics: Conference Series 1237(3):032009. Retrieved from https://iopscience.iop.org/article/10.1088/1742-6596/1237/3/032009 .
Mathieu M, Couprie C, LeCun Y. 2015. Deep Multi-Scale Video Prediction Beyond Mean Square Error. arXiv:1511.05440. Retrieved from https://arxiv.org/abs/1511.05440 .
Mi Z, Zhou H, Zheng Y, Wang M. 2016;Single Image Dehazing via Multiscale Gradient Domain Contrast Enhancement. IET Image Processing 10(3):206–214. https://doi.org/10.1049/iet-ipr.2015.0112 .
Mobley C. 1994. Light and Water: Radiative Transfer in Natural Waters San Diego: Academic Press.
Panetta K, Agaian S, Zhou Y, Wharton EJ. 2011;Parameterized Logarithmic Framework for Image Enhancement. IEEE Transactions on Systems, Man, and Cybernetics, Part B 41(2):460–473. https://doi.org/10.1109/TSMCB.2010.2058847 .
Panetta K, Gao C, Agaian S. 2016;Human-Visual-System- Inspired Underwater Image Quality Measures. IEEE Journal of Oceanic Engineeringm 41(3):541–551. https://doi.org/10.1109/JOE.2015.2469915 .
Park EP, Sim J. 2017. Gradient-based Contrast Enhancement and Color Correction for Underwater Images. In : Proceeding of Asia-Pacific Signal and Information Processing Association Annual Summit and Conference. Kuala Lumpur, Malaysia. p. 1444–1447. https://doi.org/10.1109/APSIPA.2017.8282259 .
Peng Y, Cosman PC. 2017;Underwater Image Restoration Based on Image Blurriness and Light Absorption. IEEE Transactions on Image Processing 26(4):1579–1594. https://doi.org/10.1109/TIP.2017.2663846 .
Ronneberger O, Fischer P, Brox T. 2015. U-net: Convolutional Networks for Biomedical Image Segmentation. arXiv: 1505.04597. Retrieved from https://arxiv.org/abs/1505.04597 .
Roser M, Dunbabin M, Geiger A. 2014. Simultaneous Underwater Visibility Assessment, Enhancement and Improved Stereo. In : Proceeding of IEEE International Conference on Robotics and Automation. Hong Kong, China. p. 3840–3847. https://doi.org/10.1109/ICRA.2014.6907416 .
Simchony T, Chellappa R, Shao M. 1990;Direct Analytical Methods for Solving Poisson Equations in Computer Vision Problems. IEEE Transactions on Pattern Analysis and Machine Intelligence 12(5):435–446. https://doi.org/10.1109/34.55103 .
Sun X, Liu L, Li Q, Dong J, Lima E, Yin R. 2018;Deep Pixel to Pixel Network for Underwater Image Enhancement and Restoration. IET Image Processing 13(3):469–474. https://doi.org/10.1049/iet-ipr.2018.5237 .
Tarel J, Hautière N. 2009. Fast Visibility Restoration from a Single Color or Gray Level Image. In : Proceeding of IEEE International Conference on Computer Vision. Kyoto, Japan. p. 2201–2208. https://doi.org/10.1109/ICCV.2009.5459251 .
Treibitz T, Schechner YY. 2012;Turbid Scene Enhancement Using Multi-Directional Illumination Fusion. IEEE Transactions on Image Processing 21(11):4662–4667. https://doi.org/10.1109/TIP.2012.2208978 .
Uplavikar P, Wu Z, Wang Z. 2019. All-in-One Underwater Image Enhancement Using Domain-Adversarial Learning. arXiv: 1905.13342. Retrieved from https://arxiv.org/abs/1905.13342 .
Vincent L. 1993;Morphological Grayscale Reconstruction in Image Analysis: Applications and Efficient Algorithms. IEEE Transactions on Image Processing 2(2):176–201. https://doi.org/10.1109/83.217222 .
Wang Y, Ding X, Wang R, Zhang J, Fu X. 2017a. Fusion-Based Underwater Image Enhancement by Wavelet Decomposition. In : Proceeding of IEEE International Conference on Industrial Technology. Toronto, ON, Canada. p. 1013–1018. https://doi.org/10.1109/ICIT.2017.7915500 .
Wang Y, Zhang J, Cao Y, Wang Z. 2017b. A Deep CNN Method for Underwater Image Enhancement. In : Proceeding of IEEE International Conference on Image Processing. Beijing, China: p. 1382–1386. https://doi.org/10.1109/ICIP.2017.8296508 .
Wang Y, Song W, Fortino G, Qi L.-Z, Zhang W, Liotta A. 2019;An Experimental-Based Review of Image Enhancement and Image Restoration Methods for Underwater Imaging. IEEE Access 7:140233–140251. https://doi.org/10.1109/ACCESS.2019.2932130 .
Yang M, Sowmya A. 2015;An Underwater Color Image Quality Evaluation Metric. IEEE Transactions on Image Processing 24(12):6062–6071. https://doi.org/10.1109/TIP.2015.2491020 .
Yang M, Hu J, Li C, Rohde G, Du Y, Hu K. 2019;An In-Depth Survey of Underwater Image Enhancement and Restoration. IEEE Access 7:123638–123657. https://doi.org/10.1109/ACCESS.2019.2932611 .
Zetian M, Zheng L, Yafei W, Xianping F, Zhengyu C. 2018. Multiscale Gradient Domain Underwater Image Enhancement. In : Proceeding of 2018 Oceans - MTS/IEEE Kobe Techno-Oceans. Kobe, Japan. p. 1–5. https://doi.org/10.1109/OCEANSKOBE.2018.8559180 .
Zhao X, Tao J, Song Q. 2015;Deriving Inherent Optical Properties from Background Color and Underwater Image Enhancement. Ocean Engineering 94:163–172. https://doi.org/10.1016/j.oceaneng.2014.11.036 .
Zhang S, Zhang J, Fang S, Cao Y. 2014. Underwater Stereo Image Enhancement Using a New Physical Model. In : Proceeding of IEEE International Conference on Image Processing. Paris, France. p. 5422–5426. https://doi.org/10.1109/ICIP.2014.7026097 .
Zhu J, Park T, Isola P, Efros AA. 2018. Unpaired Image-to-Image Translation Using Cycle-Consistent Adversarial Networks. arXiv:1703.10593. Retrieved from https://arxiv.org/abs/1703.10593 .

Article information Continued

Fig. 1

UGAN network architecture: The generator transforms an underwater image into a fake clean image through a u-shape network consisting of an encoder and decoder. The discriminator takes the generated image as input and produces a 32×32 patch image to distinguish between the real clean image and the fake clean image.

Fig. 2

(a) Real original underwater image of 481×416, showing a dominant greenish color tone and exposure to strong light at the top, and the enhanced images by (b) DCP, (c) gradient-Tarel, (d) gradient-Peng, (e) image fusion, (f) cycleGAN, and (g) UGAN.

Fig. 3

(a) Real original underwater image of 481×416, showing a dominant greenish color tone and expose to strong light in the middle, and the enhanced images by (b) DCP, (c) gradient-Tarel, (d) gradient-Peng, (e) image fusion, (f) cycleGAN, and (g) UGAN.

Fig. 4

(a) Real original underwater image of 711×406, showing low contrast and visibility, and the enhanced images by (b) DCP, (c) Gradient-Tarel, (d) Gradient-Peng, (e) Image fusion, (f) CycleGAN, and (g) UGAN.

Table 1

Image reconstruction from a given gradient

Workflow of image recovery from a given gradient
(1) Compute divJ and ∇2B
divJ≃∇Jx(x,y) − ∇Jx(x−1,y) + ∇Jy(x,y) − Jy(x,y − 1)
2BB(x+1,y) + B(x−1,y) + B(x,y+1) + B(x,y−1)−4B(x,y)
(2) Y=divJ−∇2β
(3) Discrete sine transformation of Y: Ψuv(Y)
(4) Ψuv(T)=Ψuv(Y)2cos(πu/M))+2cos(πu/N)4
(5) Compute inverse the discrete sine transform of Ψuv(T)
(6) Reconstruction: = T+B

Table 2

Figure of merits and computational time of original underwater image in Fig. 2 and single enhancements

UIQM UCIQE GW Blur metric Computational time (s)
Original 2.20 29.20 0.052 0.36 -
DCP 2.62 33.40 0.094 0.36 13.3
Gradient-Tarel 4.29 31.06 0.043 0.31 12.3
Gradient-Peng 4.41 31.27 0.040 0.30 11.4
Image fusion 2.42 28.72 0.013 0.55 1.1
CycleGAN 2.25 29.06 0.024 0.35 1.5
UGAN 3.13 29.11 0.019 0.35 3.5

Table 3

Figure of merits and computational time of original underwater image in Fig. 3 and single enhancements

UIQM UCIQE GW Blur metric Computational time (s)
Original 1.69 22.34 0.059 0.30 -
DCP 2.50 27.40 0.076 0.31 12.7
Gradient-Tarel 4.64 26.44 0.054 0.25 14.3
Gradient-Peng 4.87 27.96 0.054 0.24 12.3
Image fusion 2.03 22.11 0.008 0.50 0.7
CycleGAN 3.34 24.40 0.019 0.29 1.4
UGAN 4.09 27.92 0.029 0.32 3.5

Table 4

Figure of merits and computational time of original underwater image in Fig. 4 and single enhancements

UIQM UCIQE GW Blur metric Computational time (s)
Original 0.48 12.12 0.008 0.44 -
DCP 1.13 22.38 0.033 0.44 18.6
Gradient-Tarel 1.27 16.61 0.005 0.43 20.0
Gradient-Peng 1.48 18.22 0.016 0.43 17.8
Image fusion 1.15 15.00 0.001 0.60 0.9
CycleGAN 1.62 26.06 0.071 0.44 1.4
UGAN 2.97 26.93 0.013 0.49 3.6