Information: Enhancement of Low Contrast Images Based On Effective Space Combined With Pixel Learning
Information: Enhancement of Low Contrast Images Based On Effective Space Combined With Pixel Learning
Article
Enhancement of Low Contrast Images Based on
Effective Space Combined with Pixel Learning
Gengfei Li 1,2, *, Guiju Li 1 and Guangliang Han 1
1 Changchun Institute of Optics, Fine Mechanics and Physics, Chinese Academy of Sciences,
#3888 Dongnanhu Road, Changchun 130033, China; [email protected] (G.L.); [email protected] (G.H.)
2 University of Chinese Academy of Science, #19 Yuquan Road, Beijing 100049, China
* Correspondence: [email protected]; Tel.: +86-177-1023-7905
Abstract: Images captured in bad conditions often suffer from low contrast. In this paper, we
proposed a simple, but efficient linear restoration model to enhance the low contrast images.
The model’s design is based on the effective space of the 3D surface graph of the image. Effective space
is defined as the minimum space containing the 3D surface graph of the image, and the proportion
of the pixel value in the effective space is considered to reflect the details of images. The bright
channel prior and the dark channel prior are used to estimate the effective space, however, they may
cause block artifacts. We designed the pixel learning to solve this problem. Pixel learning takes the
input image as the training example and the low frequency component of input as the label to learn
(pixel by pixel) based on the look-up table model. The proposed method is very fast and can restore
a high-quality image with fine details. The experimental results on a variety of images captured in
bad conditions, such as nonuniform light, night, hazy and underwater, demonstrate the effectiveness
and efficiency of the proposed method.
1. Introduction
In the image acquisition process, the low illumination in nonuniformly illuminated environment
or light scattering caused by turbid medium in the foggy/underwater environment will lead to low
image contrast. Based on the variety of bad conditions, however, it is difficult to enhance the image
of these conditions through a unified approach. Though traditional methods such as the histogram
equalization will deal with all these low contrast images, most results are show to be uncomfortable
for the human visual system. Therefore, most of them establish a specific recovery model based on the
distinctive physical environment to enhance the images.
Dealing with light compensation is usually done by using the Retinex (retina-cortex) model,
which is built on the human visual system [1–3]. The early single-scale Retinex algorithm proposed
by Jobson [4] can either provide dynamic range compression on a small scale or tonal rendition on a
large scale. Therefore, Jobson continued his research and proposed the MSR (multiscale Retinex)
algorithm [5], which has been the most widely used in recent years. Most improved Retinex
algorithms [6–15] are based on MSR. However, the Gaussian filtering used by the MSR algorithm
calculates a large number of floating data, which makes the algorithm take too much time.
Therefore, for practical use, Jiang et al. [15] used hardware acceleration to implement the MSR
algorithm. In addition, some research [16] used the dehaze model instead of the Retinex model
to deal with the negative input image for light compensation or using the bright channel prior with
the guided filter [17] for a quick lighting compensation.
Early haze removal algorithms require multiple frames or additional depth map information [18–20].
Since Fattal et al. [21] and Tan et al. [22] proposed a single-frame dehaze algorithm relying on stronger
priors or assumptions, the single-frame dehaze algorithms have become a research focus. Subsequently,
He et al. [23] proposed the DCP (dark channel prior) for haze removal, which laid the foundation
for the dehazing algorithm in recent years. Combining the DCP with the guided filter [24] is also
the most efficient method for the dehazing algorithm. Since then, the principal study of the fog
algorithm has focused on the matting technique of transmittance [24–30]. Meng et al. [25] applied a
weighted L1-norm-based contextual regularization to optimize the estimation of the unknown scene
transmission. Sung et al. [26] used a fast guided filter that was combined with the up/down samples to
optimize the performance time. Zhang et al. [28] used the five-dimensional feature vectors to recover
the transmission values by finding their nearest neighbors from the fixed points. Li et al. [30] computed
a spatially varying atmospheric light map to predict the transmission and refined it by the guided
filter [24]. The guided filter [24] is an O(n) time edge-preserving filter with the similar result of bilateral
filter. However, in the application, the well-refined methods are hard to perform on a video system due
to the time cost, and the fast local filters like guided/bilateral filters would concentrate the blurring
near the strong edges, then introducing halos.
There have been several attempts to restore and enhance the underwater image [31–33].
However, there is no general restoration model for such degraded images. Research mainly applies
the white balance [32–34] to correct the color bias at first and then uses a series of contrast stretch
processing to enhance the visibility of underwater images.
Although these studies have applied different models and methods for image enhancement,
the essential goal of all of these is to stretch the contrast. In this paper, we proposed a unified
restoration model for these low-contrast degraded images to reduce the human operations or even in
some multidegradation environments. Moreover, due to the artifacts produced by the patch-based
methods we applied in our approach, the pixel learning refinement is proposed.
2. Background
I ( x ) = L ( x ).∗ R ( x ), (1)
where x is the image coordinates and operator .* denotes the matrix point multiplication. The main
parameter of the Retinex algorithm is the Gaussian filter’s radius. A large radius can get obtain color
recovery, and a small radius can retain more details. Therefore, the most commonly used Retinex
algorithm is multiscale. Early research on Retinex was mainly for light compensation. In recent years,
a large number of studies has been undertaken to enhance the image with hazy [10] and underwater
images [8,31] using the Retinex algorithm. However, these methods are developed with their own
framework for a specific environment.
I( x ) = J( x ).∗ T( x ) + (1 − T( x )) × A, (2)
Information 2017, 8, 135 3 of 20
where I is the hazy image, J is the clear image, T is the medium transmittance image and A is the
global atmospheric light, usually constant. In the methods of estimating the transmittance image T,
the DCP is the simplest and most widely-used method, which is given by:
(a)
Figure 1. Cont.
Information 2017, 8, 135 4 of 20
Information 2017, 8, 135 4 of 20
(b)
(c)
(d)
Figure 1. Images and the projection of their 3D surface graphs on the x-z plane. In the 2D projection,
Figure 1. Images and the projection of their 3D surface graphs on the x-z plane. In the 2D projection,
the x-axis is the image width, and the y-axis is the pixel value: (a) the clear images; (b) the low
the x-axis is the image width, and the y-axis is the pixel value: (a) the clear images; (b) the low
illumination images; (c) the hazy images; (d) the underwater images.
illumination images; (c) the hazy images; (d) the underwater images.
Information 2017, 8, 135 5 of 20
As can be seen, the projection of clear images is almost filled by the domain of the pixel value.
On the contrary, the degraded images’ are compressed. Therefore, we assume that the proportion
of the pixel value in the effective space denotes the detail information of the image. Since the image
size is fixed in processing, we estimate the two smooth surfaces of the effective space that are the
upper surface U (up) and the lower surface D (down) to obtain the proportion of the pixel value in the
effective space. It should be noted that in the ideal clear image, U is a constant plane at 255, while D is
zero. According to the relationship of the proportion, we can establish a linear model by:
I − DI J − DJ J−0
= = , (5)
UI − DI UJ − DJ 255 − 0
where I represents the input degraded image, J represents the ideal clear image, UI and UJ represent
the upper surface of the effective space of I and J, respectively, and similarly, DI and DJ represent
the lower surface. Once we obtain U and D, we can enhance the low contrast images according to
Equation (5). However, the denominator in Equation (5) can be zero when the pixel value of D is equal
to U. Therefore, we introduce a small parameter in the denominator to avoid the division by zero.
Our final enhanced model is given as:
I − DI
J = U J .∗ , (6)
UI − DI + λ
where D is estimated by the block dark channel operation, according to Equation (3), U is estimated
by the block bright channel operation, which can be calculated by Equation (3) with the min function
replaced by the max function and λ is a small factor to prevent the division by zero, usually set to
0.01. UJ denotes the light intensity of the enhanced images, ideally set to the maximum gray level
(the eight-bit image is 255). The images always looks dim after haze removal based on Equation (2).
Our model can improve this phenomenon while combining the Retinex model and the dehaze model
so that our model can be adapted according to the scene requirements.
dark b (I)
T = 1− , (7)
A
where A represents the light intensity of all three channels. Putting Equation (7) into Equation (2),
we can remove the haze with the style of our model as Equation (6):
I−A I−A I − DI
J= +A = DI
+ A = A J .∗ , (8)
T 1− A A − DI
where, in our model, AJ denotes the illumination of the clear image and A is the input
image’s illumination; whereas, in the fog model, AJ = A, which both represent the input
image’s illumination. Due to the impact of the clouds, the light intensity is usually low when the
weather is rainy or foggy. This can lead to the result of haze removal looking dim. On the other hand,
most haze removal research assumes that the atmospheric light intensity is uniform, and A is estimated
as a constant. According to Equation (9), if the input image has sufficient illumination, which means
A = 255 = UJ = AJ , the proposed model is equivalent to the dehaze model. Nevertheless, the light
Information 2017, 8, 135 6 of 20
intensity of a real scene is always nonuniform. As can be seen in Equation (9), when DI remains the
same, the larger the A we give, the smaller the J we have. On the whole, a large estimated value
of the constant A will be great in the thick foggy region, but for the foreground region with only a
little mist, it is too large. This is the main cause of haze removal always becoming a dim restoration.
Moreover, if the input image I is a haze-free image, it will be the typical low contrast problem of
Retinex. According to the constant zero tendency of the block dark channel image and Equation (1),
we have a Retinex model with the style of Equation (6):
I I−0 I − DI
J = 255 × R = 255 × = 255 × = 255 × , (9)
L L−0 L − DI
where R ∈ [0, 1] represents the reflectivity in Equation (1). In order to unify the dynamic range of the
output image J, we multiply R by 255 of the eight-bit image. According to Equation (9), if we estimate
the illuminated image L by a Gaussian filter, the proposed model is equivalent to the Retinex model.
It is notable that the bright channel prior is also a good method to estimate L, and the bright channel
prior will be faster than the Gaussian filter due to the integer operation. The study of Wang et al. [17]
shows that the bright channel prior has a significant effect in terms of light compensation.
Due to the constant zero tendency of the block dark channel image, the proposed model will
automatically turn into the Retinex model to compensate for the light intensity of the image in the
haze-free region. Besides, the model mainly modified the two assumptions of the dehaze model that
AJ is equal to A, and A is a constant; so that the proposed algorithm can increase the exposure of
J when the input I is a hazy image. However, as a result of the use of the bright/dark channel prior,
a refinement process is necessary to improve the block artifacts produced by the min/max filter used
in the priors.
where S represents the process of smoothing and I represents the input image. In Equation (10),
the first penalty term kI( x ) − S( x )k forces proximity between I and S, which can preserve the sharp
gradient edges of I for S, and this will help reduce the artifacts, such as the haloes. The second penalty
term λ × k∇ S( x )k is the degree of similarity between the estimated component and the input image,
and it forces the spatial smoothness on the processed image S. In the previous research results, there
were a number of methods to solve the optimization problem of Equation (10), in which the most
practical of the methods was the guided filter [24]. However, [24] still retained the halo artifacts due to
the frequent use of the mean filter. Although, there are some improvements such as applying iteration
optimization [27] or solving the equation of higher-order [24] to optimize the model, these algorithms
are usually too time consuming for a practical use due to the large number of pixels in an input image
as training examples for an optimization problem. Therefore, we introduce the idea of online learning
to overcome the time-consuming process of optimization.
Online learning is different from batch learning. Its samples come in sequence; in other words,
a one-by-one processing. Then, the classifier is updated according to each new sample. If the input
samples are large enough, the online learning can converge in only one scanning. As an online
learning technique, the pixel learning outputs the results pixel-by-pixel while learning from the input
image to perform the iteration optimization in just one instance of scanning of the whole input image.
The input image Ip (pixel level bright/dark channel or grayscale image) is taken as the training
Information 2017, 8, 135 7 of 20
example, and the mean filter result of its block bright/dark channel image Ibm is used as the label for
learning. As for most machine learning algorithms, we use the square of the difference between the
predictor and the label as the cost function, i.e., E(Ip( x )) = kY( x ) − Ibm( x )k2 , where Y(x) represents
the estimation of the current input pixel value Ip(x). The pixel learning should obtain a convergent
result by one-time scan like the online learning. However, the convergence of learning usually starts
with a large error and converge slowly, which could produce noises during the pixel by pixel learning.
Considering Equation (10), the fusion of the low-frequency image and the original image which
contains the high-frequency information can make the initial output error smaller, so7 ofthat
Information 2017, 8, 135 20
the fast
convergence and the noise suppression can be achieved. Here α fusion is applied as the learning
represents which
iterative equation, the estimation
is the of the current
simplest inputmethod.
fusion pixel valueItIp(x). The pixel
is given as: learning should obtain a
convergent result by one-time scan like the online learning. However, the convergence of learning
usually starts with a large error and converge slowly, which∗ could produce noises during the pixel
P(Equation
by pixel learning. Considering x ).∗ the
x ) = Y((10), (1 −ofα)the
α +fusion . Ibm ( x ),
low-frequency image and the original (11)
image which contains the high-frequency information can make the initial output error smaller, so
where x isthat
thetheimage coordinates,
fast convergence noisedenotes
and theY(x) suppression the
canestimation
be achieved. result of the
Here α fusion input as
is applied pixel
the value at
learning iterative
x as the forward equation,
propagation which is Ibm(x)
process, the simplest fusion method.
denotes the pixelIt isvalue
given as:
of a low-frequency image at
x as the label, α is the fusion weight as P( xthe ( x).*α + (1 − αof
) = Ygradient ).*the
Ibm(iterative
x) , step and P(x) is (11) the learning
results as the
wherepredictor. Note
x is the image that when
coordinates, Y(x)E(Ip(x))
denotes theis large,
estimation thisresult
means there
of the inputispixel
an edge that
value at x asshould be
preserved.theSimilarly, when E(Ip(x))
forward propagation process,isIbm(x)
small,denotes
this means
the pixel there
valueisofaa texture that should
low-frequency image at be
x assmoothed.
the label, αwe
As a consequence, is the fusion the
design weight as theweight
fusion gradientαofbased
the iterative step and P(x) is the learning results as
on E(Ip(x)):
the predictor. Note that when E(Ip(x)) is large, this means there is an edge that should be preserved.
Similarly, when E(Ip(x)) is small, this means thereE(Ipis( xa))texture that should be smoothed. As a
α(weight
consequence, we design the fusion x ) = α based on E(Ip(x)): 2 , (12)
E(Ip( x )) + thred1
E (Ip( x))
α ( x) = 2
, (12)
where thred1 is used to judge whether the detailE (needs Ip( x)) + thred
to be 1 preserved, and a small thred1 preserves
(a) (b)
Figure 2. Comparison of α fusion and guided filter: (a) the α fusion, with thred1 = 10, the radius of
Figure 2. Comparison
the mean filter of fusion
α (b)
is 40; filter. filter: (a) the α fusion, with thred1 = 10, the radius of the
and guided
the guided
mean filter is 40; (b) the guided filter.
Information 2017, 8, 135 8 of 20
where I2Y
where I2Yisisour
ourmapping
mapping model, which
model, whichis aislook-up table,
a look-up and and
table, we calculate the latest
we calculate prediction
the latest result
prediction
of Ip(x)ofand
result Ip(x)theand
output. Then, we
the output. update
Then, the latest
we update themapping of pixelofvalues
latest mapping from Ip(x)
pixel values fromtoIp(x) I2Y.
in P(x)
P(x) to
The mapping model mainly plays the role of logical classification rather
in I2Y. The mapping model mainly plays the role of logical classification rather than linearthan linear transformation,
so we obtain a more
transformation, so we accurate matting
obtain a more resultmatting
accurate near theresult
depth discontinuities.
near Figure 3 shows
the depth discontinuities. Figurethe3
comparison among theamong
shows the comparison iterative
theresult of PL,
iterative α fusion
result of PL,and guided
α fusion andfilter [24].filter
guided It can be It
[24]. seen
canthat the
be seen
iterative result of pixel learning smoothed more textures of the background than
that the iterative result of pixel learning smoothed more textures of the background than the initial the initial output of
α fusionofand
output α the guided
fusion andfilter. However,
the guided there
filter. are somethere
However, sharpare edges
somethatsharp
have been
edgessmoothed
that havedue to
been
the excessively smooth label we have set near the edges of depth. To this end,
smoothed due to the excessively smooth label we have set near the edges of depth. To this end, we we modify the label,
which
modifyisthe thelabel,
low-frequency image
which is the Ibm.
low-frequency image Ibm.
(a) (b)
Figure 4. Comparison of refinement from different labels: (a) the label is Ibm, with thred1 = 10; (b)
Figure 4. Comparison of refinement from different labels: (a) the label is Ibm, with thred1 = 10; (b) the
the label is Ib, with thred1 = 10.
label is Ib, with thred1 = 10.
As can be seen, the bright region of Figure 4b leaves some block artifacts, but sharp edges near
As can be seen, the bright region of Figure 4b leaves some block artifacts, but sharp edges near
the depth discontinuities. Thus, we need the model to learn from Ibm for large pixel values, and the
the depth
small discontinuities. Thus, we
pixel values are learned fromneed the model we
Ib. Concretely, to learn fromand
fuse Ibm IbmIbfor
by large pixelbased
α fusion values, and the
on the
smallpixel
pixelvalue,
valuesso are learned
for the fusionfrom
weight, Concretely,
Ib. we have: we fuse Ibm and Ib by α fusion based on the pixel
value, so for the fusion weight, we have:
Ib.* Ibm
βD = 2
, (14)
∗ + thred 2
Ib.* Ibm
Ib. Ibm
β = , (14)
where β D is the fusion weight of αD fusion
Ib.∗for
IbmD.+It thred2
is worth
2 mentioning that the estimation of
upper surface U should be the opposite of lower surface D, which is:
where β D is the fusion weight of α fusion for D. It is worth mentioning that the estimation of upper
(255 − Ib).* (255 − Ibm )
surface U should be the opposite of βlower
U =
surface D, which is: 2
, (15)
(255 − Ib).* (255 − Ibm ) + thred 2
∗
where thred2 is set to 20 by experience.
βU = In this−
(255 Ib)we
way, can −
. (255 the ) α fusion
useIbm , to combine Ib with Ibm (15)
∗ 2
to obtain a new label image by: (255 − Ib). (255 − Ibm) + thred2
(a) (b)
(a) (b)
Figure 5. Comparison of refinement: (a) the pixel learning, where thred1 = 10, thred2 = 20; (b) the
Figure 5. Comparison of refinement: (a) the pixel learning, where thred1 = 10, thred2 = 20; (b) the guided filter.
guided
Figure filter.
5. Comparison of refinement: (a) the pixel learning, where thred1 = 10, thred2 = 20; (b) the
guided filter.
In addition, it is also important to pay attention to the failure of using the priors of the
In addition, it is also important to pay attention to the failure of using the priors of the
dark/bright
dark/brightchannel
channelin in
the sky region
regionoror the extremely darkshadow
shadow region. The failure will result in
In addition, it is the
alsosky
important thepay
to extremely dark
attention region.
to the failure of The failure
using will result
the priors in
of the
overenhancement,
overenhancement,
dark/bright channelwhichwhich can
in thecanstretch
sky stretch
regiontheornontexture
the nontexture
the extremelydetails
details such
dark such
shadowasas
thethe compression
compression
region. information,
information,
The failure will resultasin as
shown in Figure
shown 6. Therefore,
in Figure
overenhancement, 6. Therefore,
which can wewecancan
limit
stretch limit
the the
theestimation
estimation
nontexture of
ofU
details U to
to be
such be notcompression
not
as the toolow
too low(not
(nottoo
toohigh
highforfor
information, asD) by
D)
cutting
by off the
cutting
shown initialization
off the 6.
in Figure of I2Y:
initialization
Therefore, of I2Y:
we can limit the estimation of U to be not too low (not too high for D)
by cutting off the initialization of I2Y: I 2Y (i ) = min( I 2Y (i ), t )
D D D
I2YD (i ) = min( I2YD (i ), tD, ) (18)
II22YYUD((ii)) == max(
min( I 2YDU (i ), tDU ) , (18)
I2YU (Ii2)Y=(i)max ( I2Y (i ), t , )
= max( I 2YU(i ), t )U
(18)
where i is the index of the look-up table Uand tD and UtU are U
the cutoff thresholds for D and U,
where i is thei index
respectively.
where is the of the look-up
Empirically,
index we set
of the table and
tD = 150
look-up tD and
and
table and tDU and
tU = t70 aredefault
as the cutoff
tU are thethresholds
values. thefor
Oncethresholds
cutoff D and
value forUU,
of Dis respectively.
not
andtoo U,
small
Empirically, and
respectively.
wethe value
tD =of150
Empirically,
set D isand
notset
we ttoo
U t=Dlarge,
asthe
= 150
70 andskytU =region
default 70 as and
values. the values.
default
Once extremelyOnce
the value dark Ushadow
ofthe value region
is notoftooU is of the
not
small too
and the
valueimage
small
of D andwill
is notnottoo
the be large,
stretched
value of Dthe too
is not much,
skytoo soand
large,
region that the region
the sky
the useless
extremely details
and the
darkwill not bedark
extremely
shadow enhanced offor
shadow
region the the
regiondisplay.
image of the
will not
The
image result
will after
not using
be Equation
stretched too (18)
much, with
so default
that the truncation
useless are
details shown
will notin
be
be stretched too much, so that the useless details will not be enhanced for the display. The result afterFigure
enhanced6c. As
for can
the be seen,
display.
Figure 6b shows that the result (18)without
with adefault
truncation is strange in the in
prior failure regions.
can beUsing
usingThe result
Equation after using
(18) with Equation
default truncation are showntruncation are shown
in Figure Figure
6c. As can 6c. As
be seen, Figure seen,
6b shows
Equation
Figure 6b(18) showsto cut offthe
that the result
initialization
withoutmay lead to a more
a truncation comfortable
is strange in the results for theregions.
prior failure human visualUsing
that the result without a truncation is strange in the prior failure regions. Using Equation (18) to cut off
system.
Equation (18) to cut off the initialization may lead to a more comfortable results for the human visual
the initialization
system.
may lead to a more comfortable results for the human visual system.
6. Experiment
6. Experiment and
and Discussion
Discussion
The method
The methodwe weproposed
proposedwaswas implemented
implemented withwith
C++ inC++theinMFC
the (Microsoft
MFC (Microsoft
FoundationFoundation
Classes)
Classes) framework. A personal computer with an Intel Core i7 CPU at 2.5 GHz
framework. A personal computer with an Intel Core i7 CPU at 2.5 GHz and 4 GB of memory and 4 GB of memory
was
was used.
used. Experiments
Experiments werewere conducted
conducted usingusing four kinds
four kinds of lowofcontrast
low contrast
images images including
including thekinds
the three three
kinds mentioned
mentioned before before
and anand an additional
additional one forone for comparison.
comparison. We compared
We compared our approach
our approach with thewith the
typical
typical algorithms
algorithms and studies
and the latest the latest studies
on each kindonof each kind ofimage
low contrast low tocontrast image
verify the to verify and
effectiveness the
effectiveness and superiority
superiority of ours. of ours.
Table 1.
Table 1. Default configuration
configuration of
of the
the parameters.
parameters.
Parameters
Parameters
Values
Values
Equations
Equations
Radius of max/min filter 20 Equation (13)
Radius of max/min filter 20 Equation (13)
Radius
Radius of mean
of mean filter filter 4040 Equation
Equation(13)(13)
thred1
thred1 2020 Equation
Equation(12)(12)
thred2
thred2 2020 Equations (14)(14)
Equations and (15)
and (15)
tD t D 150
150 Equation
Equation (18) (18)
tU 70 Equation (18)
tU 70 Equation (18)
thredWB 50 Equation (19)
thredWB 50 Equation (19)
Information 2017, 8, 135 13 of 20
Information 2017, 8, 135 13 of 20
Information 2017, 8, 135 13 of 20
(a) (b)
Figure 10. Results of enhancement for color images: (a) the original hazy image; (b) our results.
(a) (b)
Figure 10. Results of enhancement for color images: (a) the original hazy image; (b) our results.
Figure 10. Results of enhancement for color images: (a) the original hazy image; (b) our results.
Information 2017, 8, 135 14 of 20
Information 2017, 8, 135 14 of 20
Besides,
Information
Besides, our
2017,
our8, approach
approachalso
135 alsoworks
worksforforgrayscale
grayscaleimages, such
images, as infrared
such as infraredimages. According
images. to20the
14 of
According to
different bit depths, we give another set of parameters. Figure 11 shows the video
the different bit depths, we give another set of parameters. Figure 11 shows the video screenshots screenshots from a
Besides,
theodolite our
of the naval approach
base. also works
Webase.
stretch for grayscale images, such as infrared images. According to
from a theodolite of the naval Wethe infrared
stretch the images
infraredbyimages
a linearbymodel with
a linear the max
model value
with and
the max
minthe different
value, which bitisdepths,
the wewidely
most give another
used set of parameters.
model to display Figure 11
infrared shows the video screenshots
images.
value and min value, which is the most widely used model to display infrared images.
from a theodolite of the naval base. We stretch the infrared images by a linear model with the max
value and min value, which is the most widely used model to display infrared images.
Figure 11. Results of enhancement for infrared images (14 bits): thred1 = 800, thred2 = 1600, tD = 4000,
Figure
Figure 11.11. Results
Results ofof enhancementfor
enhancement forinfrared
infraredimages
images (14
(14 bits): thred1
thred1 ==800,
800,thred2
thred2= =1600,
1600,tDtD
= 4000,
= 4000,
tU = 14,000.
tU tU = 14,000.
= 14,000.
6.3. Haze Removal
6.3.6.3. Haze Removal
Haze Removal
Next,
Next,wewecompare
compareour ourapproach
approach with
with Zhang’s [28], He’s
Zhang’s [28], He’s [23],
[23], Kim’s
Kim’s[29]
[29]and
andTan’s
Tan’s[21]
[21]
Next,
approaches we compare our approach with Zhang’s [28], He’s [23], Kim’s [29] and Tan’s [21] approaches
approachesforforhazehazeremoval.
removal. InIn Figure
Figure 12,
12, the depth edges
the depth edges between
between thetheforeground
foregroundand and
for haze removal.
background may In Figure haloes
produce 12, the by
depth
manyedges between
dehazing the foreground
algorithms. As and
can bebackground
seen, the may produce
results of other
background may produce haloes by many dehazing algorithms. As can be seen, the results of other
haloes
methods by usually
many dehazing algorithms. As can be seen, thethe results
halo,ofexcept
other methods usually look dim,
methods usuallylook
lookdim,
dim,and
and most
most ofof them
them restore
restore the halo, except Zhang’smethod.
Zhang’s method. OnOn thethe
and most
contrary,
contrary,
of them
our restore
ourapproach
the halo,
approachcompensates
except
compensates for
Zhang’s
for the
method.
the illumination
On
illumination and
the contrary,
and has
hasnot our approach
notintroduced
introducedany compensates
anysignificant
significant
for the illumination and has not introduced any significant haloes.
haloes.
haloes.
(a)
(a)
Figure 12. Cont.
Information 2017, 8, 135 15 of 20
Information 2017, 8, 135 15 of 20
Information 2017, 8, 135 15 of 20
(a)
(a)
(a)
(a)
(a)
(a)
(a)
6.7. Quantitative
6.7. Quantitative Comparison
Comparison
Based on the
Based on the above above results
results onmultidegraded
on the the multidegraded image
image in Figurein16,Figure 16, we aconducted
we conducted quantitativea
quantitative using
comparison comparison
the MSE using
(meanthesquared
MSE (mean
error)squared
and the error) and the SSIM
SSIM (structural (structural
similarity similarity
index). Table 2
shows the quantitative comparisons, in which the MSE represents the texture details of an details
index). Table 2 shows the quantitative comparisons, in which the MSE represents the texture image,
of an image, and the SSIM is used for measuring the similarity
and the SSIM is used for measuring the similarity between two images. between two images.
Table 2.
Table 2. Quantitative
Quantitativecomparison of of
comparison Figure 16 based
Figure on the
16 based on MSE and the
the MSE andStructural Similarity
the Structural index
Similarity
(SSIM).
index (SSIM).
He’s [23] Li’s [30] Lin’s [11] Ancuti’s [32] Avg Changes Ours
He’s [23] Li’s [30] Lin’s [11] Ancuti’s [32] Avg Changes Ours
MSE 0.066 0.019 0.023 0.071 0.0283 0.073
MSE
SSIM 0.066
0.48 0.019
0.73 0.023
0.88 0.071
0.56 0.0283
−0.4325 0.073
0.23
SSIM 0.48 0.73 0.88 0.56 −0.4325 0.23
From Table 2, we can observe that the MSE is inversely proportional to SSIM, meaning that the
From
greater theTable 2, webetween
difference can observe that the
the results MSE
and theisinput
inversely proportional
image, to SSIM,
the more details are meaning
restored. that
The
the greater the difference between the results and the input image, the more
Avg changes column shows the average changes in MSE and SSIM between the proposed method details are restored.
and Avg
The thosechanges
from column
the other shows the He’s
studies. average
[23]changes in MSE[32]
and Ancuti’s andapproaches
SSIM betweenhadthe proposed
a higher MSE method
and a
and
lower SSIM than Li’s [30] and Lin’s [11] approaches; in other words, He’s [23] and Ancuti’slower
those from the other studies. He’s [23] and Ancuti’s [32] approaches had a higher MSE and a [32]
SSIM than Li’s
approaches [30]less
were andsimilar
Lin’s [11] approaches;
to the in other
input image, words,
as well He’s [23]
as restored and details
more Ancuti’s
of [32]
the approaches
image. Our
were less similar
approach had thetohighest
the input
MSEimage,
and theas lowest
well asSSIM,
restored more details
meaning that theofproposed
the image. Our approach
method obtained
more texture information than the other methods. From the average changes, we can see that our
Information 2017, 8, 135 18 of 20
had the highest MSE and the lowest SSIM, meaning that the proposed method obtained more texture
information than the other methods. From the average changes, we can see that our method improves
the MSE and reduces the SSIM more than other methods. That means our approach is better suited for
this kind of low contrast images.
7. Conclusions
Image quality can be affected by the shooting of certain scenes, such as hazy, night or underwater,
which will degrade the contrast of images. In this paper, we have proposed a generic model for
enhancing the low contrast images based on the observation of the 3D surface graph, called the effective
space. The effective space is estimated by the dark and bright channel priors, which are patch-based
methods. In order to reduce the artifacts produced by the patches, we have also designed the pixel
learning for edge-preserving smoothing, which was inspired by online learning. Combining the
model with the pixel learning, low contrast image enhancement becomes simpler and has fewer
artifacts. Compared with a number of competing methods, our method shows more favorable results.
The quantitative assessment demonstrates that our approach can provide an obvious improvement to
both traditional and up-to-date algorithms. Our method has been applied to a theodolite of the naval
base and can enhance a 720 × 1080 video stream by 20 ms/f (50 fps).
Most of the parameters of our method were set empirically, which depends on some features
of an image such as the size or depth. To make a better choice of the parameters, more complex
features [36–40] should be studied, which will be figured out systematically in future work.
Besides, since the look-up table model replaced the values one by one, the spatial continuity of
the image will be destroyed and some noise will be introduced. This is a challenging problem, and an
advanced model is needed to keep the memory of spatial continuity. We leave this for future studies.
Acknowledgments: This work is supported by the National Natural Science Foundations of China (Grant No. 61401425
and Grant No. 61602432).
Author Contributions: Gengfei Li designed the algorithm, edited the source code, analyzed the experiment
results and wrote the manuscript; Guiju Li and Guangliang Han made contribution to experiments design and
revised the paper.
Conflicts of Interest: The authors declare no conflicts of interest.
References
1. Land, E.H. Recent advances in Retinex theory and some implications for cortical computations: Color vision
and the natural image. Proc. Natl. Acad. Sci. USA 1983, 80, 5163–5169. [CrossRef] [PubMed]
2. Land, E.H. An alternative technique for the computation of the designator in the Retinex theory of color
vision. Proc. Natl. Acad. Sci. USA 1986, 83, 3078–3080. [CrossRef] [PubMed]
3. Land, E.H. Recent advances in Retinex theory. Vis. Res. 1986, 26, 7–21. [CrossRef]
4. Jobson, D.J.; Rahman, Z.; Woodell, G.A. Properties and performance of a center/surround Retinex.
IEEE Trans. Image Process. 1997, 6, 451–462. [CrossRef] [PubMed]
5. Jobson, D.J.; Rahman, Z.; Woodell, G.A. A multiscale Retinex for bridging the gap between color images and
the human observation of scenes. IEEE Trans. Image Process. 1997, 6, 965–976. [CrossRef] [PubMed]
6. Kimmel, R.; Elad, M.; Shaked, D.; Keshet, R.; Sobel, I. A Variational Framework for Retinex. Int. J. Comput. Vis.
2003, 52, 7–23. [CrossRef]
7. Ma, Z.; Wen, J. Single-scale Retinex sea fog removal algorithm fused the edge information. Jisuanji Fuzhu
Sheji Yu Tuxingxue Xuebao J. Comput. Aided Des. Comput. Graph. 2015, 27, 217–225.
8. Zhang, S.; Wang, T.; Dong, J.; Yu, H. Underwater Image Enhancement via Extended Multi-Scale Retinex.
Neurocomputing 2017, 245, 1–9. [CrossRef]
9. Si, L.; Wang, Z.; Xu, R.; Tan, C.; Liu, X.; Xu, J. Image Enhancement for Surveillance Video of Coal Mining Face
Based on Single-Scale Retinex Algorithm Combined with Bilateral Filtering. Symmetry 2017, 9, 93. [CrossRef]
10. Wang, Y.; Wang, H.; Yin, C.; Dai, M. Biologically inspired image enhancement based on Retinex.
Neurocomputing 2016, 177, 373–384. [CrossRef]
Information 2017, 8, 135 19 of 20
11. Lin, H.; Shi, Z. Multi-scale retinex improvement for nighttime image enhancement. Opt. Int. J. Light
Electron Opt. 2014, 125, 7143–7148. [CrossRef]
12. Xie, S.J.; Lu, Y.; Yoon, S.; Yang, J.; Park, D.S. Intensity variation normalization for finger vein recognition
using guided filter based singe scale Retinex. Sensors 2015, 15, 17089–17105. [CrossRef] [PubMed]
13. Lan, X.; Zuo, Z.; Shen, H.; Zhang, L.; Hu, J. Framelet-based sparse regularization for uneven intensity
correction of remote sensing images in a Retinex variational framework. Opt. Int. J. Light Electron Opt. 2016,
127, 1184–1189. [CrossRef]
14. Wang, G.; Dong, Q.; Pan, Z.; Zhang, W.; Duan, J.; Bai, L.; Zhanng, J. Retinex theory based active contour
model for segmentation of inhomogeneous images. Digit. Signal Process. 2016, 50, 43–50. [CrossRef]
15. Jiang, B.; Woodell, G.A.; Jobson, D.J. Novel Multi-Scale Retinex with Color Restoration on Graphics Processing
Unit; Springer: New York, NY, USA, 2015.
16. Shi, Z.; Zhu, M.; Guo, B.; Zhao, M. A photographic negative imaging inspired method for low illumination
night-time image enhancement. Multimedia Tools Appl. 2017, 76, 1–22. [CrossRef]
17. Wang, Y.; Zhuo, S.; Tao, D.; Bu, J.; Li, N. Automatic local exposure correction using bright channel prior for
under-exposed images. Signal Process. 2013, 93, 3227–3238. [CrossRef]
18. Shwartz, S.; Namer, E.; Schechner, Y.Y. Blind Haze Separation. In Proceedings of the IEEE Computer
Society Conference on Computer Vision and Pattern Recognition, New York, NY, USA, 17–22 June 2006;
pp. 1984–1991.
19. Shen, X.; Li, Q.; Tian, Y.; Shen, L. An Uneven Illumination Correction Algorithm for Optical Remote Sensing
Images Covered with Thin Clouds. Remote Sens. 2015, 7, 11848–11862. [CrossRef]
20. Kopf, J.; Neubert, B.; Chen, B.; Cohen, M.; Cohen, D.; Deussen, O.; Uyttendaele, M.; Lischinski, D. Deep photo:
Model-based photograph enhancement and viewing. ACM Trans. Graph. 2008, 27, 1–10. [CrossRef]
21. Fattal, R. Single image dehazing. ACM Trans. Graph. 2008, 27, 1–9. [CrossRef]
22. Tan, R.T. Visibility in bad weather from a single image. In Proceedings of the IEEE Conference on Computer
Vision and Pattern Recognition, CVPR 2008, Anchorage, AK, USA, 23–28 June 2008; pp. 1–8.
23. He, K.; Sun, J.; Tang, X. Single Image Haze Removal Using Dark Channel Prior. IEEE Trans. Pattern Anal.
Mach. Intell. 2011, 33, 2341–2353. [PubMed]
24. He, K.; Sun, J.; Tang, X. Guided Image Filtering. IEEE Trans. Pattern Anal. Mach. Intell. 2013, 35, 1397–1409.
[CrossRef] [PubMed]
25. Meng, G.; Wang, Y.; Duan, J.; Xiang, S.; Pan, C. Efficient Image Dehazing with Boundary Constraint and
Contextual Regularization. In Proceedings of the IEEE International Conference on Computer Vision, Sydney,
Australia, 1–8 December 2013; pp. 617–624.
26. Jo, S.Y.; Ha, J.; Jeong, H. Single Image Haze Removal Using Single Pixel Approach Based on Dark Channel
Prior with Fast Filtering. In Computer Vision and Graphics; Springer: Cham, Switzerland, 2016; pp. 151–162.
27. Ju, M.; Zhang, D.; Wang, X. Single image dehazing via an improved atmospheric scattering model.
Vis. Comput. 2017, 33, 1613–1625. [CrossRef]
28. Zhang, S.; Yao, J. Single Image Dehazing Using Fixed Points and Nearest-Neighbor Regularization.
In Proceedings of the Asian Conference on Computer Vision, Taipei, Taiwan, 20–24 November 2016; Springer:
Cham, Switzerland, 2016; pp. 18–33.
29. Kim, J.-H.; Jang, W.-D.; Sim, J.-Y.; Kim, C.-S. Optimized contrast enhancement for real-time image and video
dehazing. J. Vis. Commun. Image Represent. 2013, 24, 410–425. [CrossRef]
30. Li, Y.; Tan, R.T.; Brown, M.S. Nighttime Haze Removal with Glow and Multiple Light Colors. In Proceedings
of the IEEE International Conference on Computer Vision, Santiago, Chile, 7–13 December 2015; IEEE
Computer Society: Washington, DC, USA, 2015; pp. 226–234.
31. Ji, T.; Wang, G. An approach to underwater image enhancement based on image structural decomposition.
J. Ocean Univ. China 2015, 14, 255–260. [CrossRef]
32. Ancuti, C.; Ancuti, C.O.; Haber, T.; Bekaert, P. Enhancing underwater images and videos by fusion.
In Proceedings of the 2012 IEEE Conference on Computer Vision and Pattern Recognition, Providence,
RI, USA, 16–21 June 2012; pp. 81–88.
33. Ma, C.; Ao, J. Red Preserving Algorithm for Underwater Imaging. In Geo-Spatial Knowledge and Intelligence;
Springer: Singapore, 2017; pp. 110–116.
34. Ebner, M. Color constancy based on local space average color. Mach. Vis. Appl. 2009, 20, 283–301. [CrossRef]
Information 2017, 8, 135 20 of 20
35. Van Herk, M. A fast algorithm for local minimum and maximum filters on rectangular and octagonal kernels.
Pattern Recognit. Lett. 1992, 13, 517–521. [CrossRef]
36. Le, P.Q.; Iliyasu, A.M.; Sanchez, J.A.G.; Hirota, K. Representing Visual Complexity of Images Using a 3D
Feature Space Based on Structure, Noise, and Diversity. Lect. Notes Bus. Inf. Process. 2012, 219, 138–151.
[CrossRef]
37. Gu, K.; Zhai, G.; Lin, W.; Liu, M. The Analysis of Image Contrast: From Quality Assessment to Automatic
Enhancement. IEEE Trans. Cybernet. 2017, 46, 284–297. [CrossRef] [PubMed]
38. Hou, W.; Gao, X.; Tao, D.; Li, X. Blind image quality assessment via deep learning. IEEE Trans. Neural Netw.
Learn. Syst. 2017, 26, 1275–1286.
39. Gu, K.; Li, L.; Lu, H.; Lin, W. A Fast Computational Metric for Perceptual Image Quality Assessment.
IEEE Trans. Ind. Electron. 2017. [CrossRef]
40. Wu, Q.; Li, H.; Meng, F.; Ngan, K.N.; Luo, B.; Huang, C.; Zeng, B. Blind Image Quality Assessment Based on
Multichannel Feature Fusion and Label Transfer. IEEE Trans. Circuits Syst. Video Technol. 2016, 26, 425–440.
[CrossRef]
© 2017 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access
article distributed under the terms and conditions of the Creative Commons Attribution
(CC BY) license (http://creativecommons.org/licenses/by/4.0/).