Next Article in Journal
Comparative Study of First Order Optimizers for Image Classification Using Convolutional Neural Networks on Histopathology Images
Previous Article in Journal
The Colour of the Night Sky
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

On the Use of Deep Active Semi-Supervised Learning for Fast Rendering in Global Illumination

by
Ibtissam Constantin
1,
Joseph Constantin
1 and
André Bigand
2,*
1
LaRRIS, Faculty of Sciences, Lebanese University, Fanar, BP 90656 Jdeidet, Lebanon
2
LISIC, University of Littoral Côte d’Opale (ULCO), Calais, BP 719, 62228 Calais CEDEX, France
*
Author to whom correspondence should be addressed.
J. Imaging 2020, 6(9), 91; https://doi.org/10.3390/jimaging6090091
Submission received: 28 July 2020 / Revised: 23 August 2020 / Accepted: 3 September 2020 / Published: 8 September 2020

Abstract

:
Convolution neural networks usually require large labeled data-sets to construct accurate models. However, in many real-world scenarios, such as global illumination, labeling data are a time-consuming and costly human intelligent task. Semi-supervised learning methods leverage this issue by making use of a small labeled data-set and a larger set of unlabeled data. In this paper, our contributions focus on the development of a robust algorithm that combines active and deep semi-supervised convolution neural network to reduce labeling workload and to accelerate convergence in case of real-time global illumination. While the theoretical concepts of photo-realistic rendering are well understood, the increased need for the delivery of highly dynamic interactive content in vast virtual environments has increased recently. Particularly, the quality measure of computer-generated images is of great importance. The experiments are conducted on global illumination scenes which contain diverse distortions. Compared with human psycho-visual thresholds, the good consistency between these thresholds and the learning models quality measures can been seen. A comparison has also been made with SVM and other state-of-the-art deep learning models. We do transfer learning by running the convolution base of these models over our image set. Then, we use the output features of the convolution base as input to retrain the parameters of the fully connected layer. The obtained results show that our proposed method provides promising efficiency in terms of precision, time complexity, and optimal architecture.

1. Introduction

In computer graphics, the goal of photo-realistic rendering is to create convincing images, given the description of a world. This process involves different areas like video production [1] and manufacturing [2]. Path-Tracing can be seen as a random walk that makes it possible to estimate the value of a pixel by randomly sampling many paths and computing the average contribution based on Monte Carlo path tracing algorithm [3]. Then, producing photo-realistic images requires computing a complex multidimensional integral of the scene function at every pixel of the image. Due to the high cost of tracing millions of paths, the cost of computing one image is excessive for computer graphics applications. Thus, a challenging problem is to dramatically reduce the computation time in order to provide a practical real-time global illumination.
Different models have been developed to simulate the behavior of the human visual system. Some of these models are based on full reference quality metrics in order to detect the difference between two images while the others use reduced reference machine learning algorithms. In case of quality metrics, the Visible Differences Predictor model was proposed to predict the probability of detecting the difference between two images [4] while the Sarnoff Visual Discrimination Model was based on a set of complex sub-models that simulate several aspects of the human visual system [5]. However, these models are complex and incomplete due to the human system’s complexity. Hence, they need the reference image which is the visually converged one and they require long computation times and are difficult to parametrize [6]. Considering the fact that in path tracing algorithm, the reference image is not initially available like in classic image processing methods, we try in this paper to apply reduced reference quality assessment techniques by searching noise attributes using deep architecture of convolution neural networks.
Machine learning algorithms have provided models for accelerating global illumination rendering by detecting stochastic noise [7,8,9]. Some of these models used support vector machine (SVM) and relevance vector machine (RVM) to train the network with labeled data images [8,9]. Another model used a deep spike neural network with a clustering dynamic algorithm for the training images [7]. The main idea behind these techniques is to detect the increasing quality of images by estimating decreasing noise level using noise attributes. However, the proposed models require large amounts of carefully chosen labeled images to tune their parameters in order to give a good precision on the testing scenes.
One of the main assets of deep learning algorithms over other machine learning algorithms is their great modeling capacity, which allows them to handle complex high-dimensional data-sets through feature representations. Convolution neural networks require large labeled data-sets to construct accurate models; however, in many real-world scenarios, labeling data are a time-consuming and costly human intelligent task. Active and semi-supervised learning algorithms are invented to improve the classification accuracy and to reduce labeling workload using both labeled and unlabeled data. An active learning algorithm takes the images that have the lowest confidence as the most informative ones [10,11]. It selects such images and asks the expert for its label. Active learning needs human involvement and aims at selecting the most useful images for training. It can improve the model’s performance and can accelerate the convergence speed. The semi-supervised learning method selects the images that have the highest confidence and adds the predicted label by the machine itself without any human involvement [12]. However, when the initial model is very weak, many labels would be wrongly predicted, and thus it would introduce mistakes into the training set.
In active learning, uncertain and erroneous portion of the training images are required to be queried, annotated manually with minimum human cost, and added to the training set to get the highest gain in classification accuracy. In the semi-supervised learning, the classifier is constructed beginning with a few labeled training images. A portion of the unlabeled training images that are not selected during active learning are labeled using the current classifier, and the most confident images among the predicted labeled images are added to the training set, repeatedly, until convergence. It can be seen that there is a very good complementarity between active and semi-supervised learning. Thus, we integrate them in global illumination rendering in order to reduce the size of the large labeled data-sets that are needed to efficiently train convolution neural networks and to improve model performance.
In this paper, we present an active semi-supervised learning algorithm based on deeply embedded feature representations by alternatively using labeled and unlabeled data images for real-time and interactive global illumination. We use a deep architecture which is a powerful machine learning technique from the field of deep learning in order to extract a rich feature representations for a wide range of images [13]. The role of deep learning is to create higher level noise prediction through the use of multiple layers of nonlinear operations. In order to further reduce manual labeling workload, the semi-supervised learning algorithm is designed to make full use of those unlabeled images that are not queried by active learning [14,15]. The model uses semi-supervised learning to select class central sub-images that are not selected during active learning based on different layers for features extraction, concatenation, and prediction [14,16]. It has also been shown that semi-supervised learning overcomes the limitations of supervised learning in many different applications [14,17]. However, to our knowledge, its performance was not well investigated in the domain of stochastic noise detection for rendering algorithms using deep convolution neural networks. The contributions of this paper are summarized in the following:
  • We used active semi-supervised learning for efficient stochastic noise detection in global illumination rendering algorithms.
  • We combined active and semi-supervised learning to minimize the effect of imbalanced image classification in order to model the capability of the deep feature distribution for efficient noise detection.
  • We extensively evaluated our algorithm on different global illumination scenes with different resolutions containing diffuse and specular surfaces. The proposed algorithm demonstrates outstanding performance compared with state-of-the-art algorithms applied for accelerating global illumination rendering.
  • We made a comparative study based on memory space and computation time. We showed that our model is computational efficient for real world applications because it yielded a small questioning time on a block of images.
The motivation behind this work is to perform a rapid calculation of the illumination in the rendered images. The convolution neural network is proposed to handle the data, but due to the typical training of the deep neural network, we propose a semi-supervised training algorithm, which improves the performance and makes the approach suitable for real-time applications. One key issue for most deep learning algorithms is that they need large amounts of labeled images to train the model. Since manual labeling is time-consuming, we have proposed to combine active learning and semi-supervised learning to reduce manual labeling workload. Realizing that active learning algorithm is only interested in images that are more likely to be on the class boundary, while ignoring the usage of the rest large amounts of unlabeled images, this paper designs a semi-supervised learning algorithm to make full use of the rest of the non-queried images. The proposed active semi-supervised algorithm uses active learning to select class boundary images, and semi-supervised learning to select class central images. By adding class central images, we are believed to better describe the class distribution, and to help active learning to find the boundary images more precisely.
The paper is structured as follows: Section 2 gives a discussion on related work on active semi-supervised learning, Section 3 describes how to design the image quality database, and Section 4 describes the architecture of the models and the learning methodologies. Section 5 shows the experimental results. Section 6 gives a discussion summarizing the advantages of the proposed approach and the outcomes and findings of the experimental analysis. Finally, the paper is summarized with some conclusions in Section 7.

2. Related Work on Active Semi-Supervised Learning

One of the main drawbacks using deep learning is about the necessity of a huge learning database. A large amount of labeled computer-generated images is expensive to acquire, but unlabeled images are easy to generate. Thus, the idea of active learning (AL) [18] and semi-supervised learning [19] to improve our classification task using unlabeled images is investigated. These two iterative procedures (AL and SSL) make it possible to improve machine learning in a lot of applications. At each iteration, AL selects the image that has the lowest confidence as the most informative one and asks the human observer for its label (noisy or not). At the same time, SSL selects the images that have the highest confidence and adds the predicted labels by the machine without any human investigation (HVS that takes about one hour per image). Thus, this procedure is very efficient in case of computer-generation of images.
The scientific literature on semi-supervised learning (SSL) is growing rapidly, showing significant performance gains in recent years. A semi-supervised feature extraction algorithm has been developed for pattern categorization [20]. However, this algorithm was tested only for classification on benchmark databases without considering experimental data for real-life applications. The performance of the inductive rules family classifiers has been improved by presenting a new technique for the presentation order of the training samples, which combines a clustering method with a density measure function [21]. The main drawback of this approach is that the convergence to a good result could be quite time-consuming especially when the training set contains thousands of samples. Thus, it cannot be applied in case of global illumination because the learning algorithm should select pertinent samples from a set containing a huge number of unlabeled images. Recently, ref. [22] proposed to use manifold embedding with deep learning architecture to improve the results using the structural information of the data.
A group of semi-supervised methods consists of three different models trained on the same dataset using bootstrap sampling. After a supervised training of every model, an unlabeled data point is added to the training set of one model only if the other two models agree on its label [23]. Another type of semi-supervised methods can be referred to as self supervised methods. These algorithms learn to exploit robustness to stochastic perturbations caused by noise or randomness in data augmentation [24,25,26]. This idea has been recently applied by [27] for image set classification with success. The idea of integrating active method along with semi-supervised learning was also introduced in the scientific literature. An active based semi-supervised SVM (SVM, support vector machine) learning algorithm was proposed in [28,29]. These algorithms used active learning to improve the performance of the initial classifier, and then the semi-supervised learning algorithm was designed to assign labels to the remaining unlabeled samples. Other research papers combined active with semi-supervised learning for image compression, speech recognition, and multi-task learning [30,31,32]. A very competitive class of semi-supervised learning algorithms was proposed to exploit robustness in case of stochastic noise [33]. However, the proposed models take a full dimension of the image as input without using a deep convolution neural network (CNN), which is shown to be successful. Indeed, CNNs can extract appropriate noise features while jointly performing discrimination. Moreover, the performance of the developed algorithms was not investigated in case of scenes with different resolutions and by using different rendering scenarios. The goal of this paper is to create a new approach based on deep active semi-supervised learning to automatically allow a stopping criterion when a perceptive convergence is reached in global illumination methods. We show the advantage of this model in terms of parameters number and high precision by comparing it with SVM.

3. Design of the Image Quality Database

Different Image Quality databases are available to test the performance of the learning algorithms with respect to the human visual thresholds. However, scenes captured using camera devices are usually afflicted by mixture of multiple distortions that do not modelize synthetic distortions found in existing databases well [34,35]. The model is built on data corresponding to images of globally illuminated scenes. The path tracing algorithm was used in order to reduce noise. This algorithm generates stochastic paths from the camera to the 3D scene. For each intersection of a path with the surface, a direction of reflection or refraction is randomly extracted. The luminance at a point x in direction w is defined by [36]:
L ( x , w ) = L e ( x , w ) + S V ( y , x ) f r ( w y x , x , w ) L i n ( y , x ) G ( y , x ) d A ( y )
where S is the scene surface, L e is the emitted luminance, V is the mutual visibility, f r is the bidirectional scattering distribution function, w y x is the direction from y to x, L i n is the incidence luminance, G ( y , x ) = cos θ y cos θ x / y x 2 , θ is the direction between incident and normal directions, and A is the area measure.
The luminance of a pixel is evaluated recursively based on a Monte Carlo technique [37]. For each pixel, the final luminance is the average of the contributions of all generated paths. We can then compute several images from the same point of view by adding equally, between two successive images, a certain number of new paths for each pixel. For each scene, several images were obtained, the first one being very noisy and the last one being the reference image. In order to get experimental data about noise perception, pairs of images were presented to the observer. One of these images, called reference image, was computed with a high number of paths per pixel. The second image, the so-called test image, was chosen from a stack of images arranged from very noisy ones above to converged ones below. During the experiments, the observer was asked to modify the quality of the noisy image by pointing to the areas where the differences were perceived between the current image and its reference one. Each operation then entailed the selection and display of the corresponding next level sub-image by reducing visually the noise in this image’s sub-part. This operation was done until the observer considered that the two images are visually identical. This operation is reversible meaning that an observer is able to go down or up into the images’ stack. Note that all the observers worked in the same conditions, the same display with identical luminance tuning, and the same illumination conditions. The results were recorded for 33 different observers and the average number of paths required for each sub-image to be perceived as identical to the reference one by 95% of the observers were computed. We tested the performances of our algorithms on global illumination scenes with 512 × 512 and 800 × 800 resolutions. For the classic scenes with 512 × 512 resolution, the images were cut into sixteen non-overlapping blocks of sub-images of size 128 × 128  pixels (Figure 1). The maximum number of paths per pixel was set to 10,100 in order to obtain non-distorted copies of the sub-images. The sub-images were computed by adding 101 paths for each pixel between two successive sub-images using the path tracing algorithm [38]. The scenes with 800 × 800 resolution were computed for diffuse and specular rendering (Figure 2). The human vision system (HVS) thresholds for each sub-image were then tested. In this case, the images were cut into 16 non-overlapping sub-images of size 200 × 200 pixels. The number of paths per pixel between two successive sub-images and the largest number of paths per pixel were set differently for each scene in order to test the performance of the deep active semi-supervised learning algorithm using different scenarios (Table 1).
The labeling process selected sub-images computed using diffuse and specular rendering and asked the observers for their qualities [8]. The average number of paths required for each block of sub-images to be perceived as identical to the reference one is shown in Figure 3.

4. The Proposed Method

4.1. Architecture of the Convolution Neural Network

Convolution neural networks are usually composed of several processing layers, each layer involving linear as well as nonlinear operators that are jointly learned in an end-to-end manner to solve a particular task [39,40]. In practice, very few people train an entire convolution neural network from scratch because it is relatively rare to have a dataset of sufficient size. Instead, it is common to take a convolution neural network that is pre-trained on a different large dataset and then use it either as a feature extractor or as an initialization for a further learning process.
Feature extraction consists of using the representations learned by a previous network to extract interesting features from new images. These features are then introduced to a new prediction model which is trained from scratch. In order to extract the noise features, we took the convolution base of a previously trained network [7], ran the data on it, and trained a new prediction model on top of the output. The network was designed using twelve convolution layers of depth and spread equal to one [41,42,43] (see Table 2). Next, we inserted a thirteen layer which applies a one stage 2D wavelet decomposition on the input sub-image in order to extract noise from the wavelet coefficients in the high frequency sub-bands [44]. The sub-image noise was estimated as a pixel subtraction between the current sub-image and the sub-image computed by each layer. The mean and the standard deviation pooling were applied to the thirteen activation layers. We performed such image analysis employing statistical noise features because these features can effectively distinguish digital images from their tampered versions [45,46]. We plot the values of the standard deviation versus each sub-image per block (Figure 4). It is shown that these features are important to the learning models because their values monotonically change until they reach stable thresholds when the sub-images are not affected by noise.
Next, the feature concatenation pooling was performed by grouping the version feature vectors in a single feature vector in order to obtain a total of 26 noise feature vector used as input to the model. The architecture of the convolution neural network is shown in Figure 5.
The input to the prediction model was obtained by computing the difference between the noise features of a quick ray traced sub-image of the scene and the current one. Finally, we ran the convolution neural network over our data-set, thus recorded its output and then used this data as input to train a support vector machine model called DSVM and an active semi-supervised learning model called DSVM AS in order to predict the quality of the global illumination sub-images (Figure 6).
Next, these models were compared with the standard support vector machine (SVM) and the active semi-supervised SVM called SVM AS by considering the full luminance component of L a b color sub-images (L) as input and applying to L four different denoising algorithms: Linear filtering with averaging and Gaussian filters, median filters and adaptive Wiener filters. Each sub-image was also denoised using Wavelet analysis [33]. A comparison has also been made with the pre-trained VGG19 deep convolution neural network in order to show the efficiency of our work. The output of all models contained only one neuron which gave the value −1 for noisy sub-images and the value + 1 for low-noisy sub-images.

4.2. Active Semi-Supervised Learning of Noise Features

4.2.1. Active Learning

Active learning has been developed to improve learning ability in complex environments [47] such as our task. Active learning is a kind of semi-supervised learning algorithm that selects the most informative images to be labeled by an expert. In active learning, an uncertain portion of the training image set is required to be queried, annotated by an expert with minimum human cost and added to the training images set to get the highest gain in prediction accuracy. Most active learning methods select only a single image at each iteration. In active learning batch mode, a batch of images are selected at each step, and instead of updating the model at each iteration, the update is performed once at each batch iteration [48]. It is shown that the images closest to the current hyperplane are informative so they can shrink the size of the version space (the version space is defined in [49]), and, in this case, active learning can accelerate the convergence of the model when the version space is symmetric. Such images are more likely to be on the class boundary; that is to say, they have a high probability to accelerate the convergence of the model.
The active learning algorithm tries to query the sub-images that are the closest to the current classification hyperplane for each iteration. However, the prerequisite that the version space should be symmetric cannot be satisfied in many real-life situations. Moreover, when the labeled training set is small, it is difficult to describe the spatial distribution of the pattern classes. As a result, the sub-images closest to the current classification hyperplane may be less likely to be on the class boundary, and then selecting them is not the best choice. To solve this problem, we try in this paper to use semi-supervised learning in order to add some sub-images that can better describe the class distribution into the labeled training set, then the class boundary would be more clear, and it could be very beneficial for active learning to precisely select the class boundary samples. We take the class central sub-images as the ones that can better describe the class distribution.

4.2.2. Deep Semi-Supervised Learning

In this paper, we adopt a flexible strategy by integrating the uncertainty from the concept of active learning and the confidence criterion from that of semi-supervised learning (see [50] for more details). Semi-supervised learning is widely used in deep learning frameworks [51], where the learning dataset is often a challenging task. The active semi-supervised learning algorithm initially trains the prediction model to extract noise features from a small set of confidently labeled sub-images, and it repeatedly retrains the prediction model by adding the batch of sub-images features selected by the active learning algorithm. Then, the confidence criterion of semi-supervised learning is applied to produce machine-labeled sub-images. Finally, we train the prediction model on the unlabeled sub-images using active learning until convergence.
The semi-supervised learning algorithm is applied for the unlabeled remaining sub-images that are not chosen during active learning [16]. This algorithm adds class central sub-images to describe the class distribution into the labeled training set, then the class boundary could be very beneficial for accelerating the convergence [33].
Class central sub-images are more likely to exist in the non-queried set of sub-images, then the updated unlabeled sub-images set U is computed as follows:
U = U x q
where x q is the set of sub-images queried during active learning. The semi-supervised learning algorithm explores the central sub-images which are at a median distance to the current classification hyperplane. In case of scene rendering, the unlabeled sub-images should be classified into two classes: the noisy sub-image class denoted as N and the low-noisy sub-image class denoted as R as follows:
N = x i | x i U , f ( x i ) < 0
R = x i | x i U , f ( x i ) > 0
where f ( x i ) is the output of the prediction model for the sub-image x i . During the iteration of active learning, the score changing rate can be computed as follows:
ϖ ( x i ) = ( t = j j + P 1 c h t ( x i ) ) / P
where P is the number of iterations of active learning and c h t ( x i ) is defined as follows:
c h t ( x i ) = 0 , if l t ( x i ) l t 1 ( x i ) 1 , if l t ( x i ) = l t 1 ( x i )
where l t ( x i ) is the label of the sub-image x i at time t given by the model itself. The proposed semi-supervised learning algorithm computes the score changing values for the unlabeled sub-images of each class, and selects the sub-image central class of which the changing value score remains unchanged during the iterations of the active learning algorithm as follows:
U N = x i | x i N , ϖ ( x i ) = 0
U R = x i | x i R , ϖ ( x i ) = 0
Then, by using U R and U N , the two sub-images that are chosen by the learning algorithm from the noisy and the low-noisy classes have the median distance to the current classification hyperplane as follows:
x N = m e d i a n x i ( d ( x i ) | x i U N )
x R = m e d i a n x i ( d ( x i ) | x i U R )
Then, by using the changing value score, the sub-image sets U R and U N would have a higher confidence on unlabeled sub-image predicted scores. The deep active semi-supervised learning for global illumination algorithms is given by Algorithm 1.
Algorithm 1 Deep active semi-supervised learning algorithm
Require:Set of scenes; P: Number of iterations of active learning; S: Number of iterations of semi-supervised learning.
1:
for all (Scenes) do
2:
  Read-Randomly(sub-image).
3:
   I ← CNN(sub-image).
4:
  if Labeled(sub-image) then
5:
    L I
6:
   else
7:
    U I
8:
   end if
9:
end for
10:
Train-Prediction-Model(L).
11:
for k { 1 , , S } do
12:
   for l { 1 , , P } do
13:
    L ← Run-Active-Learning( U ).
14:
   end for
15:
  Calculate-Labels-Changing-Rate( U )
16:
   I + ← Select-Class-Center(noisy).
17:
   I ← Select-Class-Center(low-noisy).
18:
   L + ← Add-Predicted-Labels( I + ).
19:
   L ← Add-Predicted-Labels( I ).
20:
   L ← Add( I + , L + , I , L ).
21:
   U ← Remove( I + , I ).
22:
  Retrain-Model( L ).
23:
end for
24:
repeat
25:
   L ← Run-Active-Learning( U ).
26:
until Stopping Criterion

5. Experimental Results

5.1. Experimental Setup

In order to select the initial parameters of the models, we conducted experiments on the 512 × 512 resolution scenes. The inputs for SVM and deep SVM (DSVM) were vectors of size 16,384 ( 128 × 128 ) and 26, respectively. We used a cross validation with different RBF kernels on the scenes (a)–(c) to determine the standard deviation σ and the penalty factor C [46]. It consisted of dividing the data into V groups of the same size and putting a group away. The learning was then carried out on the V 1 other groups and the product model was used to predict the excluded group. In an iterative way, the same treatment was repeated on all V groups tests, removing each time a group. The precision estimate was the total precision rate across all V group tests. The learning set was split into 303 groups each of size 16 sub-images. Table 3 shows the models optimal parameters. It gives the penalty factor C, which controls the trade-off between complexity of the model and the number of non-separable sub-images, and the standard deviation σ for the radial basis functions that verify Mercer’s condition. The Mercer’s theorem tells us only whether a kernel is actually an inner-product kernel in some space, and therefore, admissible for use in a support vector machine. In mathematics, specifically in functional analysis, Mercer’s theorem states that a symmetric, and positive-definite matrix can be represented as a sum of a convergent sequence of product functions. From Mercer’s theorem, a matrix is a Gram Matrix if and only if it is positive and semi-definite, i.e., it is an inner product matrix in some space. For a function to be a kernel, the inner product matrix created by a data-set should necessarily be positive-semi-definite [52]. We notice that SVM needs less parameters than deep SVM and gives higher precision, which is equal to 96.24 % .
Experiments were also done on the 800 × 800 resolution scenes. The SVM and DSVM used as input a vector of size equal, respectively, to 40,000 ( 200 × 200 ) and 26. The scenes (g) and (i) were used for cross-validation in order to find the optimal parameters. The set of sub-images was split into 235 and 630 groups respectively for SVM and DSVM, each of size 16 sub-images. The model optimal parameters are shown in Table 4. We find that SVM needs less number of parameters than deep SVM and gives higher precision, which is equal to 97.50 % .

5.2. Parameter Fine-Tuning

We first considered the sub-images of the scenes with 512 × 512 resolution computed by the path tracing algorithm. The initial set of labeled sub-images contained only 505 sub-images selected from the scene (a). The set of unlabeled sub-images contained 9191 sub-images selected from the remaining sub-images of the scenes.
We found that the performance of the active semi-supervised learning algorithm quickly increases at the first 10 iterations, so we used, S = 10 and P = 5 as initial values for the active semi-supervised algorithm. The active learning iterated 50 times and the semi-supervised learning added 20 sub-images centers with their predicted labels for SVM and DSVM. In each iteration, active learning took the sub-images that were closest to the current classification hyper-plane; it selected such sub-images and asked the expert for their labels. The actual contribution of the expert in the active phase was to give the active learning algorithm the correct label for each sub-image that had the lowest confidence in order to improve the performance of the model. Active learning needs human involvement and aims at selecting the most useful sub-images for training. It can greatly improve the model’s performance and can accelerate the convergence speed. During active learning, 1091 and 3212 sub-images were added to the set of labeled sub-images respectively for SVM and DSVM, in order to obtain a maximum precision equal to 99 % . In order to obtain a good stopping criterion, we computed the precision on the set of the sub-images that were not selected during active learning, given that all our global illumination sub-images were labeled by human observers as shown in Section 3. In this case, the precision was equal to the percentage of sub-images classified correctly on this set. In practice, and during online decision, it should be important to define a testing set of labeled sub-images in order to compute the precision and to find a good stopping criterion. The variation of precision for SVM and DSVM active semi-supervised learning ( SVM AS and DSVM AS ) is shown in Figure 7. It is clear that, in case of SVM AS , the precision increases after few oscillations by querying some pertinent sub-images until it reaches its maximum value after 1040 iterations. This might be due to the fact that, in this space, the two classes of sub-images overlap severely.
Figure 8 shows the nine images added at the first iterations of the active learning algorithm. It contains, for each selected sub-image, the scene name to which it belongs, its block number, and the number of paths computed by the path tracing algorithm in order to generate it. In order to accelerate the convergence, it is shown that most of the sub-images selected are noisy and have a number of paths near the boundary of decision, which is close to the human visual score.
The number of kernels was equal to 983 and 2584 for SVM AS and DSVM AS , respectively. Moreover, the DSVM AS model yielded a good precision with a number of parameters less than SVM AS . The number of parameters was equal to 26 × 2584 + 2584 = 69,768 for DSVM AS and 128 × 128 × 983 + 983 = 16,106,455 for SVM AS .
Next, we selected the sub-images of the scenes with 800 × 800 resolution computed by the path tracing algorithm. The initial set of labeled sub-images contained only 475 sub-images selected from the scene (i). The set of unlabeled sub-images contained 9605 sub-images selected from the remaining sub-images of the scenes. The semi-supervised learning program was executed with the initial parameters S = 10 and P = 5 . After 50 iterations of active learning, the semi-supervised learning program added 20 sub-images centers with their predicted labels to the set of labeled sub-images. The active learning algorithm added 2386 sub-images to the set of labeled sub-images in case of SVM AS , and 3285 in case of DSVM AS . The maximum precision is equal to 99 % and 93.4 % , respectively, for SVM AS and DSVM AS . The number of kernels for SVM AS and DSVM AS was equal, respectively, to 1402 and 2066.

5.3. Performance Comparison

5.3.1. Comparative Study with SVM

In this study, we compared the active semi-supervised learning algorithm with SVM in order to show its robustness. We tuned the parameters of the SVM and DSVM models on the 3232 sub-images of the 512 × 512 resolution scenes (a) and (b). The optimal number of kernels was equal to 1223 and 1560, respectively, for SVM and DSVM. We noticed that DSVM used a number of parameters less than DSVM AS , SVM AS and SVM. The number of parameters for this model was equal to 26 × 1560 + 1560 = 42,120 .
The models’ performances were compared on all global illumination scenes. The variation of the mean square error and the mean square error ranges are listed in Figure 9 and Table 5, respectively. It is clearly seen that SVM AS achieves higher performance on the scenes (b)–(f). The mean square error is between 0 and 0.09 for this model, except for the blocks 6 and 7 of the scene (f), where the DSVM AS model assures a better performance than other models. However, the SVM model assures higher performance than SVM AS on the scene (a) considering that this scene is used to tune its parameters. Moreover, it is shown that DSVM AS model gives better precision than DSVM does on the scenes (b)–(f) and provides mean square errors less than SVM on the scenes (d)–(f). Finally, the actual thresholds of the models and the desired ones obtained by the human visual system (HVS) are also compared to show the effectiveness of our approach (Figure 10). It is shown that the SVM AS gives similar results to those of the human visual system (HVS) on the 512 × 512 resolution scenes. We also found that the DSVM AS model gives more coherent results to those of the human visual system (HVS) than DSVM does on the 512 × 512 resolution scenes and achieves a better precision than SVM on the scenes (d)–(f). In addition, it is clearly seen that SVM AS is better in precision than DSVM AS . However, the DSVM AS model needs less parameters for fast rendering applications.
We also tested the accuracy of the models by using the F 1 score. The highest possible value of F 1 is 1, indicating perfect precision and recall [53]. Table 6 shows the variation of the F 1 score of the deep learning models for the 512 × 512 resolution scenes. It is shown that the SVM AS and DSVM AS models assure a better accuracy than other models.
Next, we considered the scenes with 800 × 800 resolution and we tuned the SVM and DSVM models on the scenes (g) and (i). The experimental learning set contained 3760 sub-images. The number of kernels for SVM and DSVM was equal, respectively, to 2367 and 1380. It was observed that the DSVM model used a number of parameters less than the other learning models. The number of parameters for this model was equal to 26 × 1380 + 1380 = 37,260 .
Finally, we applied the path rendering algorithm by adding at each iteration different number of paths to each sub-images based on its noise level (Section 2). For each block of the selected scenes, the variation of the mean square error (Figure 11), the thresholds of the learning models, as well the desired human visual thresholds are shown in Figure 12.
The mean square error ranges for the learning models are listed in Table 7. We find that SVM AS gives the best performance on the 800 × 800 resolution scenes. The mean square error is between 0 and 0.03 for all scenes with the exception of the block 5 in the scene (i) where the DSVM AS model assures a better performance than SVM AS with a mean square error equal to 0.02 .
It should be also noticed that DSVM AS provides mean square errors less than DSVM on all the 800 × 800 resolution scenes apart from the scene (g) that is selected to tune the DSVM model parameters. This study also shows that SVM AS gives similar results to those of the human visual system (HVS) on all the used scenes. However, the DSVM AS model uses a simple architecture for fast path tracing algorithms.
Table 8 shows the variation of the F 1 score of the deep learning models for the 800 × 800 resolution scenes. It is shown that the SVM AS and DSVM AS models assure a better accuracy than other models.

5.3.2. Comparison Based on Memory Space and Time Complexities

We tested the time and space complexities for our models on a PC with 2 Intel CPUs at 2.54 GHZ and 16 GB of RAM using Matlab. Table 9 shows the number of parameters needed for each model and the times in seconds needed for learning and testing on any block of the 512 × 512 resolution scene. A comparative study based on time complexity between this model and the pooling spike neural network defined in our previous research [7] shows that the testing time for the pooling spike neural network on the same number of sub-images is equal to 0.01 s; however, it is equal to 0.008 s for the deep active semi-supervised model.
It is shown that the DSVM AS model is computationally efficient for real world applications because it gives a small run-time on a block of sub-images (101 sub-images).

5.3.3. Comparison with Other State-of-the-Art Deep Learning Models

We considered first the VGG19 model which is composed of two parts. The first part, which is called the convolution base of the network, is composed of a stack of convolution and max pooling layers. The second part, which is called the classifier, is generally composed of two fully connected layers and a softmax layer. The main objective of the classifier is to classify the image according to the characteristics detected in the convolution base (Figure 13). We do transfer learning, which is a popular method in computer vision because it allows for building precise models in a considerable time saving. With transfer learning, instead of starting the learning process from scratch, deep learning models that have been learned when solving different problems are used. We first ran the convolution base of the VGG19 network over our image set composed from the scenes (a) and (c)–(f) in order to obtain good precision by using a large labeled images set to construct accurate model. Then, we used the output features of the convolution base as input to retrain the parameters of a fully connected layer with 256 sigmoid hidden units, a dropout layer, and a fully connected layer with one sigmoid output unit. The maximum precision on the evaluation scene (b) is equal to 86 % after 1000 cycles (Figure 14). The learning process stopped when the generalization performance had peaked its maximum value on the evaluation scene (b). Next, we applied the path tracing algorithm on the 512 × 512 resolution scenes. Figure 15 shows the variation of the actual thresholds of the VGG19 model, the human visual system (HVS) score, and the DSVM AS model. It is shown that DSVM AS gives more coherent results to those of the human visual system (HVS) than the VGG19 model.
Next, we evaluated our proposed methodology against MobileNet, InceptionV3, and DenseNet201 deep learning models [54]. The scenes (a) and (c)–(f) are used for training and the scene (b) is selected for the evaluation process. Table 10 shows the maximum precision for the deep learning models on the scene (b) after 1000 training cycles. It shown that the SVM AS and DSVM AS models assure a higher precision than other state-of-the-art deep learning models.

5.3.4. Comparison between Active and Semi-Supervised Learning

We further compared our active semi-supervised learning algorithms SVM AS and DSVM AS with the active learning models SVM AL and DSVM AL . In Figure 16a, we show the mean square errors of each scene over the number of images queried when performing learning on 512 × 512 resolution scenes. We find that the SVM AS model is better in performance than the SVM AL model. However, when we compared the DSVM AS model with DSVM AL , similar performance was obtained (Figure 16b). We conclude that the proposed active semi-supervised algorithm provides advantageous results compared with active learning when the two classes overlap severely. However the two algorithms give similar performance otherwise.

6. Discussion

We have proposed a novel active semi-supervised learning method, utilizing the powerful deep approach for finding a stopping criterion, when a perceptive convergence is reached in global illumination algorithms. The active learning algorithm adds labeled data to the training set at each iteration from an imperfect images-set by a selective sampling strategy followed by queries, and is expected to improve detection performance. The approximation of an image using path tracing algorithm are widely used in computer graphics, they are prone to create imperfect shadowing effects since only the visible parts of the scene are taken into account [55].
While active learning selects the most uncertain images based on the assumption of a perfect expert, a semi-supervised algorithm adds samples by taking the unlabeled images at median distance to the current classification hyper-plane and by considering the label changing rate to ensure labeling reliability. This method achieved competitive results on global illumination scenes with different resolutions containing diffuse and specular surfaces. This method did not require a large validation set to tune the parameters of the learning model. This is considered a huge advantage over many state-of-the-art deep learning algorithms, where the dependency on large validation sets to adjust parameters is a disadvantage.
These experiments lead us to several observations that will be useful in developing in the future a detection system based on imperfect data. These methods perform well when the learning and the testing processes are initially done with a Path Tracing algorithm on standard global illumination scenes. However, rendering a single image is difficult to interpret because of image artifacts which hinder structure determination. The general solution to this problem is to generate multiple images of the same data from displaced view points and view them sequentially. In this case, multiple images of complex data involve significant allocation of both time and computing resources and the depth information measure is only available while the image is in motion. This problem can be solved by generating a single stereoscopic pair. Stereoscopy gives unambiguous depth information when it is the only available depth cue, and depth discrimination is enhanced when other cues are also included.
Only two images need to be rendered, the two halves of the stereo pair, and no motion is necessary. Images can be examined with no loss of perceived structure. In future research, we will try to accelerate global illumination rendering in case of computer generated stereoscopic images. The thresholds computed in this paper were realized in LISIC laboratory based on a training database built from experimentation performed with students. These students do not have any knowledge about computer graphics. We were therefore interested in knowing whether other people would perceive noise in the same sub-images and the difference in accuracy in their perception. Thus, experiments were performed with different users. The results we obtained show that the thresholds remain coherent with the previous ones by 95 % of the users.
The results show that it is possible to achieve a detection performance that outperforms the performance obtained with fully labeled images, even when a fraction of the training data are used in the training set. As a practical matter, the experiments show that the active semi-supervised learning can be applied to an existing perception model that was originally designed for supervised training and it requires a small time during testing, so it can be used to make online decisions on photo-realistic images. Experiment results also prove that the deep active semi-supervised framework can significantly outperform supervised models for accelerating a global illumination algorithm and confirm that the proposed method can be applied on a wide variety of applications.

7. Conclusions

The path rendering algorithm generates theoretically correct images with an infinite number of paths. To render an image in a finite time, the number of paths has to be set before launching the rendering algorithm. Limiting the number of paths introduces a remaining variance, which can be seen as rendering noise in the final image. To analyze the quality of the final image, it would be interesting to have a noise model introduced by a rendering algorithm, and it is under investigation. However, such a model can be difficult to obtain and is specific to the considered algorithm, whereas many algorithms are implemented in modern renders. This paper introduces the application of deep active semi-supervised learning algorithm for stochastic noise detection in rendering algorithms. The introduced approach selects the pertinent sub-images using active semi-supervised learning and the stochastic noise features for the prediction models by using a specific deep convolution neural network. It is shown that the SVM AS model assures the best performance based on minimum square error and threshold accuracy on the different global illumination scenes; however, it needs a high questioning time which slows down its response during online rendering. It is clear that the introduced deep active semi-supervised model ( DSVM AS ) offers high prediction on the testing set when compared with SVM, DSVM, and VGG19 models. In this case, the performance of the model is improved by selecting only the pertinent sub-images during learning.
In addition to offering a good prediction on the testing base, this model gives a small questioning time on a block of sub-images so it is computationally efficient for online rendering in case of real world applications. Indeed, we think that the new paradigm we have introduced (specific CNN learning using active semi-supervised algorithm) should be of great interest for a great number of knowledge-base models that are involved in numerous applications.
In our case, a future research method could be to use deep active semi-supervised learning in case of computer generated stereoscopic images, since convolution neural network is less sensitive to the curse of dimensionality. Our technique can also be compared with other approaches such as deep spike neural networks in case of higher dimension images (HDR images). It is also important to investigate how to optimize our work on a GPU card using parallel computing. Finally, our approach also requires building a larger number of images set computed with different rendering techniques.

Author Contributions

This is to certify that all the authors have participated sufficiently in the work to take public responsibility for the content, including participation in the concept, design, analysis, writing, or revision of the manuscript. All authors have read and agreed to the published version of the manuscript.

Funding

This project has been funded with support from the Lebanese University under grant number 428/2015.

Acknowledgments

We would like to thank the LISIC laboratory at the Littoral Cote d’Opale University (ULCO) for providing us with the data used in our experiments. Special thanks to Miss Ferial Srour-Nemr for her excellent proofreading.

Conflicts of Interest

Each author certifies that this material or similar material has not been and will not be submitted to or published in any other publication.

Code and Data Availability Policies

The data that support the findings of this study are available from the LISIC Laboratory at ULCO University France but restrictions apply to the availability of these data, which were used under license for the current study, and so are not publicly available. Data are however available from the authors upon reasonable request and with permission of the LISIC Laboratory.

References

  1. Wang, B.; Meng, X.; Xu, Y.; Song, X. Fast Point Based Global Illumination. In Proceedings of the 2011 12th International Conference on Computer-Aided Design and Computer Graphics (CAD/Graphics), Jinan, China, 15–17 September 2011; pp. 93–98. [Google Scholar]
  2. Ikeda, S.; Watanabe, S.; Raytchev, B.; Tamaki, T.; Kaneda, K. Spectral Rendering of Interference Phenomena Caused by Multilayer Films Under Global Illumination Environment. ITE Trans. Media Technol. Appl. 2015, 3, 76–84. [Google Scholar] [CrossRef]
  3. Popov, S.; Ramamoorthi, R.; Durand, F.; Drettakis, G. Probabilistic connections for bidirectional path tracing. Comput. Graph. Forum 2015, 34, 75–86. [Google Scholar] [CrossRef]
  4. Takouachet, N.; Delepoulle, S.; Renaud, C. A perceptual stopping condition for global illumination computations. In Proceedings of the 23rd Spring Conference on Computer Graphics, Budmerice, Slovakia, 26–28 April 2007; pp. 61–68. [Google Scholar]
  5. Lubin, J. A visual discrimination model for imaging system design and evaluation. In Vision Models for Target Detection and Recognition; Peli, E., Ed.; World Scientific: Singapore, 1995; pp. 245–283. [Google Scholar]
  6. Longhurst, P.; Chalmers, A. User validation of image quality assessment algorithms. In Proceedings of the TPCG 04: Theory and Practice of Computer Graphics, Bournemouth, UK, 8–10 June 2004; IEEE Computer Society: Washington, DC, USA, 2004; pp. 196–202, ISBN 0-7695-2137-1. [Google Scholar]
  7. Constantin, J.; Bigand, A.; Constantin, I. Pooling spike neural network for fast rendering in global illumination. Neural Comput. Appl. 2019, 69, 49–58. [Google Scholar] [CrossRef]
  8. Constantin, J.; Bigand, A.; Constantin, I.; Hamad, D. Image Noise Detection in Global Illumination Methods based on FRVM. NeuroComputing 2015, 64, 82–95. [Google Scholar] [CrossRef]
  9. Takouachet, N.; Delepoulle, S.; Renaud, C.; Zoghlami, N.; Tavares, J.R. Perception of noise and Global Illumination: Towards an automatic stopping criterion based on SVM. Comput. Graph. 2017, 69, 49–58. [Google Scholar] [CrossRef] [Green Version]
  10. Lu, Z.; Wu, X.; Bongard, J.C. Active learning through adaptive heterogeneous ensembling. IEEE Trans. Knowl. Data Eng. 2015, 27, 368–381. [Google Scholar] [CrossRef]
  11. Xu, L.; Li, B.; Chen, E. Ensemble pruning via constrained eigenoptimization. In Proceedings of the 2012 IEEE 12th International Conference on Data Mining, ICDM, Brussels, Belgium, 10–13 December 2012; pp. 715–724. [Google Scholar]
  12. Maulik, U.; Chakraborty, D. A self-trained ensemble with semisupervised SVM: An application to pixel classification of remote sensing imagery. Pattern Recognit. 2011, 44, 615–623. [Google Scholar] [CrossRef]
  13. Bianco, S.; Celona, L.; Napoletano, P.; Schettin, R. On the use of Deep Learning for Blind Image quality Assessment. arXiv 2016, arXiv:1602.05531, 1–7, 1–7. [Google Scholar] [CrossRef]
  14. Guo, H.; Wang, W. An Active Learning-based SVM Multi-Class Classification Model. Pattern Recognit. 2015, 48, 130–135. [Google Scholar] [CrossRef]
  15. Rhee, P.K.; Erdenee, E.; Kyun, S.D.; Ahmed, M.U.; Jin, S. Active and semi-supervised learning for object detection with imperfect data. Cogn. Syst. Res. 2017, 45, 109–1231. [Google Scholar] [CrossRef]
  16. Guangwu, A.; Minggang, S. Semi-Supervised Learning Based Social Image Semantic Mining Algorithm. J. Multimed. 2014, 9, 245–252. [Google Scholar]
  17. Richarz, J.; Vajda, S.; Grzeszick, R.; Fink, G. Semi-supervised learning for character recognition in historical archive documents. J. Pattern Recognit. 2014, 47, 1011–1020. [Google Scholar] [CrossRef]
  18. Settles, B. Active Learning Litterature Survey; University of Wisconsin-Madison: Madison, WI, USA, 2009. [Google Scholar]
  19. Huang, T.; Yu, Y.; Guo, G. A classification algorithm based on local clusters with a few labeled training examples. Knowl. Based Syst. 2010, 23, 563–571. [Google Scholar] [CrossRef]
  20. Dornaika, F.; Traboulsi, Y.E.; Assoum, A. Inductive and Flexible Feature Extraction for Semi-Supervised Pattern Categorization. Pattern Recognit. 2016. [Google Scholar] [CrossRef]
  21. Ruz, G.A. Improving the performance of inductive learning classifiers through the presentation order of the training patterns. Expert Syst. Appl. 2016, 58, 1–9. [Google Scholar] [CrossRef]
  22. Zhu, R.; Dornaika, F.; Ruichek, Y. Semi-supervised elastic manifold embedding with deep learning architecture. Pattern Recognit. 2020, 107, 107425. [Google Scholar] [CrossRef]
  23. Zhou, Z.H.; Li, M. Tri-training: Exploiting unlabeled data using three classifiers. IEEE Trans. Knowl. Data Eng. 2005, 17, 1529–1541. [Google Scholar] [CrossRef] [Green Version]
  24. Rasmus, A.; Berglund, M.; Honkala, M.; Valpola, H.; Raiko, T. Semisupervised learning with ladder networks. Adv. Neural Inf. Process 2015, 14, 3546–3554. [Google Scholar]
  25. Tran, P.V. Semi-Supervised Learning with Self-Supervised Networks. Machine learning. CoRR 2019. Available online: https://128.84.21.199/abs/1906.10343v1 (accessed on 2 September 2020).
  26. Zhai, X.; Oliver, A.; Kolesnikov, A.; Beyer, L. Self-Supervised Semi-Supervised Learning. Computer Vision and Pattern Recognition. CoRR 2019. Available online: https://arxiv.org/pdf/1905.03670.pdf (accessed on 2 September 2020).
  27. Yan, W.; Sun, Q.; Sun, H.; Li, Y. Semi-supervised learning framework based on statiscal analysis for image set classification. Pattern Recognit. 2020, 107, 107500. [Google Scholar] [CrossRef]
  28. Li, C.; Wu, C. A new semi-supervised support vector machine learning algorithm based on active learning. In Proceedings of the International Conference on Future Computer and Communication, Wuhan, China, 21–24 May 2010; pp. 638–641. [Google Scholar]
  29. Leng, Y.; Xu, X.; Qi, G. Combining active learning and semi-supervised learning to construct SVM classifier. Knowl. Based Syst. 2013, 44, 121–131. [Google Scholar] [CrossRef]
  30. He, X.; Ji, M.; Bao, H. A Unified Active and Semi-supervised Learning Framework for Image Compression. In Proceedings of the International Conference on Computer Vision and Pattern Recognition, Miami, FL, USA, 20–25 June 2009; pp. 65–72. [Google Scholar]
  31. Yu, D.; Varadarajan, B.; Deng, L.; Acero, A. Active learning and semisupervised learning for speech recognition: A unified framework using the global entropy reduction maximization criterion. Comput. Speech Lang. 2010, 24, 433–444. [Google Scholar] [CrossRef]
  32. Li, H.; Liao, X.; Carin, L. Active learning for semi-supervised multitask learning. In Proceedings of the International Conference on Acoustics, Speech and Signal Processing, Taipei, Taiwan, 19–24 April 2009; pp. 1637–1640. [Google Scholar]
  33. Constantin, J.; Constantin, I.; Bigand, A.; Hamad, D. Perception of noise in Global Illumination based on Inductive Learning. In Proceedings of the 2016 International Joint Conference on Neural Networks (IJCNN), Vancouver, BC, Canada, 24–29 July 2016; pp. 5021–5028. [Google Scholar]
  34. Liu, X.; Pedersen, M.; Hardeberg, J. CID:IQ A New Image Quality Database. Image Signal Process. 2014, 8509, 193–202. [Google Scholar]
  35. Virtanen, T.; Nuutinen, M.; Vaahteranoksa, M.; Oittinen, P.; Häkkinen, J. CID2013: A Database for Evaluating No-Reference Image Quality Assessment algorithms. IEEE Trans. Image Process. 2015, 24, 390–402. [Google Scholar] [CrossRef] [PubMed]
  36. Wald, I.; Kollig, T.; Benthin, C.; Keller, A.; Slusalleki, P. Interactive Global Illumination using Fast Ray Tracing. In Proceedings of the 13th Eurographics Workshop on Rendering, Pisa, Italy, 26–28 June 2002; pp. 15–24. [Google Scholar]
  37. Kollig, T.; Keller, A. Efficient Bidirectional Path Tracing by Randomized Quasi-Monte Carlo Integration. In Monte Carlo and Quasi-Monte Carlo Methods 2000; Niederreiter, H., Fang, K., Hickernell, F., Eds.; Springer: Berlin/Heidelberg, Germany, 2002; pp. 290–305. ISBN 978-3-540-42718-6. [Google Scholar]
  38. Hedman, P.; Karras, T.; Lehtinen, J. Sequential Monte Carlo instant radiosity. In Proceedings of the 20th ACM SIGGRAPH Symposium on Interactive 3D Graphics and Games, Redmond, WA, USA, 26–28 February 2016; pp. 121–128. [Google Scholar]
  39. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef]
  40. Razavian, A.; Azizpour, H.; Sullivan, J.; Carlsson, S. CNNfeatures off-the-shelf: An astounding baseline for recognition. In Proceedings of the 2014 IEEE Conference on Computer Vision and Pattern Recognition CVPR Workshops, Columbus, OH, USA, 23–28 June 2014; pp. 806–813. [Google Scholar]
  41. Makandar, A.; Halalli, B. Image Enhancement Techniques using Highpass and Lowpass Filters. Int. J. Comput. Appl. 2015, 109, 21–27. [Google Scholar] [CrossRef]
  42. Dawood, F.; Rahmat, R.; Kadiman, S.; Abdullah, L.; Zamrin, M. Effect Comparison of Speckle Noise Reduction Filters on 2D-echocardiographic. World Acad. Sci. Eng. Technol. 2012, 6, 425–430. [Google Scholar]
  43. Biswas, P.; Sarkar, A.; Mynuddin, M. Deblurring Images using a Wiener Filter. Int. J. Comput. Appl. 2015, 109, 36–38. [Google Scholar] [CrossRef]
  44. Gao, D.; Liao, Z.; Lv, Z.; Lu, Y. Multi-scale Statistical Signal Processing of Cutting Force in Cutting Tool Condition Monitoring. Int. J. Adv. Manuf. Technol. 2015, 90, 1843–1853. [Google Scholar] [CrossRef]
  45. Gou, H.; Swaminathan, A.; Wu, M. Intrinsic Sensor Noise Features for Forensic Analysis on Scanners and Scanned Images. IEEE Trans. Inf. Forensics Secur. 2009, 4, 476–491. [Google Scholar]
  46. Fernandez-Maloigne, C.; Robert-Inacio, F.; Macaire, L. Digital Color: Acquisition, Perception, Coding and Rendering; Chapter 9: Quality Assessment Approaches, 314 Pages; Wiley-ISTE: London, UK, 2012; ISBN 978-1-848-21346-3. [Google Scholar]
  47. Abin, A.A.; Beigy, H. Active selection of clustering constraints: A sequential approach. Pattern Recognit. 2014, 47, 443–458. [Google Scholar] [CrossRef]
  48. Demir, B.; Persello, C.; Bruzzone, L. Batch-mode active learning methods for the interactive classification of remote sensing images. IEEE Trans. Geosci. Remote. Sens. 2011, 49, 1014–1031. [Google Scholar] [CrossRef] [Green Version]
  49. Tong, S.; Koller, D. Support Vector Machine Active Learning with Applications to Text Classification. J. Mach. Learn. Res. 2000, 2, 45–66. [Google Scholar]
  50. Chapelle, O.; Scholkopf, B.; Zien, A. Semi-Supervised Learning; MIT Press: Cambridge, MA, USA, 2006; 528p, ISBN 9780262033589. [Google Scholar]
  51. Amorin, W.; Rosa, G.; Thomazella, R.; Castanho, J.E.C.; Dotto, F.R.L.; Junior, U.P.R.; Marana, A.N.; Beigy, J.P.P. Semi-supervised learning with connectivity-driven convolutional neural networks. Pattern Recognit. Lett. 2019, 128, 16–22. [Google Scholar] [CrossRef]
  52. Chen, D.G.; Wang, H.Y.; Tsang, E.C. Generalized Mercer theorem and its application to feature space related to indefinite kernels. In Proceedings of the IEEE International Conference on Machine Learning and Cybernetics, Kunming, China, 12–15 July 2008; Volume 2, pp. 774–777. [Google Scholar]
  53. Lipton, Z.; Elkan, C.; Narayanaswamy, B. Thresholding Classifiers to Maximize F1 Score. arXiv 2014, arXiv:1402.1892. [Google Scholar]
  54. Chollet, F. Deep Learning with Python; Manning Publications: Greenwich, CT, USA, 2017; 384p, ISBN 9781617294433. [Google Scholar]
  55. Falk, M.; Grottel, S.; Krone, M.; Reina, G. Interactive GPU-Based Visualization of Large Dynamic Particle Data; Morgan & Claypool: San Rafael, CA, USA, 2016; 122p, ISBN 978-1627052856. [Google Scholar]
Figure 1. Rendering results for the sub-images of the scenes with 512 × 512 resolution. The scenes with 512 × 512 resolution (af) consist of diffuse surfaces with complex geometry. They contain objects with different material properties presenting rich and complicated shading variations under different lighting conditions.
Figure 1. Rendering results for the sub-images of the scenes with 512 × 512 resolution. The scenes with 512 × 512 resolution (af) consist of diffuse surfaces with complex geometry. They contain objects with different material properties presenting rich and complicated shading variations under different lighting conditions.
Jimaging 06 00091 g001
Figure 2. Rendering results for the sub-images of the scenes with 800 × 800 resolution. The scenes with 800 × 800 resolution (gl) are used to illustrate view-dependent indirect illumination effects caused by strong inter-reflections between the diffuse and specular surfaces.
Figure 2. Rendering results for the sub-images of the scenes with 800 × 800 resolution. The scenes with 800 × 800 resolution (gl) are used to illustrate view-dependent indirect illumination effects caused by strong inter-reflections between the diffuse and specular surfaces.
Jimaging 06 00091 g002
Figure 3. (a) desired thresholds for the sub-images of the scenes with resolutions 512 × 512 (a) and 800 × 800 (b).
Figure 3. (a) desired thresholds for the sub-images of the scenes with resolutions 512 × 512 (a) and 800 × 800 (b).
Jimaging 06 00091 g003
Figure 4. Standard deviation variation per block.
Figure 4. Standard deviation variation per block.
Jimaging 06 00091 g004
Figure 5. Architecture of the convolution neural network.
Figure 5. Architecture of the convolution neural network.
Jimaging 06 00091 g005
Figure 6. A schematic representation of the Path Tracing progressive algorithm including noise detection for each sub-image. At each iteration, new paths are added to the un-converged sub-images. Then, the model is tested if each new sampled sub-image is still noisy or not. According to the models answer, we decide then to add new paths or to stop computation for the corresponding sub-image as it is supposed to be visually converged.
Figure 6. A schematic representation of the Path Tracing progressive algorithm including noise detection for each sub-image. At each iteration, new paths are added to the un-converged sub-images. Then, the model is tested if each new sampled sub-image is still noisy or not. According to the models answer, we decide then to add new paths or to stop computation for the corresponding sub-image as it is supposed to be visually converged.
Jimaging 06 00091 g006
Figure 7. Variation of precision during the iterations of the active learning algorithm: (a) SVM AS (b) DSVM AS .
Figure 7. Variation of precision during the iterations of the active learning algorithm: (a) SVM AS (b) DSVM AS .
Jimaging 06 00091 g007
Figure 8. The nine sub-images added by the active learning algorithm.
Figure 8. The nine sub-images added by the active learning algorithm.
Jimaging 06 00091 g008
Figure 9. Variation of the mean square error for the 512 × 512 resolution scenes. It is shown that the SVM AS model achieves higher performance on the scenes (bf). However, the SVM model assures higher performance than SVM AS on the learning scene (a). We also found that the DSVM AS model gives better performance than DSVM does on the scenes (bf) and provides mean square errors less than SVM on the scenes (df).
Figure 9. Variation of the mean square error for the 512 × 512 resolution scenes. It is shown that the SVM AS model achieves higher performance on the scenes (bf). However, the SVM model assures higher performance than SVM AS on the learning scene (a). We also found that the DSVM AS model gives better performance than DSVM does on the scenes (bf) and provides mean square errors less than SVM on the scenes (df).
Jimaging 06 00091 g009
Figure 10. Variation of the thresholds for the 512 × 512 resolution scenes. It is shown that the SVM AS gives similar results to those of the human visual system (HVS) on the scenes (af). We also found that the DSVM AS model gives more coherent results to those of the human visual system (HVS) than DSVM does on the scenes (af) and achieves a better precision than SVM on the scenes (df).
Figure 10. Variation of the thresholds for the 512 × 512 resolution scenes. It is shown that the SVM AS gives similar results to those of the human visual system (HVS) on the scenes (af). We also found that the DSVM AS model gives more coherent results to those of the human visual system (HVS) than DSVM does on the scenes (af) and achieves a better precision than SVM on the scenes (df).
Jimaging 06 00091 g010
Figure 11. Variation of the mean square error for the sub-images of the 800 × 800 resolution scenes. It is shown that the SVM AS model gives the best performance on the scenes (hl). We also found that DSVM AS provides mean square errors less than SVM and DSVM on the scenes (hl). However, the DSVM model assures higher performance than DSVM AS on the learning scene (g).
Figure 11. Variation of the mean square error for the sub-images of the 800 × 800 resolution scenes. It is shown that the SVM AS model gives the best performance on the scenes (hl). We also found that DSVM AS provides mean square errors less than SVM and DSVM on the scenes (hl). However, the DSVM model assures higher performance than DSVM AS on the learning scene (g).
Jimaging 06 00091 g011
Figure 12. Variation of the thresholds for the sub-images of the 800 × 800 resolution scenes. We found that SVM AS gives similar results to those of the human visual system (HVS) on the scenes (gl). It is shown that the DSVM AS model achieves a better precision than SVM and DSVM on the scenes (hl).
Figure 12. Variation of the thresholds for the sub-images of the 800 × 800 resolution scenes. We found that SVM AS gives similar results to those of the human visual system (HVS) on the scenes (gl). It is shown that the DSVM AS model achieves a better precision than SVM and DSVM on the scenes (hl).
Jimaging 06 00091 g012
Figure 13. Architecture of the VGG19 Neural Network.
Figure 13. Architecture of the VGG19 Neural Network.
Jimaging 06 00091 g013
Figure 14. Variation of precision during the learning process of the VGG19 network (maximum precision equal to 86 % after 1000 cycles).
Figure 14. Variation of precision during the learning process of the VGG19 network (maximum precision equal to 86 % after 1000 cycles).
Jimaging 06 00091 g014
Figure 15. Variation of the actual thresholds of the VGG19 model, the human visual system score, and the DSVM AS model for the sub-images of the 512 × 512 resolution scenes. We found that DSVM AS gives more coherent results to those of the human visual system (HVS) than the VGG19 model on the scenes (af).
Figure 15. Variation of the actual thresholds of the VGG19 model, the human visual system score, and the DSVM AS model for the sub-images of the 512 × 512 resolution scenes. We found that DSVM AS gives more coherent results to those of the human visual system (HVS) than the VGG19 model on the scenes (af).
Jimaging 06 00091 g015
Figure 16. The mean square error over the number of queries when performing learning on the 512 × 512 resolution scenes. (a) SVM AS and SVM AL (b) DSVM AS and DSVM AL .
Figure 16. The mean square error over the number of queries when performing learning on the 512 × 512 resolution scenes. (a) SVM AS and SVM AL (b) DSVM AS and DSVM AL .
Jimaging 06 00091 g016
Table 1. Number of paths per pixel between two successive sub-images and largest number of paths per pixel for the scenes with 800 × 800 resolution.
Table 1. Number of paths per pixel between two successive sub-images and largest number of paths per pixel for the scenes with 800 × 800 resolution.
Scenes with
800 × 800 Resolution
Number of Paths between
Two Successive Sub-Images
Largest Number
of Paths
(g)5700
(h)505000
(i)10950
(j)505000
(k)10950
(l)505000
Table 2. Deep architecture.
Table 2. Deep architecture.
LayersC1C2C3C4C5C6C7C8C9C10C11C12
Size 3 × 3 5 × 5 3 × 3 5 × 5 3 × 3 5 × 5 3 × 3 5 × 5 3 × 3 5 × 5 3 × 3 5 × 5
Padding121212121212
Table 3. Optimal parameters for the models for the scenes with 512 × 512 resolution. The bold value simply indicates the maximum precision.
Table 3. Optimal parameters for the models for the scenes with 512 × 512 resolution. The bold value simply indicates the maximum precision.
Learning
Model
Penality
Factor C
Standard
Deviation
PrecisionMean Number
of Kernels
SVM85096.24%1750
DSVM87090.7%1970
Table 4. Optimal parameters for the models for the 800 × 800 resolution scenes. The bold value simply indicates the maximum precision.
Table 4. Optimal parameters for the models for the 800 × 800 resolution scenes. The bold value simply indicates the maximum precision.
Learning
Model
Penality
Factor C
Standard
Deviation
PrecisionMean Number
of Kernels
SVM87097.50%1770
DSVM812083.22%5350
Table 5. Mean square error range for the 512 × 512 resolution scenes. The bold values simply indicate the minimum Mean square error range.
Table 5. Mean square error range for the 512 × 512 resolution scenes. The bold values simply indicate the minimum Mean square error range.
ScenesSVM SVM AS DSVM DSVM AS
(a)0[0–0.01][0–0.15][0–0.13]
(b)[0–0.07][0–0.02][0–0.14][0–0.14]
(c)[0–0.07]0[0–0.13][0–0.11]
(d)[0–0.06]0[0–0.07][0–0.05]
(e)[0–0.07][0–0.01][0–0.1][0-0.07]
(f)[0.04–0.13][0–0.09][0–0.09][0–0.07]
Table 6. F 1 score for the 512 × 512 resolution scenes. The bold values simply indicate the maximum F 1 score.
Table 6. F 1 score for the 512 × 512 resolution scenes. The bold values simply indicate the maximum F 1 score.
ScenesSVM SVM AS DSVM DSVM AS
(a)10.990.920.93
(b)0.950.990.940.95
(c)0.9510.910.95
(d)0.9610.940.97
(e)0.950.990.940.96
(f)0.860.970.920.95
Table 7. Mean square error range on the 800 × 800 resolution scenes. The bold values simply indicate the minimum Mean square error range.
Table 7. Mean square error range on the 800 × 800 resolution scenes. The bold values simply indicate the minimum Mean square error range.
ScenesSVM SVM AS DSVM DSVM AS
(g)0[0–0.03][0–0.09][0.01–0.15]
(h)[0.15–0.27][0–0.03][0.07–0.19][0.03–0.15]
(i)[0–0.15][0–0.15][0–0.13][0–0.07]
(j)[0.05–0.13][0–0.01][0-0.11][0–0.11]
(k)[0.07–0.35][0–0.03][0-0.11][0–0.11]
(l)[0–0.05][0–0.03][0–0.09][0–0.05]
Table 8. F 1 score for the 800 × 800 resolution scenes. The bold values simply indicate the maximum F 1 score.
Table 8. F 1 score for the 800 × 800 resolution scenes. The bold values simply indicate the maximum F 1 score.
ScenesSVM SVM AS DSVM DSVM AS
(g)10.990.900.90
(h)0.630.990.770.85
(i)0.910.980.930.97
(j)0.900.990.910.95
(k)0.670.990.930.96
(l)0.970.990.960.97
Table 9. Number of parameters and the times in seconds needed for learning and testing on the 512 × 512 resolution scenes. The bold values simply indicate the minimum values obtained by the models.
Table 9. Number of parameters and the times in seconds needed for learning and testing on the 512 × 512 resolution scenes. The bold values simply indicate the minimum values obtained by the models.
ModelNo. of ParametersLearning Time
(s)
Testing Time
(s)
SVM20,038,85540.17
DSVM42,12010.008
SVM AS 16,106,45580530.15
DSVM AS 69,76820950.008
Table 10. Maximum precision for the deep learning models after 1000 training cycles. The bold values simply indicate the higher precision.
Table 10. Maximum precision for the deep learning models after 1000 training cycles. The bold values simply indicate the higher precision.
ModelPrecision (%)
InceptionV376
MobileNet75
Resnet21081
VGG1986
SVM AS 99
DSVM AS 91

Share and Cite

MDPI and ACS Style

Constantin, I.; Constantin, J.; Bigand, A. On the Use of Deep Active Semi-Supervised Learning for Fast Rendering in Global Illumination. J. Imaging 2020, 6, 91. https://doi.org/10.3390/jimaging6090091

AMA Style

Constantin I, Constantin J, Bigand A. On the Use of Deep Active Semi-Supervised Learning for Fast Rendering in Global Illumination. Journal of Imaging. 2020; 6(9):91. https://doi.org/10.3390/jimaging6090091

Chicago/Turabian Style

Constantin, Ibtissam, Joseph Constantin, and André Bigand. 2020. "On the Use of Deep Active Semi-Supervised Learning for Fast Rendering in Global Illumination" Journal of Imaging 6, no. 9: 91. https://doi.org/10.3390/jimaging6090091

APA Style

Constantin, I., Constantin, J., & Bigand, A. (2020). On the Use of Deep Active Semi-Supervised Learning for Fast Rendering in Global Illumination. Journal of Imaging, 6(9), 91. https://doi.org/10.3390/jimaging6090091

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop