Next Article in Journal
Self-Sensing Rubber for Bridge Bearing Monitoring
Previous Article in Journal
An Integrated Single-Beam Three-Axis High-Sensitivity Magnetometer
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Semantically Adaptive JND Modeling with Object-Wise Feature Characterization, Context Inhibition and Cross-Object Interaction

1
School of Communication Engineering, Hangzhou Dianzi University, No. 2 Street, Xiasha, Hangzhou 310018, China
2
Lishui Institute of Hangzhou Dianzi University, Nanmingshan Street, Liandu, Lishui 323000, China
3
Hangzhou Arcvideo Technology Co., Ltd., No. 3 Xidoumen Road, Xihu, Hangzhou 310012, China
*
Author to whom correspondence should be addressed.
Sensors 2023, 23(6), 3149; https://doi.org/10.3390/s23063149
Submission received: 15 December 2022 / Revised: 12 January 2023 / Accepted: 15 January 2023 / Published: 15 March 2023
(This article belongs to the Topic Advances in Perceptual Quality Assessment of User Generated Contents)
(This article belongs to the Section Intelligent Sensors)

Abstract

:
Performance bottlenecks in the optimization of JND modeling based on low-level manual visual feature metrics have emerged. High-level semantics bear a considerable impact on perceptual attention and subjective video quality, yet most existing JND models do not adequately account for this impact. This indicates that there is still much room and potential for performance optimization in semantic feature-based JND models. To address this status quo, this paper investigates the response of visual attention induced by heterogeneous semantic features with an eye on three aspects, i.e., object, context, and cross-object, to further improve the efficiency of JND models. On the object side, this paper first focuses on the main semantic features that affect visual attention, including semantic sensitivity, objective area and shape, and central bias. Following that, the coupling role of heterogeneous visual features with HVS perceptual properties are analyzed and quantified. Second, based on the reciprocity of objects and contexts, the contextual complexity is measured to gauge the inhibitory effect of contexts on visual attention. Third, cross-object interactions are dissected using the principle of bias competition, and a semantic attention model is constructed in conjunction with a model of attentional competition. Finally, to build an improved transform domain JND model, a weighting factor is used by fusing the semantic attention model with the basic spatial attention model. Extensive simulation results validate that the proposed JND profile is highly consistent with HVS and highly competitive among state-of-the-art models.

1. Introduction

Regarding the human visual system (HVS) as a communication system with limited bandwidth and processing capacity, it is continuously receiving data input. As a result, the HVS can only sense variations in signal strength above a certain threshold, which is termed as JND [1]. Research on JND can be traced back to the experimental psychology of Ernst Weber [2] in the 19th century and was transferred to the field of digital multimedia at the end of the 20th century. Visual JND can build computational models by virtue of relevant physiology, psychology and neural research, combined with feature detection. Over the past decades, JND has proven to be a multi-factorial problem, including contrast sensitivity function (CSF), luminance adaptation (LA), masking effects and visual attention. An overview of these factors can be found in the survey by researchers in [3]. Using the computational domain as a classification criterion, there are two branches of existing JND models, i.e., the pixel domain (where JND thresholds are computed directly for each pixel) [4,5,6,7,8,9,10,11,12,13] and the transform domain (where the image is first transformed into a subband domain, and then JND thresholds are calculated for each subband) [14,15,16,17,18,19,20,21,22,23,24]. Nevertheless, both model types generally follow the same design philosophy of simulating the visual-masking effects of several elements before combining (multiplying or adding) them to obtain an overall JND estimation. The pixel-domain JND model has undergone a protracted evolution. Based on the study of luminance and contrast, the pioneering work of Chou et al. [4] computed luminance adaptation and contrast masking (CM), using the winner as the JND thresholds. Yang et al. [5] considered the overlap effect between LA and CM, and proposed a nonlinear additivity model of masking (NAMM). Furthermore, the image can be decomposed into plain, texture and edge regions for more accurate modeling of CM [6]. Since the visual sensitivity decreases with increasing retinal eccentricity, the fovea masking was introduced [7]. Inspired by the internal generation mechanism, the masking effects of ordered and disordered components were quantified based on free energy theory [8], and the pattern masking was extended sequentially [9]. In addition, the RMS contrast was used as the spatial CSF of JND in the pixel domain [10]. Based on hierarchical predictive coding theory, self-information and information entropy were utilized to calculate the perceptual surprise and perceptual suppression effects at different levels to improve the accuracy of the JND model [12].
Transform coding is a widespread means of mainstream image/video coding, and transform domain JND modeling is also a significant research area. In 1992, Ahumada [14] proposed the first DCT domain JND model combining spatial CSF and LA. Based on [14], Watson [15] introduced the CM effect and proposed the DCTune model, which laid the research foundation for the DCT domain JND modeling. Subsequently, a more realistic LA function was proposed to improve JND estimation by integrating a block classification (plain, edge, and texture) strategy [16]. For video, corresponding video JND models for the DCT domain were proposed considering spatiotemporal CSF and eye movement compensation [17], LA effect based on gamma correction and CM effect based on more accurate block classification [18], texture complexity and frequency, visual sensitivity and visual attention [19], the various sizes of DCT blocks (from 4 × 4 to 32 × 32) [20], motion direction [21], fovea masking [22], temporal duration and residual fluctuations [23].
In recent years, machine-/deep-learning-based modeling of visual perception has become a new research trend due to the rapid development of deep neural networks [25,26,27,28,29,30,31,32,33]. To fill the vacancy of JND databases for image and video compression, several scholars have suggested MCL-JCI [34], JND-pano [35], MCL-JCV [36], VideoSet [37], etc. Based on these datasets, various modeling techniques for JND eventually emerged, most notably subjective data regression [25,27], binary classification [26], picture/video-wise JND (PWJND/VWJND) or satisfied user ratio (SUR) modeling [27,28,29,30], and finding appropriate weighting factors for the JND models [25,26]. In particular, motivated by the smoothing performance (non-smoothed regions have better capability for hiding noise than smoothed regions), Wu et al. [31] proposed an unsupervised learning method for generating JND images in the pixel domain. Considering that JND should be evaluated in the human brain perception domain, Jin et al. [33] presented an HVS-based signal degradation network and employed visual attention loss to further regulate JND estimation. In addition, JND modeling also extends to machine vision [38,39]; however, this specific topic is outside the scope of this paper. Visual attention is a key attribute of HVS [40], and models combined with visual attention can improve the accuracy of JND threshold. In a study by Chen et al. [7], attention-induced fixation was detected and a scheme was designed to reflect the increase in JND with increasing retinal eccentricity. Video JND estimation for smooth pursuit eye movements (SPEM) and non-SPEM situations was accomplished by investigating the relationship between JND thresholds and parameters (spatial frequency, eccentricity, retinal motion velocity) [22]. Considering the special focus of HVS on people, Wang et al. [41] combined Itti’s [42] and Judd’s [43] attention model (with face detector and person detector) to adjust the JND threshold.
The main feature of the HVS attention mechanism is the information-selection process, in which only a small number of visual signals are transmitted to the brain for processing. Indeed, the object-based attention theory suggests that humans are drawn to objects and advanced concepts [44]. Although Wang at al. [41] considered the advanced semantics of faces, there are often more advanced semantics in images than in plain faces. In any case, traditional visual attention models mainly consider low-level features such as color, luminance, and orientation, while underestimating high-level semantic information, which leads to reduced JND model accuracy under the attention mechanism of HVS. This is mainly due to the difficulty of semantic feature extraction and fusion. Meanwhile, it is well known that deep neural networks are full of semantic information [45], but current JND models based on deep learning mainly try to construct PWJND/VWJND or SUR without fully considering semantic character behaviors.
In addition, neurons transmit and express visual information by consuming energy [46]. The structure and function of neural networks follow the essential principles of resource allocation and minimization of energy constraints. Therefore, not all stimuli trigger neuronal responses, i.e., biased competition for visual attention influences neuronal activity [47]. The theory of biased competition suggests that selectivity of one (or more) points in the visual process is caused by interaction between visual objects for neural representations. More interestingly, with the limited attentional resources, there will always be certain items and their visual features that win out in the competition. Therefore, it is inevitable to consider such interaction while constructing a semantics-based attention model.
From the above discussion, it is clear that semantics is crucial for accurate estimation of JND thresholds. Furthermore, it should be noted that this paper concentrates on the tuning scheme for DCT-based JND estimation because the majority of image/video processing, especially in image/video compression, is conducted in the DCT domain. Particularly, this paper argues that there are three semantic features to be considered in the attention allocation problem of videos, namely, object, context, and cross-object. Since it is troublesome to extract semantic attributes manually, this paper uses grad-CAM [48] and VNext [49] to extract semantic objects and features through deep learning. Based on this, attention effects induced by object-based semantic features are analyzed, including the sensitivity and size of semantic objects and central bias. Second, contextual information is another vital element that enhances instance differentiation and is considered as a form of center-surround contrast. Furthermore, considering limited attentional resources and infinite information, this paper describes the attentional competition effects of different semantic features (including focus intensity, spatial distance, and relative area) in terms of cross-objects to achieve an accurate model of semantic attention.
Specifically, inspired by the framework of [23], this paper proposes a statistical probability model corresponding to the feature parameters and quantifies the response of visual attention in a perceptual sense using information theory. Then, the attentional competition factor for calibrating the semantic attention model is proposed. Finally, the adaptive semantic attention weight is obtained by unifying and fusing the perceptual feature parameters, and the JND model is adjusted using this weight.
The rest of this paper is organized as follows. Section 2 presents three features and describes the quantification strategies for these parameters. Section 3 analyzes the interaction between the objects and context. Section 4 depicts the fusion approaches of these parameters together with the competition properties of attention. Section 5 details the proposed JND profile. Simulation results and conclusions are given in the Section 6 and Section 7, respectively.

2. Object-Wise Semantics Parameter Extraction and Quantification

As stated, the literature lacks a clear understanding of the mechanism of interaction between semantic perception and semantic features. This motivates us to formulate an effectual JND profile leveraging the HVS semantic traits, which is subject to challenges. The first issue is how to precisely extract and quantify the feature parameters that thoroughly depict the perceptual response of semantic HVS features in videos, such as semantic sensitivity, objective area, central bias.

2.1. Semantic Instances Extraction

Accurate extraction of high-level semantic features of videos helps to better visual attention modeling. On that basis, this paper uses grad-CAM for the extraction of high-level semantic features. In this network, the further the layer, the ampler the semantics is and the more spatial information can be preserved. In this respect, grad-CAM contains all the semantics of our target of interest. For more details, please see [48]. With grad-CAM, we can draw heat maps which are shown in Figure 1 below. Moreover, to effectively utilize the various properties of semantic objects, this paper uses VNext [49] to extract semantic objects, and the results are shown in Figure 2.

2.2. Semantic Sensitivity Quantification

The research [50] has shown that attention is focused on informative content. One definition of attention describes it as a limited resource for information processing, with the area of interest carrying the bulk of the information transmitted to the brain. Hence, Bruce et al. [51] pointed out that Shannon self-information can be used to outline attention. Whereupon, the attention value of a pixel located at ( x , y ) in a video frame is defined below.
A ( x , y ) = log 2 P ( F = f ( x , y ) )
where F denotes a random variable of the feature of pixel ( x , y ) , f ( x , y ) represents the image features of pixel ( x , y ) , P ( F = f ( x , y ) ) indicates the probability of feature F.
The concept of constructing statistically sparse representations of images appears to be fundamental to the primate visual system, as evidenced by a large body of research. That is, to some extent, computing the probability distribution of features is essentially a matter of statistical feature sparsity. If a pixel associated with a feature has a lower probability of appearing, it has a better chance of attracting attention because it conveys a greater amount of information. In view of this, the probability density function (PDF) of feature F is measured as follows by performing histogram statistics on the whole picture.
PDF = h i s t ( F = f ( x , y ) ) = P ( F = f ( x , y ) )
Accordingly, the initial attention based on semantic sensitivity (m) can be calculated as follows.
I ( m ) = log 2 h i s t ( f ( m ) )
It is worth noting that instead of using a uniform-fitting function, to better suit the image content, the histogram of each frame is adaptively fitted. Figure 3 shows the histogram of the 2nd frame in the HEVC standard test sequence “BasketballDrill” and its PDF fitting curve, separately.

2.3. Objective Area Quantification

The literature demonstrates that the larger the area, the higher the probability of attention, i.e., the probability of attention of semantic objects is proportional to size, however, this property becomes smooth after a certain saturation threshold [52]. This relationship can be summarized below using a piecewise function [53], with values within [0, 1].
log 2 ( p ( η j ) ) = ln ( 1 η j ) c 1 0 η j t 1 1 exp ( c 2 · η j ) t 1 < η j t 2 s 2 + c 3 ln ( 1 η j + t 2 ) t 2 < η j t 3 s 3 · exp ( c 4 ( η j t 3 ) ) t 3 < η j t 4 0 t 4 < η j 1.0
where 0 < j N , N is the number of semantic objects in an image frame, η j represents the size of semantic object, which is calculated as: η j = B j / ( W × H ) , B j represents the number of objective pixels, and W , H refer to the width and height of the image frame. Figure 4 shows the relationship between area and attention.
Likewise, for the same size semantic objects, attention will be affected by the aspect ratio. As is known to all, the aspect ratio of current popular displays is 16:9. It follows that the closer the aspect ratio of a semantic object is to our vision field, the more it attracts attention.
Figure 5 is an illustration for the aspect ratio. For a more clear description, the blue rectangle on the left is labeled as A, the one on the right is B, and the container containing these two rectangles is C. Despite having the same area, A and B have various aspect ratios. The aspect ratios of A and C are 16:9, while B is 9:16. It is obvious from the example that rectangle A attracts more attention than rectangle B. Note that the rectangle is used merely for convenience of observation, and the results hold true other shapes as well.
Therefore, the joint effect of semantic object size, the aspect ratio ( R r ) is used to measure the attention of semantic object size.
I ( η j ) = log 2 [ p ( η j ) ] α 1 log 2 R r j β 1 β 2
where α 1 , β 1 , β 2 are all adjustment control parameters, r j denotes the aspect ratio of the j-th semantic object and is normalized to [0, 1].
The model is consistent with HVS subjective perception that HVS perceptual attention intensity increases with the size of the object but there is a saturation point. The reason is that when the object area is too large, the target and background are non-separable, i.e., the target is likely to be the background component at this point. In addition, targets with an aspect ratio closer to 16:9 attract greater HVS attention, which raises perceptual sensitivity and lowers JND thresholds.

2.4. Central Bias Quantification

It is stated that most images have the foreground object in the center of the image frame [43]. It follows that when a person looks at an image, he or she habitually looks at the center of the image first. Eye-movement experiments have shown that HVS tends to focus on the center of the image, and the deviation from the target center to the image center, i.e., the central bias, has been considered a noteworthy prior in attention modeling [44]. Based on this prior knowledge, many attention modeling algorithms based on central bias use various methods to increase the attention value at the image center location as a way to highlight the salient targets in the image.
Nevertheless, the feedback brought by the central bias is not always favorable. For some unconventional cases where the target is off-center in the image, attention enhancement at the image center can lead to over/under estimation of the target far from the image center. In general, the problem with the classical center-priority based attention modeling algorithm is that it tends to lack flexibility, which leads to significant inaccuracies when the foreground targets of the image are not positioned as expected.
The formation of a top-notch attention map is the ultimate goal of attention modeling. The complexity of attention modeling is greatly reduced and the quality of the final attention map is improved if the locations of the salient targets are approximated before proceeding to formal attention modeling.
In this letter, the object (M) closest to the image center is calculated, and the central bias is constructed with M as the experienced center of vision. The closer an object is to the visual center, the higher probability it is to be noticed, i.e., the attention probability of semantic objects is inversely proportional to the central bias distance, which is denoted as follows.
P ( A | d j ) = 1 log 2 1 + d j ω 1 ω 2
where d j is the distance from the semantic object to M, and normalized to [0, 1].
d j = m e a n ( f O j ( x f , y f ) ( x M , y M ) 2 )
here, O j is a semantic object, f denotes the pixel of O j , ( x f , y f ) represents the coordinates corresponding to f, ( x M , y M ) means the center coordinates of M, · is the euclidean distance, ω 1 , ω 2 serve as the adjustment parameters.
As illustrated in Figure 6, the modified central bias result is better in line with HVS perception. Since the target M in the figure is closest to the image center, it obviously draws the highest attention, which is shown by the brightest brightness. In this instance, the target M stands in for the image center as the experienced center of vision and the closer the other objects are to the target M, the more attention they attract, and the brighter they are presented in the figure.

3. Context-Wise Attentional Inhibition Quantification

Given that HVS still beats the state-of-the-art computer vision systems, modeling visual attention in natural scenes is a hot topic of research today. How is such great effectiveness achieved? The ability of HVS to utilize the context of a scene to direct attention before most items are recognized appears to be a key factor compared to artificial systems [54].
How does the context of a scene serves to focus attention on objects in the scene? Davenport and Potter [55] suggest that there is an interaction between objects and context during scene processing, which is supported by research. On this basis, this paper conceptualizes the local spatial arrangement of the image background as an alternative to the context and focuses primarily on the complexity of the patterns.
In most situations, the region between the target and its background is heterogeneous and has a complex organization. The density of the visual information presented is normally used to describe the complexity of the visual background. In general, a scene is complex when its background displays a large amount of information and the unpredictability of this information is considerable [56]. Typically, the complexity of the visual context usually has a detrimental impact on the visual task. The higher the visual complexity, the longer the response latency of the nerve cells and the stronger the degree of inhibition of the target [57].
As the contextual complexity of the target suppresses the attention strength, the more complex the pattern of the background, the stronger the suppression, while this is going on, the findings of numerous human visual field experiments indicate that the human binocular visual field is approximately circular (oval), with the proximate form depicted in Figure 7 [58]. Hence, to represent the inhibition of each semantic target background, we utilize the average of the contextual complexity in the circumcircle of semantic targets to simplify the calculation. Since entropy reflects the chaotic degree of a system, we estimate contextual complexity by local entropy. As a result, attention and contextual complexity are related in the following way.
P ( A ¯ | u j ) = log 2 1 + u j ξ 1 ξ 2
where u j is the contextual complexity, and normalized to [0, 1].
u j = m e a n ( e n t f i l t ( c i r c l e ( O j ) O j ) )
here, ξ 1 and ξ 2 serve as the scaling factors, c i r c l e ( · ) denotes circumcircle, e n t f i l t ( · ) is the function to calculate the local entropy.
Figure 8 shows the inhibitory effect of contextual complexity on semantics.

4. Fusion Strategies and Cross-Object-Wise Attentional Competition

Furthermore, given elaborately chosen feature parameters, the second sore point is how to quantify the interaction among these feature parameters, i.e., how to fuse these heterogeneous feature parameters.
Consolidating the semantics-based attention model is the aim of this part. In general, different feature maps do not have the same attributes, and different features contain complementary global contextual information and local detailed information between them. Consequently, to obtain reliable results for the attention distribution, it is essential to select the appropriate weights for each feature map.
In this paper, to estimate the weights of the feature maps above, the gradient change value of each feature map is first calculated, where the greater the variation of the data, the more information can be gleaned. The gradient variation is then used to estimate the relative weights of these feature maps. Finally, utilizing their individual weights, the feature maps are normalized and combined into a single attention map.
a t t p r e = λ 1 · I ( m ) + λ 2 · I ( η ) + λ 3 · P ( A | d ) P ( A ¯ | u )
here, λ i ( i = 1 , 2 , 3 ) serves as the weighting factor, and is calculated as follows.
λ i = v g ( z i ) i = 1 3 v g ( z i )
v g ( z i ) = max ( g ( z i ) ) min ( g ( z i ) ) m e a n ( g ( z i ) )
g ( z i ( O m ) ) = m = 1 , m n N z i ( O m ) z i ( O n )
where z i { I ( m ) , I ( η ) , P ( A | d ) } is the attention map of different semantic features.
Figure 9 displays the initial attention map with different weighting factors. As illustrated, compared to map (a), the map (b) with the gradient variation based weights is more consistent with the HVS subjective perception. This is partly due to the fact that we fully take the content characteristics of different feature maps into account.
However, there is a competitive mechanism that manifests itself as a relative enhancement of responses to task-relevant objects or a relative inhibition of neglected objects within the brain [59]. That is, when modeling the semantic attention model, it is not sufficient to consider only the semantic object itself, but also to take full account of the biased competition between different objects.
  • Focused Intensity: when multiple stimuli are present in the visual field at the same time, inhibitory competition in the visual cortex affects the allocation of attention, which probably stem from the inability to bias the interaction toward a particular object [60]. In reality, the brain’s representation of information is essential for human vision, and regardless of the distance between neurons or neuronal populations, they do not operate independently, but constantly interact with each other. Because their activities are interconnected and competing, they cannot provide “independent attentional resources” [61]. That is, when there are multiple semantic objects in an image, attention to each semantic object is impacted by the other objects. The higher the intensity of the focus, the more pronounced the suppression of the other objects;
  • Spatial Distance: event-related potentials, functional MRI investigations as well as single-cell recordings in monkeys [62] and humans [63] have pointed out that neural enhancement in attentional focus may be followed by neural inhibition in peripheral regions. Detection is slower and discrimination performance is poorer at interference locations close to the target compared to interference locations far from the target [64]. In the receptive fields of cells, shifting attention from one stimulus to another can have a strong effect when two stimuli are in close proximity to each other. For stimuli that are far apart, the effect is much less [59]. In other words, the more spatially distant the objects are from each other, the smaller the inhibitory impact;
  • Relative Area: since an object generally has several neighbors, the relative area of an object shows with its neighbors affects its attention competition. The rationale for selecting relative area is that, even if the object has an attractive intrinsic area, it may not stand out unless it exhibits the greatest contrast if all of its neighbors also have attractive areas [53]. In particular, the higher the area contrast, the more strongly other objects will be suppressed.
Based on the points discussed above, the attentional competition weight is defined as follows:
C A ( k ) = k = 1 , k l N D τ ( k , l ) t i m e s D ν ( k , l ) · S δ ( k )
in which, D τ ( k , l ) , D ν ( k , l ) are the average focused intensity distance and the spatial proximity distance between the k-th, l-th semantic objects, respectively.
D τ ( k , l ) = exp F ( O k ) F ( O l ) F ( O k ) + F ( O l )
F ( O k ) = m e a n ( a t t p r e ( O k ) )
where F ( O k ) denotes the average focused intensity fo the k-th object.
D ν ( k , l ) = exp μ ( x k , y k ) ( x l , y l ) 2
where μ is a scaling factor, ( x k , y k ) and ( x l , y l ) are the coordinates of the center point of the k-th, l-th semantic objects.
S δ ( k ) = 1 exp ( 12 · G ( k ) )
where S δ ( k ) represents the relative area competition weight of the k-th semantic object, G ( k ) is the relative area of the k-th object with respect to its neighbors, see [53] for more details.
As a result, the semantics-based attention model is constructed by melting the attentional competition weight.
S e m a t t = a t t p r e · C A
Figure 10a–c takes the 2nd frame of the ”Basketball Drill” sequence as an illustration, showing the attention map of semantic sensitivity, objective area and central bias. Figure 10d displays the inhibition effect of the contextual complexity. Figure 10e shows the attentional competition result. Figure 10f exhibits the final attention map. Intuitively, brighter areas indicate higher attention, inhibition and competition.

5. Semantic-Based Spatio-Temporal Transform Domain JND Profile

As mentioned above, this paper combines semantic sensitivity, objective area, central bias, contextual complexity and attentional competition to measure semantic perceptual attention. Then, based on the model in [65], considering the impact of semantics, this study modifies the spatial attention factor, aiming for a more accurate JND model. The framework of the proposed JND profile is shown in Figure 11.
By introducing the weighting factor of the semantics-based attention, we propose the following JND model.
J N D ( t , n , i , j ) = J N D s t ( t , n , i , j ) · w s ( t , n )
here, J N D s t ( t , n , i , j ) is the spatio-temporal JND threshold of coefficient ( i , j ) of the n-th block in the t-th frame, considering the LA, CSF and masking effects [23].
Taking the spatio-temporal JND threshold J N D s t as the basis, this work proposes a patch-wise weighting factor, the attentional weight w s ( t , n ) , accounting for the visual perceptual attention of semantics, aiming at developing a more accurate JND model. Patch level w s ( t , n ) is determined as the mean of pixel-wise adjustment factors in the spatial domain, w s ( t , i , j ) , of the n-th image block in the t-th frame.
To be specific, by following [65], we estimate spatial attention A s ( x ) , and combine semantic attention S e m a t t ( x ) to get semantics-based spatial attention A F ( x ) using NAMM [5].
A F ( x ) = A s ( x ) + S e m a t t ( x ) 0.3 · min ( A s ( x ) , S e m a t t ( x ) )
In general, the larger the visual attention A F ( x ) , the smaller the JND threshold. Intuitively, it seems sensible to apply the sigmoid-like function to normalize A F ( x ) to [0, 1] and measure the spatial attention weight w s ( t , n ) below.
w s ( t , n ) = ϑ 1 · 1 1 1 + exp ( ϑ 2 · A F ( t , n ) )
where ϑ 1 , ϑ 2 are normal constants, A F ( t , n ) is the mean of the n-th block of A F ( x ) . According to subjective experiments, ϑ 1 and ϑ 2 are set to 2.5 and 2, respectively.
Figure 12a,b display the obtained attention map and the corresponding weight map. Figure 12c shows the final spatio-temporal JND threshold map. In Figure 12a,c, brighter areas indicate higher visual attention or masking, while the opposite is true for Figure 12b.

6. Experimental Results

6.1. Comparison of Model Performance

An ideal JND model should, in a sense, distribute noise more fairly, concealing more of it, with acceptable perceptual quality. Thus, we add coefficient-wise JND-guided noise to video sequences, as described in [22], to assess the effectiveness.
R ^ ( t , n , i , j ) = R ( t , n , i , j ) + ρ · r a n d ( t , n , i , j ) · J N D ( t , n , i , j )
where R ( t , n , i , j ) is the transform coefficients of original sequence, R ^ ( t , n , i , j ) is the JND noise contaminated coefficients, ρ regulates the energy of JND noise, and r a n d ( t , n , i , j ) { + 1 , 1 } is a bipolar random noise.
We evaluate the effectiveness of the proposed work in comparison to four benchmark models, namely Bae 2017 [22], Zeng 2019 [66], Wang 2020 [12], Xing 2021 [23], and Li 2022 [67]. Ten videos with diverse semantics were chosen from the HEVC standard sequences in order to meet the requirements of different resolutions. Four of them are videos with 1920 × 1080 full HD resolution, namely, “Kimono1”, “ParkScene”, “Basketball Drive” and “BQTerrace”, as shown in Figure 13a–d. Three of them, “FourPeople”, “Johnny”, “KristenAndSara”, are 1280 × 720 resolution videos, as shown in Figure 13e–g. The remaining three, “Basketball Drill”, “PartyScene”, and “RaceHorses”, are videos with 832 × 480 resolution, as shown in Figure 13h–j.
Using Peak Signal-to-Noise Ratio (PSNR) as an objective evaluation criterion, a lower PSNR indicates a better ability to mask noise. However, despite that PSNR is the most popular objective quality evaluation metric, the results of a large number of empirical-psychological studies have demonstrated that PSNR scores fall short of properly capturing HVS perception due to the involvement of visual physiological and psychological mechanisms. Furthermore, the human eye is the receiver of the final visual signal, and therefore, subjective quality metrics must be taken into account when evaluating the proposed JND model.
In this paper, the visual quality of JND-contaminated videos was assessed by recruiting 17 subjects with normal or corrected normal vision, using the subjective viewing test in [68]. Specifically, the monitor used for the display was a 27-in LED monitor, while the view distance was six times the height of the video frame. The difference between the original and processed sequences was observed using the Double Stimulus Continuous Quality Scale (DSCQS) method [23].
Figure 14 illustrates the testing procedure of the DSCQS. In the experiment, for each presentation, the reference and test sequences are arranged in a pseudo-randomized form. During the voting period, participants are expected to rate the quality of each of the two videos. The visual quality of the JND-contaminated videos is then measured by calculating the MOS difference (DMOS) between the original and matched processing sequences. The calculation is described below.
DMOS = M O S J N D M O S O R I
where M O S O R I and M O S J N D are the measured average opinion score values for the original and test videos, respectively. Five quality scales are used, i.e., excellent (80–100), good (60–80), fair (40–60), poor (20–40), and bad (0–20). The smaller the value of DMOS, the better the visual quality of the JND-polluted video.
The detailed performance results of different JND models are shown in Table 1. From the panoramic viewpoint, it can be found that the proposed JND model has the best performance on PSNR and DMOS for all videos except the “RaceHorses” sequence. This is mainly due to the large proportion of foreground semantic objects in the “RaceHorses” sequence, while our model guides only a small amount of noise into the foreground semantic object region. Therefore, the PSNR value of the proposed JND model is slightly higher than this of Xing, 2021. In addition, as shown in Table 2, the Video Multimethod Assessment Fusion (VMAF) scores [69] of the noise-contaminated videos of different JND models are measured, and the proposed model obtains optimal or suboptimal scores. The larger VMAF scores indicate that the subjective quality of the noise-contaminated sequences is closer to that of the original sequences. The results in Table 2 further verify the superiority of the proposed model in the VMAF metric scenario. Meanwhile, the average PSNR values of Bae 2017, Zeng 2019, Wang 2020, Xing 2021, Li 2022 and the proposed JND profile are 28.36 dB, 30.08 dB, 28.90 dB, 27.83 dB, 30.95 dB, and 27.03 dB, respectively. The average DMOS values of these are 12.49, 23.47, 24.08, 12.51, 19.59 and 9.42, respectively. The average VMAF values of these are 90.09, 87.71, 90.98, 97.30, 88.96 and 97.39, respectively. To visualize the data, Figure 15 displays the bar graphs of the average PSNR, VMAF and accompanying DMOS results for test sequences of various resolutions. Apparently, the noise-contaminated video generated by our model achieves the best perceptual quality (smallest DMOS score) and the largest distortion (lowest PSNR value) compared to the other four models. These experimental results validate the superiority of our JND profile in guiding noise injection.
In order to compare these five JND models more clearly, Figure 16 takes the 4th frame of the “Kimono1” sequence as an example. In general, HVS tends to pay more attention to the semantic targets [65]. In Figure 16a, the woman is prominent relative to the background. Therefore, a lower JND value should be applied in this region to inject less noise. We mainly focus on the head and body to allow for facilitate visual comparison. Apparently, there is considerable visible noise in Figure 16d–g, while the eyes and ears in Figure 16c are blurred. In contrast, Figure 16h is significantly clearer with the lowest PSNR value. As far as the body is concerned, the noise in Figure 16j–m is plainly apparent without exception, while in Figure 16o it is almost undetectable. As a consequence, the above objective and subjective evaluations validate that the proposed semantic attention-based weighting model is effective and excellent.
As mentioned in the introduction, high-level semantic objects other than humans often appear in image and video frames. Frame 70 of the “Basketball Drill” sequence is depicted in Figure 17 as an illustration. Clearly, this is a scene depicting basketball training. Naturally, in this case, the basketball and the ball frame are obviously of interest in addition to the basketball players. With respect to the basketball frame, Figure 17c–g show significant distortion, while Figure 17h exhibits strong performance. For the basketball, Figure 17j–l,n depict considerable distortion, Figure 17m shows significant distortion at the lower right of the basketball, while the noise in Figure 17o is barely visible.

6.2. Ablation Experiment and Analysis

To verify the effectiveness of each module in this algorithm, this paper performed ablation experiments from three perception modules: object, context, and cross-object, and evaluated their impacts on the perception results.
The experimental results in Table 3 demonstrate that a combination of the three modules achieves optimal results, while the other combinations do not achieve the desired effect. In other words, when considering the influence of semantics on JND, one of the three aspects of object, context and cross-object features is indispensable. It is worth noting that since the context and cross-object modules are calculated based on object, combinations that do not contain object module are discarded. In practical applications, different kinds of videos focus on different scales of semantic features, and the combination of three modules achieves better generality.

7. Conclusions

A novel semantic-based JND model was proposed in this paper by thoroughly mining and characterizing the semantic feature parameters that affect attention. The interaction between semantic visual features and HVS was investigated from object, contextual and cross-object perspectives. This analysis included HVS responses induced by semantic sensitivity, objective area and central bias, as well as perceptual suppression brought about by contextual complexity and cross-object competition for attention. In conjunction with underlying attention in the spatial domain, perceptual attention to stimuli was measured with information-theoretic support and incorporated into a patch-level weighting factor. Using the semantic-based attentional weight, the JND model for the spatiotemporal transform domain was modified. The experimental results demonstrate the effectiveness of the proposed JND model with superior performance and stronger distortion-hiding ability compared to the state-of-the-art JND model.
Existing JND models only consider the effects of unimodal signals, while the study of cross-modal JND remains an open problem. However, there exists great incentive to transfer this issue from the laboratory to real-world adoption. In the future, we will further investigate multimodal asynchronous perception, seek a more unified approach to fusing multi-modal feature parameters, and obtain more accurate JND thresholds.

Author Contributions

Conceptualization, X.W. and H.Y.; data curation, X.W.; funding acquisition, H.Y.; methodology, X.W., Y.L. and Y.C.; project administration, H.Y.; software, X.W. and S.Z.; Supervision, H.Y.; writing—original draft, X.W.; writing—review and editing, H.Y., Y.L., S.Z. and Y.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by “Pioneer” and “Leading Goose” R&D Program of Zhejiang Province (2022C01068), NSFC 61972123, and NSFC 62031009.

Conflicts of Interest

The authors declare that they have no known competing financial interests or personal relationships that could have appeared to influence the work reported in this paper.

References

  1. Macknik, S.L.; Livingstone, M.S. Neuronal Correlates of Visibility and Invisibility in the Primate Visual System. Nat. Neurosci. 1998, 1, 144–149. [Google Scholar] [CrossRef] [PubMed]
  2. Carlson, N.R. Psychology: The Science of Behavior; McGraw-Hill: New York, NY, USA, 1987. [Google Scholar]
  3. Wu, J.; Shi, G.; Lin, W. Survey of Visual Just Noticeable Difference Estimation. Front. Comput. Sci. 2019, 13, 4–15. [Google Scholar] [CrossRef]
  4. Chou, C.H.; Li, Y.C. A Perceptually Tuned Subband Image Coder Based on the Measure of Just-noticeable-distortion Profile. IEEE Trans. Circuits Syst. Video Technol. 1995, 5, 467–476. [Google Scholar] [CrossRef]
  5. Yang, X.; Lin, W.; Lu, Z.; Ong, E.P.; Yao, S. Just-noticeable-distortion Profile with Nonlinear Additivity Model for Perceptual Masking in Color Images. IEEE Int. Conf. Acoust. Speech Signal Process. 2003, 3, III-609. [Google Scholar]
  6. Liu, A.; Lin, W.; Paul, M.; Deng, C.; Zhang, F. Just Noticeable Difference for Images with Decomposition Model for Separating Edge and Textured Regions. IEEE Trans. Circuits Syst. Video Technol. 2010, 20, 1648–1652. [Google Scholar] [CrossRef]
  7. Chen, Z.; Guillemot, C. Perceptually-Friendly H.264/AVC Video Coding Based on Foveated Just-Noticeable-Distortion Model. IEEE Trans. Circuits Syst. Video Technol. 2010, 20, 806–819. [Google Scholar] [CrossRef]
  8. Wu, J.; Shi, G.; Lin, W.; Liu, A.; Qi, F. Just Noticeable Difference Estimation for Images with Free-energy Principle. IEEE Trans. Multimed. 2013, 15, 1705–1710. [Google Scholar] [CrossRef]
  9. Wu, J.; Li, L.; Dong, W.; Shi, G.; Lin, W.; Kuo, C.C.J. Enhanced Just Noticeable Difference Model for Images with Pattern Complexity. IEEE Trans. Image Process. 2017, 26, 2682–2693. [Google Scholar] [CrossRef]
  10. Jakhetiya, V.; Lin, W.; Jaiswal, S.; Gu, K.; Guntuku, S.C. Just Noticeable Difference for Natural Images Using RMS Contrast and Feed-back Mechanism. Neurocomputing 2018, 275, 366–376. [Google Scholar] [CrossRef]
  11. Chen, Z.; Wu, W. Asymmetric foveated just-noticeable-difference model for images with visual field inhomogeneities. IEEE Trans. Circuits Syst. Video Technol. 2019, 30, 4064–4074. [Google Scholar] [CrossRef]
  12. Wang, H.; Yu, L.; Liang, J.; Yin, H.; Li, T.; Wang, S. Hierarchical Predictive Coding-based JND Estimation for Image Compression. IEEE Trans. Image Process. 2020, 30, 487–500. [Google Scholar] [CrossRef] [PubMed]
  13. Cui, X.; Peng, Z.; Chen, F.; Jiang, G.; Yu, M. Perceptual ultra-high definition video coding based on adaptive just noticeable distortion model. Displays 2022, 75, 102301. [Google Scholar] [CrossRef]
  14. Ahumada, A.J., Jr.; Peterson, H.A. Luminance-model-based DCT Quantization for Color Image Compression. Hum. Vis. Vis. Process. Digit. Disp. III 1992, 1666, 365–374. [Google Scholar]
  15. Watson, A.B. DCTune: A Technique for Visual Optimization of DCT Quantization Matrices for Individual Images. Soc. Inf. Disp. Dig. Tech. Pap. 1993, 24, 946. [Google Scholar]
  16. Zhang, X.; Lin, W.; Xue, P. Improved Estimation for Just-noticeable Visual Distortion. Signal Process. 2005, 85, 795–808. [Google Scholar] [CrossRef]
  17. Jia, Y.; Lin, W.; Kassim, A.A. Estimating Just-noticeable Distortion for Video. IEEE Trans. Circuits Syst. Video Technol. 2006, 16, 820–829. [Google Scholar] [CrossRef]
  18. Wei, Z.; Ngan, K.N. Spatio-Temporal Just Noticeable Distortion Profile for Grey Scale Image/Video in DCT Domain. IEEE Trans. Circuits Syst. Video Technol. 2009, 19, 337–346. [Google Scholar]
  19. Wang, H.; Wang, L.; Hu, X.; Tu, Q.; Men, A. Perceptual video coding based on saliency and just noticeable distortion for H. 265/HEVC. In Proceedings of the International Symposium on Wireless Personal Multimedia Communications, Sydney, NSW, Australia, 7–10 September 2014; pp. 106–111. [Google Scholar]
  20. Bae, S.H.; Kim, J.; Kim, M. HEVC-based perceptually adaptive video coding using a DCT-based local distortion detection probability model. IEEE Trans. Image Process. 2016, 25, 3343–3357. [Google Scholar] [CrossRef]
  21. Wan, W.; Wu, J.; Xie, X.; Shi, G. A Novel Just Noticeable Difference Model via Orientation Regularity in DCT Domain. IEEE Access 2017, 5, 22953–22964. [Google Scholar] [CrossRef]
  22. Bae, S.H.; Kim, M. A DCT-Based Total JND Profile for Spatiotemporal and Foveated Masking Effects. IEEE Trans. Circuits Syst. Video Technol. 2017, 27, 1196–1207. [Google Scholar] [CrossRef]
  23. Xing, Y.; Yin, H.; Zhou, Y.; Chen, Y.; Yan, C. Spatiotemporal Just Noticeable Difference Modeling with Heterogeneous Temporal Visual Features. Displays 2021, 70, 102096. [Google Scholar] [CrossRef]
  24. Jiang, Q.; Liu, Z.; Wang, S.; Shao, F.; Lin, W. Towards Top-Down Just Noticeable Difference Estimation of Natural Images. IEEE Trans. Image Process. 2022, 31, 3697–3712. [Google Scholar] [CrossRef] [PubMed]
  25. Ki, S.; Bae, S.H.; Kim, M.; Ko, H. Learning-based Just-noticeable-quantization-distortion Modeling for Perceptual Video Coding. IEEE Trans. Image Process. 2018, 27, 3178–3193. [Google Scholar] [CrossRef] [PubMed]
  26. Hadizadeh, H.; Heravi, A.R.; Bajić, I.V.; Karami, P. A Perceptual Distinguishability Predictor for JND-noise-contaminated Images. IEEE Trans. Image Process. 2018, 28, 2242–2256. [Google Scholar] [CrossRef] [PubMed]
  27. Zhang, X.; Yang, C.; Wang, H.; Xu, W.; Kuo, C.C.J. Satisfied-user-ratio Modeling for Compressed Video. IEEE Trans. Image Process. 2020, 29, 3777–3789. [Google Scholar] [CrossRef] [PubMed]
  28. Liu, H.; Zhang, Y.; Zhang, H.; Fan, C.; Kwong, S.; Kuo, C.C.J.; Fan, X. Deep Learning-based Picture-wise Just Noticeable Distortion Prediction Model for Image Compression. IEEE Trans. Image Process. 2019, 29, 641–656. [Google Scholar] [CrossRef]
  29. Lin, H.; Hosu, V.; Fan, C.; Zhang, Y.; Mu, Y.; Hamzaoui, R.; Saupe, D. SUR-FeatNet: Predicting the Satisfied User Ratio Curve for Image Compression with Deep Feature Learning. Qual. User Exp. 2020, 5, 1–23. [Google Scholar] [CrossRef]
  30. Wu, Y.; Wang, Z.; Chen, W.; Lin, L.; Wei, H.; Zhao, T. Perceptual VVC quantization refinement with ensemble learning. Displays 2021, 70, 102103. [Google Scholar] [CrossRef]
  31. Wu, Y.; Ji, W.; Wu, J. Unsupervised Deep Learning for Just Noticeable Difference Estimation. In Proceedings of the 2020 IEEE International Conference on Multimedia & Expo Workshops (ICMEW), London, UK, 6–10 July 2020; pp. 1–6. [Google Scholar]
  32. Jin, J.; Yu, D.; Lin, W.; Meng, L.; Wang, H.; Zhang, H. Full RGB Just Noticeable Difference (JND) Modelling. arXiv 2022, arXiv:2203.00629. [Google Scholar]
  33. Jin, J.; Xue, Y.; Zhang, X.; Meng, L.; Zhao, Y.; Lin, W. HVS-Inspired Signal Degradation Network for Just Noticeable Difference Estimation. arXiv 2022, arXiv:2208.07583. [Google Scholar]
  34. Jin, L.; Lin, J.Y.; Hu, S.; Wang, H.; Wang, P.; Katsavounidis, I.; Aaron, A.; Kuo, C.C.J. Statistical Study on Perceived JPEG Image Quality via MCL-JCI Dataset Construction and Analysis. Electron. Imaging 2016, 2016, 1–9. [Google Scholar] [CrossRef] [Green Version]
  35. Liu, X.; Chen, Z.; Wang, X.; Jiang, J.; Kowng, S. JND-Pano: Database for Just Noticeable Difference of JPEG Compressed Panoramic Images. In Pacific Rim Conference on Multimedia, Proceedings of the 19th Pacific-Rim Conference on Multimedia, Hefei, China, 21–22 September 2018; Springer: Berlin/Heidelberg, Germany, 2018; pp. 458–468. [Google Scholar]
  36. Wang, H.; Gan, W.; Hu, S.; Lin, J.Y.; Jin, L.; Song, L.; Wang, P.; Katsavounidis, I.; Aaron, A.; Kuo, C.C.J. MCL-JCV: A JND-based H.264/AVC Video Quality Assessment Dataset. In Proceedings of the 2016 IEEE International Conference on Image Processing (ICIP), Phoenix, AZ, USA, 25–28 September 2016; pp. 1509–1513. [Google Scholar]
  37. Wang, H.; Katsavounidis, I.; Zhou, J.; Park, J.; Lei, S.; Zhou, X.; Pun, M.O.; Jin, X.; Wang, R.; Wang, X.; et al. VideoSet: A Large-scale Compressed Video Quality Dataset Based on JND Measurement. J. Visual Commun. Image Represent. 2017, 46, 292–302. [Google Scholar] [CrossRef]
  38. Jin, J.; Zhang, X.; Fu, X.; Zhang, H.; Lin, W.; Lou, J.; Zhao, Y. Just Noticeable Difference for Deep Machine Vision. IEEE Trans. Circuits Syst. Video Technol. 2021, 32, 3452–3461. [Google Scholar] [CrossRef]
  39. Lin, W.; Ghinea, G. Progress and Opportunities in Modelling Just-Noticeable Difference (JND) for Multimedia. IEEE Trans. Multimed. 2021, 24, 3706–3721. [Google Scholar] [CrossRef]
  40. Liu, H.; Shi, S.; Bai, R.; Liu, Y.; Lian, X.; Shi, T. A brain-inspired computational model for extremely few reference image quality assessment. Displays 2023, 76, 102331. [Google Scholar] [CrossRef]
  41. Wang, H.; Yu, L.; Wang, S.; Xia, G.; Yin, H. A Novel Foveated-JND Profile Based on an Adaptive Foveated Weighting Model. In Proceedings of the 2018 IEEE Visual Communications and Image Processing (VCIP), Taichung, Taiwan, 9–12 December 2018; pp. 1–4. [Google Scholar]
  42. Itti, L.; Koch, C. A Saliency-based Search Mechanism for Overt and Covert Shifts of Visual Attention. Vision Res. 2000, 40, 1489–1506. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  43. Judd, T.; Ehinger, K.; Durand, F.; Torralba, A. Learning to Predict Where Humans Look. In Proceedings of the 2009 IEEE 12th International Conference on Computer Vision, Kyoto, Japan, 29 September–2 October 2009; pp. 2106–2113. [Google Scholar]
  44. Borji, A.; Sihite, D.N.; Itti, L. What Stands out in a Scene? A Study of Human Explicit Saliency Judgment. Vision Res. 2013, 91, 62–77. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  45. Wang, C.; Wang, C.; Li, W.; Wang, H. A brief survey on RGB-D semantic segmentation using deep learning. Displays 2021, 70, 102080. [Google Scholar] [CrossRef]
  46. Wang, R.; Zhang, Z. Energy Coding in Biological Neural Networks. Cognit. Neurodyn. 2007, 1, 203–212. [Google Scholar] [CrossRef] [Green Version]
  47. Feldman, H.; Friston, K.J. Attention, Uncertainty, and Free-energy. Front. Hum. Neurosci. 2010, 4, 1–23. [Google Scholar] [CrossRef] [Green Version]
  48. Selvaraju, R.; Cogswell, M.; Das, A.; Vedantam, R.; Parikh, D.; Batra, D. Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization. In Proceedings of the 2017 IEEE International Conference on Computer Vision (ICCV 2017), Venice, Italy, 22–29 October 2017; pp. 618–626. [Google Scholar]
  49. Wu, J.; Jiang, Y.; Zhang, W.; Bai, X.; Bai, S. SeqFormer: A Frustratingly Simple Model for Video Instance Segmentation. arXiv 2021, arXiv:2112.08275. [Google Scholar]
  50. Long, Y.; Jin, D.; Wu, Z.; Zuo, Z.; Wang, Y.; Kang, Z. Accurate Identification of Infrared Ship in Island-Shore Background Based on Visual Attention. In Proceedings of the 2022 IEEE Asia-Pacific Conference on Image Processing, Electronics and Computers (IPEC), Dalian, China, 14–16 April 2022; pp. 800–806. [Google Scholar]
  51. Bruce, N.D.; Tsotsos, J.K. Saliency, Attention, and Visual Search: An Information Theoretic Approach. J. Vis. 2009, 9, 1–24. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  52. Shen, Z.; Zhang, L.; Li, R.; Hou, J.; Liu, C.; Hu, W. The effects of color combinations, luminance contrast, and area ratio on icon visual search performance. Displays 2021, 67, 101999. [Google Scholar] [CrossRef]
  53. Syeda-Mahmood, T.F. Attentional Selection in Object Recognition; Technical Report 1420; MIT Computer Science & Artificial Intelligence Laboratory: Cambridge, MA, USA, 1993. [Google Scholar]
  54. Wu, C.C.; Wick, F.A.; Pomplun, M. Guidance of Visual Attention by Semantic Information in Real-world Scenes. Front. Psychol. 2014, 5, 54. [Google Scholar] [CrossRef] [Green Version]
  55. Davenport, J.L.; Potter, M.C. Scene Consistency in Object and Background Perception. Psychol. Sci. 2004, 15, 559–564. [Google Scholar] [CrossRef]
  56. Beck, M.R.; Lohrenz, M.C.; Trafton, J.G. Measuring Search Efficiency in Complex Visual Search Tasks: Global and Local Clutter. J. Exp. Psychol. Appl. 2010, 16, 238. [Google Scholar] [CrossRef] [PubMed]
  57. Caroux, L.; Mouginé, A. Influence of Visual Background Complexity and Task Difficulty on Action Video Game Players’ Performance. Entertain. Comput. 2022, 41, 100471. [Google Scholar] [CrossRef]
  58. Dimond, S.J.; Farrington, L.; Johnson, P. Differing Emotional Response from Right and Left Hemispheres. Nature 1976, 261, 690–692. [Google Scholar] [CrossRef]
  59. Duncan, J. EPS Mid-Career Award 2004: Brain Mechanisms of Attention. Q. J. Exp. Psychol. 2006, 59, 2–27. [Google Scholar] [CrossRef]
  60. Scalf, P.E.; Basak, C.; Beck, D. Attention Does More than Modulate Suppressive Interactions: Attending to Multiple Items. Exp. Brain Res. 2011, 212, 293–304. [Google Scholar] [CrossRef]
  61. Scalf, P.E.; Torralbo, A.; Tapia, E.; Beck, D.M. Competition Explains Limited Attention and Perceptual Resources: Implications for Perceptual Load and Dilution Theories. Front. Psychol. 2013, 4, 243. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  62. Schall, J.D.; Sato, T.R.; Thompson, K.G.; Vaughn, A.A.; Juan, C.H. Effects of Search Efficiency on Surround Suppression During Visual Selection in Frontal Eye Field. J. Neurophysiol. 2004, 91, 2765–2769. [Google Scholar] [CrossRef] [PubMed]
  63. Müller, N.G.; Kleinschmidt, A. The Attentional ‘Spotlight’s’ Penumbra: Center-surround Modulation in Striate Cortex. Neuroreport 2004, 15, 977–980. [Google Scholar] [CrossRef]
  64. Mounts, J.R. Evidence for Suppressive Mechanisms in Attentional Selection: Feature Singletons Produce Inhibitory Surrounds. Percept. Psychophys. 2000, 62, 969–983. [Google Scholar] [CrossRef] [Green Version]
  65. Yan, Y.; Ren, J.; Sun, G.; Zhao, H.; Han, J.; Li, X.; Marshall, S.; Zhan, J. Unsupervised Image Saliency Detection with Gestalt-laws Guided Optimization and Visual Attention Based Refinement. Pattern Recognit. 2018, 79, 65–78. [Google Scholar] [CrossRef] [Green Version]
  66. Zeng, Z.; Zeng, H.; Chen, J.; Zhu, J.; Zhang, Y.; Ma, K.K. Visual Attention Guided Pixel-Wise Just Noticeable Difference Model. IEEE Access 2019, 7, 132111–132119. [Google Scholar] [CrossRef]
  67. Li, J.; Yu, L.; Wang, H. Perceptual redundancy model for compression of screen content videos. IET Image Process. 2022, 16, 1724–1741. [Google Scholar] [CrossRef]
  68. Yang, X.; Lin, W.; Lu, Z.; Ong, E.P.; Yao, S.S. Just Noticeable Distortion Model and Its Applications in Video Coding. Signal Process. Image Commun. 2005, 20, 662–680. [Google Scholar] [CrossRef]
  69. Li, Z.; Aaron, A.; Katsavounidis, I.; Moorthy, A.; Manohara, M. Toward a practical perceptual video quality metric. Netflix Tech Blog 2016, 6, 2. [Google Scholar]
Figure 1. The grad-CAM heat-map results.
Figure 1. The grad-CAM heat-map results.
Sensors 23 03149 g001
Figure 2. The VNext extraction results.
Figure 2. The VNext extraction results.
Sensors 23 03149 g002
Figure 3. (a) A semantic sensitivity histogram; (b) A semantic sensitivity PDF curve.
Figure 3. (a) A semantic sensitivity histogram; (b) A semantic sensitivity PDF curve.
Sensors 23 03149 g003
Figure 4. The function map of Equation (4).
Figure 4. The function map of Equation (4).
Sensors 23 03149 g004
Figure 5. The illustration for the aspect ratio.
Figure 5. The illustration for the aspect ratio.
Sensors 23 03149 g005
Figure 6. (a) The traditional central bias map; (b) The modified central bias map.
Figure 6. (a) The traditional central bias map; (b) The modified central bias map.
Sensors 23 03149 g006
Figure 7. Areas of vision field.
Figure 7. Areas of vision field.
Sensors 23 03149 g007
Figure 8. (a) The 212th original frame of the “Basketball Drill”; (b) The complexity map of the whole frame; (c) The contextual complexity map of the semantics; (d) The inhibitory effect of contextual complexity on semantics.
Figure 8. (a) The 212th original frame of the “Basketball Drill”; (b) The complexity map of the whole frame; (c) The contextual complexity map of the semantics; (d) The inhibitory effect of contextual complexity on semantics.
Sensors 23 03149 g008
Figure 9. (a) The initial attention map with the identical weights; (b) The initial attention map with the gradient variation based weights.
Figure 9. (a) The initial attention map with the identical weights; (b) The initial attention map with the gradient variation based weights.
Sensors 23 03149 g009
Figure 10. (ac) The attention map of the semantic sensitivity, objective area, central bias; (d) The inhibition effect of the contextual complexity; (e) The attentional competition map; (f) The semantic attention map.
Figure 10. (ac) The attention map of the semantic sensitivity, objective area, central bias; (d) The inhibition effect of the contextual complexity; (e) The attentional competition map; (f) The semantic attention map.
Sensors 23 03149 g010
Figure 11. A briefoverview of the proposed spatio-temporal JND model.
Figure 11. A briefoverview of the proposed spatio-temporal JND model.
Sensors 23 03149 g011
Figure 12. (a) The semantic-based spatial attention A F ( x ) map; (b) The spatial attention weight w s ( x ) map; (c) The proposed JND threshold map.
Figure 12. (a) The semantic-based spatial attention A F ( x ) map; (b) The spatial attention weight w s ( x ) map; (c) The proposed JND threshold map.
Sensors 23 03149 g012
Figure 13. The HEVC standard test sequences. (a) Kimono1; (b) ParkScene; (c) Basketball Drive; (d) BQTerrace; (e) FourPeople; (f) Johnny; (g) KristenAndSara; (h) Basketball Drill; (i) PartyScene; (j) RaceHorses.
Figure 13. The HEVC standard test sequences. (a) Kimono1; (b) ParkScene; (c) Basketball Drive; (d) BQTerrace; (e) FourPeople; (f) Johnny; (g) KristenAndSara; (h) Basketball Drill; (i) PartyScene; (j) RaceHorses.
Sensors 23 03149 g013
Figure 14. DSCQS method, where the original sequence (ORI) and test sequence (TEST) are pseudo-randomly ordered for each presentation.
Figure 14. DSCQS method, where the original sequence (ORI) and test sequence (TEST) are pseudo-randomly ordered for each presentation.
Sensors 23 03149 g014
Figure 15. (a) Average PSNR values in Table 1 for 832 × 480, 1280 × 720, and 1920 × 1080; (b) Average DMOS values in Table 1 for 832 × 480, 1280 × 720, and 1920 × 1080; (c) Average VMAF scores in Table 2 for 832 × 480, 1280 × 720, and 1920 × 1080 [12,22,23,66,67].
Figure 15. (a) Average PSNR values in Table 1 for 832 × 480, 1280 × 720, and 1920 × 1080; (b) Average DMOS values in Table 1 for 832 × 480, 1280 × 720, and 1920 × 1080; (c) Average VMAF scores in Table 2 for 832 × 480, 1280 × 720, and 1920 × 1080 [12,22,23,66,67].
Sensors 23 03149 g015
Figure 16. The 4th frame of the “Kimono1” sequence. (a) The original image; (bh) are the enlarged images of the sub-region ①; (io) are the enlarged images of the sub-region ②. In turn, they are the original patch and the distorted versions processed by Bae 2017, Zeng 2019, Wang 2020, Xing 2021, Li 2022 and the proposed JND model, respectively. The PSNR values of the distorted versions are 27.53 dB, 28.17 dB, 27.76 dB, 27.10 dB, 27.98 dB and 26.69 dB, respectively.
Figure 16. The 4th frame of the “Kimono1” sequence. (a) The original image; (bh) are the enlarged images of the sub-region ①; (io) are the enlarged images of the sub-region ②. In turn, they are the original patch and the distorted versions processed by Bae 2017, Zeng 2019, Wang 2020, Xing 2021, Li 2022 and the proposed JND model, respectively. The PSNR values of the distorted versions are 27.53 dB, 28.17 dB, 27.76 dB, 27.10 dB, 27.98 dB and 26.69 dB, respectively.
Sensors 23 03149 g016
Figure 17. The 70th frame of the “Basketball Drill” sequence. (a) The original image; (bh) are the enlarged images of the sub-region ①; (io) are the enlarged images of the sub-region ②. In turn, they are the original patch and the distorted versions processed by Bae 2017, Zeng 2019, Wang 2020, Xing 2021, Li 2022 and the proposed JND model, respectively. The PSNR values of the distorted versions are 27.24 dB, 32.12 dB, 30.63 dB, 26.16 dB, 33.42 dB and 25.42 dB, respectively.
Figure 17. The 70th frame of the “Basketball Drill” sequence. (a) The original image; (bh) are the enlarged images of the sub-region ①; (io) are the enlarged images of the sub-region ②. In turn, they are the original patch and the distorted versions processed by Bae 2017, Zeng 2019, Wang 2020, Xing 2021, Li 2022 and the proposed JND model, respectively. The PSNR values of the distorted versions are 27.24 dB, 32.12 dB, 30.63 dB, 26.16 dB, 33.42 dB and 25.42 dB, respectively.
Sensors 23 03149 g017
Table 1. Performance comparison of different JND models.
Table 1. Performance comparison of different JND models.
SequencesBae 2017 [22]Zeng 2019 [66]Wang 2020 [12]Xing 2021 [23]Li 2022 [67]Proposed
PSNRDMOSPSNRDMOSPSNRDMOSPSNRDMOSPSNRDMOSPSNRDMOS
BasketballDrill27.2414.2431.9322.1830.0522.9426.1911.5933.3219.8225.2210.88
PartyScene29.4010.1830.4323.8828.9124.5327.9412.2432.0419.3527.1010.06
RaceHorses31.279.5930.5822.7629.5723.4729.5010.4731.4821.0629.968.71
FourPeople26.8614.4129.0624.4128.3824.9425.9014.8829.8619.2925.3011.94
Johnny27.7613.6530.8921.7629.2022.4728.9811.7131.2617.9426.917.29
KristenAndSara27.2812.4729.3024.1228.0125.5924.9412.9429.6920.1224.567.71
Kimono127.5413.9428.1725.0627.8625.6426.5514.7627.9323.2925.947.41
ParkScene26.6113.0028.2523.1827.6024.0024.9413.1228.6220.5324.568.18
BasketballDrive31.1010.7632.2922.4730.5623.2930.5210.6533.4216.2429.159.88
BQTerrace28.5410.5329.8923.6528.9023.8829.7312.7131.8818.2928.548.12
Average28.3612.4930.0823.4728.9024.0827.8312.5130.9519.5927.039.42
Table 2. Video multimethod assessment fusion scores.
Table 2. Video multimethod assessment fusion scores.
SequencesBae 2017 [22]Zeng 2019 [66]Wang 2020 [12]Xing 2021 [23]Li 2022 [67]Proposed
PSNRVMAFPSNRVMAFPSNRVMAFPSNRVMAFPSNRVMAFPSNRVMAF
BasketballDrill27.2487.9831.9387.0730.0592.3726.1997.9933.3288.4025.2298.78
PartyScene29.4087.6430.4384.0628.9188.8627.9496.3632.0485.4027.1096.03
RaceHorses31.2793.3630.5890.4029.5798.6529.5099.9131.4891.6129.9699.92
FourPeople26.8689.6629.0687.4828.3887.6325.9093.0729.8688.5425.3093.10
Johnny27.7689.9230.8989.1829.2089.2128.9895.8031.2690.8326.9195.56
KristenAndSara27.2890.0129.3088.5628.0187.0724.9495.6229.6990.1124.5695.84
Kimono127.5490.0128.1785.4427.8685.7126.5599.8227.9385.5325.9499.71
ParkScene26.6186.7628.2583.1527.6089.1124.9495.1528.6283.2924.5695.66
BasketballDrive31.1093.2532.2991.1530.5695.8030.5299.9233.4292.5229.1599.93
BQTerrace28.5492.0829.8990.5628.9095.4029.7399.4031.8893.4028.5499.35
Average28.3690.0930.0887.7128.9090.9827.8397.3030.9588.9627.0397.39
Table 3. Performance Comparison of Different Combinations of Modules.
Table 3. Performance Comparison of Different Combinations of Modules.
ModulesPSNR
ObjectContextCross-Object832 × 4801280 × 7201920 × 1080
26.6727.8528.41
29.9532.0727.91
31.4529.5131.09
29.80
25.6626.0726.53
28.3029.2625.02
30.3227.4830.19
29.12
25.5626.1026.32
28.2228.7825.67
30.2828.2430.70
28.97
25.2225.3025.94
27.1026.9124.56
29.9624.5629.15
28.54
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, X.; Yin, H.; Lu, Y.; Zhao, S.; Chen, Y. Semantically Adaptive JND Modeling with Object-Wise Feature Characterization, Context Inhibition and Cross-Object Interaction. Sensors 2023, 23, 3149. https://doi.org/10.3390/s23063149

AMA Style

Wang X, Yin H, Lu Y, Zhao S, Chen Y. Semantically Adaptive JND Modeling with Object-Wise Feature Characterization, Context Inhibition and Cross-Object Interaction. Sensors. 2023; 23(6):3149. https://doi.org/10.3390/s23063149

Chicago/Turabian Style

Wang, Xia, Haibing Yin, Yu Lu, Shiling Zhao, and Yong Chen. 2023. "Semantically Adaptive JND Modeling with Object-Wise Feature Characterization, Context Inhibition and Cross-Object Interaction" Sensors 23, no. 6: 3149. https://doi.org/10.3390/s23063149

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop