Previous Article in Journal
HISF: Hierarchical Interactive Semantic Fusion for Multimodal Prompt Learning
error_outline You can access the new MDPI.com website here. Explore and share your feedback with us.
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Eye-Tracking and Emotion-Based Evaluation of Wardrobe Front Colors and Textures in Bedroom Interiors

College of Furniture and Industrial Design, Nanjing Forestry University, Nanjing 210037, China
*
Author to whom correspondence should be addressed.
Multimodal Technol. Interact. 2026, 10(1), 7; https://doi.org/10.3390/mti10010007
Submission received: 24 November 2025 / Revised: 26 December 2025 / Accepted: 3 January 2026 / Published: 6 January 2026

Abstract

Wardrobe fronts form a major visual element in bedroom interiors, yet material selection for their colors and textures often relies on intuition rather than evidence. This study develops a data-driven framework that links gaze behavior and affective responses to occupants’ preferences for wardrobe front materials. Forty adults evaluated color and texture swatches and rendered bedroom scenes while eye-tracking data capturing attraction, retention, and exploration were collected. Pairwise choices were modeled using a Bradley–Terry approach, and visual-attention features were integrated with emotion ratings to construct an interpretable attention index for predicting preferences. Results show that neutral light colors and structured wood-like textures consistently rank highest, with scene context reducing preference differences but not altering the order. Shorter time to first fixation and longer fixation duration were the strongest predictors of desirability, demonstrating the combined influence of rapid visual capture and sustained attention. Within the tested stimulus set and viewing conditions, the proposed pipeline yields consistent preference rankings and an interpretable attention-based score that supports evidence-informed shortlisting of wardrobe-front materials. The reported relationships between gaze, affect, and choice are associative and are intended to guide design decisions within the scope of the present experimental settings.

1. Introduction

Wardrobe fronts and other large cabinet surfaces significantly shape the visual and emotional experience of domestic interiors. Yet, in contemporary custom home-furnishing practice, color and texture selections are often guided by designer intuition or small expert panels, with limited input from user data. This ad hoc approach leads to inconsistent outcomes and poor reproducibility, hindering the advancement of human-centric, evidence-based interior design [1].
Existing research indicates that early visual attention and gaze dynamics are closely linked to choice behavior, where initial fixations can predict preferences in decision-making tasks [2,3]. Meanwhile, the Pleasure–Arousal–Dominance (PAD) framework offers a reliable basis for quantifying emotional responses to materials and environments [4,5]. Color preferences are known to be influenced by ecological and contextual factors [6,7], while material properties such as gloss and roughness also affect aesthetic and environmental appraisal [8]. However, two key gaps impede design application: first, eye-movement metrics are seldom integrated with affective measures in a unified analytic framework that links gaze behavior, PAD responses, and preference outcomes, while not claiming causal identification under the current experimental design; second, few studies translate experimental findings into reproducible rankings for real-world design alternatives, particularly in the context of custom furniture like wardrobes—frequently used elements in home environments.
Compared with prior interior eye-tracking work that typically reports gaze–preference correlations or single-model predictors, we contribute a design-ready pipeline that (a) produces calibrated preference strengths (Bradley–Terry) with uncertainty [9], (b) explains and generalizes those ranks through an interpretable gaze-based Integrated Attention Index (IAI) derived from predictive models, and (c) links attention mechanisms to affect (PAD) and Desire through multivariate association (CCA) and a parsimonious path model. For designers, this combination turns laboratory measurements into reproducible material shortlists and actionable heuristics, rather than standalone statistical effects.
By integrating eye-tracking and PAD measures within a unified framework, we aim to establish an interpretable, data-driven pipeline that characterizes the associations between visual attention, affective appraisal, and preference within controlled interior-design stimuli. Specifically, we examine whether eye-tracking features predict user preferences for wardrobe colors and textures, how PAD mediates this relationship, and how a data-driven pipeline can generate applicable design rankings.
Our methodology combines isolated evaluations of color and texture with scene-based selections reflecting real wardrobe layouts. Using pairwise choices, we compute Bradley–Terry preference strengths with bootstrap confidence intervals, while single-item eye tracking informs an attention index trained via leave-one-subject-out validation. We further investigate how scene context modulates baseline preferences and whether attentional and affective mechanisms explain these shifts [10,11].
This paper introduces a Gaze–Emotion–Preference (GEP) framework that unifies early and sustained viewing measures with PAD and choice data in a reproducible model. It provides convergent validation through BT rankings, an attention index, and an external questionnaire, and offers practical tools for palette screening, texture shortlisting, and scene composition. Ultimately, the framework connects data analytics with design practice, supporting scalable, emotion-informed material selection and evidence-based decision-making in custom home-furnishing design. Here, “preference” is operationalized by (i) forced-choice selections modeled via Bradley–Terry strengths and (ii) Desire ratings, while “emotion” is quantified using the PAD dimensions. We emphasize that the framework is evaluated on a discretized set of representative materials and controlled renderings; external validity to broader material catalogs and naturalistic bedroom environments requires additional validation.
The research framework is illustrated in Figure 1.

2. Materials and Methods

2.1. Extraction of Visual Physical Properties

This step ensures that all visual inputs are perceptually standardized and comparable across experimental conditions. We derived device-independent descriptors of the stimuli to support analysis and replication. Color swatches were converted to CIE L*a*b* coordinates to approximate perceptual uniformity [12]; texture patches were summarized with rotation-invariant statistics. All computations were scripted to ensure identical preprocessing across isolated items and their scene instantiations.
The CIE L*a*b* and texture descriptors are device-independent representations used for analysis. The visual presentation of rendered stimuli remains display-dependent. To reduce variability, all experiments were conducted on the same monitor with fixed brightness settings, and the rendering and export pipeline used a consistent sRGB color space with gamma 2.2. No hardware colorimeter calibration was performed; instead, presentation was controlled by using a single display device and fixed display settings across all sessions.

2.1.1. Color Data Extraction and Adaptive Palette Coverage

Color samples were measured in CIE L*a*b* coordinates under CIE standard illuminant D65 and a 2° standard observer. Measurements were obtained using an X-Rite i1Pro 2 spectrophotometer with a 45°/0° measurement geometry under controlled laboratory conditions. For each color swatch, three repeated measurements were taken from a central region of the surface to reduce edge effects and local surface non-uniformity.
Measurement replicates were screened using the CIEDE2000 color-difference metric (ΔE00). Replicates exceeding a tolerance of ΔE00 = 1.0 were re-acquired, and swatches exhibiting unstable measurements after repetition were excluded from further analysis [13,14].
To cover the perceptual color space without redundancy, the measured L*a*b* coordinates were clustered using perceptual distances computed by ΔE00. Cluster medoids were selected as representative palette entries for subsequent experimental stimuli.

2.1.2. Texture Feature Extraction

To characterize texture structure while minimizing chromatic influences, texture images were converted to grayscale prior to feature extraction. Rotation-averaged gray-level co-occurrence matrix (GLCM) statistics and local binary pattern (LBP) histograms were computed on these grayscale representations to capture spatial regularity and micro-pattern structure.
This processing choice isolates structural texture information from color cues in the feature space, ensuring that texture descriptors reflect surface pattern characteristics rather than chromatic variation.

2.1.3. Preparation of Interior-Scene Stimuli

Bedroom scenes were rendered with a fixed camera, neutral D65 illumination, and a standardized layout [15]. Regions of interest on wardrobe façades and background were pre-defined to map one-to-one to later eye-tracking AOIs [16,17]. Images were gamma-corrected, and background luminance was constrained to a narrow band to limit global-contrast confounds [18,19]. Elemental colors/textures were then embedded into the same ROIs with identical transforms, ensuring consistent interpretation between isolated items and their scene instantiations.
In both the element and scene conditions, AOIs were restricted to the wardrobe-front façade surfaces (door panels) and did not include adjacent walls, floor, ceiling, or decorative objects. Because the 3D model, camera pose, and wardrobe geometry were fixed, the AOI boundaries were identical across all stimuli within each condition; only the mapped color/texture of the wardrobe-front AOI changed. A separate background AOI (all non-wardrobe regions) was retained only for exposure normalization (e.g., fixation-share) and for data-quality checks.

2.2. Experimental

This section details participants, apparatus, and the unified procedure for the color–texture element task and the scene task. Unless noted, settings were identical across tasks.

2.2.1. Participants

Forty adults (20 female, 20 male; 20–30 years) took part. All reported normal or corrected-to-normal vision; myopic participants wore their habitual glasses. Individuals with diagnosed color-vision deficiency, clinically significant astigmatism, or large inter-ocular differences were excluded [20]. Written informed consent was obtained. Two trained researchers administered the sessions, one responsible for recording data and the other for operating the equipment. The study was conducted in accordance with institutional ethical guidelines and approved by the relevant ethics committee.

2.2.2. Experimental Equipment

The eye-tracking experiment utilized a Tobii Pro Fusion remote eye tracker, with a sampling rate of 250 Hz and a head movement range of 30 cm × 25 cm. Stimuli were presented on a Dell P2422H LCD monitor (23.8 inch, 1920 × 1080 resolution, 60 Hz refresh rate). Color management was handled at the operating-system level using the default sRGB ICC profile. The monitor was not hardware-calibrated with a colorimeter; however, all experiments were conducted on the same display with fixed brightness and contrast settings. The display was set to an sRGB color mode with a nominal D65 white point, gamma 2.2, and a target luminance verified at approximately 120 cd/m2 using the built-in system calibration utility. These settings were held constant across all experimental sessions to minimize presentation variability. An MSI laptop was used for stimulus preparation, questionnaire presentation, and data export/inspection. Calibration followed the manufacturer’s multi-point routine and was repeated if accuracy fell below acceptable bounds [21].
Eye-movement events were parsed using a velocity-threshold identification method (I-VT) as implemented in the ErgoLAB–Tobii processing pipeline. A saccade velocity threshold of 30°/s was applied, and fixations shorter than 60 ms were excluded. All parameters were held constant across experimental conditions to ensure the comparability of fixation-based metrics, including time to first fixation, fixation count, mean fixation duration, scanpath length, and fixation distribution ratio.
Raw gaze samples marked as invalid by the Tobii SDK were removed prior to event detection. No participant-specific parameter tuning was performed.

2.2.3. Experimental Procedure and Task

Participants completed two forced-choice tasks in a counterbalanced order:
  • Material-element task: On each trial, 6 swatches were presented. Participants indicated the most preferred option using a keypad or mouse;
  • Bedroom-scene task: On each trial, 6 wardrobe scenes were presented. Participants indicated the most preferred configuration.
Each trial followed the same timing logic: central fixation 500 ms, stimulus display self-paced with a maximum of 6 s, response, and an inter-trial interval of 500–800 ms [22]. Before the formal session, participants completed 10–12 practice trials; successful calibration results are shown in Figure 2. To enable later fusion of attention and affect, brief PAD ratings were collected on nine-point scales using standardized SAM visuals [23]. The order of tasks and the spatial arrangement of alternatives were counterbalanced across participants to reduce position and sequence biases. The detailed experimental flow is illustrated in Figure 3.
The stimulus display was self-paced (maximum 6 s) to reduce time pressure and allow participants to complete multi-alternative comparisons. We note that self-paced exposure can couple viewing time with preference; therefore, fixation-duration-related predictors are interpreted as correlates of preference under this task design rather than causal determinants.

2.3. Data Acquisition and Analysis

The experiment comprised two choice tasks: a Material-element task (six isolated alternatives per trial) and a Bedroom-scene task (six rendered bedroom scenes per trial). For gaze-feature extraction, we additionally employed three single-element viewing conditions:
  • Color-Single, presenting isolated color swatches;
  • Color-Texture-Single, presenting the original texture samples;
  • DeColor-Texture-Single, presenting the same texture samples after removing chromatic information to emphasize structural texture cues.
These single-element conditions were used to derive participant-level attention profiles and to construct interpretable gaze features for prediction.
All experimental records were consolidated into a long-format dataset indexed by participant, stimulus and task condition. Eye-tracking data from the three single-element tasks (Color-Single, DeColor-Texture-Single and Color-Texture-Single) were processed to extract five standardized metrics: fixation count (FC), time to first fixation (TTFF), mean fixation duration (MFD), saccade count (SC) and fixation-duration ratio (FDR) [24,25]. Time to first fixation was inverted so that higher values indicated faster orienting. Trials with missing or implausible gaze data were removed, and repeated presentations were averaged. Scene-choice data were encoded as binary labels for each trial. All features were z-scored using training-fold statistics to avoid information leakage [26].

2.3.1. Clustering of Visual-Attention Patterns

To characterize individual viewing strategies, each participant was represented by a vector containing the five gaze features concatenated across the three single-element tasks. These vectors were standardized and grouped using k-means clustering.
Cluster quality was evaluated using the Silhouette coefficient:
s ( i ) = b ( i ) a ( i ) max { a ( i ) , b ( i ) } ,
where a ( i ) is the mean intra-cluster distance for sample i and b ( i ) is the minimum mean inter-cluster distance, and the Calinski–Harabasz index:
CH ( k ) = tr ( B k ) / ( k 1 ) tr ( W k ) / ( n k ) ,
where B k and W k denote between- and within-cluster dispersion matrices. Cluster membership was not used as a covariate in the main predictive models; instead, it served to ensure that model performance and feature-importance patterns remained qualitatively similar across attention-profile subgroups. Clustering was used as a descriptive check of heterogeneity in viewing strategies rather than as a primary explanatory variable. We report cluster sizes and verify that key modeling results are not driven by any single cluster through sensitivity analyses.
The clustering procedure identified three attention-profile clusters. Cluster sizes were n = 14, n = 13, and n = 13, respectively. The solution achieved a mean Silhouette coefficient of 0.41 and a Calinski–Harabasz index of 162.7, indicating moderate but stable separation among viewing-strategy groups.
Sensitivity analyses confirmed that removing any single cluster did not materially change the direction or magnitude of the main predictive and ranking results.

2.3.2. Preference Prediction Modeling

To assess whether gaze metrics predict material choice, each candidate in a trial was assigned a binary label ( y = 1 if chosen). A penalized logistic regression model was fitted using standardized FC, inverse TTFF, MFD, SC, FDR and, when available, relative fixation share. The choice probability was modeled as:
P ( y = 1 | x ) = σ ( β 0 + β T x ) , σ ( t ) = 1 1 + e t ,
with parameters estimated by maximizing the l 2 -regularized log-likelihood [27]
L ( β ) λ | | β | | 2 2 .
Generalization was evaluated using leave-one-subject-out cross-validation. Model performance was summarized using AUC, accuracy, F1 score and top-1 hit rate [28]. A non-linear reference classifier was also fitted to assess robustness, and feature-importance patterns were compared qualitatively [29].

2.3.3. Bradley–Terry Pairwise Ranking

To obtain calibrated preference strengths, trial-level choices were transformed into winner–loser pairs and analyzed using the Bradley–Terry model [30]. Let θ i denote the latent utility of stimulus i . The Bradley–Terry probability that stimulus i is preferred over stimulus j is given by
P ( i j ) = exp ( θ i ) exp ( θ i ) + exp ( θ j ) .
Utilities are identifiable up to an additive constant; for identifiability, we impose the constraint i θ i = 0 .
Each six-alternative trial was converted into winner–loser comparisons by pairing the selected option (winner) with each of the remaining five options (losers), yielding five comparisons per trial. Because these derived comparisons are not statistically independent within a trial, Bradley–Terry estimates are interpreted as aggregate preference strengths resulting from a multinomial-to-pairwise reduction. Uncertainty was therefore quantified using bootstrap resampling at the trial level.
Because each participant contributed multiple choices, we additionally conducted a robustness check using a mixed-effects Bradley–Terry specification with participant-level random intercepts. The resulting rank orders and top-tier materials were unchanged relative to the standard BT fit, indicating that clustering by observer did not materially affect the conclusions.
All robustness analyses were conducted in Python 3.13. Standard Bradley–Terry models were estimated using the choix package, and within-participant dependence was examined via mixed-effects logistic regression with participant-level random intercepts implemented in statsmodels. Rank-order correlations between the standard and mixed-effects models exceeded ρ = 0.95, and the top-ranked materials were identical across model specifications, indicating that participant-level clustering did not materially affect the preference conclusions.
Parameters were estimated using a majorization–minimization algorithm with a log-scale identifiability constraint. Bootstrap resampling provided percentile confidence intervals for θ i . Preference strengths were primarily estimated and reported separately for each task condition. Where a summary across conditions was reported, it was computed only for the common stimulus set shared across the corresponding tasks and was treated as a descriptive index rather than a standalone inferential estimate.

2.3.4. Integrated Attention Index

To produce an interpretable material score, feature weights were derived from the logistic model coefficients and normalized as
w j = | β j | m | β m |
For each material i the Integrated Attention Index (IAI) was computed as
IAI i = j w j z i j
where z i j is the standardized value of feature j [31,32].
The index provides a continuous ranking reflecting how perceptual behavior contributes to preference. When data coverage was insufficient for stable estimation, equal weights were used. Agreement with Bradley–Terry strengths and questionnaire results was used to assess consistency.

2.3.5. Linking Attention and Emotion

Eye-tracking features were aligned with Pleasure, Arousal and Dominance ratings for each participant–stimulus pair. Canonical correlation analysis quantified the shared structure between gaze variables X and PAD scores Y by maximizing
ρ = max a , b corr ( a T X , b T Y )
revealing how attentional patterns co-vary with emotional appraisals [33].
Analyses were performed separately for color and texture datasets at the participant × stimulus level (color: n = 280 observations; texture: n = 240 observations). X comprised five standardized gaze variables (FC, TTFF−1, MFD, SC, FDR), and Y comprised three affective variables (Pleasure, Arousal, Dominance). Canonical correlations were tested using Wilks’ Lambda. The first canonical correlation was statistically significant in both datasets (p < 0.001), and subsequent dimensions were not interpreted. The path model was estimated as a standardized SEM (maximum likelihood with robust standard errors) with 5000 bootstrap resamples to obtain confidence intervals for key paths; all reported β values are standardized coefficients. Models were fitted in R (lavaan), and fit indices (CFI/TLI/RMSEA/SRMR) are reported in the Results.
For interpretability, three composite constructs—Attraction, Retention and Exploration—were created from standardized gaze features and entered into a sequential path model. In this model, the three constructs predicted Pleasure, Pleasure predicted Arousal and Arousal predicted Desire. Standardized path coefficients and bootstrap confidence intervals were used to evaluate direct and indirect effects. This integration provides a mechanistic bridge between visual behavior, affective response and material preference.
In this study, Desire is treated as an approach-oriented preference outcome reflecting willingness to select or adopt a material option. Desire is not a substitute for Dominance. PAD dimensions were collected and analyzed in the CCA stage; the subsequent path model focuses on a parsimonious Pleasure–Arousal–Desire sequence to connect affective appraisal to preference intention under our experimental setting.

3. Results

3.1. Stimulus Set

Materials were supplied by Shanghai Muli Industrial Development Co., Ltd. (Shanghai, China). Color swatches were transformed to CIE L*a*b* and clustered with a density-adaptive method; seven perceptually stable groups emerged and were represented by k-medoids, labeled by dominant hue as: Black, Grey, White, Wax white, Yellow, Red, and Green (Figure 4). Texture patches, summarized by rotation-averaged GLCM and LBP descriptors, formed six separable classes—Line Finish, Leather Finish, Mountain Grain, Straight Grain, Special Grain, and Horizontal Grain—with class medoids used as exemplars (Figure 5).
A standard linear wardrobe layout was adopted for presentation. A three-dimensional bedroom model was built in 3 ds Max. To limit extraneous influences, illumination was fixed to a uniformly bright level (≈90 ± 5 cd/m2), and decorative objects were omitted. All room elements other than the wardrobe were held constant. Wardrobe fronts were manipulated in two stimulus sets: a color set (levels as in Figure 4) and a texture set (categories as in Figure 5). Example stimuli are shown in Figure 6 and Figure 7. Because black exhibited very low fixation and choice rates in the single-color condition, it was excluded from the interior-scene color set. As a result, any cross-condition aggregation was limited to the common alternatives shared by the corresponding tasks to preserve comparability.

3.2. Eye-Tracking Preprocessing and Subject Stratification

Raw gaze streams were reduced to a long table keyed by subject–stimulus–condition with five metrics per item: fixation count (FC), time to first fixation (TTFF), mean fixation duration (MFD), saccade count (SC), and fixation-duration ratio (FDR). Latency was inverted to emphasize faster orienting. All variables were z-standardized; where relevant, we normalized exposure by the within-set fixation share. Trials with missing AOIs, zero dwell across items, or implausible latencies were excluded; repeated presentations were averaged.
For downstream analyses, we focused on single-element conditions (Color-Single; DeColor-Texture-Single; Color-Texture-Single). Each subject’s standardized features from these three conditions were concatenated to form an attention profile. Subjects were then stratified via k-means; the number of clusters was chosen by maximizing the mean of Silhouette and Calinski–Harabasz indices. Between-cluster differences in FC, TIFFinv, MFD, SC, and FDR were screened using Mann–Whitney tests with Holm correction. This yields a clean, comparable feature set and a subject stratification used in the subsequent prediction, ranking, weighting, and PAD-fusion analyses.

3.3. Preference Prediction Performance

Choices were modeled as binary outcomes from eye-tracking features aggregated at the subject–stimulus level. Six z-scored predictors were used: fixation count (zFC), inverse time-to-first-fixation (zTIFFinv), mean fixation duration (zMFD), saccade count (zSC), fixation-duration ratio (zFDR), and within-condition fixation share (zFR). Negatives were completed within each subject’s candidate set. Evaluation used leave-one-subject-out cross-validation. On each fold we trained L2-regularized logistic regression and a 200-tree random forest. We report AUC, accuracy, F1, Top-1 hit rate, standardized coefficients, and feature importances.
The results reveal a consistent landscape of gaze features across task conditions, as illustrated by the feature-mean heatmap (Figure 8). In the Color-Single condition, participants exhibited more active visual search, characterized by higher standardized fixation counts and saccade frequencies, together with shorter mean fixation durations. This pattern suggests rapid scanning behavior when evaluating isolated colors. In contrast, both Color-Texture-Single and DeColor-Texture-Single showed feature values close to zero on most axes, with modest increases in fixation-duration ratio, indicating more concentrated dwell within relevant regions during isolated texture appraisal. The Texture-Scene condition displayed a distinct profile marked by substantially longer mean fixation durations and markedly lower fixation-duration ratios, along with slightly reduced saccade counts. These patterns imply slower, more deliberate appraisal in the richer bedroom context, accompanied by more diffuse allocation of attention across the scene. Time to first fixation showed minimal variation across conditions, suggesting that the initial orienting response remained relatively stable despite differences in task structure and stimulus complexity.

3.4. Pairwise Preference Ranking

Bradley–Terry analyses of all pairwise choices yield clear, stable orders. Across isolated swatches and bedroom scenes, the rank order is White > Ivory white > Gray, with these achromatic options sitting above the normalized mean (geometric mean = 1) and showing limited overlap in 95% CIs. In scenes, strengths contract toward 1, indicating that context attenuates color contrasts while preserving the ordering (Figure 9a).
Mountain Grain ranks highest, followed by Straight Grain and Line Finish; this leading tier is separated from Leather Grain, Special Grain, and Horizontal Grain, which cluster around the mean with broader, overlapping CIs. Scene textures again compress toward the mean, mirroring the color pattern (Figure 9b).
Single-question checks reproduce the same structure: isolated color keeps White > Ivory white > Gray with wide separation, and isolated textures keep Mountain Grain > Straight Grain > Line Finish as the top tier; scene versions narrow the gaps. Taken together, achromatic colors and orderly/fine textures have the highest win probabilities, whereas contextual scenes systematically homogenize preferences.

3.5. Integrating Gaze Features for Preference Scoring

To explain BT ranks and project them to new designs, we quantified each eye-movement feature’s contribution to scene choices and combined them into an Integrated Attention Index (IAI). For color, mean fixation duration (MFD) dominates (44.0%), followed by inverse time-to-first-fixation (TIFFinv, 35.5%); saccade count (10.3%), fixation-duration ratio (6.9%), and fixation count (3.3%) add modest signal. For texture, MFD is even more dominant (62.6%), with TIFFinv (17.0%), SC (9.1%), FDR (9.4%), and FC (1.8%) contributing smaller increments. Thus, sustained viewing and rapid first looks carry most predictive power; simple counts contribute little.
Applying these weights to single-item sets yields IAI rankings that support and refine BT outcomes. For color, neutral lights lead—White first, then Ivory White and Gray—while saturated warms trail (Table 1). For texture, Mountain Grain, Straight Grain, and Line Finish score highest, with Leather Grain next and Special Grain lower (Table 2). Minor re-orderings versus BT align with the dominant role of MFD: items eliciting longer fixations and faster first looks gain IAI advantage when BT strengths are close.
The two analyses align well: the Bradley–Terry model establishes the preference hierarchy, and the Integrated Attention Index explains it through gaze behavior while enabling predictions for new materials. Options that attract faster initial fixations and sustain longer viewing receive higher IAI scores and are more likely to align with the top-performing neutral light colors and fine-grained textures. Conversely, materials that disperse attention or delay the first look tend to score lower and rank less favorably.

3.6. Linking Gaze to Affect

Eye-tracking metrics were standardized and organized into three attention constructs: Attraction, represented by inverse time to first fixation and fixation count; Retention, represented by mean fixation duration and the fixation-duration ratio; and Exploration, represented by saccade count. These constructs were first related to affective responses using canonical correlation analysis, and the resulting associations were then examined through a structural path model reflecting a sequential Pleasure–Arousal–Desire process.
Canonical correlation analysis revealed a significant association between gaze behavior and affective responses. The first canonical correlation was r = 0.62 for the color dataset and r = 0.58 for the texture dataset. Wilks’ Lambda tests indicated that the first canonical dimension was statistically significant in both cases (p < 0.001), whereas subsequent dimensions were not significant and were therefore not interpreted.
The proposed path model demonstrated an acceptable to good fit to the data. For the color dataset, model fit indices were CFI = 0.96, TLI = 0.94, RMSEA = 0.045, and SRMR = 0.038. Similar fit was observed for the texture dataset.
Standardized path coefficients from attention constructs to Pleasure and Desire were all statistically significant (p < 0.01). Bootstrap-based 95% confidence intervals confirmed the robustness of the estimated effects (Table 3).
For both color and texture, the first canonical variate loaded positively on TTFF−1, fixation count and FDR, and negatively on mean fixation duration. Thus, stimuli that are found quickly and receive dense early fixations align with higher PAD states, whereas prolonged viewing reflects a distinct retention process (Figure 10).
For color, attraction strongly predicts Pleasure (β ≈ 0.9), while Retention and Exploration are near zero. Pleasure in turn raises Arousal (β ≈ 1.1), and Arousal predicts Desire (β ≈ 0.7). Hence early attentional capture dominates color preference, with effects propagated along the Pleasure–Arousal–Desire pathway (Figure 11a).
For texture, a similar pathway emerged but with more differentiation among constructs. Attraction again strongly predicted Pleasure (β ≈ 1.0); Retention showed a small positive effect (≈0.1); Exploration is negative (≈−0.2 to −0.3). Pleasure again elevates Arousal (β ≈ 1.1), which increases Desire (β ≈ 0.65). Textures that concentrate attention early and suppress scanning obtain higher PAD and are chosen more often (Figure 11b).
These mechanisms help clarify the preference rankings. Neutral light colors such as white, ivory white and gray, along with the most orderly wood-grain textures, achieve higher desirability because they attract early and concentrated visual attention. When these materials are placed within full bedroom scenes, contextual factors reduce the magnitude of differences between options, yet the overall preference order remains unchanged.

3.7. Agreement with the Subjective Questionnaire

The independent questionnaire reproduced the BT hierarchy and IAI projections. For color, the top tier remained White, Ivory white, and Gray in both element and scene contexts; warmer chromatic hues consistently trailed. For texture, Mountain Grain, Straight Grain, and Line Finish formed the leading group, followed by Leather Finish, with Special Grain and Horizontal Grain near the middle-to-lower range. Minor reorderings were confined to mid-ranked items and did not alter tier boundaries. Overall, the survey yielded the same leaders and the same context-driven contraction observed in BT and IAI, supporting the external validity of the gaze–affect pipeline.

4. Discussion

4.1. Stable Preference Structures Across Interior Contexts

Across all analyses, wardrobe-front material preferences exhibited a clear and stable hierarchical structure that persisted from isolated swatches to full bedroom scenes. Neutral light colors—particularly white, ivory white, and gray—consistently occupied the top tier, while saturated hues and darker tones ranked lower. Similarly, structured wood-like textures such as mountain grain, straight grain, and line finish were preferred over more irregular or visually fragmented patterns.
When materials were embedded within realistic bedroom scenes, preference differences became less pronounced, yet the relative ordering remained unchanged. This pattern suggests a contextual compression effect rather than a reversal of preference: surrounding spatial elements, lighting, and furniture reduce contrast between options without altering their fundamental appeal. This finding is particularly relevant for interior practice, as it indicates that material rankings derived from controlled evaluations can remain informative when transferred to more complex spatial contexts.

4.2. Linking Visual Attention to Affective Evaluation and Choice

The eye-tracking results indicate that time to first fixation and mean fixation duration are the most influential predictors of wardrobe-front preference. Materials that attract attention rapidly and sustain longer visual engagement are more likely to be chosen, reflecting the combined importance of early perceptual salience and sustained evaluative processing.
By integrating PAD emotion ratings, the analysis further clarifies how attentional patterns translate into affective responses and ultimately into preference. Materials associated with higher Pleasure and Arousal tended to exhibit more efficient attentional profiles, suggesting a coherent mechanism linking perceptual processing, emotional appraisal, and choice behavior. This finding aligns with prior work in environmental psychology and decision science, while extending it to furniture-specific material evaluation in interior settings.
Importantly, the distinction between attraction, retention, and exploration highlights that not all gaze activity contributes equally to desirability. Rapid orienting and concentrated viewing appear to be more strongly associated with positive affect than extensive scanning, providing a more nuanced understanding of how visual behavior supports preference formation.

4.3. Implications for Bedroom Interior Design and Material Selection

The present findings are consistent with recent eye-tracking studies demonstrating that façade and interior surface attributes shape visual comfort and dwell patterns. At the same time, this study extends those insights by focusing on a decision-critical furniture component—wardrobe fronts—within controlled bedroom scenes, thereby bridging architectural-scale research and furniture-scale design decisions.
From a practical perspective, the results suggest clear design heuristics for bedroom interiors targeting young adult users. In compact or multipurpose bedrooms, combining neutral light-colored wardrobe fronts with orderly, fine-grained wood textures is likely to maximize perceived visual comfort and desirability. When materials are evaluated within full scenes, designers should expect preference differences to narrow rather than reverse; consequently, mid-ranked options may be differentiated more effectively through subtle adjustments in texture regularity, luminance, or surface finish rather than through large shifts in hue.
Beyond specific material recommendations, the integrated framework proposed in this study supports early-stage decision making in customized interior design. By combining calibrated pairwise rankings with an interpretable attention-based index, designers can efficiently screen large material sets and focus on options that consistently elicit favorable perceptual and affective responses.

4.4. Limitations and Directions for Future Research

Several limitations should be noted. The study examined a finite set of colors and textures, and the number of participants, although sufficient for controlled eye-tracking research, does not reflect the diversity of occupants encountered in real housing projects. Although grayscale descriptors reduce chromatic confounding in texture modeling, residual color cues in the original samples may still influence preference under real viewing conditions. All evaluations were conducted under laboratory conditions with standardized renderings and luminance, which differ from the varied lighting and furnishings of lived-in bedrooms.
Generalizability is therefore bounded: our participants were young adults (20–30 years), and material preferences can vary with age, culture, household composition, and life stage. Accordingly, the reported ranks and mechanisms should be interpreted as most representative of younger adult users under controlled viewing conditions, and future studies should replicate the pipeline with older and more diverse cohorts.
Because trials were self-paced and followed by repeated affect ratings, fatigue or attentional drift may influence gaze dynamics and judgment consistency. Future studies will therefore adopt fixed-duration presentations and block-wise temporal stability checks to disentangle preference effects from exposure-time endogeneity. In addition, alternative path specifications will be examined that explicitly incorporate the Dominance dimension, in order to test whether perceived control adds explanatory power for Desire across different contexts and populations.
More broadly, future work should expand the range of materials and include more diverse user groups to capture cultural, age-related, and lifestyle-specific differences in perception [34]. Field studies in furnished homes or immersive virtual reality environments may further clarify how dynamic lighting and spatial complexity shape attention and emotion. Beyond static gaze metrics, incorporating temporal transitions between fixations may yield richer indicators of perceptual strategies. Finally, embedding the proposed framework into digital design platforms could help demonstrate how data-driven evaluation supports large-scale customization and improves the consistency of interior material selection.

5. Conclusions

This study demonstrates a reproducible, data-driven approach to understanding how wardrobe front materials shape visual comfort and occupant preference in bedroom interiors. By integrating eye-tracking metrics with affective evaluations and calibrated pairwise modeling, the framework clarifies both the structure of material preferences and the mechanisms that drive them.
Three conclusions stand out. First, preference orders are stable across evaluation formats. Neutral light colors and structured wood-like textures consistently ranked highest, and scene context moderated but did not overturn these patterns. Second, visual attention plays a central role in material desirability. Rapid initial fixation combined with longer sustained viewing emerged as the most influential predictors of choice, linking perceptual efficiency with positive affective appraisal. Third, the integrated attention index provides a practical tool for screening new materials without requiring repeated pairwise tests, offering designers a scalable method for identifying visually comfortable and widely appealing wardrobe front options.
Overall, the findings contribute to human-centred interior design by offering quantitative evidence of how material attributes guide perception and preference. The proposed pipeline strengthens the connection between perceptual research and interior practice, supporting more consistent, evidence-based decision making in the design of bedroom environments.

Author Contributions

Conceptualization, W.X.; methodology, W.X.; software, W.X.; validation, X.M.; formal analysis, W.X.; investigation, W.X.; resources, X.M.; data curation, W.X.; writing—original draft preparation, W.X.; writing—review and editing, Y.C. and X.M.; visualization, W.X.; supervision, Y.C. and X.M.; project administration, Y.C.; funding acquisition, Y.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

This study protocol was approved by the Institutional Ethics Committee of Nanjing Forestry University (Approval No. 2025-D-017, Approval Date: 15 January 2025).

Informed Consent Statement

Informed consent was obtained from all subjects involved in the study.

Data Availability Statement

Data are contained within the article.

Conflicts of Interest

The authors declare no conflicts of interest.

References

  1. Aktan Ábrahám, S.; Deniz, D. Rethinking the Interior Design Process: Advancing Circularity through Evidence-Based Design. In Proceedings of the EKSIG 2025: DATA AS EXPERIENTIAL KNOWLEDGE AND EMBODIED PROCESSES; Design Research Society: Budapest, Hungary, 2025. [Google Scholar]
  2. Orquin, J.L.; Mueller Loose, S. Attention and Choice: A Review on Eye Movements in Decision Making. Acta Psychol. 2013, 144, 190–206. [Google Scholar] [CrossRef] [PubMed]
  3. Shimojo, S.; Simion, C.; Shimojo, E.; Scheier, C. Gaze Bias Both Reflects and Influences Preference. Nat. Neurosci. 2003, 6, 1317–1322. [Google Scholar] [CrossRef]
  4. Pan, S. Emotional Analysis of Broadcasting and Hosting Speech by Integrating Grid PSO-SVR and PAD Models. Int. J. Cogn. Comput. Eng. 2025, 6, 55–64. [Google Scholar] [CrossRef]
  5. Schloss, K.B.; Palmer, S.E. An Ecological Valence Theory of Human Color Preferences. J. Vis. 2010, 9, 358. [Google Scholar] [CrossRef]
  6. Jonauskaite, D.; Camenzind, L.; Parraga, C.A.; Diouf, C.N.; Mercapide Ducommun, M.; Müller, L.; Norberg, M.; Mohr, C. Colour-Emotion Associations in Individuals with Red-Green Colour Blindness. PeerJ 2021, 9, e11180. [Google Scholar] [CrossRef]
  7. Palmer, S.E.; Schloss, K.B. Color Preference. In Encyclopedia of Color Science and Technology; Luo, R., Ed.; Springer: Berlin/Heidelberg, Germany, 2015; pp. 1–7. ISBN 978-3-642-27851-8. [Google Scholar]
  8. Banaei, M.; Ahmadi, A.; Gramann, K.; Hatami, J. Emotional Evaluation of Architectural Interior Forms Based on Personality Differences Using Virtual Reality. Front. Archit. Res. 2020, 9, 138–147. [Google Scholar] [CrossRef]
  9. Mehrabian, A. Pleasure-Arousal-Dominance: A General Framework for Describing and Measuring Individual Differences in Temperament. Curr. Psychol. 1996, 14, 261–292. [Google Scholar] [CrossRef]
  10. Hall, M.; Elliott, K.; Meng, J. Using the PAD (Pleasure, Arousal, and Dominance) Model to Explain Facebook Attitudes and Use Intentions. J. Soc. Media Soc. 2017, 6, 144–169. [Google Scholar]
  11. Zhao, Y.; Xie, D.; Zhou, R.; Wang, N.; Yang, B. Evaluating Users’ Emotional Experience in Mobile Libraries: An Emotional Model Based on the Pleasure-Arousal-Dominance Emotion Model and the Five Factor Model. Front. Psychol. 2022, 13, 942198. [Google Scholar] [CrossRef]
  12. Cebrián, P.; Pérez-Sienes, L.; Sanz-Vicente, I.; López-Molinero, Á.; de Marcos, S.; Galbán, J. Solving Color Reproducibility between Digital Devices: A Robust Approach of Smartphones Color Management for Chemical (Bio)Sensors. Biosensors 2022, 12, 341. [Google Scholar] [CrossRef]
  13. Ikotun, A.M.; Habyarimana, F.; Ezugwu, A.E. Benchmarking Validity Indices for Evolutionary K-Means Clustering Performance. Sci. Rep. 2025, 15, 21842. [Google Scholar] [CrossRef]
  14. Tejada-Casado, M.; Herrera, L.J.; Carrillo-Perez, F.; Ruiz-López, J.; Ghinea, R.I.; Pérez, M.M. Exploring the CIEDE2000 Thresholds for Lightness, Chroma, and Hue Differences in Dentistry. J. Dent. 2024, 150, 105327. [Google Scholar] [CrossRef]
  15. ISO/CIE 11664-2:2022; Colorimetry—Part 2: CIE Standard Illuminants. International Organization for Standardization: Geneva, Switzerland, 2022.
  16. Huang, D.; Gong, W.; Wang, X.; Liu, S.; Zhang, J.; Li, Y. A Cognition–Affect–Behavior Framework for Assessing Street Space Quality in Historic Cultural Districts and Its Impact on Tourist Experience. Buildings 2025, 15, 2739. [Google Scholar] [CrossRef]
  17. Wang, Z.; Shen, M.; Huang, Y. Exploring the Impact of Facade Color Elements on Visual Comfort in Old Residential Buildings in Shanghai: Insights from Eye-Tracking Technology. Buildings 2024, 14, 1758. [Google Scholar] [CrossRef]
  18. Shi, W.; Zhou, M.; Ono, K. Cognitive Style and Visual Attention in Multimodal Museum Exhibitions: An Eye-Tracking Study on Visitor Experience. Buildings 2025, 15, 2968. [Google Scholar] [CrossRef]
  19. Wlazły, W.; Bonenberg, A. Modulating Perception in Interior Architecture through Décor: An Eye-Tracking Study of a Living Room Scene. Buildings 2025, 15, 48. [Google Scholar] [CrossRef]
  20. Fanlo-Zarazaga, A.; Echevarría, J.I.; Pinilla, J.; Alejandre, A.; Pérez-Roche, T.; Gutiérrez, D.; Ortín, M.; Pueyo, V. Validation of a New Digital and Automated Color Perception Test. Diagnostics 2024, 14, 396. [Google Scholar] [CrossRef] [PubMed]
  21. Wang, K.; Hou, W.; Ma, H.; Hong, L. Eye-Tracking Characteristics: Unveiling Trust Calibration States in Automated Supervisory Control Tasks. Sensors 2024, 24, 7946. [Google Scholar] [CrossRef] [PubMed]
  22. Echeverria-Altuna, I.; Nobre, A.C.; Boettcher, S.E.P. Goal-Dependent Use of Temporal Regularities to Orient Attention under Spatial and Action Uncertainty. J. Cogn. 2024, 7, 37. [Google Scholar] [CrossRef] [PubMed]
  23. Sainz-de-Baranda Andujar, C.; Gutiérrez-Martín, L.; Miranda-Calero, J.Á.; Blanco-Ruiz, M.; López-Ongil, C. Gender Biases in the Training Methods of Affective Computing: Redesign and Validation of the Self-Assessment Manikin in Measuring Emotions via Audiovisual Clips. Front. Psychol. 2022, 13, 955530. [Google Scholar] [CrossRef]
  24. Guo, R.; Kim, N.; Lee, J. Empirical Insights into Eye-Tracking for Design Evaluation: Applications in Visual Communication and New Media Design. Behav. Sci. 2024, 14, 1231. [Google Scholar] [CrossRef]
  25. Hu, H.; Li, H.; Wang, B.; Zhang, M.; Wu, B.; Wu, X. Application of Eye-tracking in Nursing Research: A Scoping Review. Nurs. Open 2024, 11, e2108. [Google Scholar] [CrossRef]
  26. Lones, M.A. Avoiding Common Machine Learning Pitfalls. Patterns 2024, 5, 101046. [Google Scholar] [CrossRef]
  27. Fridgeirsson, E.A.; Williams, R.; Rijnbeek, P.; Suchard, M.A.; Reps, J.M. Comparing Penalization Methods for Linear Models on Large Observational Health Data. J. Am. Med. Inform. Assoc. 2024, 31, 1514–1521. [Google Scholar] [CrossRef]
  28. Hijazi, H.; Gomes, M.; Castelhano, J.; Castelo-Branco, M.; Praça, I.; de Carvalho, P.; Madeira, H. Dynamically Predicting Comprehension Difficulties through Physiological Data and Intelligent Wearables. Sci. Rep. 2024, 14, 13678. [Google Scholar] [CrossRef] [PubMed]
  29. Hermens, F.; Krucien, N.; Ryan, M. The Use of Machine Learning to Understand the Role of Visual Attention in Multi-Attribute Choice. Acta Psychol. 2024, 251, 104581. [Google Scholar] [CrossRef] [PubMed]
  30. Seymour, R.G.; Sirl, D.; Preston, S.P.; Dryden, I.L.; Ellis, M.J.A.; Perrat, B.; Goulding, J. The Bayesian Spatial Bradley–Terry Model: Urban Deprivation Modelling in Tanzania. J. R. Stat. Soc. Ser. C Appl. Stat. 2022, 71, 288–308. [Google Scholar] [CrossRef]
  31. Gao, Y.; Zhang, Z.; Yu, W. Preference Learning for Multi-Criteria Sorting with Interacting Criteria: A Framework Integrating Threshold-Based Value-Driven Sorting Procedure with Attention Network. Inf. Fusion 2026, 125, 103443. [Google Scholar] [CrossRef]
  32. Li, Y.; Tong, Z.; Tong, S.; Westerdahl, D. A Data-Driven Interval Forecasting Model for Building Energy Prediction Using Attention-Based LSTM and Fuzzy Information Granulation. Sustain. Cities Soc. 2022, 76, 103481. [Google Scholar] [CrossRef]
  33. Cruz-Cano, R.; Cohen, A.; Mead-Morse, E. Canonical Correlation Analysis of Survey Data: The SurveyCC R Package. R J. 2025, 16, 141–157. [Google Scholar] [CrossRef]
  34. Berquet, S.; Aleem, H.; Grzywacz, N.M. A Fisher Information Theory of Aesthetic Preference for Complexity. Entropy 2024, 26, 901. [Google Scholar] [CrossRef] [PubMed]
Figure 1. Research methods framework.
Figure 1. Research methods framework.
Mti 10 00007 g001
Figure 2. Calibration results showing fixation accuracy and precision at five points.
Figure 2. Calibration results showing fixation accuracy and precision at five points.
Mti 10 00007 g002
Figure 3. Experimental procedure.
Figure 3. Experimental procedure.
Mti 10 00007 g003
Figure 4. Representative sample colors.
Figure 4. Representative sample colors.
Mti 10 00007 g004
Figure 5. Representative chromatic and achromatic texture samples. (a) Representative chromatic texture samples; (b) Representative achromatic texture samples.
Figure 5. Representative chromatic and achromatic texture samples. (a) Representative chromatic texture samples; (b) Representative achromatic texture samples.
Mti 10 00007 g005
Figure 6. Color group of interior scene stimulus materials. ① White; ② Wax white; ③ Yellow; ④ Grey; ⑤ Red; ⑥ Green.
Figure 6. Color group of interior scene stimulus materials. ① White; ② Wax white; ③ Yellow; ④ Grey; ⑤ Red; ⑥ Green.
Mti 10 00007 g006
Figure 7. Texture group of interior scene stimulus materials. ① Line Finish; ② Leather Finish; ③ Mountain Grain; ④ Straight Grain; ⑤ Special Grain; ⑥ Horizontal Grain.
Figure 7. Texture group of interior scene stimulus materials. ① Line Finish; ② Leather Finish; ③ Mountain Grain; ④ Straight Grain; ⑤ Special Grain; ⑥ Horizontal Grain.
Mti 10 00007 g007
Figure 8. Feature-mean heatmap.
Figure 8. Feature-mean heatmap.
Mti 10 00007 g008
Figure 9. Bradley–Terry model results. (a) Color preferences; (b) Texture preferences. Dots indicate estimated Bradley–Terry strengths, and horizontal solid lines represent 95% confidence intervals.
Figure 9. Bradley–Terry model results. (a) Color preferences; (b) Texture preferences. Dots indicate estimated Bradley–Terry strengths, and horizontal solid lines represent 95% confidence intervals.
Mti 10 00007 g009
Figure 10. CCA biplots. (a) Color; (b) Texture.
Figure 10. CCA biplots. (a) Color; (b) Texture.
Mti 10 00007 g010
Figure 11. SEM Path Diagram. (a) Color; (b) Texture. Arrows indicate hypothesized directional relationships between constructs, and solid lines represent estimated structural paths with standardized coefficients.
Figure 11. SEM Path Diagram. (a) Color; (b) Texture. Arrows indicate hypothesized directional relationships between constructs, and solid lines represent estimated structural paths with standardized coefficients.
Mti 10 00007 g011
Table 1. IAI rankings for color stimuli.
Table 1. IAI rankings for color stimuli.
RankStimulusIAI
1White1.349
2Ivory White1.339
3Gray0.010
4Red−0.059
5Yellow−0.073
6Green−0.144
7Black−0.183
Table 2. IAI rankings for texture stimuli.
Table 2. IAI rankings for texture stimuli.
RankStimulusIAI
1Mountain Grain1.180
2Straight Grain1.170
3Line Finish0.170
4Leather Finish0.078
5Special Grain−0.013
6Horizontal Grain−0.238
Table 3. Canonical correlation and path model summary.
Table 3. Canonical correlation and path model summary.
AnalysisStatisticColorTexture
CCAFirst canonical r0.620.58
Wilks’ Lambda (p)<0.001<0.001
SEMCFI0.960.95
TLI0.940.93
RMSEA0.0450.048
RMR0.0380.041
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Chen, Y.; Xu, W.; Ma, X. Eye-Tracking and Emotion-Based Evaluation of Wardrobe Front Colors and Textures in Bedroom Interiors. Multimodal Technol. Interact. 2026, 10, 7. https://doi.org/10.3390/mti10010007

AMA Style

Chen Y, Xu W, Ma X. Eye-Tracking and Emotion-Based Evaluation of Wardrobe Front Colors and Textures in Bedroom Interiors. Multimodal Technologies and Interaction. 2026; 10(1):7. https://doi.org/10.3390/mti10010007

Chicago/Turabian Style

Chen, Yushu, Wangyu Xu, and Xinyu Ma. 2026. "Eye-Tracking and Emotion-Based Evaluation of Wardrobe Front Colors and Textures in Bedroom Interiors" Multimodal Technologies and Interaction 10, no. 1: 7. https://doi.org/10.3390/mti10010007

APA Style

Chen, Y., Xu, W., & Ma, X. (2026). Eye-Tracking and Emotion-Based Evaluation of Wardrobe Front Colors and Textures in Bedroom Interiors. Multimodal Technologies and Interaction, 10(1), 7. https://doi.org/10.3390/mti10010007

Article Metrics

Back to TopTop