Next Article in Journal
Research on the Quantitative Relationship Between Positioning Error and Coherent Synthesis Success Rate in a Moving Platform Distributed Coherent Synthesis System
Previous Article in Journal
Multi-Scale Adaptive Modulation Network for Efficient Image Super-Resolution
Previous Article in Special Issue
A Hybrid Frequency Decomposition–CNN–Transformer Model for Predicting Dynamic Cryptocurrency Correlations
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

From Unstructured Feedback to Structured Insight: An LLM-Driven Approach to Value Proposition Modeling

Department of Management of Technology, Gyeongsang National University, Jinju-daero 501, Jinju 52828, Republic of Korea
*
Author to whom correspondence should be addressed.
Electronics 2025, 14(22), 4407; https://doi.org/10.3390/electronics14224407 (registering DOI)
Submission received: 21 October 2025 / Revised: 10 November 2025 / Accepted: 10 November 2025 / Published: 12 November 2025

Abstract

Online customer reviews contain rich signals about product value but are difficult to convert into strategy-ready evidence. This study proposes an end-to-end framework that maps review text to the Value Proposition Canvas (VPC) and quantifies alignment between user needs and product performance. Using customer reviews for three Samsung Galaxy Watch generations, an LLM extracts six dimensions (Customer Jobs, Pains, Gains, Feature Gaps, Emotions, Usage Context). Extracted phrases are embedded with a transformer model, clustered via K-means with data-driven k selection, and labeled by an LLM to form an interpretable taxonomy. Subsequently, the analysis derives frequency profiles, a gap density indicator, a context–gap matrix, and a composite Product–Market Fit (PMF) score that balances gain rate, gap rate, and coverage with sensitivity analysis to alternative weights. The findings show predominantly positive affect, with unmet needs concentrated in battery endurance and interaction stability. Productivity- and interaction-centric jobs attain the highest PMF score, while several monitoring-centric jobs are comparatively weaker. Significant cross-generation differences in job composition indicate evolving usage priorities across successive releases. The framework provides a scalable, reproducible path from unstructured VOC to decision support, enabling data-driven prioritization for product and UX management while advancing theory-grounded analysis of customer value.

1. Introduction

In the era of digital commerce, online customer reviews have emerged as a crucial source of getting customer feedback, preferences and latent unmet needs [1,2]. More than mere sentiment labels or star ratings, these reviews can capture nuanced user experiences, encompassing functional, emotional, and contextual dimensions. They also provide valuable insights for product improvement and marketing strategy formulation [3]. Bridging such rich unstructured feedback with actionable product innovation remains a key challenge for both academic researchers and practitioners [4]. A seminal dataset in large-scale user review analysis is the Amazon Reviews dataset curated by the research group of McAuley. The 2023 version of this dataset spans over 571 million reviews across 33 categories and deepens earlier collections in both scale and variety [5]. Earlier versions have already facilitated extensive research in recommendation systems, sentiment analysis, and user behavior modeling [6,7,8]. The availability of such massive public review corpora has enabled scholars to analyze consumer feedback at previously unattainable scales, but translating raw textual data into innovation-relevant insight remains nontrivial [9].
Traditionally, review analysis has focused on techniques such as bag-of-words [10], topic modeling [11], and sentiment analysis [12], sometimes augmented with aspect-based sentiment analysis [13]. These methods help identify frequent themes or polarity distributions but often fall short in capturing the contextual nuance, consumer motivations, or unarticulated needs embedded in narrative text. In their survey, Malik & Bilal (2024) noted that many existing NLP-based review analysis methods rely heavily on keyword co-occurrence, topic modeling, or shallow sentiment classification, which limits their ability to capture deeper contextual nuance and latent user motivations [14]. Such limitations hinder the extraction of in-depth insights for product enhancement and data-driven strategic decisions. Consequently, organizations face delays in capturing timely feedback, leading to missed opportunities for practical product improvements. Moreover, the contextual and narrative characteristics of review data are often insufficiently reflected, resulting in analyses overly dependent on surface-level statistics.
Recent advances in large language models (LLMs) have opened new possibilities for overcoming these challenges. Because LLMs are pretrained on massive corpora and capable of decoding contextual meanings through human-like summarization and classification, they can be used to infer implicit user goals rather than relying purely on keyword co-occurrence. Some recent studies have begun to explore this capacity for extracting customer needs (CNs) from product reviews and user interviews. For instance, Timoshenko et al. (2025) demonstrated that, under appropriate prompt-engineering and fine-tuning conditions, an LLM can match or even outperform expert human analysts in formulating well-structured CNs [15]. Similarly, Wei et al. (2025) proposed an LLM-based framework to extract user priorities and attribute-level insights from user-generated content, demonstrating that prompt-driven models can identify nuanced consumer expectations and latent service attributes with greater interpretability than conventional topic modeling [16]. Another recent model, InsightNet, proposes an end-to-end architecture to extract structured insights from customer feedback by constructing multi-level taxonomies and performing topic-level extraction via a fine-tuned LLM pipeline [17].
Nevertheless, few studies have connected such extraction to formal frameworks of product strategy or value creation. That makes it harder for product managers or R&D teams to operationalize the output. Other approaches require extensive labeled data or human-in-the-loop correction, limiting adaptability to new products or domains [18]. Bridging this gap requires integrating the analytical capabilities of LLMs with established theories of value creation, a direction that holds both academic significance and managerial relevance by connecting consumer insight generation with actionable product development and innovation strategy.
Therefore, this study proposes an integrative method that leverages large-scale online customer reviews through LLM-based analysis and maps the extracted insights onto the Value Proposition Canvas (VPC), a widely adopted framework distinguishing Customer’s Jobs, Pains, and Gains [19]. In this framework, “Jobs” refer to the functional, social, and emotional tasks customers aim to accomplish; “Pains” represent the constraints, inconveniences, and risks encountered during task performance; and “Gains” denote the benefits, outcomes, and values expected from success. The approach automates this process by designing tailored prompts and extraction routines that identify these elements from review text, embed them into a vector space, cluster them into higher-order themes, and link the results to product–market fit evaluation.
This study distinguishes itself from prior research on customer review analysis through three primary contributions. First, it advances beyond conventional sentiment scoring or topic extraction by generating structured, actionable insights—specifically, the Jobs, Pains, and Gains elements—that are directly applicable to product strategy and innovation planning. Second, by leveraging LLM, the proposed framework captures complex categorical structures without requiring extensive labeled data, thereby ensuring scalability and adaptability across diverse product domains. Third, it integrates unstructured review information with VPC, thereby constructing a quantitative–qualitative bridge that advances data-driven customer value modeling and provides a practical decision-support instrument for product managers and R&D teams. Collectively, this integrative approach bridges LLM-based text analytics with customer value theory, contributing to both the theoretical advancement of AI-enabled consumer insight research and the practical domain of evidence-based product innovation.
The remainder of this paper is organized as follows: Section 2 summarizes prior research on customer need identification. Section 3 outlines the proposed methodology and analytical procedures. Section 4 presents empirical validations and case studies. Section 5 discusses the implications of the research findings. Section 6 concludes with a summary of key findings and addresses research limitations and directions for future work.

2. Related Works

Properly understanding the voice of customer (VOC) has long been a critical foundation for innovation and product strategy, as it translates customer experiences and perceptions into actionable knowledge for design and development. However, extracting structured insights from vast, narrative-rich customer feedback has remained challenging due to its unstructured and context-dependent nature [20]. Early research on VOC analysis relied mainly on qualitative methods, such as manual coding and small-sample content analysis, to identify recurring needs and user frustrations [21]. With the rise of large-scale online review platforms and publicly available datasets such as the Amazon Reviews corpus, the field gradually transitioned toward quantitative, data-driven NLP methods capable of handling massive volumes of user-generated texts [22].
This quantitative wave introduced techniques such as bag-of-words (BOW) modeling, topic modeling, and sentiment or aspect-based sentiment analysis, enabling scalable identification of frequent themes and polarity patterns. For example, Zhang et al. (2011), as well as Fang and Zhan (2015), demonstrated how BOW and TF–IDF features could classify review polarity, thereby laying a foundation for automated VOC mining [23,24]. Building on these text-based representations, Aguwa et al. (2012) further quantified the voice of the customer through a customer-satisfaction ratio model that translated textual VOC indicators into measurable performance metrics, representing one of the earliest attempts to link review analytics to managerial decision frameworks [25]. Subsequent studies, such as Schouten and Frasincar (2016) and Qiu et al. (2018), demonstrated how aspect-level sentiment analysis can predict or reconstruct product ratings by aggregating aspect-specific polarity [7,26]. These approaches improved scalability but still relied on surface co-occurrence patterns, offering only limited capacity to capture contextual nuance and latent user motivations [6,8]. In parallel, Yazıcı and Ozansoy Çadırcı (2024) and Barravecchia et al. (2022) compared topic modeling algorithms for extracting actionable marketing insights from large review corpora [11,27]. Building on these, subsequent research has sought to embed review analysis within theory-linked requirement frameworks. For example, automated Kano classification models have been developed to categorize product attributes based on customer reviews, while Quality Function Deployment pipelines translate review-level requirements into engineering or design characteristics [28,29].
As a remedy, Lee et al. (2023) introduced a context-aware customer needs identification framework that leverages linguistic pattern mining to extract functional and contextual signals from online product reviews, improving interpretability over generic topic or sentiment pipelines [4]. Even such context-enriched NLP approaches have yet to be fully integrated with strategic decision frameworks that explicitly link customer expressions to value creation constructs. These studies represent an important methodological step toward bridging textual analytics and product design; however, they largely remain confined to attribute-level mappings and do not capture the broader structure of customer value creation.
Despite the growing sophistication of NLP techniques, no established framework yet exists for automatically inferring the VPC signals (Customer Jobs, Pains, and Gains) from large-scale online reviews. With the rapid progress of deep representation learning and large language models (LLMs), VOC analysis has entered a new methodological phase. Timoshenko et al. (2025) demonstrated that LLMs can match expert analysts in generating structured CN statements when appropriately prompted [15], while Wei et al. (2025) and Praveen et al. (2024) proposed end-to-end pipelines that automatically build multi-level taxonomies of user needs and service attributes [16,30]. Nonetheless, most prior LLM-based studies end at the extraction or taxonomy construction stage and rarely connect their outputs with formal value creation frameworks or quantitative fit metrics relevant to product decision-making. The present study therefore aims to operationalize VPC-consistent signals—Customer Jobs, Pains, and Gains—from unstructured customer reviews and link them to Product–Market Fit diagnostics, providing an end-to-end, theory-grounded, and data-driven bridge from unstructured feedback to actionable strategic insight.

3. Data and Methods

3.1. Data

The dataset used in this study was constructed from publicly available customer reviews posted on Walmart.com, one of the largest retail e-commerce platforms in the United States. The data consist of user-generated textual feedback and associated star ratings on Samsung Galaxy Watch products, collected to examine variations in customer experience and perceived value across different product generations. Specifically, three distinct models (Galaxy Watch 5, Galaxy Watch 6, and Galaxy Watch 7) were selected as representative devices in Samsung’s smartwatch lineup, reflecting incremental advances in design and functionality. Data collection was conducted in August 2025.
For each model, approximately 2000 customer reviews were sampled and downloaded from Walmart’s review repository covering the period from September 2022 to August 2025, ensuring comparability across product generations. Each record contained the following fields: (1) model name (e.g., “Galaxy Watch 5”), (2) review date, (3) customer identifier (anonymized), (4) star rating expressed on a five-point ordinal, (5) review title, and (6) customer review text. More specifically, 1995 customer reviews were collected for Watch 5, 2000 for Watch 6, and 2005 for Watch 7. On average, each review contained approximately 293 characters, corresponding to a mean of 54.9 words per entry.
Prior to analysis, all reviews were subjected to a data cleaning process designed to preserve the linguistic and contextual integrity of the original user expressions. Entries containing missing values or incomplete text fields were excluded. “Review title” and “Customer review text” were subsequently concatenated into a single composite field to serve as the unified textual input for subsequent analysis. Given that LLMs rely heavily on contextual cues for semantic interpretation, no lexical-level preprocessing, such as stop word removal, stemming, or lemmatization, was performed. Only minimal adjustments were applied: unnecessary whitespace, line breaks and formatting artifacts—if they were present—were removed. Moreover, review entries consisting solely of promotional disclaimers were excluded (e.g., [This review was collected as part of a promotion.]). This approach ensured that the textual nuances, sentiment markers, and stylistic features—often critical for inferring consumer intent—remained intact for subsequent LLM-based analysis. The resulting dataset comprised 6000 cleaned reviews, with an average length of approximately 293 characters (55 words) per entry.

3.2. Analysis Framework

This study develops an end-to-end analysis pipeline that converts raw customer reviews into strategy-ready customer value signals aligned with the VPC. The framework integrates LLM-based information extraction, embedding-based semantic clustering, and quantitative synthesis to derive interpretable insights from customer feedback. The overall workflow, summarized in Figure 1, consists of five sequential analytical steps.
First, customer reviews are analyzed through an LLM-driven extraction module that identifies six key dimensions (Customer Jobs, Customer Pains, Customer Gains, Feature Gaps, Emotions, and Usage Context) in accordance with the VPC structure. Second, each extracted construct is converted into a fixed-length semantic representation using a transformer-based embedding model, enabling the measurement of similarity among concepts. Third, unsupervised cluster discovery is performed for each field to group semantically related phrases, and optimal cluster numbers are determined through joint inspection of inertia and silhouette coefficients. Fourth, the resulting clusters are assigned interpretable category names generated by an LLM, yielding a conceptual taxonomy that summarizes consumer perceptions and unmet needs. Finally, quantitative aggregation and visualization produce frequency distributions, gap density matrices, and product–market fit indicators that reveal priority areas for product improvement. All procedures were implemented in Python 3.12.7 with fixed random seeds, explicit prompts, and periodic checkpoints to ensure reproducibility. To further enhance methodological transparency, an additional graphical workflow diagram is provided in Appendix A (Figure A4), which visually traces the transformation from raw reviews to structured VPC signals and, subsequently, to quantitative PMF metrics via the sequential stages of extraction, embedding, clustering, synthesis and reporting. This illustration shows the end-to-end transformation from input data to analytical outputs.

3.3. LLM-Based Extraction of VPC Signals

Each customer review is analyzed by an instruction-tuned or an inference-optimized LLM that returns a structured JSON object with six defined fields: Customer Jobs, Customer Pains, Customer Gains, Feature Gaps, Emotions, and Usage Context (Note: The extracted JSON object can be found in the Appendix A). In this study, Meta’s Llama 3.3-70B-Instruct-Turbo model was employed as the inference engine due to its strong contextual understanding. The model is an instruction-tuned, text-only LLM optimized for multilingual dialog tasks and has been shown to outperform many existing open-source and proprietary chat models on common industry benchmarks. Its stability in structured text generation makes it particularly suitable for extracting semantically organized information from customer reviews [31]. Two distinct prompts—system prompt and user prompt—were designed to capture both explicit statements and implied cues as well as to return empty lists when evidence is absent. Inference was conducted through a hosted API (Together AI, https://www.together.ai/ (accessed on 17 October 2025)) for Llama-3.3-70B-Instruct-Turbo. Hence, no local accelerator configuration is required. We used conservative decoding (temperature ≈ 0.0) with schema-constrained prompting and strict JSON validation to ensure stability and recoverability. Missing keys were replaced with empty lists, and a checkpointing mechanism recorded intermediate results at fixed intervals to ensure safe recovery and reproducibility across large corpora. Downstream components were implemented in Python 3.12.7 with fixed random seeds. The core library stack for embedding and clustering is listed as follows: scikit-learn 1.6.1, numpy 2.0.2, together 1.5.30, pandas 2.2.2.
These six analytical dimensions jointly capture both the functional and experiential aspects of user perception embedded in product reviews. “Customer Jobs” refer to the functional, emotional, or social objectives that users seek to achieve through product use, while “Customer Pains” represent the difficulties or frustrations encountered in accomplishing those objectives. “Customer Gains” denote the benefits or positive outcomes realized or desired by users. “Feature Gaps” identify the discrepancies between expected and actual product performance. “Emotions” reflect affective responses expressed or implied in the reviews, categorized into ten classes: love, joy, trust, surprise, frustration, disgust, anger, fear, sadness, and neutral. This taxonomy extends the Parrott (2001) framework of six primary human emotions (love, joy, surprise, anger, sadness, fear) to incorporate domain-relevant affective states such as trust and frustration [32]. Finally, “Usage Context” captures situational clues regarding when, where or how the product is used (e.g., “during workouts,” “while traveling,” or “at night”), providing essential information for contextualizing customer experiences. Table 1 presents an example of the extracted VPC signal, showing how unstructured review text can be systematically parsed into six VPC dimensions. It captures both the functional and emotional aspects of user experience embedded in the narrative while leaving “Customer Pains”, “Feature Gaps” and “Usage Context” fields empty when no explicit evidence is found in the original text.
To assess the reliability of the LLM-based extraction, a random subset of 1000 cases was independently reviewed by domain experts. The resulting expert–model agreement exceeded 90% across all six analytical dimensions, demonstrating strong consistency between human judgment and model inference.

3.4. Embedding Generation

Following the extraction process, each textual element within the six analytical dimensions was converted into a continuous vector representation to standardize the extracted outputs and enable semantic comparison and clustering, since conceptually similar VPC signals often appear in varied linguistic forms. Embedding generation was performed using a transformer-based retrieval/encoding model that encodes each phrase into a fixed-length vector (e.g., 1024 dimensions). This representation preserves contextual and syntactic nuances, allowing conceptually similar expressions (e.g., “battery lasts long” and “strong battery life”) to occupy neighboring regions in the embedding space.
The embedding model “BAAI/bge-m3”, developed by Beijing Academy of Artificial Intelligence (BAAI), was employed to obtain dense embedding vectors. It provides state-of-the-art representational quality across multiple retrieval settings and document granularities. Unlike conventional embedding models optimized for a single task or short English texts, BGE-M3 unifies dense, sparse, and multi-vector retrieval under a single architecture, enabling robust performance on both sentence-level and long-document inputs (up to 8192 tokens). Its self-knowledge distillation framework jointly optimizes these retrieval modes, producing embeddings that are more discriminative and semantically coherent than those from prior models such as E5, Contriever, or OpenAI’s text-embedding series [33]. These features make BGE-M3 particularly well-suited for analyzing heterogeneous, user-generated review data that vary in linguistic style, length, and expression.
All embeddings were computed independently for the five major dimensions—Customer Jobs, Customer Pains, Customer Gains, Feature Gaps, and Usage Context—to maintain field-specific semantic structures. The “Emotions” dimension was excluded from embedding analysis because it was already represented by discrete categorical labels. The resulting embedding matrices served as the basis for unsupervised clustering and subsequent quantification. To ensure methodological robustness, embeddings were generated under consistent model parameters, with batch processing applied to large text sets, and intermediate checkpoints and fixed random seeds used to maintain traceability and reproducibility. The resulting high-dimensional embeddings provide a scalable foundation for discovering latent semantic patterns within large volumes of customer feedback.

3.5. Clustering of Extracted Concepts

The generated embeddings were subjected to unsupervised clustering to identify semantically coherent groups of customer concepts within each analytical dimension. K-means clustering was employed to partition the embedding space into discrete clusters, enabling the discovery of latent structures without the need for manual supervision [34,35]. To determine the optimal number of clusters, the inertia (elbow) criterion and the silhouette coefficient were jointly examined across a predefined range of k values. This dual-metric approach ensures a balance between compactness within clusters and separability across clusters, thereby enhancing interpretability [36].
To maintain methodological consistency and avoid distortions introduced by non-linear transformations, clustering was performed directly in the normalized embedding space without any prior dimensionality reduction. All K-means clustering procedures were executed with fixed random seeds and multiple initializations to ensure stability and reproducibility. The initialization parameter (n_init), which defines the number of random centroid initializations before selecting the solution with the lowest within-cluster variance, was set to 10. The number of clusters was systematically explored within the range of k = 5 to 20 at increments of 1 to capture an interpretable range of clustering granularity. The optimal k-value was determined by comparing the results from the Kneedle-based elbow method and the silhouette coefficient, selecting the smaller k to enhance interpretability. The Kneedle algorithm locates the “knee” point of the inertia curve by identifying where the marginal gain in intra-cluster compactness exhibits the most pronounced decline relative to a linear reference line. Formally, the knee point k satisfies the condition Δ(k) = f(k) − L(k) = max(f(k) − L(k)), where f(k) denotes the normalized inertia curve and L(k) represents the line connecting its endpoints [37]. This method offers a mathematically grounded and noise-robust approach for automatic elbow detection, making it well-suited for large-scale, high-dimensional embedding analyses. In this study, the value of k for each VPC signal is presented in Table 2. Figure 2 illustrates the evaluation of cluster compactness and separation across varying k values, used to determine the optimal number of clusters for the “Customer Jobs” dimension. The elbow point, indicated by the gray line, represents the value of k beyond which additional clusters yield marginal improvements in cohesion while increasing model complexity.

3.6. LLM-Based Labeling of Clusters

For each cluster obtained from the embedding-based K-means procedure, a concise, human-interpretable category label was generated using an LLM under a structured instruction prompt. For this purpose, the same LLM described in 3.3 (Llama 3.3-70B-Instruct-Turbo) was employed to generate high-level semantic abstractions and produce coherent category names. To ensure both semantic representativeness and computational efficiency, a subset of cluster phrases was selected based on their cosine similarity to the cluster centroid, using the top 50 most representative phrases rather than random or arbitrary samples. This centroid-based selection method prioritizes phrases that lie closest to the semantic core of each cluster, thereby enhancing the fidelity of the LLM’s summarization process while maintaining token efficiency. The prompt constrained the output to a single, 3–5-word noun phrase without punctuation, thereby encouraging consistent and semantically meaningful labels across clusters. Temperature was fixed at 0.3 to minimize variance in labeling outcomes.
The resulting LLM-generated labels collectively form a hierarchical conceptual taxonomy that merges heterogeneous expressions into coherent, higher-order categories aligned with customer value constructs. This taxonomy enables the projection of raw phrases onto standardized conceptual labels via a mapping function (Customer Jobs, Pains, Gains, Feature Gaps, and Usage Context), which in turn supports subsequent quantitative analyses—including frequency profiling, co-occurrence structures, job–gain/gap heatmaps, and product–market-fit scoring—while preserving interpretability for managerial decision-making.

3.7. Quantitative Analysis and Decision Metrics

Building upon the generated category label, quantitative analyses were conducted to evaluate the relative prominence of customer needs, the degree of unmet demand, and the overall alignment between product performance and user expectations. These analyses aim to establish an interpretive bridge between qualitative insights derived from LLM-based extraction and measurable indicators of product–market fit.
First, a frequency-based analysis was performed across the standardized dimensions of the VPC signals—namely Customer Jobs, Pains, Gains, Feature Gaps, Usage Context, and Emotions. Each label within these dimensions was aggregated and counted to capture the distributional structure of customer feedback expressions, thereby quantifying how frequently specific needs or perceptions were articulated in the review corpus. This frequency profiling serves as a descriptive foundation for subsequent analyses by highlighting the most relevant functional, emotional, and contextual attributes observed across user segments.
Second, a gap density analysis was performed to assess the extent to which each customer job was accompanied by corresponding gain or gap signals. For every review instance, the occurrence of gains and gaps was cross-referenced with the associated job categories to capture alignment patterns between customer expectations and realized outcomes. The proportion of “gap-only” cases, where a Customer Job was mentioned without a corresponding gain but with an identified gap, served as an indicator of latent dissatisfaction and unmet functional needs. This ratio-based measure quantifies the intensity of performance shortfalls across job categories, highlighting areas in which targeted product refinement or feature enhancement could generate the greatest incremental value.
Third, a context–gap matrix analysis was performed to examine how unmet needs are distributed across different usage scenarios. A two-dimensional matrix was constructed to quantify the co-occurrence intensity between usage contexts and unmet needs (high gaps), with the resulting heatmap illustrating where performance deficiencies are most concentrated. This can provide actionable guidance for product improvement priorities.
Finally, a Product–Market Fit (PMF) assessment was implemented to evaluate the overall balance between perceived benefits and performance deficiencies across different customer jobs. “Product–Market Fit” refers to the degree to which a product satisfies latent customer demand by delivering value that matches or exceeds expectations, such that customers adopt, use, and advocate the product [38]. This concept is conceptually aligned with value proposition design methods that emphasize matching product offers to customer segments [19]. In essence, achieving PMF implies a state in which a product’s value proposition is aligned with the market’s needs and enables sustainable growth. For each job j, a composite PMF score was computed as follows:
P M F j   s c o r e   = 0.4 × G a i n j M e n t i o n s j 0.4 × G a p j M e n t i o n s j + 0.2 × M e n t i o n s j T o t a l   R e v i e w s
Here, G a i n j and G a p j represent the number of gain- and gap-related signals associated with job j, and M e n t i o n s j indicates the total number of occurrences of that job within the corpus. A higher PMF score indicates stronger alignment between customer expectations and delivered value, whereas a lower score reveals latent dissatisfaction or structural weaknesses in product–market alignment. This composite measure provides a quantitative yet interpretable framework for evaluating how effectively the product fulfills core customer needs across heterogeneous usage conditions. PMF scores were computed for the full dataset and separately for each product generation.

4. Results

4.1. Distributional Patterns of Customer Value Signals

A frequency-based analysis was conducted to examine the distributional patterns of standardized customer value signals extracted across the six analytical dimensions of the VPC. The resulting frequency distributions summarize how frequently each extracted customer value category (Customer Jobs, Pains, Gains, etc.) appeared in the review corpus. In the case of “Customer Jobs”, 14 different category labels were identified, reflecting the diverse functional purposes associated with the smartwatch use. The top 3 most frequently observed jobs were productivity management (n = 2370), device usage experience (n = 1923), and social connection needs (n = 1498). These results indicate that customers primarily engage with the smartwatch to enhance personal efficiency, manage device-related interactions, and maintain social connectivity, highlighting its dual role as both a productivity facilitator and a lifestyle integration tool. (See Appendix B Table A1, Table A2, Table A3, Table A4 and Table A5 for the corresponding tabular distributions.)
Within the “Customer Gains” dimension, 11 distinct category labels were identified, reflecting various forms of added value and user satisfaction derived from the smartwatch. The top 3 most prevalent gain categories were information accessibility (n = 2452), product attributes (n = 2012), and product convenience (n = 1756). These findings indicate that customers derive the greatest value from the smartwatch’s ability to provide instant information access and deliver reliable, appealing performance, underscoring its role as an integrated tool that enhances both functional efficiency and lifestyle optimization.
Analysis of the “Customer Pains” data revealed 10 distinct category labels capturing the primary sources of dissatisfaction in product usage. The top 3 most frequently mentioned pain categories were battery life problems (n = 984), user experience issues (n = 752), and device usability problems (n = 629). These results indicate that customer dissatisfaction is primarily driven by concerns over limited battery endurance, inconsistent system performance, and functional malfunctions in specific device features, highlighting the need for technical refinements that enhance reliability, efficiency, and overall user experience.
Analysis of the “Feature Gaps” data identified 14 distinct category labels that represent areas where customers perceived functional or design-related shortcomings. The 3 most frequently mentioned gap categories were battery life concerns (n = 638), user experience features (n = 370), and functionality issues (n = 265). These findings suggest that customer-perceived gaps are primarily associated with insufficient battery performance, limited or unstable feature functionality, and shortcomings in interface design, emphasizing the necessity for targeted improvements in performance stability and user-centered feature enhancement.
Regarding the “Usage Context” dimension, 11 distinct category labels were identified, describing the diverse situations and routines in which the smartwatch is utilized. The three most frequently mentioned contexts were daily life concerns (n = 1029), personal device usage (n = 940), and fitness activities (n = 928). These findings suggest that smartwatch use is deeply embedded in users’ everyday habits, personal device interactions, and fitness routines, indicating that the product functions as a seamlessly integrated companion supporting both daily convenience and fitness-oriented lifestyles.
For “Emotions”, the distribution revealed a predominantly positive emotional profile with a strong bias toward joy and love. As shown in Table 3, joy was the most frequently observed emotion (n = 4265), followed by love (n = 2399), frustration (n = 1298) and trust (n = 148). Negative emotions such as anger (n = 114), surprise (n = 83), sadness (n = 39), disgust (n = 27), and fear (n = 4) appeared far less frequently, while neutral expressions accounted for a minor share (n = 106). This distribution indicates that overall customer sentiment toward the product was strongly positive, with only limited instances of frustration and negative affect reflecting localized dissatisfaction rather than widespread discontent. The relative dominance of joy and love further implies a highly favorable user experience, whereas the presence of frustration points to specific usability or performance gaps that merit closer examination in subsequent analyses. This overall positivity is consistent with the average star rating of 4.54 out of 5 across the review corpus, demonstrating that affective tone and quantitative satisfaction metrics converge to depict a favorable customer perception of the product.

4.2. Gap Density Analysis

According to the gap density analysis, a gap density ratio for each “Customer Jobs” dimension was calculated. Figure 3 visualizes the proportion of “gap-only” instances, in which a customer job was mentioned without a corresponding gain signal across all job categories. The results show that the product usage experience and device usage experience exhibit the highest gap densities, indicating that customers most frequently reported unmet needs related to operational performance and device interaction quality. Moderate levels of gap density were also observed for categories such as time management tools and notification management, suggesting persistent limitations in functionality integration and responsiveness. In contrast, health tracking and sleep tracking features presented relatively low gap ratios, implying that core wellness-related functions are generally meeting user expectations. Overall, the gap distribution underscores that perceived deficiencies are concentrated in interaction- and usability-oriented dimensions rather than in fundamental monitoring capabilities. Although the overall gap density ratios were relatively low, their consistent presence across multiple interaction-related job categories underscores latent opportunities for performance and usability enhancement.

4.3. Context–Gap Matrix Analysis

As shown in Figure 4, the visualization highlights specific contexts in which functional shortcomings are repeatedly mentioned, providing insights into which usage situations require targeted product refinement. The color scale indicates the frequency of co-occurrence between each usage context and unmet need, with deeper color intensity representing higher concentrations of reported deficiencies. Notably, the highest gap intensities are observed in “device usage habits”, “fitness activities” and “daily life concerns”, primarily driven by persistent issues related to battery life and feature reliability. This pattern suggests that customers most frequently experience dissatisfaction during prolonged or performance-intensive use, where power consumption and functional consistency become critical factors. Conversely, contexts such as “home environment needs” and “mobile lifestyle needs” exhibit relatively low gap frequencies, indicating that environment-dependent interactions contribute less to overall dissatisfaction. Overall, the distribution underscores that perceived deficiencies are predominantly associated with usage contexts involving sustained device operation and active engagement, emphasizing the importance of optimizing power efficiency and performance stability.

4.4. Product–Market Fit Analysis

As shown in Figure 5, the visualization presents the comparative distribution of gain and gap signals across various customer jobs, illustrating the relative balance between perceived benefits and performance deficiencies. The results indicate that “productivity management” and “device usage experience” exhibit the highest overall gain counts, suggesting that customers consistently recognize these areas as primary sources of value and satisfaction. However, these same categories also display relatively high gap counts, implying that while they contribute strongly to perceived utility, they simultaneously represent key domains where functional or performance-related issues are most frequently encountered. By contrast, jobs such as “music experience,” “personal style preferences,” and “phone call management” show both lower gain and gap frequencies, reflecting their limited impact on overall user satisfaction. Collectively, this pattern suggests that the smartwatch’s product–market fit is strongest in productivity- and interaction-oriented contexts. Yet sustained improvements in reliability and responsiveness remain critical for reinforcing perceived value in high-engagement use cases.
Next, PMF score following Equation (1) was calculated and shown in Table 4. The resulting distribution of PMF values revealed discernible variation across customer jobs, reflecting differing levels of alignment between user expectations and perceived performance. Customer Jobs such as “productivity management”, “social connection needs” and “health monitoring tools” achieved the highest PMF scores, indicating a strong alignment between functional value and user satisfaction. Conversely, lower PMF values were observed for categories such as “watch usage experience” and “time management tools”, suggesting that these domains contribute less to perceived overall fit or occupy a secondary position in users’ value hierarchy. Collectively, the PMF score analysis underscores that the smartwatch achieves its strongest product–market alignment in use cases emphasizing productivity enhancement, continuous interaction, and health-related functionality, while further improvements in peripheral features could strengthen the breadth of its market appeal. In particular, within the health monitoring family of jobs, “heart rate monitoring” exhibits a comparatively modest PMF (0.302), noticeably below “health monitoring tools” (0.329) and “health tracking features” (0.326), indicating residual concerns about sensing reliability and measurement stability. Overall, the PMF scores span a relatively narrow band (0.201–0.393), yet the dispersion is sufficient to set priorities at the lower end—especially “watch usage experience” (0.201) and “time management tools” (0.271). Targeted improvements in power efficiency, interaction latency, and the consistency of core sensing functions are therefore expected to yield the largest marginal gains in perceived product–market fit.
Because the baseline PMF score employed fixed weights, we conducted a sensitivity analysis using alternative weight combinations for the gain rate, gap rate, and job coverage terms. As shown in Figure 6, varying the weights shifts absolute PMF levels but does not significantly alter the relative ordering of customer jobs across the tested scenarios. High-ranked categories (e.g., productivity management, social connection needs) remain consistently above the rest, whereas low-ranked categories (e.g., watch usage experience, time management tools) persist at the lower end. This pattern indicates that the main conclusions are robust to reasonable changes in weighting, and that the inferred product–market fit hierarchy is not an artifact of a particular weight specification.
Additionally, to evaluate whether the relative ordering of customer jobs is stable across the alternative weighting schemes, the ranks from each weight scenario were aggregated and assessed using Kendall’s coefficient of concordance (W). Kendall’s W quantifies agreement among multiple ranked lists (0 = no agreement; 1 = complete agreement). Across all tested scenarios, the analysis returned W = 0.921 (p < 0.001), indicating very strong and statistically significant concordance of the rankings across scenarios.
Moreover, although the relative ordering is invariant across weight scenarios, emphasizing the gap rate lowers the absolute PMF levels of gap-heavy jobs and increases their spread from higher-performing categories. In practice, a higher gap weight reduces the absolute PMF score for gap-heavy jobs, bringing them closer to any intervention threshold and potentially triggering earlier remediation, even if the rank order is unchanged. Accordingly, the relative priorities can be confidently maintained for future product planning, while absolute PMF score shifts under alternative weights may still inform threshold-based resourcing and timing decisions.

4.5. Variations Across Product Generations

Lastly, as the analyzed data span three product generations (Watch 5–7), we examined cross-generation shifts in the prominence of customer jobs and the associated PMF scores. A chi-square test of independence indicated a significant association between product generation and job distributions (χ2(26) = 147.37, p < 0.001), implying that the mix of jobs discussed by customers changes across releases.
Figure 7 depicts proportional shifts across job category labels. “Device usage experience”, “watch usage experience”, and “time management tools” increase monotonically from Watch 5 to Watch 7, whereas “activity tracking”, “phone call management” and “notification management” exhibit a downward trend. “Productivity management” remains the single most prominent job across all generations, with only minor fluctuations, and “sleep tracking features” is comparatively stable.
Table 5 summarizes the PMF scores for each customer job across product generations. “Productivity management” remains the top-ranked job in all models, while “device usage experience” and “social connection needs” show positive Δ (7–5) values, indicating improved alignment in later generations. In contrast, “sleep tracking features” and “activity tracking” exhibit a modest decline in PMF by Watch 7, suggesting that monitoring-centric functionality has not uniformly benefited from generational updates.
Figure 8 visualizes the changes in PMF scores across different product generations to facilitate the comparison of trends and generational differences. In addition, the derived metrics (e.g., Gap density and PMF score) offer practical implications for data-driven product and UX management. These indicators can assist decision-makers in identifying feature areas with high improvement potential and quantifying how effectively current products address user needs. For instance, a high gap density in specific functional domains may signal priorities for feature enhancement, while PMF trends across generations can inform future product roadmap adjustments. Moreover, these measures can serve as quantitative inputs for setting performance indicators (KPIs) related to user satisfaction or innovation outcomes. By integrating such metrics into the product development workflow, corporations can more systematically align design and investment decisions with empirically grounded customer insights. Accordingly, this framework advances theoretical understanding while reinforcing its practical relevance for strategic decision-making.

5. Discussion

Building on the empirical results, this section synthesizes the study’s contributions and practical takeaways. The analysis results jointly characterize how customers experience and evaluate the smartwatch. A predominantly positive effect, concentrated unmet needs around battery life and interaction quality, and the persistence of the PMF ranking under alternative weight schemes were observed. Simultaneously, the cross-generation PMF table functions as a compact diagnostic, with row-wise comparisons showing generation-specific alignment and Δ columns revealing directional change, thereby translating the empirical patterns into actionable priorities for product refinement.
In terms of theoretical implication, this study advances a computational bridge between unstructured review text and formal value proposition theory. By operationalizing “Jobs–Pains–Gains” as extractable signals via LLMs and mapping them to clusterable semantic spaces, the framework moves beyond sentiment or topic prevalence to recover actionable value constructs at scale. The integration of gap density diagnostics and a composite PMF index (balancing gain rate, gap rate, and coverage) yields a quantitative–qualitative coupling that ties narrative evidence to strategy-ready metrics across product generations. Methodologically, the sensitivity analysis of PMF weights and concordance testing of rankings demonstrated that the inferred value hierarchy is robust rather than weight-dependent, strengthening external validity. The cross-generation distributional tests further corroborate that the framework can detect structurally meaningful shifts over successive releases. Collectively, these elements contribute a generalizable pipeline for theory-driven text analytics that can be applied across product domains without extensive labeled data.
In terms of managerial implications, for product and UX management, the results prioritize where and how to intervene. First, high-engagement jobs (notably productivity management and health monitoring) deliver the greatest perceived value (high PMF) yet still exhibit salient gaps around battery endurance and interaction stability in everyday and fitness contexts. Hence, targeted work on power efficiency, interaction latency, and core sensing reliability is most likely to yield outsized gains. Second, the context–gap matrix pinpoints usage situations (e.g., device usage habits, daily life, fitness activities) in which deficiencies concentrate, guiding scenario-specific improvements and test plans rather than generic feature tweaks. Third, the generation-wise PMF table enables roadmap sequencing by highlighting which jobs strengthened or weakened in alignment, thereby supporting cohort- and release-specific prioritization. Fourth, jobs with consistently low PMF (e.g., watch usage experience, time management tools) should be considered for scope reduction, redesign, or bundling with higher-value flows to avoid diluting perceived fit. Finally, because the PMF ranking remains stable under plausible weighting schemes, cross-functional product and UX teams—together with data science and engineering—can rely on these priorities to set KPIs (e.g., gap-only ratio targets) and to design experiments (e.g., A/B tests tied to job-level gains and gaps), thereby increasing confidence that decisions are robust to modeling choices.
Moreover, the practical utility of the proposed framework is constrained by several methodological and contextual dependencies. As the analysis relies on customer reviews collected from a single e-commerce platform, representativeness may be affected by demographic or platform-specific biases. Future efforts should explore reviewing data across multiple sources (e.g., Amazon) to mitigate sampling bias and enhance generalizability. Simultaneously, the framework’s reliance on a specific LLM introduces inherent variability related to model configuration, training corpus composition, and prompt sensitivity. To reduce this dependency, the present study adopted deterministic temperature settings, schema validation, and checkpointing mechanisms to ensure output stability. Nonetheless, future research could employ ensemble-based approaches that integrate multiple LLMs to enhance robustness and reduce single-model bias. For instance, outputs derived from different LLM architectures (e.g., Llama or GPT) could be aggregated using majority voting or agreement-based filtering, ensuring that only consistently inferred insights are retained. Such cross-model validation would allow the framework to identify stable semantic patterns across architectures, thereby improving confidence in extracted customer signals and mitigating idiosyncratic model effects.
In sum, the study demonstrates how LLM-enabled value extraction, coupled with clustering and PMF-based synthesis, can translate raw customer voice into defensible product strategy. It can advance theory on value modeling while providing managers with precise, context-aware levers for improvement. By incorporating cross-generation analyses, the framework not only diagnoses current pain points and value drivers but also tracks how alignment evolves across releases, thereby informing evidence-based product planning.

6. Conclusions

Using LLMs, this study extracted and classified information related to jobs, pains, and gains from real consumer reviews, thereby introducing a new quantitative approach for interpreting large-scale VOC data. Practically, the derived outputs (e.g., frequency profiles, context–gap matrices, and generation-wise PMF scores) serve as compact diagnostics that help prioritize where and how to intervene, supporting KPI design and experiment planning in a data-driven manner. The proposed approach is model-agnostic and scalable, requiring no extensive task-specific labels, which facilitates transfer to adjacent product domains.
Despite its contributions, this study comes with certain limitations. First, the dataset is drawn from a single retailer platform and a limited time window, which may introduce selection and platform-specific biases. Second, extraction and labeling remain sensitive to prompt design, LLM configuration, and embedding choices. Although the analysis and experimental procedures are documented in detail, alternative prompt designs and model configurations should be systematically evaluated. Third, several operational choices (e.g., binary treatment of gain/gap at the review level, heuristic PMF weights) impose simplifying assumptions that may compress nuance in multi-signal reviews. Fourth, while expert–model agreement provided an initial validation of reliability, the study did not include comparative baselines, external validation on additional datasets, or quantitative performance metrics due to the absence of standardized benchmarks for VPC-aligned extraction. Building upon this research, future studies could establish comparative validation frameworks using larger and more diverse review corpora, enabling systematic evaluation across alternative model architectures and dataset conditions to further substantiate the robustness and generalizability of the proposed approach. Additionally, cross-model benchmarking lies beyond the present scope but is acknowledged as a limitation and designated as a priority for future research.
Accordingly, future research should test different varieties of LLMs and link the proposed metrics to downstream outcomes—such as A/B tests, churn/return rates, service logs, or telemetry—to support causal claims and to close the loop between insight generation and measurable product impact.

Author Contributions

Conceptualization, C.H.S.; Methodology, C.H.S.; Software, C.H.S.; Validation, J.L. and C.H.S.; Formal analysis, J.L. and C.H.S.; Investigation, J.L. and C.H.S.; Data curation, J.L. and C.H.S.; Writing—original draft preparation, J.L. and C.H.S.; Writing—review and editing, J.L. and C.H.S.; Visualization, C.H.S.; Supervision, C.H.S.; Project administration, C.H.S. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Data Availability Statement

The data presented in this study are available on request from the corresponding author due to legal reasons.

Acknowledgments

The author would like to thank the anonymous reviewers and the editors for their insightful comments and feedback on this research.

Conflicts of Interest

The authors declare no conflicts of interest.

Abbreviations

The following abbreviations are used in this manuscript:
BAAIBeijing Academy of Artificial Intelligence
BOWBag-of-words
CNCustomer Needs
JSONJavaScript Object Notation
KPIKey Performance Indicator
LLMLarge Language Model
NLPNatural Language Processing
PMFProduct–Market Fit
VOCVoice of Customer
VPCValue Proposition Canvas

Appendix A

Figure A1. Prompt for VPC signal extraction.
Figure A1. Prompt for VPC signal extraction.
Electronics 14 04407 g0a1
Figure A2. Prompt for labeling of clusters.
Figure A2. Prompt for labeling of clusters.
Electronics 14 04407 g0a2
Figure A3. Example structure of an extracted JSON object returned by the LLM.
Figure A3. Example structure of an extracted JSON object returned by the LLM.
Electronics 14 04407 g0a3
Figure A4. Graphical workflow of the LLM-based VPC extraction and synthesis pipeline.
Figure A4. Graphical workflow of the LLM-based VPC extraction and synthesis pipeline.
Electronics 14 04407 g0a4

Appendix B

Table A1. Distribution of category labels for “Customer Jobs”.
Table A1. Distribution of category labels for “Customer Jobs”.
Category LabelFrequencies
productivity management2370
device usage experience1923
social connection needs1498
health monitoring tools1421
activity tracking1370
health tracking features1273
time management tools1134
notification management1089
sleep tracking features845
watch usage experience753
phone call management732
heart rate monitoring625
personal style preferences526
music experience408
Table A2. Distribution of category labels for “Customer Gains”.
Table A2. Distribution of category labels for “Customer Gains”.
Category LabelFrequencies
information accessibility2452
product attributes2012
product convenience1756
enhanced product capabilities1282
health tracking effectiveness1254
customer satisfaction1129
user experience979
battery performance923
product quality issues819
seamless integration532
product customization options393
Table A3. Distribution of category labels for “Customer Pains”.
Table A3. Distribution of category labels for “Customer Pains”.
Category LabelFrequencies
battery life problems984
user experience issues752
device usability problems629
product disappointment405
battery life issues294
limited capabilities285
feature shortcomings274
watch design issues252
phone dependency245
tracking inaccuracies209
Table A4. Distribution of category labels for “Feature Gaps”.
Table A4. Distribution of category labels for “Feature Gaps”.
Category LabelFrequencies
battery life concerns638
user experience features370
functionality issues265
reliability concerns218
product enhancements193
watch band options173
battery life expectations140
device compatibility106
activity tracking accuracy105
health monitoring features101
battery life concerns 293
limited functionality85
galaxy ai features64
bezel preferences53
Table A5. Distribution of category labels for “Usage Context”.
Table A5. Distribution of category labels for “Usage Context”.
Category LabelFrequencies
daily life concerns1029
personal device usage940
fitness activities928
outdoor activities859
device usage habits737
mobile phone interaction608
work environment needs484
sleeping issues474
mobile lifestyle needs347
home environment needs166
wrist placement143

References

  1. Zhai, M.; Wang, X.; Zhao, X. The Importance of Online Customer Reviews Characteristics on Remanufactured Product Sales: Evidence from the Mobile Phone Market on Amazon.com. J. Retail. Consum. Serv. 2024, 77, 103677. [Google Scholar] [CrossRef]
  2. Chung, J.; Lee, J.; Yoon, J. Understanding Music Streaming Services via Text Mining of Online Customer Reviews. Electron. Commer. Res. Appl. 2022, 53, 101145. [Google Scholar] [CrossRef]
  3. Qi, J.; Zhang, Z.; Jeon, S.; Zhou, Y. Mining Customer Requirements from Online Reviews: A Product Improvement Perspective. Inf. Manag. 2016, 53, 951–963. [Google Scholar] [CrossRef]
  4. Lee, J.; Jeong, B.; Yoon, J.; Song, C.H. Context-Aware Customer Needs Identification by Linguistic Pattern Mining Based on Online Product Reviews. IEEE Access 2023, 11, 71859–71872. [Google Scholar] [CrossRef]
  5. Hou, Y.; Li, J.; He, Z.; Yan, A.; Chen, X.; McAuley, J. Bridging Language and Items for Retrieval and Recommendation. arXiv 2024, arXiv:2403.03952. [Google Scholar] [CrossRef]
  6. Elahi, M.; Khosh Kholgh, D.; Kiarostami, M.S.; Oussalah, M.; Saghari, S. Hybrid Recommendation by Incorporating the Sentiment of Product Reviews. Inf. Sci. 2023, 625, 738–756. [Google Scholar] [CrossRef]
  7. Qiu, J.; Liu, C.; Li, Y.; Lin, Z. Leveraging Sentiment Analysis at the Aspects Level to Predict Ratings of Reviews. Inf. Sci. 2018, 451–452, 295–309. [Google Scholar] [CrossRef]
  8. Chua, A.Y.K.; Banerjee, S. Analyzing Review Efficacy on Amazon.com: Does the Rich Grow Richer? Comput. Hum. Behav. 2017, 75, 501–509. [Google Scholar] [CrossRef]
  9. Chatterjee, I.; Zhou, M.; Abusorrah, A.; Sedraoui, K.; Alabdulwahab, A. Statistics-Based Outlier Detection and Correction Method for Amazon Customer Reviews. Entropy 2021, 23, 1645. [Google Scholar] [CrossRef]
  10. Ali, H.; Hashmi, E.; Yayilgan Yildirim, S.; Shaikh, S. Analyzing Amazon Products Sentiment: A Comparative Study of Machine and Deep Learning, and Transformer-Based Techniques. Electronics 2024, 13, 1305. [Google Scholar] [CrossRef]
  11. Yazıcı, G.; Ozansoy Çadırcı, T. Creating Meaningful Insights from Customer Reviews: A Methodological Comparison of Topic Modeling Algorithms and Their Use in Marketing Research. J. Mark. Anal. 2024, 12, 865–887. [Google Scholar] [CrossRef]
  12. Kim, R.Y. Using Online Reviews for Customer Sentiment Analysis. IEEE Eng. Manag. Rev. 2021, 49, 162–168. [Google Scholar] [CrossRef]
  13. Zhang, W.; Li, X.; Deng, Y.; Bing, L.; Lam, W. A Survey on Aspect-Based Sentiment Analysis: Tasks, Methods, and Challenges. IEEE Trans. Knowl. Data Eng. 2023, 35, 11019–11038. [Google Scholar] [CrossRef]
  14. Malik, N.; Bilal, M. Natural Language Processing for Analyzing Online Customer Reviews: A Survey, Taxonomy, and Open Research Challenges. PeerJ Comput. Sci. 2024, 10, e2203. [Google Scholar] [CrossRef] [PubMed]
  15. Timoshenko, A.; Mao, C.; Hauser, J.R. Can Large Language Models Extract Customer Needs as Well as Professional Analysts? arXiv 2025, arXiv:2503.01870. [Google Scholar] [CrossRef]
  16. Wei, W.; Hao, C.; Wang, Z. User Needs Insights from UGC Based on Large Language Model. Adv. Eng. Inform. 2025, 65, 103268. [Google Scholar] [CrossRef]
  17. Mukku, S.S.; Soni, M.; Rana, J.; Aggarwal, C.; Yenigalla, P.; Patange, R.; Mohan, S. InsightNet: Structured Insight Mining from Customer Feedback. arXiv 2024, arXiv:2405.07195. [Google Scholar] [CrossRef]
  18. Mosqueira-Rey, E.; Hernández-Pereira, E.; Alonso-Ríos, D.; Bobes-Bascarán, J.; Fernández-Leal, Á. Human-in-the-Loop Machine Learning: A State of the Art. Artif. Intell. Rev. 2023, 56, 3005–3054. [Google Scholar] [CrossRef]
  19. Osterwalder, A.; Pigneur, Y.; Bernarda, G.; Smith, A. Value Proposition Design: How to Create Products and Services Customers Want; John Wiley & Sons: Hoboken, NJ, USA, 2015; ISBN 1118968069. [Google Scholar]
  20. Timoshenko, A.; Hauser, J.R. Identifying Customer Needs from User-Generated Content. Mark. Sci. 2019, 38, 1–20. [Google Scholar] [CrossRef]
  21. Griffin, A.; Hauser, J.R. The Voice of the Customer. Mark. Sci. 1993, 12, 1–27. [Google Scholar] [CrossRef]
  22. Ma, Y.; Chen, G.; Wei, Q. Finding Users Preferences from Large-Scale Online Reviews for Personalized Recommendation. Electron. Commer. Res. 2017, 17, 3–29. [Google Scholar] [CrossRef]
  23. Zhang, W.; Yoshida, T.; Tang, X. A Comparative Study of TF*IDF, LSI and Multi-Words for Text Classification. Expert Syst. Appl. 2011, 38, 2758–2765. [Google Scholar] [CrossRef]
  24. Fang, X.; Zhan, J. Sentiment Analysis Using Product Review Data. J. Big Data 2015, 2, 5. [Google Scholar] [CrossRef]
  25. Aguwa, C.C.; Monplaisir, L.; Turgut, O. Voice of the Customer: Customer Satisfaction Ratio Based Analysis. Expert Syst. Appl. 2012, 39, 10112–10119. [Google Scholar] [CrossRef]
  26. Schouten, K.; Frasincar, F. Survey on Aspect-Level Sentiment Analysis. IEEE Trans. Knowl. Data Eng. 2016, 28, 813–830. [Google Scholar] [CrossRef]
  27. Barravecchia, F.; Mastrogiacomo, L.; Franceschini, F. Digital Voice-of-Customer Processing by Topic Modelling Algorithms: Insights to Validate Empirical Results. Int. J. Qual. Reliab. Manag. 2021, 39, 1453–1470. [Google Scholar] [CrossRef]
  28. Chen, D.; Zhang, D.; Liu, A. Intelligent Kano Classification of Product Features Based on Customer Reviews. CIRP Ann. 2019, 68, 149–152. [Google Scholar] [CrossRef]
  29. Jin, J.; Ji, P.; Liu, Y.; Johnson Lim, S.C. Translating Online Customer Opinions into Engineering Characteristics in QFD: A Probabilistic Language Analysis Approach. Eng. Appl. Artif. Intell. 2015, 41, 115–127. [Google Scholar] [CrossRef]
  30. Praveen, S.V.; Gajjar, P.; Ray, R.K.; Dutt, A. Crafting Clarity: Leveraging Large Language Models to Decode Consumer Reviews. J. Retail. Consum. Serv. 2024, 81, 103975. [Google Scholar] [CrossRef]
  31. Wihl, J.; Rosenkranz, E.; Schramm, S.; Berberich, C.; Griessmair, M.; Woźnicki, P.; Pinto, F.; Ziegelmayer, S.; Adams, L.C.; Bressem, K.K.; et al. Data Extraction from Free-Text Stroke CT Reports Using GPT-40 and Llama-3.3-70B: The Impact of Annotation Guidelines. Eur. Radiol. Exp. 2025, 9, 61. [Google Scholar] [CrossRef]
  32. Parrott, W.G. Emotions in Social Psychology: Essential Readings; Parrott, W.G., Ed.; Psychology Press: New York, NY, USA, 2001; ISBN 0-86377-682-5. (Hardcover); 0-86377-683-3 (Paperback). [Google Scholar]
  33. Chen, J.; Xiao, S.; Zhang, P.; Luo, K.; Lian, D.; Liu, Z. BGE M3-Embedding: Multi-Lingual, Multi-Functionality, Multi-Granularity Text Embeddings Through Self-Knowledge Distillation. arXiv 2024, arXiv:2402.03216. [Google Scholar] [CrossRef]
  34. Hartigan, J.A.; Wong, M.A. Algorithm AS 136: A K-Means Clustering Algorithm. J. R. Stat. Soc. Ser. C (Appl. Stat.) 1979, 28, 100–108. [Google Scholar] [CrossRef]
  35. Riaz, S.; Fatima, M.; Kamran, M.; Nisar, M.W. Opinion Mining on Large Scale Data Using Sentiment Analysis and K-Means Clustering. Clust. Comput. 2019, 22, 7149–7164. [Google Scholar] [CrossRef]
  36. Mohamed Nafuri, A.F.; Sani, N.S.; Zainudin, N.F.A.; Rahman, A.H.A.; Aliff, M. Clustering Analysis for Classifying Student Academic Performance in Higher Education. Appl. Sci. 2022, 12, 9467. [Google Scholar] [CrossRef]
  37. Satopaa, V.; Albrecht, J.; Irwin, D.; Raghavan, B. Finding a “Kneedle” in a Haystack: Detecting Knee Points in System Behavior. In Proceedings of the 2011 31st International Conference on Distributed Computing Systems Workshops, Minneapolis, MN, USA, 20–24 June 2011; pp. 166–171. [Google Scholar]
  38. Feinleib, D. Poor Product-Market Fit. In Why Startups Fail: And How Yours Can Succeed; Feinleib, D., Ed.; Apress: Berkeley, CA, USA, 2011; pp. 3–13. ISBN 978-1-4302-4141-6. [Google Scholar]
Figure 1. Overall analytical workflow of the proposed end-to-end analysis pipeline.
Figure 1. Overall analytical workflow of the proposed end-to-end analysis pipeline.
Electronics 14 04407 g001
Figure 2. Detection of optimal number of clusters for Customer Jobs.
Figure 2. Detection of optimal number of clusters for Customer Jobs.
Electronics 14 04407 g002
Figure 3. Illustration of gap density ratio.
Figure 3. Illustration of gap density ratio.
Electronics 14 04407 g003
Figure 4. Illustration of context–gap matrix analysis.
Figure 4. Illustration of context–gap matrix analysis.
Electronics 14 04407 g004
Figure 5. Product–market fit heatmap.
Figure 5. Product–market fit heatmap.
Electronics 14 04407 g005
Figure 6. Differing weight scenario for calculating PMF scores.
Figure 6. Differing weight scenario for calculating PMF scores.
Electronics 14 04407 g006
Figure 7. Distribution of Customer Jobs by Product Generation.
Figure 7. Distribution of Customer Jobs by Product Generation.
Electronics 14 04407 g007
Figure 8. PMF score trends across Watch generation.
Figure 8. PMF score trends across Watch generation.
Electronics 14 04407 g008
Table 1. Example of a structured VPC signal extracted from a customer review. (Note: Empty outputs occur when the corresponding dimensions cannot be detected).
Table 1. Example of a structured VPC signal extracted from a customer review. (Note: Empty outputs occur when the corresponding dimensions cannot be detected).
Original Customer ReviewExtracted FieldOutput
First smart watch. Very happy with it. Very smooth and fast interface with so much customization. I even like swapping out my bands or putting different colored caps on the watch. I love using the Samsung fitness and health app on the watch too.Customer Jobstracking fitness and health,
customizing appearance
Customer Gainssmooth and fast interface, customization options
Customer Pains-
Feature Gaps-
Emotionsjoy, love
Usage Context-
Table 2. Optimal number of k determined via K-means clustering.
Table 2. Optimal number of k determined via K-means clustering.
VPC Signalk-Value
Customer Jobs14
Customer Gains11
Customer Pains10
Feature Gaps 14
Usage Context11
Table 3. Distribution of category labels for “Emotions”.
Table 3. Distribution of category labels for “Emotions”.
Category LabelFrequencies
joy4265
love2399
frustration1298
trust148
anger114
neutral106
surprise83
sadness39
disgust27
fear4
Table 4. Product–Market Fit (PMF) scores by customer job category.
Table 4. Product–Market Fit (PMF) scores by customer job category.
Category LabelPMF Score
productivity management0.393
social connection needs0.343
health monitoring tools0.329
activity tracking0.328
health tracking features0.326
phone call management0.323
notification management0.319
music experience0.314
personal style preferences0.312
device usage experience0.311
sleep tracking features0.308
heart rate monitoring0.302
time management tools0.271
watch usage experience0.201
Table 5. PMF scores by customer job and product generation (Watch 5–7).
Table 5. PMF scores by customer job and product generation (Watch 5–7).
Category LabelPMF ScoreΔ (Delta)
Watch 5Watch 6Watch 7Watch 7–5
productivity management0.3490.3440.337−0.012
social connection needs0.3020.3110.3230.021
phone call management0.3030.3110.3110.008
music experience0.3040.3050.3060.002
health tracking features0.2980.2980.3040.005
personal style preferences0.2960.3080.2980.003
device usage experience0.2680.2550.2890.021
activity tracking0.3080.3030.286−0.022
health monitoring tools0.3020.3120.283−0.019
heart rate monitoring0.2720.3160.2820.01
notification management0.2990.3070.28−0.019
sleep tracking features0.3020.3070.261−0.041
time management tools0.2440.2530.2460.002
watch usage experience0.1810.1760.1980.018
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Lee, J.; Song, C.H. From Unstructured Feedback to Structured Insight: An LLM-Driven Approach to Value Proposition Modeling. Electronics 2025, 14, 4407. https://doi.org/10.3390/electronics14224407

AMA Style

Lee J, Song CH. From Unstructured Feedback to Structured Insight: An LLM-Driven Approach to Value Proposition Modeling. Electronics. 2025; 14(22):4407. https://doi.org/10.3390/electronics14224407

Chicago/Turabian Style

Lee, Jinkyu, and Chie Hoon Song. 2025. "From Unstructured Feedback to Structured Insight: An LLM-Driven Approach to Value Proposition Modeling" Electronics 14, no. 22: 4407. https://doi.org/10.3390/electronics14224407

APA Style

Lee, J., & Song, C. H. (2025). From Unstructured Feedback to Structured Insight: An LLM-Driven Approach to Value Proposition Modeling. Electronics, 14(22), 4407. https://doi.org/10.3390/electronics14224407

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop