You are currently viewing a new version of our website. To view the old version click .
Behavioral Sciences
  • Article
  • Open Access

Published: 10 February 2025

Research on the Impact of an AI Voice Assistant’s Gender and Self-Disclosure Strategies on User Self-Disclosure in Chinese Postpartum Follow-Up Phone Calls

,
,
and
1
School of Design Art and Media, Nanjing University of Science and Technology, Nanjing 210094, China
2
Faculty of Innovation and Design, City University of Macau, Macau 999074, China
*
Author to whom correspondence should be addressed.
This article belongs to the Special Issue Paving the Path to Well-Being Through Human Behavior Analysis with Data Science

Abstract

This study examines the application of AI voice assistants in Chinese postpartum follow-up phone calls, with particular focus on how interaction design strategies influence users’ self-disclosure intention. A 2 (voice gender: female/male) × 3 (self-disclosure strategies: normal conversation without additional disclosure/objective factual disclosure/emotional and opinion-based disclosure) mixed experimental design (n = 395) was conducted to analyze how the gender and self-disclosure strategies of voice assistants affect users’ stereotypes (perceived warmth and competence), and how these stereotypes, mediated by privacy calculus dimensions (perceived risks and perceived benefits), influence self-disclosure intention. The experiment measured various indicators using a 7-point Likert scale and performed data analysis through analysis of variance (ANOVA) and structural equation modeling (SEM). The results demonstrate that female voice assistants significantly enhance users’ perceived warmth and competence, while emotional self-disclosure strategies significantly improve perceived warmth. Stereotypes about the voice assistant positively affect users’ self-disclosure intention through the mediating effects of perceived risk and benefit, with perceived benefit exerting a stronger effect than perceived risk. These findings provide valuable insights for the design and application of AI voice assistants in healthcare, offering actionable guidance for enhancing user interaction and promoting self-disclosure in medical contexts.

1. Introduction

The emergence of Artificial Intelligence (AI) voice assistants has progressively liberated individuals from tedious and complex tasks. One of the most crucial areas for AI voice assistant applications is the healthcare sector. The demanding workload, specialized knowledge requirements, and the service-oriented nature of healthcare environments make the integration of AI to replace certain manual tasks an inevitable trend (). A growing body of research has begun to examine the communication strategies and effectiveness of AI voice assistants in medical guidance, disease diagnosis, and psychological counseling, and significant progress has been achieved (; ). In contemporary healthcare systems, follow-up phone calls have become an essential component of post-care services (). This modality is particularly suitable for AI voice assistants, especially in postpartum follow-up, which is both a frequent and stable occurrence in hospitals, especially in China (). Implementing AI voice assistants in this context can significantly enhance hospitals’ operational efficiency while reducing costs (). Additionally, hospitals seek to acquire insights into patients’ authentic experiences during the medical process through follow-up calls. They need to extract reliable and in-depth information to improve service quality, identify issues in healthcare delivery, and refine management processes (). Therefore, encouraging users to disclose personal information more honestly and comprehensively when using AI voice assistants is crucial for hospitals. A study conducted in the United States () involved 528 pregnant women and utilized an AI-powered voice assistant to predict pregnancy-related risks and engage in conversations with the participants. Compared to a control group, which listed risks and suggested that participants discuss them with their clinicians, the use of the AI voice assistant led to a 16% increase in the rate of actions taken to reduce those risks. In China, a study () designed a chat framework for AI-driven health assistants within health communities. It reduced participants’ negative emotions following AI-generated replies, with a 36% decrease. The prevalence of positive emotions was found to be 15% higher in participants who received AI-generated replies than in participants who conversed with human operators. Furthermore, Xunfei Healthcare Technology has implemented proactive patient management through AI-driven follow-up phone calls and other methods. The platform was first launched in Anhui and Shandong, covering over 30 major diseases across various departments and reaching 85% of discharged patients. Because of this management, patient compliance has significantly increased ().
Follow-up phone calls have been identified as a beneficial intervention for patients transitioning from the hospital to home (). During these calls, hospital staff can offer care recommendations, improve communication regarding information exchange, assist in managing symptoms and complications, and help ease the transition from hospital to home (). The telephone is an ideal tool for using a set of structured questions to assess clearly defined outcomes, and it is more readily accepted by women (). Additionally, follow-up phone calls represent a cost-effective approach to improving patient satisfaction, enhancing health outcomes, and reducing readmission rates (). In China, hospital staff typically conduct postpartum follow-up phone calls within 1–2 weeks following maternal discharge. These follow-ups begin with inquiries about the health of both the mother and infant, addressing any concerns the mother may have. They continue with surveys on aspects such as medical service quality, nursing care, post-care support, and medical ethics. The information gathered is then reported back to the hospital (). For hospitals, besides ensuring medical standards, service quality often becomes a decisive factor in competitive success (). This feedback mechanism not only supports improvements in healthcare service quality but also contributes to the optimization of hospital management. Sufficient user feedback can help hospitals respond promptly to patients’ health conditions and offer recommendations for follow-up services. Furthermore, the ongoing improvement of medical quality, nursing quality, logistics support, and medical ethics relies heavily on user feedback. A good number of healthcare institutions have made notable progress in management and service optimization through follow-up phone call programs ().
The personalization paradox suggests that in order to receive personalized services that align with their needs, users must provide detailed personal information, often including sensitive data (). However, obtaining truthful user information in the healthcare sector presents challenges, particularly when dealing with sensitive information. Numerous patients may be reluctant to make a completely honest disclosure for a variety of reasons. Studies indicate that patients often withhold details because of fear of self-disclosure or a desire to manage impressions when interacting with healthcare providers (; ). Fear of self-disclosure refers to concerns about revealing private, sensitive, or stigmatized information. Impression management involves patients intentionally concealing or altering information to present themselves positively to healthcare professionals (). Lucas () argues that AI voice assistants have a unique advantage in reducing both self-disclosure anxiety and impression management behaviors. The anonymity afforded by AI voice assistants allows patients to express themselves more freely, enabling hospitals to collect more honest and detailed feedback. Moreover, the non-contact nature of AI voice assistant interactions addresses the needs of users who prefer to avoid public device contact while still receiving convenient and high-quality service (; ).
The CASAs (Computers as Social Actors) framework offers valuable insights into the design strategies for AI voice assistants. This framework, which is central to understanding human disclosure behaviors toward AI, posits that when people interact with intelligent agents possessing human-like features, people adopt disclosure strategies similar to those used in human-to-human interaction (). This theory has been widely applied in business, healthcare, online communities, and other domains. Based on the CASA framework, research on human communication can be extended to human–AI interactions and subsequently verified (). In particular, voice characteristics, derived from stereotype theory, are considered to be a critical factor influencing disclosure intention. Studies have also shown that the self-disclosure of one part during communication affects the disclosure behavior of the other. This reciprocal disclosure effect has been demonstrated in interactions between humans and intelligent agents (; ). This study asserts that investigating voice characteristics and disclosure strategies can provide guidance for designing effective AI voice assistants.
The application of AI technology in follow-up phone calls in hospital has shown immense potential, not only alleviating the burden on medical workers but also offering users a more flexible and secure mode of interaction. This provides a viable method for enhancing users’ trust in hospitals and encouraging greater willingness to disclose information. In order to make better use of AI voice assistants for postpartum phone calls, further research is needed to optimize interaction between AI and users. This study aims to explore how different AI voice genders and self-disclosure strategies affect users’ perceptions of AI voice assistants, thereby influencing their intention to self-disclose accurately.

3. Methods and Procedure

3.1. Study Design

A mixed experimental method was used in this study: 2 (voice gender of the telephone follow-up assistant: male voice/female voice) × 3 (information disclosure strategies of the follow-up phone call assistant: normal communication without additional disclosure/intentional disclosure of objective facts/intentional disclosure of subjective feelings and opinions) to explore how different voice genders and disclosure strategies impact users’ intention to self-disclose in AI-driven postpartum follow-up phone calls. Participants were randomly assigned to one of six experimental groups, in which they interacted with the corresponding voice assistant and engaged in the telephone follow-up.

3.2. Material Design

3.2.1. Experimental Content Design

This study is based on interviews with hospital staff and relevant postpartum follow-up training literature. The hospital’s postpartum follow-up consists of five main sections: topic introduction, inquiry about the mother and baby’s health status, survey on medical service quality, investigation of medical ethics issues with regard to hospital staff, and finally, responding to user feedback and concluding the follow-up phone call. In the topic introduction phase, the voice assistant greets the participant, introduces itself, and explains the purpose of the call. In the health status inquiry phase, considering common postpartum health issues consists of psychological (e.g., postpartum anxiety) and physiological (e.g., postoperative recovery) aspects (), this study follows relevant postpartum health follow-up guidelines and selects one question related to each aspect to ask the participant (). In the medical care quality survey phase, the voice assistant inquiries about the user’s satisfaction with the medical services and requests suggestions for improvement. In the medical ethics survey phase, the voice assistant encourages the user to disclose any unreasonable experiences they have had during the medical service process and assures the participant that their feedback will be addressed. Finally, the voice assistant responds to the user’s feedback, expresses gratitude, and concludes the follow-up phone call.
In the self-disclosure strategy variable section, the self-disclosure measurement developed by Archer and Berg () is adopted to evaluate and manipulate the voice assistant’s self-disclosure content. This method classifies disclosed information into four categories: basic information, simple and visible information, attitudes and opinions, and strong affections and basic values, assigning corresponding scores to assess the intimacy of disclosure (see Appendix A, Table A1). In the experiment, based on the first voice follow-up script, which only disclosed basic information, the second script added objective factual disclosures and simple opinion outputs. The third script disclosure revealed more intense emotions and values from the voice assistant’s perspective. The word count across the phases of the three scripts is consistent (seen in Appendix B, Figure A1).
Regarding sound feature control, this experiment examines the impact of voice gender (male/female) on user self-disclosure intention. The average fundamental frequency (F0) for male Mandarin speakers is approximately 120 Hz, and for female speakers, it is around 210 Hz (). To control voice gender, the study employs iFLYTEK’s AI speech synthesis platform, generating audio with standard male (120 Hz) and female (210 Hz) Mandarin voices. The audio rate was 275 words per minute, with pauses strategically placed at the end of each question to allow participants time to think and respond. Each audio clip lasted approximately 1 min and 20 s (±2 s). To ensure participants could recognize the voice content, standard Mandarin pronunciation was used throughout the experiment.

3.2.2. Reciprocal Self-Disclosure Manipulation

To ensure that participants could clearly distinguish between the different disclosure strategies, a pre-test was conducted with three groups of 30 Chinese women. All of the participants had given birth within the past two years. The pre-test was conducted using the Credamo online platform. Participants were randomly assigned to different disclosure content scripts, and the self-disclosure level in each script was rated on a 1–7 scale based on the self-disclosure measurement by (). The results indicated that participants could effectively differentiate between the disclosure strategies used in the scripts (N = 30, M1 = 1.900; M2 = 3.733; M3 = 5.467, F = 90.214, p < 0.001). This suggests that the manipulation of disclosure strategies was successful in achieving the intended effect.

3.2.3. Variable Measurement

The measurements for perceived warmth and perceived competence were adapted from the SCM dimensions proposed by () and the scale applied by () for voice assistants. To measure perceived risk, perceived benefit, and disclosure intention, the privacy calculus theory measurements were primarily based on the privacy calculation scale used by () in health communities, the modified version by (), and Wheeless’s self-disclosure scale (). Given the experimental context, the scale was carefully screened, translated, and adjusted to ensure relevance. This process resulted in a final version consisting of 25 items, rated on a 7-point Likert scale, where a higher score indicates stronger agreement, as can be seen in Table 1.
Table 1. Research variables and corresponding measurement items.

3.3. Experiment Process

The study was conducted via the Credamo online platform. We recruited Chinese female participants aged 18–45 who had at least one child. Among them, participants who had given birth within the past two years were selected (; ; ). Upon starting the experiment, participants were randomly assigned one of six voice materials to listen to. After each segment, they were asked to provide a brief response. After the entire voice material had finished playing and the follow-up phone call concluded, participants filled out the questionnaire provided. The experiment concluded once the questionnaire was completed. Participants who exited the experiment prematurely or failed to listen to the voice material in its entirety were excluded from the data analysis. The experimental process is shown in Figure 2.
Figure 2. The experimental process.
The male/female voices here are based on specific voice characteristics derived from SCM and do not represent particular groups. The voice samples are used solely to stimulate the formation of stereotypes. Given that voice characteristics of other genders lack of clear definitions and do not show significant differences, this experiment does not consider voice characteristics of other genders. The data collection for the experiment started in July 2024 and continued until September 2024, with supplementary data collected in October 2024. All participants signed informed consent agreements. The ethics review of the experiment was conducted by the School of Design Art & Media Nanjing University of Science & Technology.

4. Result

4.1. Descriptive Statistics

A total of 407 valid responses were initially collected for this study. However, 27 respondents were excluded for failing the manipulation check, leaving a final sample size of 385. The participants were women aged 18 to 45, who had at least one child and had given birth within the past two years. Prior to testing the research hypotheses, a reliability analysis was performed on the 385 valid responses. The results indicated that all five variable factors had Cronbach’s α coefficients exceeding the recommended threshold of 0.7, confirming the good reliability of the measurement instruments. KMO = 0.823 (>0.8), and Bartlett’s test of sphericity was statistically significant (p < 0.01), demonstrating that the questionnaire possessed robust structural validity. All statistical analyses were conducted using IBM SPSS Statistics 27 software.

4.2. The Effect of Voice Assistant Gender and Self-Disclosure Strategies on Perceived Warmth Descriptive

A two-way ANOVA was conducted to explore the effects of voice assistant gender and self-disclosure strategies on users’ perceived warmth, as shown in Figure 3 and Figure 4. The analysis showed no statistically significant interaction between gender and self-disclosure strategies, F(2, 379) = 0.446, p = 0.640, indicating that these factors independently influenced perceived warmth. A simple effects analysis revealed that the gender of the voice assistant had a significant effect on perceived warmth. Users reported higher perceived warmth when listening to follow-up phone calls voiced by female assistants (M = 5.783, SD = 0.051) compared to male assistants (M = 5.489, SD = 0.054), F(1, 379) = 15.743, p < 0.001, supporting H5a. Self-disclosure strategies also significantly influenced perceived warmth (F(2, 379) = 4.512, p = 0.012), providing support for H6a. Specifically, users in the third self-disclosure strategy group (emotional and opinion-based disclosure) reported significantly higher perceived warmth (M = 5.790, SD = 0.062) compared to those in the first group (disclosure as usual) (M = 5.562, SD = 0.065, p = 0.039) and the second group (objective based disclosure) (M = 5.555, SD = 0.066, p = 0.037). However, no significant difference was observed between the first and second groups (p = 1).
Figure 3. Interaction effects between voice assistant gender and disclosure strategies on users’ perceived warmth: the disclosure strategy is on the X-axis.
Figure 4. Interaction effects between voice assistant gender and disclosure strategies on users’ perceived warmth: gender is on the X-axis.

4.3. The Effect of Voice Assistant Gender and Self-Disclosure Strategies on Perceived Competence

A two-way ANOVA was conducted to explore the effects of voice assistant gender and self-disclosure strategies on users’ perceived competence, as shown in Figure 5. The results indicated no statistically significant interaction between gender and self-disclosure strategies, F(2, 379) = 2.239, p = 0.108, supporting H5b. Further analysis of the main effects showed that voice assistant gender had a significant effect on perceived competence. Users perceived higher competence when listening to female-voiced follow-up phone calls (M = 5.896, SD = 0.042) compared to male-voiced calls (M = 5.767, SD = 0.044), F(2, 379) = 4.393, p = 0.037. However, self-disclosure strategies did not have a statistically significant effect on perceived competence (p = 0.125), leading to the rejection of H6b.
Figure 5. Interaction effects between voice assistant gender and disclosure strategies on users’ perceived competence.

4.4. The Direct Effects of Voice Assistant Gender and Self-Disclosure Strategies on Other Variables

A one-way ANOVA was conducted to examine the potential direct effects of voice assistant gender and self-disclosure strategies on users’ self-disclosure behaviors, including perceived benefits, perceived risks, and self-disclosure intention. The analysis revealed no statistically significant direct effects of voice assistant gender or self-disclosure strategies on any of these variables (p > 0.05), suggesting that these factors do not independently influence users’ perceptions or intention to disclose information.

4.5. Use Structural Equation Modeling (SEM) to Analyze the Relationship Between Stereotypes, Privacy Calculus, and Disclosure Intention

Based on the theoretical model, a structural equation modeling (SEM) path is constructed, as illustrated in Figure 6. Each factor has five variables measured by Likert scales. The study hypothesizes that perceived warmth and perceived competence directly influence perceived risks and perceived benefits. Furthermore, perceived risks and perceived benefits directly influence the user’s disclosure intention. Additionally, the influence of perceived competence and perceived warmth on disclosure intention was also explored.
Figure 6. Path diagram of the structural equation model (SEM).
Calculate the factor loadings for each variable, as shown in Table 2. For the factor warmth, all variables exhibit significant positive loadings, with standardized load factors ranging from 0.593 to 0.693. All variables of factor competence exhibit significant positive loadings, with standardized load factors ranging from 0.547 to 0.606. For the factor risks, all variables exhibit significant positive loadings, with standardized load factors ranging from 0.605 to 0.795. For the factor benefits, all variables exhibit significant positive loadings, with standardized load factors ranging from 0.605 to 0.795. For the factor intention, all variables exhibit significant positive loadings, with standardized loading coefficients ranging from 0.528 to 0.700. The z-values for every variable are far above the critical value, with p < 0.01, indicating that all variables meet the factor requirements.
Table 2. Standard load factor for each variable.
Model regression coefficient of each pathway was calculated, as can be seen in Table 3. The direct effect of perceived warmth on perceived risks (warmth → risks) is not significant (reject H3a) (B = −0.136, SE = 0.167, p = 0.222). The direct effect of perceived warmth on perceived benefits (warmth → benefits) is shown to be significantly positive (H3b) (B = 0.550, SE = 0.082, p < 0.01). The direct effect of perceived competence on perceived risks (competence → risks) is shown to be significantly negative (H4a) (B = −0.359, SE = 0.226, p < 0.01). The direct effect of perceived competence on perceived benefits (competence → benefits) shows a significant positive effect (H4b) (B = 0.403, SE = 0.105, p < 0.01). The direct effect of perceived risks on self-disclosure intention (risks → intention) shows a significant negative effect (H1) (B = −0.341, SE = 0.033, p < 0.01). The direct effect of perceived benefits on self-disclosure intention (benefits → intention) shows a significant positive effect (H2) (B = 0.817, SE = 0.232, p < 0.01). Moreover, the direct effects of perceived warmth and perceived competence on self-disclosure intention (warmth → intention; competence → intention) are not significant (B = −0.208, SE = 0.138, p = 0.135; B = 0.138, SE = 0.153, p = 0.256). The model demonstrates good fit indices (X2/df = 2.463 < 3, GFI = 0.833, RMSEA = 0.0062 < 0.08). The Figure 7 shows the path diagram of the structural equation model (SEM) with coefficients.
Table 3. Model regression coefficient of each pathway.
Figure 7. Path diagram of the structural equation model (SEM) with coefficients.

5. Discussion

5.1. Main Findings

This study examined the effects of the gender of AI voice assistants and their self-disclosure strategies on users’ perceived warmth, competence, risks, benefits, and self-disclosure intention in postpartum follow-up phone calls in China. Using two-way ANOVA, the results showed that the gender of the voice assistant significantly influenced perceived warmth and competence. Female voice assistants were rated significantly higher in terms of perceived warmth than their male counterparts, indicating that female voices are more effective at reducing psychological distance during follow-up phone calls. This effect aligns with previous research suggesting that female voices in robots are often associated with greater warmth and emotional resonance (). For perceived competence, female voice assistants also received higher ratings, suggesting that users perceived them as more effective and trustworthy in postpartum follow-up phone calls. This result contrasts with prior findings that associate male voices with competence and trustworthiness. The discrepancy may be attributed to the study’s participants, who were women aged 18 to 45 with childbirth experience—a demographic likely to attribute higher competence to females. Additionally, stereotype content models suggest that perceptions of warmth and competence vary across social contexts. Previous studies have demonstrated that in healthcare settings, users tend to prefer female robots, perceiving them as better suited for these tasks (). In this study, both the scenario of postpartum follow-up phone call and the healthcare context likely reinforced the perception that female assistants are more competent for such roles. The unique status of postpartum women is another factor that influences stereotypes. A study conducted in Hong Kong () revealed that postpartum women prefer services provided by groups that understand their specific needs. This may help explain why both postpartum follow-up tasks and healthcare-related scenarios lead users to associate female gender with a better performance in such roles in this study.
Regarding self-disclosure strategies, the results indicated that strategies involving emotional and opinion-based disclosures significantly enhanced perceived warmth compared to normal conversation or factual disclosures. High-intimacy self-disclosures fostered closer psychological connections with users, eliciting emotional resonance and enhancing perceptions of empathy, friendliness, and trustworthiness. Furthermore, emotional interactions may have increased the assistant’s perceived anthropomorphism, as suggested by (), who found that anthropomorphism in robots enhances perceived warmth. () further highlighted that when users perceive AI assistants as human-like, personal self-disclosures enhance intimacy and trust, while factual disclosures have similar effects when the assistants are perceived as machines. This implies that users in this study were more likely to perceive the follow-up assistants as human-like. However, self-disclosure strategies did not significantly influence perceived competence. This result indicates that competence judgments are more strongly associated with the assistant’s gender than with disclosure strategies. () suggested that, while perceived risk and perceived warmth are discussed as two dimensions of a theory, warmth judgments are the primary judgments, whereas competence judgments require more substantial evidence to change. Users’ initial perceptions of warmth are shaped by context, identity, and voice characteristics, while self-disclosure content primarily plays the role of reinforcing these initial impressions. Since competence judgments are less malleable than warmth perceptions, this explains why self-disclosure strategies significantly affected perceived warmth but not perceived competence.
Although the gender and self-disclosure strategies of voice assistants significantly influenced stereotype formation, they did not directly affect users’ self-disclosure, including perceived risks, benefits and intention. This indicates that while gender and disclosure strategies affect users’ perception, they do not directly determine whether users disclose information. Through structural equation modeling (SEM) analysis, the study found that perceived warmth and perceived competence do not directly influence participants’ self-disclosure intention; instead, their effects are mediated by perceived risks and perceived benefits. Specifically, perceived warmth and perceived competence are positively correlated with perceived benefits, while perceived competence is negatively correlated with perceived risks. However, the impact of perceived warmth on perceived risks was found to be non-significant. Users who perceived the assistant as warm or competent were less likely to perceive risks and more likely to expect benefits.
Notably, the effect of perceived benefits was significantly stronger than that of perceived risks, suggesting that users prioritized benefit evaluations over risks concerns when deciding to disclose information. This finding aligns with Social Exchange Theory, which posits that users are more willing to disclose personal information when they anticipate high benefits (e.g., better services or emotional support), even in the presence of certain risks. In the scenario of postpartum follow-up phone calls in China, the relatively low perceived risks of disclosure may reflect users’ trust in medical institutions and the cooperative nature of doctor–patient relationships. In Chinese culture, doctors are generally associated with positive stereotypes (). A study in China on people’s perceptions of doctors’ stereotypes showed that the public holds a positive stereotype of doctors at a rate of 68.97%. However, 31.03% of respondents stereotyped doctors negatively, citing stress, attitude toward patients, medical competence, and ethical concerns as their reasons for this view; privacy-related issues were not mentioned (). In this experiment, perceived warmth did not significantly affect perceived risks. This could be attributed to the fact that Chinese users are likely to associate the negative stereotypes about doctors (such as attitude toward patients, medical competence, and ethics) with a lack of professional ability. Given the positive stereotype of doctors in China, users are more likely to associate the perceived warmth of the voice with personal benefits, while judging the risk of disclosing information based on the perceived competence of the voice.
On one hand, it is widely acknowledged that China is experiencing a low fertility rate (). In 2022, 9.56 million births were recorded, corresponding to a birth rate of 6.77%. On the other hand, pregnant women in China are proving increasingly willing to pay for higher-quality medical services (). The number of hospital discharges from primary care institutions decreased from 4.795 million in 2009 to 1.473 million in 2021, a decline of 69.3%. The low fertility rate, high single-child rate, and the growing demand for high-quality healthcare have led to behaviors driven by the desire for perceived benefits. Furthermore, with Confucianism as its cultural foundation, China is experiencing a shift toward modernized thinking. Women are often expected to maintain composure for the sake of their children, frequently suppressing their own needs and emotions, which can result in isolation, anxiety, and depression (). The process of adapting to the role of a postpartum mother, coupled with external social pressures and the demands of newborn care, can contribute to significant psychological stress (). This explains why postpartum women are more likely to engage in proactive behaviors to seek perceived benefits, such as social and psychological support.

5.2. Significance

5.2.1. Academic Value

This study contributes to the CASA framework by exploring the interactions between SCM and privacy calculus theory—two widely acknowledged theories whose relationship has been seldom explored. By applying SCM and privacy calculus theory to the postpartum follow-up scenario, this study affirms the relevance of these theories in this specialized setting. It provides empirical evidence that these theories interact within specific scenario with AI voice assistants, extending the applicability of the CASA framework.
This study emphasizes the importance of voice assistant gender and self-disclosure strategies in shaping how users interact with the AI technology. In particular, it shows that female voice assistants and emotionally oriented disclosure strategies significantly enhance users’ perception of warmth, which indirectly increases their self-disclosure intention. Specifically, emotionally and opinion-based disclosures improve perceived warmth and indirectly affect self-disclosure intention among target users. The disclosure strategy of AI voice assistants has rarely been examined as a factor influencing stereotype formation, making this study a valuable addition to this field.
Furthermore, the results indicate that users’ disclosure behavior is largely motivated by perceived benefits. It suggests that users are more focused on the potential advantages of interacting with the system rather than on associated risks. This finding underscores the unique, context-dependent influences of social environment and specific interaction settings on self-disclosure intentions with voice assistants. By revealing these contextual characteristics, this study provides empirical support for expanding theories on self-disclosure and privacy calculus. This study helps enhance the understanding of human–machine interaction dynamics and disclosure mechanisms in technology-mediated communication.

5.2.2. Practical Value

From a design standpoint, these findings offer important guidance for optimizing voice assistant functionality. In postpartum follow-up phone calls, both perceived warmth and competence positively influence users’ self-disclosure intention. This suggests that any design strategies that enhance these qualities merit exploration. Specifically, when designing voice assistants for postpartum follow-up, it is recommended to prioritize female voice characteristics and emotional and opinion-based disclosure strategies to increase users’ perceived warmth and perceived benefits.
Additionally, this study reveals that, within the Chinese healthcare context, users are more likely to be motivated by perceived benefits, providing hospitals with insights for refining patient communication approaches. This is a useful finding for healthcare services, implying that emphasizing service benefits may be more effective in encouraging user disclosure than focusing solely on risk reduction. This finding also reflects aspects of Chinese unique doctor–patient relationship dynamics, offering valuable reference points for researching healthcare communication in this specific social context. Finally, this study contributes to characterizing the user profile of medical service recipients in China. In particular, it emphasizes that there are many postpartum users who tend to overlook perceived risks in healthcare interactions. This insight suggests that healthcare propaganda should focus on this group to raise their awareness of the risks when disclosure and protect their interests.
By examining user interactions and preferred communication modes, this study aims to facilitate the adoption of AI voice assistants in healthcare scenarios. Since the interaction context plays a crucial role in shaping user stereotypes, future research could further investigate how gender and emotional expression in voice assistants may be optimized across different scenarios to enhance user engagement and trust in AI technologies.
Finally, although this study was conducted in China, the findings hold significant application value in contexts with different cultural backgrounds. Many developing countries are in the same situation as China. They are undergoing a shift toward modernization and the diversification of medical services, which is influenced by traditional cultural stereotypes (). Therefore, the role of perceived benefits in driving self-disclosure should be emphasized. The diversity of the medical workforce is increasing. According to data from the American Medical Association, 30% of practicing physicians belong to minority groups, and 36% are women (). These findings suggest that stereotypes related to women’s voices in specific scenarios may lead to heightened perceived competence. In low-income countries, the medical resources are insufficient to provide comprehensive healthcare. Follow-up phone calls can be a cost-effective approach to help alleviate this issue. The application of AI also addresses the labor-intensive challenges associated with telephone follow-up (). Further research into the application of voice assistant self-disclosure strategies in these regions could help doctors acquire information more efficiently.

6. Conclusions

This study underscores the significance of gender characteristics and self-disclosure strategies in shaping users’ experiences interacting with voice assistants during postpartum follow-up phone calls. The findings reveal that female voice assistants enhance users’ perceived warmth and competence, thereby effectively increasing their self-disclosure intention. Emotional self-disclosure strategies also have a positive effect on perceived warmth to encourage users to disclose. This study highlights the pivotal role of users’ perceived warmth benefits in making self-disclosure decisions.
The study has several limitations. First, the experimental design did not account for scenarios where users refuse to answer or interrupt calls because of self-protection mechanisms, reflecting a broader limitation of the privacy calculus theory. Second, the study primarily focused on postpartum users across China, resulting in a relatively broad target population. Given the potential variations in stereotype formation across different social and cultural contexts, the findings may exhibit regional differences within China and the limited generalizability of global contexts. Moreover, postpartum women’s physiological changes over time may influence the outcomes. Validating these findings requires a larger dataset. Future studies could investigate postpartum women’s medical service experiences from a longitudinal perspective or at different time intervals.
In future research, more specific social contexts need to be investigated and the scenarios need to be better tailored, ensuring the practical applicability of the AI voice assistant. Additionally, attention should be paid to the phenomena of call refusal and interruption. These scenarios deserve further investigation to understand their causes and effects. Future studies should focus on optimizing the design of voice assistants to improve user experience and trust across diverse cultural and healthcare environments. By implementing these strategies, user needs can be better met, and medical services can be further improved and developed.

Author Contributions

X.S.: Conceptualization, Investigation, Methodology, Writing—review and editing, Project administration, Investigation; T.S.: Writing—original draft, Writing—review and editing, Investigation, Data curation; Q.J.: Validation, Resources, Investigation; B.J.: Funding acquisition, Project administration, Supervision. All authors have read and agreed to the published version of the manuscript.

Funding

The work is supported by Major Projects in Philosophy and Social Sciences of Higher Education Institutions in Jiangsu Province (No. 2024SJZD101).

Institutional Review Board Statement

The study was conducted in accordance with the Declaration of Helsinki and approved by the Institutional Review Board of School of Design Art & Media Nanjing University of Science & Technology (protocol code NO.NJUST SDAM-2024-0023 on 3 April 2024).

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors on request.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A. Rating Scheme for Self-Disclosure

Rating Scheme for Self-Disclosure ()

Table A1. Rating scheme for self-disclosure (in Chinese and English).
Table A1. Rating scheme for self-disclosure (in Chinese and English).
自我披露种类 (Self-Disclosure Category)分值 (Rating)
基础 (Base)1
简单易得的信息 (Simple and visible information)0到1 (0 to 1)
态度和观点 (Attitudes and opinions)0到2 (0 to 2)
强烈的情感,价值观和难以取得的信息
(Strong affections, basic values, and less visible information)
0到3 (0 to 3)
总计(Total) 1到7 (1 to 7)
注释;评分基于分类中自我披露的亲密程度有所变化. (Note: The ratings varied depending on the level of self-disclosure intimacy within the category.).

Appendix B. Postpartum Follow-Up Voice Script

Appendix B.1. Postpartum Follow-Up Voice Script (In Chinese)

Figure A1. Postpartum follow-up voice script (in Chinese).

Appendix B.2. Postpartum Follow-Up Voice Script (In English)

Figure A2. Postpartum follow-up voice script (in English).

Appendix C. Interview Outline

  • Introduction
Self-Introduction: Briefly introduce the study purpose, background, and structure of the interview.
Confidentiality and Informed Consent: Clarify that the interview content will only be used for research purposes and that the respondent’s personal information will be kept strictly confidential.
Interview Duration: Inform the respondent about the estimated interview time and compensation (30 min, RMB 100).
2.
Collection of Basic Information about the Doctor
What is your name and age?
How long have you been working in obstetrics and gynecology?
Have you ever participated in postpartum follow-up phone calls? What other related work have you done?
3.
Implementation of Postpartum Follow-up phone calls
What is the process for postpartum telephone follow-ups at your medical institution? Are there standardized procedures and guidelines?
What are the main contents of postpartum follow-up calls? What aspects are covered (e.g., physical recovery, emotional support, parenting advice)? What information needs to be obtained?
Are postpartum follow-up calls typically conducted by doctors directly or by other staff members (e.g., nurses, dedicated personnel)?
Have you received any training related to postpartum telephone follow-ups? Are there any details that need special attention?
What should be the tone, speed, and attitude during follow-up calls?
4.
Evaluation of the Effectiveness of Postpartum Follow-up phone calls
What objectives do you think postpartum telephone follow-ups should achieve?
Based on your observations, do postpartum women better understand and manage postpartum issues such as physical recovery and emotional changes?
Have you received feedback from postpartum women regarding follow-up calls? What are the main points of this feedback?
Have any postpartum women mentioned that the follow-up calls failed to address their concerns effectively? Which areas do these complaints typically focus on?
5.
Continuous Improvement and Challenges
In your opinion, what aspects of postpartum follow-up phone calls need to be strengthened or improved? Are there any urgent issues that require attention?
What is your opinion on using AI voice assistants for follow-up phone calls? What do you think are the advantages and challenges?
6.
Conclude the Interview and Express Gratitude

References

  1. Altman, I., & Taylor, D. A. (1973). Social penetration: The development of interpersonal relationships. Available online: https://psycnet.apa.org/record/1973-28661-000 (accessed on 15 July 2024).
  2. Archer, R. L., & Berg, J. H. (1978). Disclosure reciprocity and its limits: A reactance analysis. Journal of Experimental Social Psychology, 14(6), 527–540. [Google Scholar] [CrossRef]
  3. Augustin, Y., Carolus, A., & Wienrich, C. (2022). Privacy of AI-based voice assistants: Understanding the users’ perspective: A purposive review and a roadmap for elevating research on privacy from a user-oriented perspective. In International Conference on Human-Computer Interaction (pp. 309–321). Springer International Publishing. [Google Scholar] [CrossRef]
  4. Bansal, G., & Gefen, D. (2010). The impact of personal dispositions on information sensitivity, privacy concern and trust in disclosing health information online. Decision Support Systems, 49(2), 138–150. [Google Scholar] [CrossRef]
  5. Brodribb, W., Zadoroznyj, M., & Dane, A. (2013). The views of mothers and GPs about postpartum care in Australian general practice. BMC Family Practice, 14(1), 139. [Google Scholar] [CrossRef]
  6. Bryant, A., Blake-Lamb, T., Hatoum, I., & Kotelchuck, M. (2016). Women’s use of health care in the first 2 years postpartum: Occurrence and correlates. Maternal and Child Health Journal, 20(S1), 81–91. [Google Scholar] [CrossRef] [PubMed]
  7. Chan, Z. C., Wong, K. S., Lam, W. M., Wong, K. Y., & Kwok, Y. C. (2013). An exploration of postpartum women’s perspective on desired obstetric nursing qualities. Journal of Clinical Nursing, 23(1–2), 103–112. [Google Scholar] [CrossRef] [PubMed]
  8. Choi, B., Wu, Y., Yu, J., & Land, L. (2018). Love at First Sight: The interplay between privacy dispositions and privacy calculus in online social connectivity management. Journal of the Association for Information Systems, 19(3), 124–151. [Google Scholar] [CrossRef]
  9. Cloarec, J. (2020). The personalization–privacy paradox in the attention economy. Technological Forecasting and Social Change, 161, 120299. [Google Scholar] [CrossRef]
  10. Collins, N. L., & Miller, L. C. (1994). Self-disclosure and liking: A meta-analytic review. Psychological Bulletin, 116(3), 457. [Google Scholar] [CrossRef] [PubMed]
  11. Cuddy, A. J., Fiske, S. T., & Glick, P. (2008). Warmth and competence as universal dimensions of social perception: The stereotype content model and the BIAS map. Advances in Experimental Social Psychology, 40, 61–149. [Google Scholar] [CrossRef]
  12. Culnan, M. J., & Armstrong, P. K. (1999). Information privacy concerns, procedural fairness, and impersonal trust: An empirical investigation. Organization Science, 10(1), 104–115. [Google Scholar] [CrossRef]
  13. Dwyer, T., Hoit, G., Burns, D., Higgins, J., Chang, J., Whelan, D., Kiroplis, I., & Chahal, J. (2023). Use of an artificial intelligence conversational agent (Chatbot) for hip arthroscopy patients following surgery. Arthroscopy Sports Medicine and Rehabilitation, 5(2), e495–e505. [Google Scholar] [CrossRef] [PubMed]
  14. Ernst, C.-P. H., Herm-Stapelberg, N., Association for Information Systems & AIS Electronic Library (AISeL). (2020). The impact of gender stereotyping on the perceived likability of virtual assistants. AMCIS 2020 Proceedings, 4. Available online: https://core.ac.uk/download/pdf/326836357.pdf (accessed on 2 July 2024).
  15. Esposito, A., Amorese, T., Cuciniello, M., Riviello, M. T., Esposito, A. M., Troncone, A., Torres, M. I., Schlögl, S., & Cordasco, G. (2019). Elder user’s attitude toward assistive virtual agents: The role of voice and gender. Journal of Ambient Intelligence and Humanized Computing, 12(4), 4429–4436. [Google Scholar] [CrossRef]
  16. Fan, Z. (2023). Research on measurement and psychological formation mechanism of professional stigma in contemporary Chinese doctors [Doctoral dissertation, Changchun University of Traditional Chinese Medicine]. (In Chinese). [Google Scholar]
  17. Fiske, S. T., Cuddy, A. J. C., & Glick, P. (2006). Universal dimensions of social cognition: Warmth and competence. Trends in Cognitive Sciences, 11(2), 77–83. [Google Scholar] [CrossRef] [PubMed]
  18. Fiske, S. T., Cuddy, A. J. C., Glick, P., & Xu, J. (2002). A model of (often mixed) stereotype content: Competence and warmth respectively follow from perceived status and competition. Journal of Personality and Social Psychology, 82(6), 878–902. [Google Scholar] [CrossRef] [PubMed]
  19. Frith, C. D., & Singer, T. (2008). The role of social cognition in decision making. Philosophical Transactions of the Royal Society B: Biological Sciences, 363(1511), 3875–3886. [Google Scholar] [CrossRef]
  20. Fu, J., Bian, X., Zhang, H., & Pian, X. (2011). Intervention impact of phone follow-up on hospital quality promotion. Chinese Hospitals, 15(9), 38–39. (In Chinese) [Google Scholar] [CrossRef]
  21. Gerlach, J., Buxmann, P., & Dinev, T. (2019). “They’re all the same!” stereotypical thinking and systematic errors in users’ privacy-related judgments about online services. Journal of the Association for Information Systems, 20(6), 4. [Google Scholar] [CrossRef]
  22. Grimmelikhuijsen, S., Herkes, F., Leistikow, I., Verkroost, J., De Vries, F., & Zijlstra, W. G. (2019). Can decision transparency increase citizen trust in regulatory agencies? Evidence from a representative survey experiment. Regulation & Governance, 15(1), 17–31. [Google Scholar] [CrossRef]
  23. Hadi, A. A., Paliwoda, M., Dawson, J., Walker, K., & New, K. (2022). Women’s utilisation, experiences and satisfaction with postnatal follow-up care. Sultan Qaboos University Medical Journal, 22(4), 455–471. [Google Scholar] [CrossRef]
  24. Handler, S. M., Boyce, R. D., Ligons, F. M., Perera, S., Nace, D. A., & Hochheiser, H. (2013). Use and perceived benefits of mobile devices by physicians in preventing adverse drug events in the nursing home. Journal of the American Medical Directors Association, 14(12), 906–910. [Google Scholar] [CrossRef] [PubMed]
  25. Herzog, O., & Rögner, K. (2022, August 29–September 2). Let me introduce myself-using self-disclosure as a social cue for health care robots. 2022 31st IEEE International Conference on Robot and Human Interactive Communication (RO-MAN) (pp. 1358–1364), Napoli, Italy. [Google Scholar] [CrossRef]
  26. Ho, A., Hancock, J., & Miner, A. S. (2018). Psychological, relational, and emotional effects of self-disclosure after conversations with a chatbot. Journal of Communication, 68(4), 712–733. [Google Scholar] [CrossRef] [PubMed]
  27. Igarashi, T., Nihei, M., Inoue, T., Sugawara, I., & Kamata, M. (2022). Eliciting a user’s preferences by the self-disclosure of socially assistive robots in local households of older adults to facilitate verbal human–robot interaction. International Journal of Environmental Research and Public Health, 19(18), 11319. [Google Scholar] [CrossRef] [PubMed]
  28. Iqbal, M. Z., & Campbell, A. G. (2021). From luxury to necessity: Progress of touchless interaction technology. Technology in Society, 67, 101796. [Google Scholar] [CrossRef]
  29. Jack, B. W., Bickmore, T., Yinusa-Nyahkoon, L., Reichert, M., Julce, C., Sidduri, N., Martin-Howard, J., Zhang, Z., Woodhams, E., Fernandez, J., Loafman, M., & Cabral, H. J. (2020). Improving the health of young African American women in the preconception period using health information technology: A randomised controlled trial. The Lancet Digital Health, 2(9), e475–e485. [Google Scholar] [CrossRef] [PubMed]
  30. Jain, P. (2020). The stereotype content model as an explanation of biased perceptions in a medical interaction: Implications for Patient-Provider relationship. Health Communication, 37(1), 64–73. [Google Scholar] [CrossRef] [PubMed]
  31. Jin, S. V., & Youn, S. (2021). Why do consumers with social phobia prefer anthropomorphic customer service chatbots? Evolutionary explanations of the moderating roles of social phobia. Telematics and Informatics, 62, 101644. [Google Scholar] [CrossRef]
  32. Keith, M., Thompson, S., Hale, J., & Greer, C. (2012). Examining the rationality of location data disclosure through mobile devices. Available online: https://aisel.aisnet.org/icis2012/proceedings/ISSecurity/8/ (accessed on 4 July 2024).
  33. Kim, S. Y., Schmitt, B. H., & Thalmann, N. M. (2019). Eliza in the uncanny valley: Anthropomorphizing consumer robots increases their perceived warmth but decreases liking. Marketing Letters, 30, 1–12. [Google Scholar] [CrossRef]
  34. Kim, Y. (2024). AI health counselor’s self-disclosure: Its impact on user responses based on individual and cultural variation. International Journal of Human–Computer Interaction, 40, 1–14. [Google Scholar] [CrossRef]
  35. Kurzban, R., Tooby, J., & Cosmides, L. (2001). Can race be erased? Coalitional computation and social categorization. Proceedings of the National Academy of Sciences, 98(26), 15387–15392. [Google Scholar] [CrossRef]
  36. Lass, N. J., Hughes, K. R., Bowyer, M. D., Waters, L. T., & Bourne, V. T. (1976). Speaker sex identification from voiced, whispered, and filtered isolated vowels. The Journal of the Acoustical Society of America, 59(3), 675–678. [Google Scholar] [CrossRef] [PubMed]
  37. Laufer, R. S., & Wolfe, M. (1977). Privacy as a concept and a social issue: A multidimensional developmental theory. Journal of social Issues, 33(3), 22–42. [Google Scholar] [CrossRef]
  38. Li, H., Sarathy, R., & Xu, H. (2011). The role of affect and cognition on online consumers’ decision to disclose personal information to unfamiliar online vendors. Decision Support Systems, 51(3), 434–445. [Google Scholar] [CrossRef]
  39. Li, M., Guo, F., Ren, Z., & Wang, X. (2023). Users’ affective preference for humanoid robot in service area and its relations with voice acoustic parameters. Industrial Engineering and Management, 28(6), 145–153. (In Chinese) [Google Scholar] [CrossRef]
  40. Liu, L. (2016). Effect analysis of maternal character, pregnancy timing and mode of delivery on postpartum depression. Chinese Medical Sciences, 6(10), 60–63. (In Chinese). [Google Scholar]
  41. Lucas, G. M., Gratch, J., King, A., & Morency, L. P. (2014). It’s only a computer: Virtual humans increase willingness to disclose. Computers in Human Behavior, 37, 94–100. [Google Scholar] [CrossRef]
  42. Martins, M. V., Peterson, B. D., Costa, P., Costa, M. E., Lund, R., & Schmidt, L. (2013). Interactive effects of social support and disclosure on fertility-related stress. Journal of Social and Personal Relationships, 30(4), 371–388. [Google Scholar] [CrossRef]
  43. Mikulincer, M., & Nachshon, O. (1991). Attachment styles and patterns of self-disclosure. Journal of Personality and Social Psychology, 61(2), 321–331. [Google Scholar] [CrossRef]
  44. Moon, Y. (2000). Intimate exchanges: Using computers to elicit self-disclosure from consumers. Journal of Consumer Research, 26(4), 323–339. [Google Scholar] [CrossRef]
  45. Nass, C., Moon, Y., & Green, N. (1997). Are machines gender neutral? Gender-stereotypic responses to computers with voices. Journal of Applied Social Psychology, 27(10), 864–876. [Google Scholar] [CrossRef]
  46. Painuly, S., Sharma, S., & Matta, P. (2023, March 23–25). Artificial intelligence in e-healthcare supply chain management system: Challenges and future trends. 2023 International Conference on Sustainable Computing and Data Communication Systems (ICSCDS) (pp. 569–574), Erode, India. [Google Scholar] [CrossRef]
  47. Pentina, I., Hancock, T., & Xie, T. (2022). Exploring relationship development with social chatbots: A mixed-method study of replika. Computers in Human Behavior, 140, 107600. [Google Scholar] [CrossRef]
  48. Pepper, M., Campbell, O. M., & Woodd, S. L. (2024). Current approaches to following up women and newborns after discharge from childbirth facilities: A scoping review. Global Health Science and Practice, 12(2), e2300377. [Google Scholar] [CrossRef]
  49. Pietraszewski, D., Curry, O. S., Petersen, M. B., Cosmides, L., & Tooby, J. (2015). Constituents of political cognition: Race, party politics, and the alliance detection system. Cognition, 140, 24–39. [Google Scholar] [CrossRef] [PubMed]
  50. Post Discharge Follow-Up Phone Call Documentation Form. (n.d.). Available online: https://www.ahrq.gov/patient-safety/settings/hospital/red/toolkit/redtool5.html (accessed on 19 July 2024).
  51. Roter, D. L., & Hall, J. A. (1987). Physicians’ interviewing styles and medical information obtained from patients. Journal of General Internal Medicine, 2(5), 325–329. [Google Scholar] [CrossRef]
  52. Saffarizadeh, K., Keil, M., Boodraj, M., & Alashoor, T. (2024). “My name is alexa. what’s your name?” The impact of reciprocal self-disclosure on post-interaction trust in conversational agents. Journal of the Association for Information Systems, 25(3), 528–568. [Google Scholar] [CrossRef]
  53. Schuetzler, R. M., Grimes, G. M., Giboney, J. S., & Nunamaker, J. F., Jr. (2018, January 3–6). The influence of conversational agents on socially desirable responding. 51st Hawaii International Conference on System Sciences (p. 283), Waikoloa Village, HI, USAAvailable online: https://digitalcommons.unomaha.edu/isqafacpub/61 (accessed on 1 July 2024).
  54. Sezgin, E., Huang, Y., Ramtekkar, U., & Lin, S. (2020). Readiness for voice assistants to support healthcare delivery during a health crisis and pandemic. NPJ Digital Medicine, 3(1), 122. [Google Scholar] [CrossRef]
  55. Sprecher, S., Treger, S., Wondra, J. D., Hilaire, N., & Wallpe, K. (2013). Taking turns: Reciprocal self-disclosure promotes liking in initial interactions. Journal of Experimental Social Psychology, 49(5), 860–866. [Google Scholar] [CrossRef]
  56. Su, F., Wang, Y., Wu, Q., Wang, P. J., & Chang, X. (2022). The influence of stereotypes on trust in doctors from patients’ perspective: The mediating role of communication. Psychology Research and Behavior Management, 15, 3663–3671. [Google Scholar] [CrossRef]
  57. Sung, B., Im, H., & Duong, V. C. (2023). Task type’s effect on attitudes towards voice assistants. International Journal of Consumer Studies, 47(5), 1772–1790. [Google Scholar] [CrossRef]
  58. Tam, T., Hewstone, M., Harwood, J., Voci, A., & Kenworthy, J. (2006). Intergroup contact and grandparent–grandchild communication: The effects of self-disclosure on implicit and explicit biases against older people. Group Processes & Intergroup Relations, 9(3), 413–429. [Google Scholar] [CrossRef]
  59. Tang, D., & Zhang, S. (2023). The research status and implications of medical insurance impact on neonatal health in a low-fertility setting. Chinese Health Economics, 42(6), 26–29, 39. (In Chinese). [Google Scholar]
  60. Traunmüller, H., & Eriksson, A. (1995). The frequency range of the voice fundamental in the speech of male and female adults [Unpublished manuscript]. Available online: https://citeseerx.ist.psu.edu/document?repid=rep1&type=pdf&doi=dcf068d9e9604d7266152b620f1c5dfa7fa45349 (accessed on 2 July 2024).
  61. Umar, N., Hill, Z., Schellenberg, J., Tuncalp, Ö., Muzigaba, M., Sambo, N. U., Shuaibu, A., & Marchant, T. (2023). Women’s perceptions of telephone interviews about their experiences with childbirth care in Nigeria: A qualitative study. PLoS Global Public Health, 3(4), e0001833. [Google Scholar] [CrossRef]
  62. Vcbeat. (2024). 2024 Medical Artificial Intelligence Report: Generative AI explodes and medical AI comes to a new crossroads. Available online: https://www.vbdata.cn/reportCard/72385 (accessed on 17 January 2025). Available online.
  63. Vogel, D. L., & Wester, S. R. (2003). To seek help or not to seek help: The risks of self-disclosure. Journal of Counseling Psychology, 50(3), 351. [Google Scholar] [CrossRef]
  64. Wang, L., Wang, D., Tian, F., Peng, Z., Fan, X., Zhang, Z., Yu, M., Ma, X., & Wang, H. (2021). CASS: Towards building a social-support chatbot for online health community. Proceedings of the ACM on Human-Computer Interaction, 5(CSCW1), 1–31. [Google Scholar] [CrossRef]
  65. Wang, T., Duong, T. D., & Chen, C. C. (2016). Intention to disclose personal information via mobile applications: A privacy calculus perspective. International Journal of Information Management, 36(4), 531–542. [Google Scholar] [CrossRef]
  66. Wang, Y., Zhou, Y., & Liao, Z. (2021). Health privacy information self-disclosure in online health community. Frontiers in Public Health, 8, 602792. [Google Scholar] [CrossRef]
  67. Wen, F., Hang, S., Ju, Y., & Wang, J. (2023). Cross-category Stereotype Content between Intimate Relationship Status and Gender. Chinese Social Psychology Review, 24, 66–90. Available online: https://xianxiao.ssap.com.cn/catalog/6739960.html (accessed on 1 July 2024). (In Chinese).
  68. Wheeless, L. R., & Grotz, J. (1976). Conceptualization and measurement of reported self-disclosure. Human Communication Research, 2(4), 338–346. [Google Scholar] [CrossRef]
  69. Williams, C. (2023). Effectiveness of follow up phone calls on postpartum women after discharge: A program evaluation [Doctoral dissertation, University of Missouri-Saint Louis]. Available online: https://search.proquest.com/openview/9bd3b9abcf8f6bb08e5d971cb58f0a8d/1?pq-origsite=gscholar&cbl=18750&diss=y (accessed on 19 July 2024).
  70. Wojciszke, B., Bazinska, R., & Jaworski, M. (1998). On the dominance of moral categories in impression formation. Personality and Social Psychology Bulletin, 24(12), 1251–1263. [Google Scholar] [CrossRef]
  71. Woods, C. E., Jones, R., O’Shea, E., Grist, E., Wiggers, J., & Usher, K. (2019). Nurse-led postdischarge telephone follow-up calls: A mixed study systematic review. Journal of Clinical Nursing, 28(19–20), 3386–3399. [Google Scholar] [CrossRef]
  72. Xu, H., Teo, H. H., Tan, B. C., & Agarwal, R. (2009). The role of push-pull technology in privacy calculus: The case of location-based services. Journal of Management Information Systems, 26(3), 135–174. [Google Scholar] [CrossRef]
  73. Yao, Y. (2009). Analysis on the satisfaction of discharged patients by telephone follow-up. Journal of Nursing Science, 24(1), 62–63. (In Chinese) [Google Scholar] [CrossRef]
  74. Zhang, Q., Wang, S., & Chen, L. (2021). The application of intelligent speech technology in electronic medical records of outpatient. China Digital Medicine, 16(8), 12–16. (In Chinese) [Google Scholar] [CrossRef]
  75. Zhang, X., Liu, S., Chen, X., Wang, L., Gao, B., & Zhu, Q. (2017). Health information privacy concerns, antecedents, and information disclosure intention in online health communities. Information & Management, 55(4), 482–493. [Google Scholar] [CrossRef]
  76. Zhang, X., Ye, Y., Cao, X., Liu, Y., Peng, B., & Zhu, K. (2024). Maternal health care services in primary medical institutions in six regionally representative provinces of China, 2021: A submitted data-based analysis. Chinese Journal of Public Health, 40(2), 169–174. [Google Scholar] [CrossRef]
  77. Zhang, Y. (2024). Gender stereotypes and mental health among Chinese people: The moderating role of marital status. Asia Pacific Journal of Social Work and Development, 34, 1–16. [Google Scholar] [CrossRef]
  78. Zhu, X. (2017). Research progress of multiple postpartum visits. World Latest Medicine Information (CD-ROM), 17(30), 36–38. Available online: https://med.wanfangdata.com.cn/Paper/Detail?id=PeriodicalPaper_sjzxyy-e201730016&dbid=WF_QK (accessed on 17 July 2024). (In Chinese).
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Article Metrics

Citations

Article Access Statistics

Multiple requests from the same IP address are counted as one view.