Next Article in Journal
Visual Threats and Visual Efficacy: Ideas of Image Reception in the Arguments of Lucas Tudense about the Changes in the Crucifixion (c.1230)
Next Article in Special Issue
Gratitude to God: Brief Prompts Do Not Increase It, Wording of Questions Matters, and Belief in a Loving, Powerful, Gift-Giving God Remains Central
Previous Article in Journal
New Age Healing: Origins, Definitions, and Implications for Religion and Medicine
Previous Article in Special Issue
Building on a Solid Foundation: Conceptual Mapping Informs Schemas for Relating to God
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Using Social Media to Assess Expressions of Gratitude to God: Issues for Consideration

1
Department of Psychological Sciences, Purdue University, West Lafayette, IN 47907, USA
2
Department of Psychiatry and Behavioral Sciences, University of California, San Francisco, CA 94107, USA
3
School of Interactive Computing, Georgia Institute of Technology, Atlanta, GA 30332, USA
*
Author to whom correspondence should be addressed.
Religions 2022, 13(9), 778; https://doi.org/10.3390/rel13090778
Submission received: 17 June 2022 / Revised: 17 August 2022 / Accepted: 22 August 2022 / Published: 25 August 2022
(This article belongs to the Special Issue Gratitude to God)

Abstract

:
With the proliferation of technology-based communication, public expressions of gratitude to God on social media have become more pervasive. At the same time, data science approaches are increasingly being applied to social media language data to assess positive human attributes. We elucidate critical considerations in assessing public expressions of gratitude to God, including language variability and comparability, degree of authenticity, machine learning language analysis, and aggregation approaches that could affect assessment accuracy.

1. Introduction

Expressing gratitude to God is deeply rooted in many different religious traditions that call for adherents to develop and cultivate it as moral virtue (Emmons and Crumpler 2000). Recent work also points to how gratitude to God is associated with well-being outcomes such as better health, greater subjective well-being, and lower depressive symptoms (Aghababaei and Tabik 2013; Krause et al. 2014; Krause et al. 2015; Rosmarin et al. 2011).
Research on gratitude to God has been built upon the self-report survey paradigm (e.g., Krause 2006; McCullough et al. 2002), which has unique strengths such as allowing individuals to report on their own experiences and intentions. The administration of self-report questionnaires is fairly routine, and many participants are familiar with this method. Moreover, self-report assessments provide researchers with the flexibility and ease to examine precise measures and specific psychological mechanisms. Because the tradition of self-report has existed for many years, numerous resources that provide recommendations and guidance on construct validity exist (Cronbach and Meehl 1955; Loevinger 1957; Messick 1980; Messick 1995; Cronbach and Meehl 1955; Loevinger 1957; Messick 1980, 1995). Self-report methods can also be used in a variety of different methods, such as experiments, cross-sectional studies, longitudinal designs, and daily diary and ecological momentary assessment methods.
Nevertheless, self-report surveys have several weaknesses. For example, they do not provide observable expressions of gratitude. As aligned with the Theory of Planned Behavior (Ajzen 1991), intentions to express gratitude do not always veridically emerge as gratitude expression behaviors. More broadly, self-ratings have multiple problems stemming from a lack of self-awareness, misremembering, or the use of heuristics that bias their accuracy (Schwarz 1999). Desirable responding may also influence participants’ responses, resulting in reports that may not match their true beliefs or behaviors (Nederhof 1985). These examples highlight a broader principle, namely that all methods are flawed in various manners and have their own strengths and weaknesses (McGrath 1981). Therefore, the best approach in moving gratitude to God research forward is to rely on various, potentially complementary methods and acknowledge their own strengths and weaknesses.
One method that can be used to complement self-report methods is the use of text analytic techniques of social media language. This appears to be a promising candidate because social media applications are widely used by people across the world. It is estimated that 4.48 billion people worldwide use social media as of 2021 (Deen 2021). Importantly, from a cursory glance at social media posts, people often express gratitude to God on these platforms (e.g., ‘thank god’, ‘praise god’), and these instances provide directly observable behaviors that happen in a naturalistic setting without researcher solicitation. Our team applied search terms (20 key terms, 10 God-focused and 10 interpersonal/general) to find relevant tweets, which resulted in a large quantity of observable behavior of gratitude towards God. Limiting it to the year 2019–2020, we found 1.2 million Tweets referencing gratitude; of the subsample of 105 k Tweets that were initially studied, 29.3% of the Tweets referenced gratitude toward God. We note that researchers more generally can use existing libraries such as the Linguistic Inquiry and Word Count (LIWC; Pennebaker et al. 2015) to create a gratitude dictionary such as gratef*, grati*, thank*, or appreciat* (the “*” symbol means that all words that start with the same letters are included) (e.g., Anicich et al. 2022) and also consider specific gratitude terms (e.g., praying hands emoji, tgif). Such an approach would still need domain-specific validation, as dictionaries do not necessarily capture nuances in language. A natural language processing-based approach on the other hand is able to capture and learn from linguistic patterns (e.g., sequences of words), thus with the potential to provide better validity to gratitude measurements. For readers interested in text analytics in general, and how they can apply it to assess psychological phenomena via social media, we refer readers to reviews by Eichstaedt et al. (2021) and Tay et al. (2020).
Another strength is that the application of machine learning text analysis can provide not only individual but also communal assessments of gratitude to God in a potentially scalable, longitudinal, and cost-effective fashion. Achieving accurate assessments enables measurements at the levels of communities and geographic regions. Indeed, it has been proposed that gratitude can serve as a moral “barometer” for society (McCullough and Tsang 2004) and can serve to motivate prosocial behavior (Grant and Gino 2010) and civic engagement (Panagopoulos 2011), which are vital for communities. Further, this opens up possibilities to investigate antecedents and outcomes at different levels of analysis, such as determining the national conditions that predict expressions of gratitude to God, and how these expressions may evolve over time when punctuated by local or global events (such as the Coronavirus pandemic).
While there are strengths to using social media to assess expressions of gratitude to God, there are also critical considerations to ensure its validity and reliability. This paper elucidates these issues so that researchers can identify potential pitfalls and amelioration strategies when applying this approach.

2. Language Variability and Comparability

Given social media usage across different cultures and nations, there is an opportunity to capture different instantiations of gratitude to God expressions. On the linguistic front, they include formal or informal language (González Bermúdez 2015; Utami et al. 2019), the use of non-standard abbreviations, short forms, deliberate misspellings and grammar (Eisenstein 2013), the use of dialects (Huang et al. 2016), the use of emojis (Guntuku et al. 2019), the use of slang and colloquialisms (Reyes et al. 2012), the use of word lengthenings (Brody and Diakopoulos 2011), and the use of different types of languages (Li et al. 2020). Multi-language and cross-cultural investigations (e.g., De Choudhury et al. 2017) can inform us of differences in gratitude expressions that need to be taken into consideration, along with similarities in patterns of expressions (e.g., Guntuku et al. 2019; Li et al. 2020). This is made more complex by the different perspectives that religions hold on the concept of the divine: there is wide variation in the “God” worshipped (e.g., Allah, Christ, etc.). Similarly, the nature of gratitude can also vary based on language norms within a religion (e.g., “it is a blessing”; “praise God”). For instance, in Arabic, word pairs like “ALHAMD LELLAH” which means thanks God and “ALLAH AKBAR” which means God is the greatest are used to express gratitude to God, but some also write them as a single word (Rabie and Sturm 2014). Cultural knowledge is necessary to understand these language norms and variations to properly measure them for cross-cultural studies. Clearly, while language on social media is not constrained like self-report ratings, researchers have to work through a vast multiplicity of gratitude expressions. These need to be carefully curated based on experience and expertise of the religion, culture, and language. Some examples include forming a cross-cultural team of researchers (Tam and Milfont 2020) who are involved in the community, use of qualitative or survey responses to understand context-specific knowledge (Broesch et al. 2020), as well as incorporating mixed methods that allow researchers to measure and compare these cross-cultural differences (Schrauf 2018).
The breadth of possibilities also raises the question of comparability across languages, cultures, and religions. At one level, a frequency approach, or counting the number of instances (or posts) where gratitude to God is expressed, may appear to have fewer issues in terms of comparability, providing the threshold for what counts as gratitude to God. At another level, an intensity approach (e.g., Madisetty and Desarkar 2017), where one seeks to capture the intensity of gratitude to God across expressions (e.g., “Thank God!!!!” Versus “Praise God”) can be more challenging. This is compounded by the possibility of different languages, dialects, and the like. There are also additional issues in how aggregation is done, as discussed later.
In general, we recommend that the research team should set parameters around what language, region, nation, culture, and religion are being examined and obtain the appropriate expertise to determine what expressions are typically considered gratitude to God. Another possibility is to recruit active social media users who describe different ways in which they express gratitude to God, which provides researchers with exemplars of social media expressions.

3. Authenticity of Expressions

As expressions of gratitude to God on many social media platforms are public, it may be difficult to discern whether these expressions are genuine. Performative use of social media, in Erving Goffman’s terms (Goffman 1959), has been observed in multiple contexts, and on these platforms, people are known to self-present or self-enhance (Hogan 2010), which would question their authenticity. For example, people may express mere rhetoric to look good in the eyes of other religious individuals. In addition, some expressions of gratitude may be subtle attempts to boast about themselves (e.g., “I truly can’t explain why, but I find beauty in the ugliest/darkest things. #ThankYouGod”) or pretenses to celebrate positive news (e.g., “I won this award for the 4th time running!! God is always on my side and I cannot thank His blessings enough”) rather than sincere gratitude to God.
By extension, a lack of such expressions may reflect unease in publicly expressing gratitude to God. In fact, many religions instruct people to express gratitude to God through private prayer, which means that many expressions of gratitude to God may not be found on social media posts. Due to this, expressions of gratitude to God that exist on public social media may not be representative of the typical expressions of gratitude to God that occur naturally in daily life. That is, they may lack ecological validity. Nevertheless, the expressions of gratitude to God on public social media may still hold great value in predicting other types of experiences that may occur exclusively on such platforms. The idea is that this is the new reality in which people communicate and experience life and so expressions of gratitude to God measured through social media can predict outcomes such as expressions of well-being or engagement on social media.
Similarly, posts that reference gratitude to God also need to be differentiated from actual gratitude behavior. This can include things like advice for others (e.g., “Thanking your close ones every day is necessary. Appreciate people you have in your life!”), scripture quotes (“Rejoice evermore. Pray without ceasing. In everything give thanks: for this is the will of God in Christ Jesus concerning you. 1 Thessalonians 5: 16–18”), or gratitude additions to unrelated posts (“#selfie #GodIsGood”). Such considerations can be made in the annotation phase so that gratitude detectors or classifiers can also make such distinctions (more below).
Social media is also a place where people tend to engage in trolling behavior in which individuals seek to provoke others through their posts (Hannan 2018). There are good possibilities that gratitude to God expressions are in the context of trolling or, more broadly, inauthentic expressions. For example, depending on the context, audience, and political ideology of the expresser, a post such as “thank God for Trump!” or “praise God for Biden” may be insincere and sarcastic. Given this nuance, there needs to be additional care to examine the context of the language used beyond simple word counts (e.g., the number of times “thank God” is used). We note that the detection of sarcasm, irony, humor, and flippant remarks on social media is challenging, and it is an active area of research (Joshi et al. 2018).
On this issue of authenticity, we encourage researchers to be mindful of the limitations of counting every expression of gratitude to God on social media as genuine heartfelt expressions. There are likely differences in authenticity based on the normative culture of expressing gratitude to God online and the social context of the expression. More generally, expressions of gratitude to God on social media may not accurately capture a person’s internal states, although they may still be very useful for assessing perceptions and reactions of others witnessing these gratitude expressions; or to understand why people express gratitude online publicly. In short, gratitude expressions online and self-reported gratitude may have only some degree of overlap and not have high convergence, but they could have good divergent validities in predicting different outcomes.

4. Machine Learning Process

One popular process for distilling information or inferring latent attributes and behaviors from social media language is to use machine learning. For instance, such methods can enable automatically identifying and/or assessing the rate or the intensity of a phenomenon of interest (Kern et al. 2016). A major advantage is the ability to scale the assessment of gratitude to God expressions to millions of social media posts in an efficient manner. Nevertheless, the building of these text classifiers (i.e., an algorithm that identifies whether a post or sentence expresses gratitude to God, in this case) comes with its own set of challenges that also require careful consideration to ensure accuracy and validity.
Due to the variability of language, supervised machine learning is often used where human annotators provide the “ground truth” (Tay et al. 2020). Human annotators will typically rate posts on whether it expresses gratitude to God, which is then used to train machine learning algorithms. This requires careful training of annotators and the development of a replicable process for how collective decisions are made. For instance, one needs to provide examples and practice with feedback to ensure that they accurately classify posts that reference gratitude expression to God. It is also important to determine the extent annotators agree with one another on each post, or inter-rater agreement (e.g., Krippendorff’s alpha-reliability) (Hayes and Krippendorff 2007). One also needs to be mindful that the algorithms developed may inherit the possible accuracies and biases of the annotators collectively (Tay et al. 2022). Finally, construct validity issues stemming from training data bias or dataset shift (i.e., training data for the machine learning model has a different distribution from the test data) may further paralyze the practical use of machine learning models (Ernala et al. 2019), while under-specified or opaque machine learning models may present ethical issues (Chancellor and De Choudhury 2020).
In the light of these challenges, one way that supervised machine learning can be implemented to study gratitude to God is by having social media users provide self-report ratings of the extent to which they express gratitude to God. These ratings can be treated as individual differences of gratitude to God. Using predictive modeling, the algorithms developed could then seek to predict which types of social media language reflect individual differences in gratitude to God. In this regard, the social media language will be studied not as a display of direct expressions of gratitude to God, rather, the process will capture all the different types of words used by people who generally express gratitude to God. This can inform a qualitative understanding of how people publicly express gratitude to God.
In general, researchers need to be aware that training algorithms to detect expressions of gratitude to God is conceptually distinct, though possibly related, to training algorithms to detect personal subjective ratings of their own gratitude to God. The former emphasizes observable language behaviors, whereas the latter emphasizes the trait of gratitude to God in individuals.

5. Aggregation Approaches

Another key consideration in the use of social media language is how one aggregates the data to make inferences. It is possible to aggregate within an individual to assess the proportion of times an individual expresses gratitude to God (vs. not). In this regard, one needs to determine whether there needs to be a base number of posts an individual should have in order to reduce sampling biases. This is because any single post will have greater weight for people who have very few posts on a platform. For example, someone who has a single post on a platform, and if it so happens to be one expressing gratitude to God, this individual will be counted as 100% expressing gratitude to God. Similarly, principles behind the sampling biases often considered in survey research, such as selection bias, nonresponse bias, attrition, etc., (Olsen 2006) will also be relevant here; appropriate adjustments may be considered, such as weighting (based on overall number of posts) (Royal 2019) or propensity score matching (when making group-level comparisons) (Caliendo and Kopeinig 2008).
Due to the widespread use of social media, one may also seek to infer community- or geographic-level gratitude to God expressions. In this case, one needs to consider whether the aggregation will be done at the post or individual level. If done at the post level, one takes a count of the gratitude to God posts within a community or geographic region. However, because an individual can contribute to multiple posts, superusers may disproportionately be represented in such an approach. Another approach is to aggregate posts at the individual level first to obtain individuals’ level of gratitude to God expressions; then, one proceeds to aggregate individual levels to the community or geographic level (Giorgi et al. 2018). While this approach limits the problem of superusers being disproportionately represented, there needs to be a sufficient number of active users on a platform to aggregate accurately. The issue of active users also raises the issue of the representativeness of a social media sample. For example, commonly used social media in text mining research such as Twitter has a skewed distribution with 42% of the userbase being ages 18–29 while 65+ only comprise 7% of the userbase (Pew Research Center 2022). Additional methods need to be adopted in this case to counter the digital divide (Van Dijk 2020), the gap between demographics at different socioeconomic levels in their access to information and communication technologies (ICT) and digital media (DiMaggio et al. 2001), as well as disproportionate levels of social media use in different communities and geographic regions.
As we have shown, the way aggregation is done from social media posts can be meaningfully different. Researchers who seek to extract social media language to index expressions of gratitude to God will need to make decisions on how best to perform aggregation. Beyond aggregation, there are also analyses that can be done to account for the structure of the data (i.e., individual posts nested within individuals which are in turn nested within communities), such as multilevel modeling (Raudenbush and Bryk 2002); this enables researchers to determine the level-specific predictors of these posts (e.g., what predicts the occurrence of the post within an individual; what predicts who tends to express gratitude to God; what communal factors predict communities that express more gratitude to God).

6. Conclusions

As researchers seek to move toward more observable behavioral approaches to capture gratitude to God expressions, social media language has become a prominent candidate as there are establishing machine learning techniques to harness such data. Nevertheless, there are also different issues that need to be considered when applying this source of data and the techniques associated with it. We hope that as the challenges are identified and made transparent, this will help researchers take steps to address or acknowledge limitations.

Author Contributions

Conceptualization, L.T., S.T., D.B.N. and M.D.C.; writing—original draft preparation, L.T.; writing—review and editing, S.T., D.B.N. and M.D.C.; funding acquisition, L.T., D.B.N. and M.D.C. All authors have read and agreed to the published version of the manuscript.

Funding

This research was funded by the John Templeton Foundation, grant number 61513.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Aghababaei, Naser, and Mohammad Taghi Tabik. 2013. Gratitude and mental health: Differences between religious and general gratitude in a Muslim context. Mental Health. Religion & Culture 16: 761–66. [Google Scholar] [CrossRef]
  2. Ajzen, Icek. 1991. The theory of planned behavior. Organizational Behavior and Human Decision Processes 50: 179–211. [Google Scholar] [CrossRef]
  3. Anicich, Eric M., Alice J. Lee, and Shi Liu. 2022. Thanks, but No Thanks: Unpacking the Relationship Between Relative Power and Gratitude. Personality and Social Psychology Bulletin 48: 1005–23. [Google Scholar] [CrossRef] [PubMed]
  4. Brody, Samuel, and Nicholas Diakopoulos. 2011. Cooooooooooooooollllllllllllll!!!!!!!!!!!!!! Using Word Lengthening to Detect Sentiment in Microblogs. Paper presented at the 2011 Conference on Empirical Methods in Natural Language Processing, Edinburgh, UK, July 27–31. [Google Scholar]
  5. Broesch, Tanya, Alyssa N. Crittenden, Bret A. Beheim, Aaron D. Blackwell, John A. Bunce, Heidi Colleran, Kristin Hagel, Michelle Kline, Richard McElreath, Robin G. Nelson, and et al. 2020. Navigating cross-cultural research: Methodological and ethical considerations. Proceedings of the Royal Society B: Biological Sciences 287: 20201245. [Google Scholar] [CrossRef] [PubMed]
  6. Caliendo, Marco, and Sabine Kopeinig. 2008. Some practical guidance for the implementation of propensity score matching. Journal of Economic Surveys 22: 31–72. [Google Scholar] [CrossRef]
  7. Chancellor, Stevie, and Munmun De Choudhury. 2020. Methods in predictive techniques for mental health status on social media: A critical review. NPJ Digital Medicine 3: 43. [Google Scholar] [CrossRef]
  8. Cronbach, Lee. J., and Paul. E. Meehl. 1955. Construct validity in psychological tests. Psychological Bulletin 52: 281–302. [Google Scholar] [CrossRef]
  9. De Choudhury, Munmun, Sanket S. Sharma, Tomaz Logar, Wouter Eekhout, and René Clausen Nielsen. 2017. Gender and Cross-Cultural Differences in Social Media Disclosures of Mental Illness. Paper presented at the 2017 ACM Conference on Computer Supported Cooperative Work and Social Computing, Portland, OR, USA, February 25–March 1; pp. 353–69. [Google Scholar]
  10. Deen, Brian. 2021. Social Network Usage & Growth Statistics: How Many People Use Social Media in 2022? Available online: https://backlinko.com/social-media-users (accessed on 11 May 2022).
  11. DiMaggio, Paul, Eszter Hargittai, W. Russell Neuman, and John P. Robinson. 2001. Social implications of the internet. Annual Review of Sociology 27: 307–36. [Google Scholar] [CrossRef]
  12. Eichstaedt, Johannes C., Margaret L. Kern, David B. Yaden, H. Andrew Schwartz, Salvatore Giorgi, Gregory Park, Courtney A. Hagan, Victoria A. Tobolsky, Laura K. Smith, Anneke Buffone, and et al. 2021. Closed- and open-vocabulary approaches to text analysis: A review, quantitative comparison, and recommendations. Psychological Methods 26: 398–427. [Google Scholar] [CrossRef]
  13. Eisenstein, Jacob. 2013. What to do about bad language on the internet. Paper presented at the 2013 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Atlanta, GA, USA, June 9–14. [Google Scholar]
  14. Emmons, Robert A., and Cheryl A. Crumpler. 2000. Gratitude as a human strength: Appraising the evidence. Journal of Social and Clinical Psychology 19: 56–69. [Google Scholar] [CrossRef]
  15. Ernala, Sindhu Kiranmai, Michael L. Birnbaum, Kristin A. Candan, Asra F. Rizvi, William A. Sterling, John M. Kane, and Munmun De Choudhury. 2019. Methodological Gaps in Predicting Mental Health States from Social Media. Paper presented at the 2019 CHI Conference on Human Factors in Computing Systems, Glasgow, UK, May 4–9; pp. 1–16. [Google Scholar]
  16. Giorgi, Salvatore, Daniel Preoţiuc-Pietro, Anneke E. K. Buffone, Daniel Rieman, Lyle H. Ungar, and H. Andrew Schwartz. 2018. The Remarkable Benefit of User-Level Aggregation for Lexical-based Population-Level Predictions. Paper presented at the 2018 Conference on Empirical Methods in Natural Language Processing, Brussels, Belgium, October 31–November 4; pp. 1167–72. [Google Scholar]
  17. Goffman, Erving. 1959. The Presentation of Self in Everyday Life. New York: Anchor. [Google Scholar]
  18. González Bermúdez, Meritxell. 2015. An analysis of twitter corpora and the differences between formal and colloquial tweets. Paper presented at the Tweet Translation Workshop 2015, Alicante, Spain, September 5; pp. 1–7. [Google Scholar]
  19. Grant, Adam M., and Francesca Gino. 2010. A little thanks goes a long way: Explaining why gratitude expressions motivate prosocial behavior. Journal of Personality and Social Psychology 98: 946–55. [Google Scholar] [CrossRef] [PubMed]
  20. Guntuku, Sharath Chandra, Mingyang Li, Louis Tay, and Lyle H. Ungar. 2019. Studying Cultural Differences in Emoji Usage across the East and the West. Paper presented at the International AAAI Conference on Web and Social Media, Münich, Germany, June 11–14; vol. 13, pp. 226–35. [Google Scholar]
  21. Hannan, Jason. 2018. Trolling ourselves to death? Social media and post-truth politics. European Journal of Communication 33: 214–26. [Google Scholar] [CrossRef]
  22. Hayes, Andrew F., and Klaus Krippendorff. 2007. Answering the Call for a Standard Reliability Measure for Coding Data. Communication Methods and Measures 1: 77–89. [Google Scholar] [CrossRef]
  23. Hogan, Bernie. 2010. The Presentation of Self in the Age of Social Media: Distinguishing Performances and Exhibitions Online. Bulletin of Science, Technology & Society 30: 377–86. [Google Scholar] [CrossRef]
  24. Huang, Yang, Diansheng Guo, Alice Kasakoff, and Jack Grieve. 2016. Understanding U.S. regional linguistic variation with Twitter data analysis. Computers, Environment and Urban Systems 59: 244–55. [Google Scholar] [CrossRef]
  25. Joshi, Aditya, Pushpak Bhattacharyya, and Mark J. Carman. 2018. Automatic Sarcasm Detection: A Survey. ACM Computing Surveys 50: 1–22. [Google Scholar] [CrossRef]
  26. Kern, Margaret L., Gregory Park, Johannes C. Eichstaedt, H. Andrew Schwartz, Maarten Sap, Laura K. Smith, and Lyle H. Ungar. 2016. Gaining insights from social media language: Methodologies and challenges. Psychological Methods 21: 507–25. [Google Scholar] [CrossRef]
  27. Krause, Neal. 2006. Gratitude toward God, stress, and health in late life. Research on Aging 28: 163–83. [Google Scholar] [CrossRef]
  28. Krause, N., R. David Hayward, Deborah Bruce, and Cynthia Woolever. 2014. Gratitude to God, self-rated health, and depressive symptoms. Journal for the Scientific Study for Religion 53: 341–55. [Google Scholar] [CrossRef]
  29. Krause, N., R. A. Emmons, and G. Ironson. 2015. Benevolent Images of God, Gratitude, and Physical Health Status. Journal of Religion and Health 54: 1503–19. [Google Scholar] [CrossRef]
  30. Li, Mingyang, Louis Hickman, Louis Tay, Lyle H. Ungar, and Sharath Chandra Guntuku. 2020. Studying Politeness across cultures using English Twitter and Mandarin Weibo. Paper presented at the AMC on Human-Computer Interaction, Oldenburg, Germany, October 5–9; pp. 1–15. [Google Scholar]
  31. Loevinger, Jane. 1957. Objective tests as instruments of psychological theory. Psychological Reports 3: 635–94. [Google Scholar] [CrossRef]
  32. Madisetty, Sreekanth, and Maunendra Sankar Desarkar. 2017. An Ensemble Based Method for Predicting Emotion Intensity of Tweets. Paper presented at the International Conference on Mining Intelligence and Knowledge Exploration, Hyderabad, India, December 13–15; pp. 359–70. [Google Scholar]
  33. McCullough, Michael E., and Jo-Ann Tsang. 2004. Parent of the virtues? In The Psychology of Gratitude. Edited by Robert A. Emmons and Michael E. McCullough. Oxford: Oxford University Press, pp. 123–44. [Google Scholar]
  34. McCullough, Michael E., Robert A. Emmons, and Jo-Ann Tsang. 2002. The grateful disposition: A conceptual and empirical topography. Journal of Personality and Social Psychology 82: 112–27. [Google Scholar] [CrossRef] [PubMed]
  35. McGrath, Joseph E. 1981. The study of research choices and dilemmas. American Behavioral Scientist 25: 179–210. [Google Scholar] [CrossRef]
  36. Messick, Samuel. 1980. Test validity and the ethics of assessment. American Psychologist 35: 1012–27. [Google Scholar] [CrossRef]
  37. Messick, Samuel. 1995. Validity of psychological assessment: Validation of inferences from persons’ responses and performances as scientific inquiry into score meaning. American Psychologist 50: 741–49. [Google Scholar] [CrossRef]
  38. Nederhof, Anton J. 1985. Methods of coping with social desirability bias: A review. European Journal of Social Psychology 15: 263–80. [Google Scholar] [CrossRef]
  39. Olsen, Kristen. 2006. Survey Participation, Nonresponse Bias, Measurement Error Bias, and Total Bias. Public Opinion Quarterly 70: 737–58. [Google Scholar] [CrossRef]
  40. Panagopoulos, Costas. 2011. Thank You for Voting: Gratitude Expression and Voter Mobilization. The Journal of Politics 73: 707–17. [Google Scholar] [CrossRef]
  41. Pennebaker, James W., Ryan L. Boyd, Kayla Jordan, and Kate Blackburn. 2015. The development and psychometric properties of LIWC2015. Available online: http://hdl.handle.net/2152/31333 (accessed on 1 August 2022).
  42. Pew Research Center. 2022. 10 Facts about Americans and Twitter. Available online: https://www.pewresearch.org/fact-tank/2022/05/05/10-facts-about-americans-and-twitter/ (accessed on 15 July 2022).
  43. Rabie, Omneya, and Christian Sturm. 2014. Feel the Heat: Emotion Detection in Arabic Social Media Content. Paper presented at the International Conference on Data Mining, Internet Computing, and Big Data, Kuala Lumpur, Malaysia, November 17–19. [Google Scholar]
  44. Raudenbush, Stephen. W., and Anthony S. Bryk. 2002. Hierarchical Linear Models: Applications and Data Analysis Methods, 2nd ed. Thousand Oaks: Sage. [Google Scholar]
  45. Reyes, Antonio, Paolo Rosso, and Davide Buscaldi. 2012. From humor recognition to irony detection: The figurative language of social media. Data & Knowledge Engineering 74: 1–12. [Google Scholar]
  46. Rosmarin, David H., Steven Pirutinsky, Adam B. Cohen, Yardana Galler, and Elizabeth J. Krumrei. 2011. Grateful to God or just plain grateful? A comparison of religious and general gratitude. Journal of Positive Psychology 6: 389–96. [Google Scholar] [CrossRef]
  47. Royal, Kenneth D. 2019. Survey research methods: A guide for creating post-stratification weights to correct for sample bias. Education in the Health Professions 2: 48–50. [Google Scholar] [CrossRef]
  48. Schrauf, Robert W. 2018. Mixed Methods Designs for Making Cross-Cultural Comparisons. Journal of Mixed Methods Research 12: 477–94. [Google Scholar] [CrossRef]
  49. Schwarz, Nobert. 1999. Self-reports: How the questions shape the answer. American Psychologist 54: 93–105. [Google Scholar] [CrossRef]
  50. Tam, Kim-Pong, and Taciano L. Milfont. 2020. Towards cross-cultural environmental psychology: A state-of-the-art review and recommendations. Journal of Environmental Psychology 71: 1–18. [Google Scholar] [CrossRef]
  51. Tay, Louis, Sang Eun Woo, Louis Hickman, and Rachel M. Saef. 2020. Psychometric and Validity Issues in Machine Learning Approaches to Personality Assessment: A Focus on Social Media Text Mining. European Journal of Personality 34: 826–44. [Google Scholar] [CrossRef]
  52. Tay, Louis, Sang Eun Woo, Louis Hickman, Brandon M. Booth, and Sidney D’Mello. 2022. A Conceptual Framework for Investigating and Mitigating Machine-Learning Measurement Bias (MLMB) in Psychological Assessment. Advances in Methods and Practices in Psychological Science 5. [Google Scholar] [CrossRef]
  53. Utami, Ema, Anggit Dwi Hartanto, Sumarni Adi, Rahardyan Bisma Setya Putra, and Suwanto Raharjo. 2019. Formal and non-formal Indonesian word usage frequency in twitter profile using non-formal affix rule. Paper presented at the 2019 1st International Conference on Cybernetics and Intelligent System (ICORIS), Denpasar, Indonesia, August 22–23; pp. 173–76. [Google Scholar]
  54. Van Dijk, Jan. 2020. The Digital Divide. Cambridge: Polity. [Google Scholar]
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Tay, L.; Thapa, S.; Newman, D.B.; Choudhury, M.D. Using Social Media to Assess Expressions of Gratitude to God: Issues for Consideration. Religions 2022, 13, 778. https://doi.org/10.3390/rel13090778

AMA Style

Tay L, Thapa S, Newman DB, Choudhury MD. Using Social Media to Assess Expressions of Gratitude to God: Issues for Consideration. Religions. 2022; 13(9):778. https://doi.org/10.3390/rel13090778

Chicago/Turabian Style

Tay, Louis, Stuti Thapa, David B. Newman, and Munmun De Choudhury. 2022. "Using Social Media to Assess Expressions of Gratitude to God: Issues for Consideration" Religions 13, no. 9: 778. https://doi.org/10.3390/rel13090778

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop