Building a Multimodal Classifier of Email Behavior: Towards a Social Network Understanding of Organizational Communication
Abstract
:1. Introduction
Research Objectives
- RQ1: How do the sender’s (a) professional influence and (b) personal influence in the organizational network affect their likelihood of receiving a reply?
- RQ2: How well does a pretrained transformer finetuned with multimodal features predict email replies?
- RQ3: How does its performance vary with (a) different feature sets in the model and (b) different training dataset sizes?
- RQ4: How does its performance vary for different datasets?
2. Related Work
2.1. Email Behavior
2.2. The Role of the Social Network
3. Method
3.1. Framework
3.1.1. Text Encoder Layer
3.1.2. Gating Layer
3.2. Experimental Setup
- Contribution analysis: To address RQ1a and RQ1b, we explored the association of individual measures with reply prediction by fitting quasibinomial generalized linear models to our dataset after standardization and feature selection and visualizing the feature importance in terms of the magnitude of the coefficients in the final fitted model.
- Cross-validation: To address RQ2, we used the Luxury Standard dataset to train models on the stylistic, professional, personal, and accommodation features just discussed for the email reply prediction task using a cross-validation setup with a 20% held-out test set. We benchmarked the performance of a multimodal transformer against several non-transformer and transformer baselines.
- Ablation analysis: To address RQ3 and establish the robustness of EMMA, we conducted an ablation analysis with different input feature sets (RQ3a) and dataset sizes (RQ3b).
- External validity: To address RQ4, we also evaluated EMMA on the Enron and Avocado datasets for the same problem, thus ensuring that EMMA generalizes to new data and social contexts.
3.3. Baselines
3.3.1. Non-Transformer Models
- CNNs: Convolutional neural networks (CNNs) are a class of neural networks that are used primarily for image recognition and classification and were first developed for optical character recognition (OCR)-related tasks [62]. Each layer in a CNN has filters that slide on the data from the input/previous layer. In the case of text, CNN’s sliding window captures patterns in the sequence of words, which become more complex with more convolution layers. In our experiments, in the CNN framework [56], we applied convolutional filters followed by max-over-time pooling to the word vectors for a single email.
- RNNs: Recurrent neural networks (RNN) are a class of neural networks that is primarily used for sequential data, such as text, audio, and time series data. While neurons in a vanilla DNN can look at input data or data from the previous layer, the neurons in an RNN can go one step further. They can look at both neurons from the last layer and the neuron output from the previous timestep. This small change in RNNs allows them to retain information from the past while calculating the output at a given timestep. To update weights in RNNs, backpropagation through time, or BPTT, is used. The derivation of the weight update rule in RNNs shows that for an input sequence length = N, the derivative term contains a weight raised to the power N. This means that if the weight term is either very small or very large, the gradient will tend to zero or explode, respectively. Furthermore, the model performance deteriorates with increasing input length.
- LSTMs: Architectures such as long short-term memory (LSTM) addressed and corrected the shortcomings in RNNs related to the vanishing gradient at very small or very large weights [57]. Apart from the primary recurrence relation present in RNNs, LSTMs also have multiple gating mechanisms, allowing them to choose the amount of information from the past that is carried forward. An LSTM comprises a forget gate, an input gate, and an output gate. The forget gate controls the amount of information that is forgotten, the update gate controls the amount of information carried forward, and the output gate controls the amount of information passed on to the next layer.
- LSTM + Attention: Attention mechanisms address the problems with RNNs that occur with increasing input length. The multi-head self-attention mechanism used in the transformer architecture [63] involves a self-attention component that looks at each word in the input sequence and computes how important other words in the input are when computing the representation for that particular word. A dot product computes the similarity between the current word and prior words in the input sequence, while a key matrix scales down the dot product so that the gradient does not become unstable for large inputs. A softmax function converts the similarity to a probability distribution, adding up to 1. The final multiplication with the value matrix results in those words retaining their embedding whose dot product score is high, which results in meaningful and contextual embeddings. Transformers have multiple such heads, the outputs of which are concatenated and transformed into a compatible dimension. Another feature of transformer models is that they use multiple encoders and decoders stacked on top of each other as their encoder and decoder blocks.
- Training hyperparameters: We trained the models on 15 epochs with a batch size of 1024 and early stopping based on validation loss. The training data were split into training and validation sets during each epoch with a 0.8:0.2 split. This was done to monitor the validation loss. Early stopping was enabled, which stopped the training if the validation loss did not improve for three consecutive epochs. All the DL-based baseline models were trained with following configuration:
- -
- Batch size = 1024;
- -
- Vocabulary size = 2000;
- -
- Input length = 200;
- -
- Word embedding dimension = 50;
- -
- Number of epochs = 15;
- -
- Train:test split = 0.8:0.2.
3.3.2. Transformer Models
- BERT: Bidirectional Encoder Representations from Transformers (BERT) was originally published by Google AI Language in 2018, where they used the transformer architecture for the task of language modeling [64]. One of the key features of BERT is that it stacks only encoder blocks on top of each other. BERT is deeply bidirectional, i.e., the model learns from left-to-right and right-to-left while going over input sentences. Also, the self-attention combined with multi-head attention in each transformer encoder block helps to generate more contextual embeddings by taking into account other words in the sentence when encoding a particular word. BERT models have multiple attention heads to focus on different aspects of the input text simultaneously. Their deep bidirectional nature allows them to infer context from before and after a word within an input text. Furthermore, the combined training task of masked language modeling and next-sentence prediction is also one of the reasons why the BERT language model has a good understanding of the structure and semantics of the English language. Due to these reasons, a single dense layer on top of the output of the BERT language model performs exceptionally well on many NLP tasks.
- RoBERTa: Robustly Optimized BERT Pretraining Approach (RoBERTa) is an alternate approach to training BERT that not only improves its performance but also results in easier training [65]. BERT masks tokens randomly during the preprocessing stage, resulting in a static mask. RoBERTa applies a dynamic masking scheme, meaning masking takes place before a sentence is selected into a minibatch to be fed into the model. This approach avoids the pitfall of using the same masked sequence every epoch.
- Training hyperparameters: For BERT and RoBERTa, the following training arguments were used:
- -
- Batch size = 150;
- -
- Input length = 200;
- -
- Number of epochs = 10;
- -
- Initial learning rate = 5e;
- -
- Warmup ratio = 0.07;
- -
- Weight decay = 0.5.
4. Data Collection
4.1. Feature Extraction
4.1.1. Stylistic Features (93 Features)
- Cognitive processes: Measures of cognitive activity through evidence of words that denote insight (think, know), causation (because, effect), discrepancy (should, would), tentative (maybe, perhaps), certainty (always, never), differentiation (hasn’t, but, else) and perception (look, heard, feeling).
- Emotional features: Measures of affect in writing, comprising measures of positive emotion (love, nice, sweet) and negative emotion (terrible, worried, sad).
- Drive features: Measures denoting what motivates people, offering insights into perspectives on achievement (win, success, better), affiliation (ally, friend, social), a need for domination (superior, better), and finally, the reward-(take, prize, benefit) or risk-orientation (danger, doubt) of the sender.
- Informal language features: Measures of the casualness of the email body, denoted by the use of swear words (damn, shit, hell), netspeak (lol, thx, lmao), nonfluencies (err, hmm, umm), and fillers (well, you know, I mean).
4.1.2. Professional Influence (2 Features)
- PageRank: This measures the number of times the sender s is encountered in a random walk over the social network:
- Betweenness centrality: This measures the number of the shortest paths connecting nodes that pass through a particular node. It identifies the degree to which senders act as conduits of information:
4.1.3. Personal Influence (10 Features)
4.1.4. Linguistic Accommodation (3 Features)
5. Results
5.1. Contribution Analysis
5.2. Cross-Validation
5.3. Ablation Analysis
5.4. External Validity
6. Discussion
7. Conclusions
Author Contributions
Funding
Data Availability Statement
Conflicts of Interest
References
- Grammarly, B. The State of Business Communication. Grammarly Harris Poll 2022. Available online: https://www.grammarly.com/business/Grammarly_The_State_Of_Business_Communication.pdf (accessed on 2 December 2023).
- Sarrafzadeh, B.; Hassan Awadallah, A.; Lin, C.H.; Lee, C.J.; Shokouhi, M.; Dumais, S.T. Characterizing and predicting email deferral behavior. In Proceedings of the Twelfth ACM International Conference on Web Search and Data Mining, Melbourne, Australia, 11–15 February 2019; pp. 627–635. [Google Scholar]
- Song, H.; Eveland, W.P., Jr. The structure of communication networks matters: How network diversity, centrality, and context influence political ambivalence, participation, and knowledge. Political Commun. 2015, 32, 83–108. [Google Scholar] [CrossRef]
- Kwon, J.; Han, I.; Kim, B. Effects of source influence and peer referrals on information diffusion in Twitter. Ind. Manag. Data Syst. 2017, 117, 896–909. [Google Scholar] [CrossRef]
- Fagan, J. How Organizational Turbulence Shapes the Broker Vision Advantage. Ph.D. Thesis, University of Kentucky, Lexington, KY, USA, 2017. [Google Scholar]
- Brass, D.J. Being in the right place: A structural analysis of individual influence in an organization. Adm. Sci. Q. 1984, 29, 518–539. [Google Scholar] [CrossRef]
- Rodan, S.; Galunic, C. More than network structure: How knowledge heterogeneity influences managerial performance and innovativeness. Strateg. Manag. J. 2004, 25, 541–562. [Google Scholar] [CrossRef]
- Rodan, S. Structural holes and managerial performance: Identifying the underlying mechanisms. Soc. Netw. 2010, 32, 168–179. [Google Scholar] [CrossRef]
- Lungeanu, A.; Carter, D.R.; DeChurch, L.A.; Contractor, N.S. How team interlock ecosystems shape the assembly of scientific teams: A hypergraph approach. Commun. Methods Meas. 2018, 12, 174–198. [Google Scholar] [CrossRef]
- Sauer, N.C.; Kauffeld, S. Meetings as networks: Applying social network analysis to team interaction. Commun. Methods Meas. 2013, 7, 26–47. [Google Scholar] [CrossRef]
- Ghani, N.A.; Hamid, S.; Hashem, I.A.T.; Ahmed, E. Social media big data analytics: A survey. Comput. Hum. Behav. 2019, 101, 417–428. [Google Scholar] [CrossRef]
- Dinh, R.; Gildersleve, P.; Blex, C.; Yasseri, T. Computational courtship understanding the evolution of online dating through large-scale data analysis. J. Comput. Soc. Sci. 2022, 5, 401–426. [Google Scholar] [CrossRef]
- Shah, D.V.; Sun, Z.; Bucy, E.P.; Kim, S.J.; Sun, Y.; Li, M.; Sethares, W. Building an ICCN Multimodal Classifier of Aggressive Political Debate Style: Towards a Computational Understanding of Candidate Performance Over Time. Commun. Methods Meas. 2023, 1–18. Available online: https://www.tandfonline.com/doi/full/10.1080/19312458.2023.2227093 (accessed on 2 December 2023). [CrossRef]
- On, B.W.; Lim, E.P.; Jiang, J.; Purandare, A.; Teow, L.N. Mining interaction behaviors for email reply order prediction. In Proceedings of the International Conference on Advances in Social Networks Analysis and Mining, Odense, Denmark, 9–11 August 2010; pp. 306–310. [Google Scholar]
- Aral, S.; Van Alstyne, M. The diversity-bandwidth trade-off. Am. J. Sociol. 2011, 117, 90–171. [Google Scholar] [CrossRef]
- On, B.W.; Lim, E.P.; Jiang, J.; Teow, L.N. Engagingness and responsiveness behavior models on the enron email network and its application to email reply order prediction. In The Influence of Technology on Social Network Analysis and Mining; Springer: Berlin/Heidelberg, Germany, 2013; pp. 227–253. [Google Scholar]
- Jones, J.M.; Carter, D.R.; Contractor, N.S. A Network Approach to Studying Team Functioning. In The Emerald Handbook of Group and Team Communication Research; Emerald Publishing Limited: Bingley, UK, 2021. [Google Scholar]
- Kleinbaum, A.M.; Stuart, T.E. Inside the black box of the corporate staff: Social networks and the implementation of corporate strategy. Strateg. Manag. J. 2014, 35, 24–47. [Google Scholar] [CrossRef]
- Kleinbaum, A.M.; Stuart, T.E. Network responsiveness: The social structural microfoundations of dynamic capabilities. Acad. Manag. Perspect. 2014, 28, 353–367. [Google Scholar] [CrossRef]
- Goldberg, A.; Srivastava, S.B.; Manian, V.G.; Monroe, W.; Potts, C. Fitting in or standing out? The tradeoffs of structural and cultural embeddedness. Am. Sociol. Rev. 2016, 81, 1190–1222. [Google Scholar] [CrossRef]
- Jin, D.; Heimann, M.; Safavi, T.; Wang, M.; Lee, W.; Snider, L.; Koutra, D. Smart roles: Inferring professional roles in email networks. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, Anchorage, AK, USA, 4–8 August 2019; pp. 2923–2933. [Google Scholar]
- Wang, W.; Hosseini, S.; Awadallah, A.H.; Bennett, P.N.; Quirk, C. Context-aware intent identification in email conversations. In Proceedings of the 42nd International ACM SIGIR Conference on Research and Development in Information Retrieval, Paris, France, 21–25 July 2019; pp. 585–594. [Google Scholar]
- Mukherjee, S.; Jiang, K. A Content-Based Approach to Email Triage Action Prediction: Exploration and Evaluation. arXiv 2019, arXiv:1905.01991. [Google Scholar]
- Jörgensen, P. Incorporating context in text analysis by interactive activation with competition artificial neural networks. Inf. Process. Manag. 2005, 41, 1081–1099. [Google Scholar] [CrossRef]
- Zajic, D.M.; Dorr, B.J.; Lin, J. Single-document and multi-document summarization techniques for email threads using sentence compression. Inf. Process. Manag. 2008, 44, 1600–1610. [Google Scholar] [CrossRef]
- Peterson, K.; Hohensee, M.; Xia, F. Email formality in the workplace: A case study on the Enron corpus. In Proceedings of the Workshop on Language in Social Media (LSM 2011), Portland, OR, USA, 23 June 2011; pp. 86–95. [Google Scholar]
- Sayed, M.F.; Cox, W.; Rivera, J.L.; Christian-Lamb, C.; Iqbal, M.; Oard, D.W.; Shilton, K. A test collection for relevance and sensitivity. In Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval, Virtual, 25–30 July 2020; pp. 1605–1608. [Google Scholar]
- Jabbari, S.; Allison, B.; Guthrie, D.; Guthrie, L. Towards the Orwellian nightmare: Separation of business and personal emails. In Proceedings of the COLING/ACL 2006 Main Conference Poster Sessions, Sydney, Australia, 17–18 July 2006; pp. 407–411. [Google Scholar]
- HaCohen-Kerner, Y.; Dilmon, R.; Hone, M.; Ben-Basan, M.A. Automatic classification of complaint letters according to service provider categories. Inf. Process. Manag. 2019, 56, 102102. [Google Scholar] [CrossRef]
- Sappelli, M.; Pasi, G.; Verberne, S.; de Boer, M.; Kraaij, W. Assessing e-mail intent and tasks in e-mail messages. Inf. Sci. 2016, 358, 1–17. [Google Scholar] [CrossRef]
- Shu, K.; Mukherjee, S.; Zheng, G.; Awadallah, A.H.; Shokouhi, M.; Dumais, S. Learning with weak supervision for email intent detection. In Proceedings of the 43rd International ACM SIGIR Conference on Research and Development in Information Retrieval, Virtual, 25–30 July 2020; pp. 1051–1060. [Google Scholar]
- Aberdeen, D.; Pacovsky, O.; Slater, A. The Learning Behind Gmail Priority Inbox. In Proceedings of the LCCC: NIPS 2010 Workshop on Learning on Cores, Clusters and Clouds, Vancouver, BC, Canada, 6 December 2010. [Google Scholar]
- Deitrick, W.; Miller, Z.; Valyou, B.; Dickinson, B.; Munson, T.; Hu, W. Author Gender Prediction in an Email Stream Using Neural Networks. J. Intell. Learn. Syst. Appl. 2012, 4, 169–175. [Google Scholar] [CrossRef]
- Robertson, R.E.; Olteanu, A.; Diaz, F.; Shokouhi, M.; Bailey, P. “I can’t reply with that”: Characterizing problematic email reply suggestions. In Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems, Yokohama, Japan, 8–13 May 2021; pp. 1–18. [Google Scholar]
- Buschek, D.; Zürn, M.; Eiband, M. The impact of multiple parallel phrase suggestions on email input and composition behaviour of native and non-native english writers. In Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems, Yokohama, Japan, 8–13 May 2021; pp. 1–13. [Google Scholar]
- Oard, D.; Webber, W.; Kirsch, D.; Golitsynskiy, S. Avocado Research Email Collection; Linguistic Data Consortium: Philadelphia, PA, USA, 2015. [Google Scholar]
- Alkhereyf, S.; Rambow, O. Work Hard, Play Hard: Email Classification on the Avocado and Enron Corpora. In Proceedings of the TextGraphs-11: The Workshop on Graph-based Methods for Natural Language Processing, Vancouver, BC, Canada, 6 June 2017; pp. 57–65. [Google Scholar] [CrossRef]
- Woehler, M.; Floyd, T.M.; Shah, N.; Marineau, J.E.; Sung, W.; Grosser, T.J.; Fagan, J.; Labianca, G.J. Turnover during a corporate merger: How workplace network change influences staying. J. Appl. Psychol. 2021, 106, 1939–1949. [Google Scholar] [CrossRef]
- Russell, E.; Woods, S.A. Personality differences as predictors of action-goal relationships in work-email activity. Comput. Hum. Behav. 2020, 103, 67–79. [Google Scholar] [CrossRef]
- Reinke, K.; Chamorro-Premuzic, T. When email use gets out of control: Understanding the relationship between personality and email overload and their impact on burnout and work engagement. Comput. Hum. Behav. 2014, 36, 502–509. [Google Scholar] [CrossRef]
- Phillips, J.G.; Reddie, L. Decisional style and self-reported email use in the workplace. Comput. Hum. Behav. 2007, 23, 2414–2428. [Google Scholar] [CrossRef]
- Russell, E.; Purvis, L.M.; Banks, A. Describing the strategies used for dealing with email interruptions according to different situational parameters. Comput. Hum. Behav. 2007, 23, 1820–1837. [Google Scholar] [CrossRef]
- Chhaya, N.; Jaidka, K.; Wadbude, R. Predicting Email Opens with Domain-Sensitive Affect Detection. In Proceedings of the International Conference on Computational Linguistics and Intelligent Text Processing, Hanoi, Vietnam, 18 March 2018; Springer: Berlin/Heidelberg, Germany, 2018; pp. 71–79. [Google Scholar]
- Jaidka, K.; Goyal, T.; Chhaya, N. Predicting email and article clickthroughs with domain-adaptive language models. In Proceedings of the 10th ACM Conference on Web Science, Amsterdam, The Netherlands, 27–30 May 2018; pp. 177–184. [Google Scholar]
- Yang, L.; Dumais, S.T.; Bennett, P.N.; Awadallah, A.H. Characterizing and predicting enterprise email reply behavior. In Proceedings of the 40th International ACM Sigir Conference on Research and Development in Information Retrieval, Tokyo, Japan, 7–11 August 2017; pp. 235–244. [Google Scholar]
- Francis, L.; Holmvall, C.M.; O’brien, L.E. The influence of workload and civility of treatment on the perpetration of email incivility. Comput. Hum. Behav. 2015, 46, 191–201. [Google Scholar] [CrossRef]
- Haesevoets, T.; De Cremer, D.; McGuire, J. How the use of Cc, Bcc, forward, and rewrite in email communication impacts team dynamics. Comput. Hum. Behav. 2020, 112, 106478. [Google Scholar] [CrossRef]
- Bavelas, A. Communication patterns in task-oriented groups. J. Acoust. Soc. Am. 1950, 22, 725–730. [Google Scholar] [CrossRef]
- Kozlowski, S.W.; Klein, K.J. A multilevel approach to theory and research in organizations: Contextual, temporal, and emergent processes. In Multilevel Theory, Research, and Methods in Organizations: Foundations, Extensions, and New Directions; Klein, K.J., Kozlowski, S.W.J., Eds.; Jossey-Bass: New York, NY, USA, 2000; pp. 3–90. [Google Scholar]
- Burke, M.; Kraut, R. Using Facebook after losing a job: Differential benefits of strong and weak ties. In Proceedings of the 2013 Conference on Computer Supported Cooperative Work, San Antonio, TX, USA, 23–27 February 2013; pp. 1419–1430. [Google Scholar]
- Granovetter, M.S. The strength of weak ties. Am. J. Sociol. 1973, 78, 1360–1380. [Google Scholar] [CrossRef]
- Sung, W.; Labianca, G.; Fagan, J.M. Executives’ network change and their promotability during a merger. Acad. Manag. Proc. 2018, 2018, 18386. [Google Scholar] [CrossRef]
- Pulles, N.J.; Hartman, P. Likeability and its effect on outcomes of interpersonal interaction. Ind. Mark. Manag. 2017, 66, 56–63. [Google Scholar] [CrossRef]
- Danescu-Niculescu-Mizil, C.; Lee, L.; Pang, B.; Kleinberg, J. Echoes of power: Language effects and power differences in social interaction. In Proceedings of the 21st International Conference on World Wide Web, Lyon, France, 16–20 April 2012; pp. 699–708. [Google Scholar]
- Rahman, W.; Hasan, M.K.; Lee, S.; Zadeh, A.; Mao, C.; Morency, L.P.; Hoque, E. Integrating multimodal information in large pretrained transformers. In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Online, 5–10 July 2020; Volume 2020, p. 2359. [Google Scholar]
- Kim, Y. Convolutional Neural Networks for Sentence Classification. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), Doha, Qatar, 25–29 October 2014; pp. 1746–1751. [Google Scholar] [CrossRef]
- Hochreiter, S.; Schmidhuber, J. Long Short-term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef]
- Graves, A.; Mohamed, A.r.; Hinton, G. Speech Recognition with Deep Recurrent Neural Networks. In Proceedings of the 2013 IEEE International Conference on Acoustics, Speech and Signal Processing, Vancouver, BC, Canada, 26–31 May 2013; Volume 38. [Google Scholar] [CrossRef]
- Luong, T.; Pham, H.; Manning, C.D. Effective Approaches to Attention-based Neural Machine Translation. In Proceedings of the 2015 Conference on Empirical Methods in Natural Language Processing, Lisbon, Portugal, 17–21 September 2015; pp. 1412–1421. [Google Scholar] [CrossRef]
- Pennington, J.; Socher, R.; Manning, C. GloVe: Global Vectors for Word Representation. In Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), Doha, Qatar, 25–29 October 2014; pp. 1532–1543. [Google Scholar] [CrossRef]
- Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. In Proceedings of the 3rd International Conference on Learning Representations, San Diego, CA, USA, 7–9 May 2015. [Google Scholar]
- Lecun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef]
- Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, L.; Polosukhin, I. Attention is All You Need. In Proceedings of the 31st International Conference on Neural Information Processing Systems, NIPS’17, Long Beach, CA, USA, 4–9 December 2017; pp. 6000–6010. [Google Scholar]
- Devlin, J.; Chang, M.W.; Lee, K.; Toutanova, K. BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers), Minneapolis, MN, USA, 2–7 June 2019; pp. 4171–4186. [Google Scholar] [CrossRef]
- Liu, Y.; Ott, M.; Goyal, N.; Du, J.; Joshi, M.; Chen, D.; Levy, O.; Lewis, M.; Zettlemoyer, L.; Stoyanov, V. RoBERTa: A Robustly Optimized BERT Pretraining Approach. arXiv 2019, arXiv:1907.11692. [Google Scholar]
- Sung, W.; Woehler, M.L.; Fagan, J.M.; Grosser, T.J.; Floyd, T.M.; Labianca, G.J. Employees’ responses to an organizational merger: Intraindividual change in organizational identification, attachment, and turnover. J. Appl. Psychol. 2017, 102, 910. [Google Scholar] [CrossRef]
- Klimt, B.; Yang, Y. The enron corpus: A new dataset for email classification research. In Proceedings of the European Conference on Machine Learning, Pisa, Italy, 20–24 September 2004; Springer: Berlin/Heidelberg, Germany, 2004; pp. 217–226. [Google Scholar]
- Lewis, D.D.; Knowles, K.A. Threading electronic mail: A preliminary study. Inf. Process. Manag. 1997, 33, 209–217. [Google Scholar] [CrossRef]
- Ireland, M.E.; Pennebaker, J.W. Language style matching in writing: Synchrony in essays, correspondence, and poetry. J. Personal. Soc. Psychol. 2010, 99, 549. [Google Scholar] [CrossRef]
- Pennebaker, J.W.; Boyd, R.L.; Jordan, K.; Blackburn, K. The Development and Psychometric Properties of LIWC2015. Technical Report, 2015. Available online: https://www.liwc.app/static/documents/LIWC-22%20Manual%20-%20Development%20and%20Psychometrics.pdf (accessed on 2 December 2023).
- Ahuja, H.; Ng, L.H.X.; Jaidka, K. Using Graph-Aware Reinforcement Learning to Identify Winning Strategies in Diplomacy Games (Student Abstract). In Proceedings of the AAAI Conference on Artificial Intelligence, Virtually, 22 February–1 March 2022; Volume 36, pp. 12899–12900. [Google Scholar]
- Jaidka, K.; Ahuja, H.; Ng, L. It Takes Two to Negotiate: Modeling Social Exchange in Online Multiplayer Games. arXiv 2023, arXiv:2311.08666. [Google Scholar]
- Tan, C.; Niculae, V.; Danescu-Niculescu-Mizil, C.; Lee, L. Winning arguments: Interaction dynamics and persuasion strategies in good-faith online discussions. In Proceedings of the 25th International Conference on World Wide Web, MontrÉal, QC, Canada, 11–15 April 2016; pp. 613–624. [Google Scholar]
- Sia, S.; Jaidka, K.; Ahuja, H.; Chhaya, N.; Duh, K. Offer a Different Perspective: Modeling the Belief Alignment of Arguments in Multi-party Debates. In Proceedings of the 2022 Conference on Empirical Methods in Natural Language Processing, Abu Dhabi, United Arab Emirates, 7–11 December 2022; pp. 11939–11950. [Google Scholar]
- Vosoughi, S.; Roy, D.; Aral, S. The spread of true and false news online. Science 2018, 359, 1146–1151. [Google Scholar] [CrossRef]
- Bonifazi, G.; Cauteruccio, F.; Corradini, E.; Marchetti, M.; Pierini, A.; Terracina, G.; Ursino, D.; Virgili, L. An approach to detect backbones of information diffusers among different communities of a social platform. Data Knowl. Eng. 2022, 140, 102048. [Google Scholar] [CrossRef]
- Garimella, K.; West, R. Hot streaks on social media. In Proceedings of the International AAAI Conference on Web and Social Media, Münich, Germany, 11–14 June 2019; Volume 13, pp. 170–180. [Google Scholar]
- Gurjar, O.; Bansal, T.; Jangra, H.; Lamba, H.; Kumaraguru, P. Effect of Popularity Shocks on User Behaviour. In Proceedings of the International AAAI Conference on Web and Social Media, Atlanta, GA, USA, 6–9 June 2022; Volume 16, pp. 253–263. [Google Scholar]
- Bonifazi, G.; Cauteruccio, F.; Corradini, E.; Marchetti, M.; Terracina, G.; Ursino, D.; Virgili, L. A framework for investigating the dynamics of user and community sentiments in a social platform. Data Knowl. Eng. 2023, 146, 102183. [Google Scholar] [CrossRef]
- Jerejian, A.C.; Reid, C.; Rees, C.S. The contribution of email volume, email management strategies and propensity to worry in predicting email stress among academics. Comput. Hum. Behav. 2013, 29, 991–996. [Google Scholar] [CrossRef]
- Kushlev, K.; Dunn, E.W. Checking email less frequently reduces stress. Comput. Hum. Behav. 2015, 43, 220–228. [Google Scholar] [CrossRef]
- Hair, M.; Renaud, K.V.; Ramsay, J. The influence of self-esteem and locus of control on perceived email-related stress. Comput. Hum. Behav. 2007, 23, 2791–2803. [Google Scholar] [CrossRef]
- Kipf, T.N.; Welling, M. Semi-Supervised Classification with Graph Convolutional Networks. In Proceedings of the International Conference on Learning Representations (ICLR), Toulon, France, 24–26 April 2017. [Google Scholar]
- Veličković, P.; Cucurull, G.; Casanova, A.; Romero, A.; Liò, P.; Bengio, Y. Graph Attention Networks. In Proceedings of the 6th International Conference on Learning Representations, Vancouver, BC, Canada, 30 April–3 May 2018. [Google Scholar]
- Davis, W. OpenAI opens GPT-3.5 Turbo up for custom tuning. Verge 2023. Available online: https://www.theverge.com/2023/8/22/23842042/openai-gpt-3-5-turbo-fine-tuning-enterprise-business-custom-chatbot-ai-artificial-intelligence (accessed on 2 December 2023).
Luxury Standard | Enron | Avocado | |
---|---|---|---|
Emails exchanged | 665,120 | 96,409 | 199,763 |
Replies received | 555,645 | 86,332 | 187,748 |
Employees | 4320 | 8403 | 1041 |
Time period in days | 443 | 730 | 730 |
E-mails per employee | 154 | 21 | 191 |
E-mails per day | 1501 | 132 | 273 |
Model | N = 665K | N = 23K | ||||
---|---|---|---|---|---|---|
Accuracy | Macro F1 | Minority F1 | Accuracy | Macro F1 | Minority F1 | |
Baseline | ||||||
CNN | 0.75 | 0.54 | 0.24 | 0.73 | 0.54 | 0.24 |
LSTM | 0.77 | 0.54 | 0.21 | 0.74 | 0.51 | 0.18 |
BiLSTM | 0.75 | 0.55 | 0.25 | 0.74 | 0.54 | 0.23 |
LSTM + Attention | 0.76 | 0.54 | 0.23 | 0.73 | 0.51 | 0.18 |
BERT | 0.80 | 0.51 | 0.12 | 0.80 | 0.54 | 0.20 |
RoBERTa | 0.81 | 0.54 | 0.19 | 0.81 | 0.54 | 0.18 |
EMMA: Email MultiModal Architecture | ||||||
EMMA with Network + Influence | 0.82 | 0.65 | 0.40 | |||
EMMA with Network + Influence + Context | 0.84 | 0.66 | 0.41 | |||
EMMA with Stylistic + Network + Influence + Context | 0.83 | 0.67 | 0.43 |
RoBERTa | 1 | 2 | 3 | 4 | |
---|---|---|---|---|---|
Stylistic features | - | ✓ | ✓ | ||
Organizational influence features | - | ✓ | ✓ | ✓ | |
Personal influence features | - | ✓ | ✓ | ✓ | |
Linguistic accommodation features | - | ✓ | ✓ | ||
Accuracy | 0.81 | 0.53 | 0.82 | 0.84 | 0.83 |
Macro F1 | 0.54 | 0.70 | 0.65 | 0.66 | 0.67 |
Minority F1 | 0.18 | 0.50 | 0.40 | 0.41 | 0.43 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Shah, H.; Jaidka, K.; Ungar, L.; Fagan, J.; Grosser, T. Building a Multimodal Classifier of Email Behavior: Towards a Social Network Understanding of Organizational Communication. Information 2023, 14, 661. https://doi.org/10.3390/info14120661
Shah H, Jaidka K, Ungar L, Fagan J, Grosser T. Building a Multimodal Classifier of Email Behavior: Towards a Social Network Understanding of Organizational Communication. Information. 2023; 14(12):661. https://doi.org/10.3390/info14120661
Chicago/Turabian StyleShah, Harsh, Kokil Jaidka, Lyle Ungar, Jesse Fagan, and Travis Grosser. 2023. "Building a Multimodal Classifier of Email Behavior: Towards a Social Network Understanding of Organizational Communication" Information 14, no. 12: 661. https://doi.org/10.3390/info14120661
APA StyleShah, H., Jaidka, K., Ungar, L., Fagan, J., & Grosser, T. (2023). Building a Multimodal Classifier of Email Behavior: Towards a Social Network Understanding of Organizational Communication. Information, 14(12), 661. https://doi.org/10.3390/info14120661