Informatics doi: 10.3390/informatics11010012
Authors: Abdulghafour Mohammad Job Mathew Kollamana
One of the main obstacles in software development projects is requirement volatility (RV), which is defined as uncertainty or changes in software requirements during the development process. Therefore, this research tries to understand the underlying factors behind the RV and the best practices to reduce it. The methodology used for this research is based upon qualitative research using interviews with 12 participants with experience in agile software development projects. The participants hailed from Austria, Nigeria, the USA, the Philippines, Armenia, Sri Lanka, Germany, Egypt, Canada, and Turkey and held roles such as project managers, software developers, Scrum Masters, testers, business analysts, and product owners. Our findings based on our empirical data revealed six primary factors that cause RV and three main agile practices that help to mitigate it. Theoretically, this study contributes to the body of knowledge relating to RV management. Practically, this research is expected to aid software development teams in comprehending the reasons behind RV and the best practices to effectively minimize it.
]]>Informatics doi: 10.3390/informatics11010011
Authors: Olga Kurasova Arnoldas Budžys Viktor Medvedev
As artificial intelligence has evolved, deep learning models have become important in extracting and interpreting complex patterns from raw multidimensional data. These models produce multidimensional embeddings that, while containing a lot of information, are often not directly understandable. Dimensionality reduction techniques play an important role in transforming multidimensional data into interpretable formats for decision support systems. To address this problem, the paper presents an analysis of dimensionality reduction and visualization techniques that embrace complex data representations and are useful inferences for decision systems. A novel framework is proposed, utilizing a Siamese neural network with a triplet loss function to analyze multidimensional data encoded into images, thus transforming these data into multidimensional embeddings. This approach uses dimensionality reduction techniques to transform these embeddings into a lower-dimensional space. This transformation not only improves interpretability but also maintains the integrity of the complex data structures. The efficacy of this approach is demonstrated using a keystroke dynamics dataset. The results support the integration of these visualization techniques into decision support systems. The visualization process not only simplifies the complexity of the data, but also reveals deep patterns and relationships hidden in the embeddings. Thus, a comprehensive framework for visualizing and interpreting complex keystroke dynamics is described, making a significant contribution to the field of user authentication.
]]>Informatics doi: 10.3390/informatics11010010
Authors: Malinka Ivanova Gabriela Grosseck Carmen Holotescu
The penetration of intelligent applications in education is rapidly increasing, posing a number of questions of a different nature to the educational community. This paper is coming to analyze and outline the influence of artificial intelligence (AI) on teaching practice which is an essential problem considering its growing utilization and pervasion on a global scale. A bibliometric approach is applied to outdraw the “big picture” considering gathered bibliographic data from scientific databases Scopus and Web of Science. Data on relevant publications matching the query “artificial intelligence and teaching” over the past 5 years have been researched and processed through Biblioshiny in R environment in order to establish a descriptive structure of the scientific production, to determine the impact of scientific publications, to trace collaboration patterns and to identify key research areas and emerging trends. The results point out the growth in scientific production lately that is an indicator of increased interest in the investigated topic by researchers who mainly work in collaborative teams as some of them are from different countries and institutions. The identified key research areas include techniques used in educational applications, such as artificial intelligence, machine learning, and deep learning. Additionally, there is a focus on applicable technologies like ChatGPT, learning analytics, and virtual reality. The research also explores the context of application for these techniques and technologies in various educational settings, including teaching, higher education, active learning, e-learning, and online learning. Based on our findings, the trending research topics can be encapsulated by terms such as ChatGPT, chatbots, AI, generative AI, machine learning, emotion recognition, large language models, convolutional neural networks, and decision theory. These findings offer valuable insights into the current landscape of research interests in the field.
]]>Informatics doi: 10.3390/informatics11010009
Authors: Angelica Lo Duca Matteo Abrate Andrea Marchetti Manuela Moretti
This paper describes the Integration of Archives and Cultural Places (IaCuP) project, which aims to integrate information about a historical cemetery, including its map and grave inventory, with genealogical and documentary knowledge extracted from relevant historical archives. The integrated data are accessible to cemetery visitors through an interactive mobile application, enabling them to navigate a graphical representation of the cemetery while exploring comprehensive visualizations of genealogical data. The basic idea stems from the desire to provide people with access to the rich context of cultural sites, which have often lost their original references over the centuries, making it challenging for individuals today to interpret the meanings embedded within them. The proposed approach leverages large language models (LLMs) to extract information from relevant documents and Web technologies to represent such information as interactive visualizations. As a practical case study, this paper focuses on the Jewish Cemetery in Pisa and the Historical Archives of the Jewish Community in Pisa, working on the genealogical tree of one of the most representative families resting in the cemetery.
]]>Informatics doi: 10.3390/informatics11010008
Authors: Margarida Mendonça Álvaro Figueira
As social media (SM) becomes increasingly prevalent, its impact on society is expected to grow accordingly. While SM has brought positive transformations, it has also amplified pre-existing issues such as misinformation, echo chambers, manipulation, and propaganda. A thorough comprehension of this impact, aided by state-of-the-art analytical tools and by an awareness of societal biases and complexities, enables us to anticipate and mitigate the potential negative effects. One such tool is BERTopic, a novel deep-learning algorithm developed for Topic Mining, which has been shown to offer significant advantages over traditional methods like Latent Dirichlet Allocation (LDA), particularly in terms of its high modularity, which allows for extensive personalization at each stage of the topic modeling process. In this study, we hypothesize that BERTopic, when optimized for Twitter data, can provide a more coherent and stable topic modeling. We began by conducting a review of the literature on topic-mining approaches for short-text data. Using this knowledge, we explored the potential for optimizing BERTopic and analyzed its effectiveness. Our focus was on Twitter data spanning the two years of the 117th US Congress. We evaluated BERTopic’s performance using coherence, perplexity, diversity, and stability scores, finding significant improvements over traditional methods and the default parameters for this tool. We discovered that improvements are possible in BERTopic’s coherence and stability. We also identified the major topics of this Congress, which include abortion, student debt, and Judge Ketanji Brown Jackson. Additionally, we describe a simple application we developed for a better visualization of Congress topics.
]]>Informatics doi: 10.3390/informatics11010007
Authors: Mariana Ávalos-Arce Heráclito Pérez-Díaz Carolina Del-Valle-Soto Ramon A. Briseño
Wireless networks play a pivotal role in various domains, including industrial automation, autonomous vehicles, robotics, and mobile sensor networks. This research investigates the critical issue of packet loss in modern wireless networks and aims to identify the conditions within a network’s environment that lead to such losses. We propose a packet status prediction model for data packets that travel through a wireless network based on the IEEE 802.15.4 standard and are exposed to five different types of interference in a controlled experimentation environment. The proposed model focuses on the packetization process and its impact on network robustness. This study explores the challenges posed by packet loss, particularly in the context of interference, and puts forth the hypothesis that specific environmental conditions are linked to packet loss occurrences. The contribution of this work lies in advancing our understanding of the conditions leading to packet loss in wireless networks. Data are retrieved with a single CC2531 USB Dongle Packet Sniffer, whose pieces of information on packets become the features of each packet from which the classifier model will gather the training data with the aim of predicting whether a packet will unsuccessfully arrive at its destination. We found that interference causes more packet loss than that caused by various devices using a WiFi communication protocol simultaneously. In addition, we found that the most important predictors are network strength and packet size; low network strength tends to lead to more packet loss, especially for larger packets. This study contributes to the ongoing efforts to predict and mitigate packet loss, emphasizing the need for adaptive models in dynamic wireless environments.
]]>Informatics doi: 10.3390/informatics11010006
Authors: Sofía Ramos-Pulido Neil Hernández-Gress Gabriela Torres-Delgado
Current research on the career satisfaction of graduates limits educational institutions in devising methods to attain high career satisfaction. Thus, this study aims to use data science models to understand and predict career satisfaction based on information collected from surveys of university alumni. Five machine learning (ML) algorithms were used for data analysis, including the decision tree, random forest, gradient boosting, support vector machine, and neural network models. To achieve optimal prediction performance, we utilized the Bayesian optimization method to fine-tune the parameters of the five ML algorithms. The five ML models were compared with logistic and ordinal regression. Then, to extract the most important features of the best predictive model, we employed the SHapley Additive exPlanations (SHAP), a novel methodology for extracting the significant features in ML. The results indicated that gradient boosting is a marginally superior predictive model, with 2–3% higher accuracy and area under the receiver operating characteristic curve (AUC) compared to logistic and ordinal regression. Interestingly, concerning low career satisfaction, those with the worst scores for the phrase “how frequently applied knowledge, skills, or technological tools from the academic training” were less satisfied with their careers. To summarize, career satisfaction is related to academic training, alumni satisfaction, employment status, published articles or books, and other factors.
]]>Informatics doi: 10.3390/informatics11010005
Authors: Charlee Kaewrat Dollaporn Anopas Si Thu Aung Yunyong Punsawad
This study presents an augmented reality application for training chest electrocardiography electrode placement. AR applications featuring augmented object displays and interactions have been developed to facilitate learning and training of electrocardiography (ECG) chest lead placement via smartphones. The AR marker-based technique was used to track the objects. The proposed AR application can project virtual ECG electrode positions onto the mannequin’s chest and provide feedback to trainees. We designed experimental tasks using the pre- and post-tests and practice sessions to verify the efficiency of the proposed AR application. The control group was assigned to learn chest ECG electrode placement using traditional methods, whereas the intervention group was introduced to the proposed AR application for ECG electrode placement. The results indicate that the proposed AR application can encourage learning outcomes, such as chest lead ECG knowledge and skills. Moreover, using AR technology can enhance students’ learning experiences. In the future, we plan to apply the proposed AR technology to improve related courses in medical science education.
]]>Informatics doi: 10.3390/informatics11010004
Authors: Aokun Chen Yunpeng Zhao Yi Zheng Hui Hu Xia Hu Jennifer N. Fishe William R. Hogan Elizabeth A. Shenkman Yi Guo Jiang Bian
It is prudent to take a unified approach to exploring how contextual social determinants of health (SDoH) relate to COVID-19 occurrence and outcomes. Poor geographically represented data and a small number of contextual SDoH examined in most previous research studies have left a knowledge gap in the relationships between contextual SDoH and COVID-19 outcomes. In this study, we linked 199 contextual SDoH factors covering 11 domains of social and built environments with electronic health records (EHRs) from a large clinical research network (CRN) in the National Patient-Centered Clinical Research Network (PCORnet) to explore the relation between contextual SDoH and COVID-19 occurrence and hospitalization. We identified 15,890 COVID-19 patients and 63,560 matched non-COVID-19 patients in Florida between January 2020 and May 2021. We adopted a two-phase multiple linear regression approach modified from that in the exposome-wide association (ExWAS) study. After removing the highly correlated SDoH variables, 86 contextual SDoH variables were included in the data analysis. Adjusting for race, ethnicity, and comorbidities, we found six contextual SDoH variables (i.e., hospital available beds and utilization, percent of vacant property, number of golf courses, and percent of minority) related to the occurrence of COVID-19, and three variables (i.e., farmers market, low access, and religion) related to the hospitalization of COVID-19. To our best knowledge, this is the first study to explore the relationship between contextual SDoH and COVID-19 occurrence and hospitalization using EHRs in a major PCORnet CRN. As an exploratory study, the causal effect of SDoH on COVID-19 outcomes will be evaluated in future studies.
]]>Informatics doi: 10.3390/informatics11010003
Authors: Sandro Pullo Remo Pareschi Valentina Piantadosi Francesco Salzano Roberto Carlini
Addressing the critical challenges of resource inefficiency and environmental impact in the agrifood sector, this study explores the integration of Internet of Things (IoT) technologies with IOTA’s Tangle, a Distributed Ledger Technology (DLT). This integration aims to enhance sustainable agricultural practices, using rice cultivation as a case study of high relevance and reapplicability given its importance in the food chain and the high irrigation requirement of its cultivation. The approach employs sensor-based intelligent irrigation systems to optimize water efficiency. These systems enable real-time monitoring of agricultural parameters through IoT sensors. Data management is facilitated by IOTA’s Tangle, providing secure and efficient data handling, and integrated with MongoDB, a Database Management System (DBMS), for effective data storage and retrieval. The collaboration between IoT and IOTA led to significant reductions in resource consumption. Implementing sustainable agricultural practices resulted in a 50% reduction in water usage, 25% decrease in nitrogen consumption, and a 50% to 70% reduction in methane emissions. Additionally, the system contributed to lower electricity consumption for irrigation pumps and generated comprehensive historical water depth records, aiding future resource management decisions. This study concludes that the integration of IoT with IOTA’s Tangle presents a highly promising solution for advancing sustainable agriculture. This approach significantly contributes to environmental conservation and food security. Furthermore, it establishes that DLTs like IOTA are not only viable but also effective for real-time monitoring and implementation of sustainable agricultural practices.
]]>Informatics doi: 10.3390/informatics11010002
Authors: Isaac Machorro-Cano José Oscar Olmedo-Aguirre Giner Alor-Hernández Lisbeth Rodríguez-Mazahua Laura Nely Sánchez-Morales Nancy Pérez-Castro
Cloud-based platforms have gained popularity over the years because they can be used for multiple purposes, from synchronizing contact information to storing and managing user fitness data. These platforms are still in constant development and, so far, most of the data they store is entered manually by users. However, more and better wearable devices are being developed that can synchronize with these platforms to feed the information automatically. Another aspect that highlights the link between wearable devices and cloud-based health platforms is the improvement in which the symptomatology and/or physical status information of users can be stored and syn-chronized in real-time, 24 h a day, in health platforms, which in turn enables the possibility of synchronizing these platforms with specialized medical software to promptly detect important variations in user symptoms. This is opening opportunities to use these platforms as support for monitoring disease symptoms and, in general, for monitoring the health of users. In this work, the characteristics and possibilities of use of four popular platforms currently available in the market are explored, which are Apple Health, Google Fit, Samsung Health, and Fitbit.
]]>Informatics doi: 10.3390/informatics11010001
Authors: Andrew Vargo Kohei Yamaguchi Motoi Iwata Koichi Kise
Vocabulary acquisition and retention is an essential part of learning a foreign language and many learners use flashcard applications to repetitively increase vocabulary retention. However, it can be difficult for learners to remember new words and phrases without any context. In this paper, we propose a system that allows users to acquire new vocabulary with media which gives context to the words. Theoretically, this use of multimedia context should enable users to practice with interest and increased motivation, which has been shown to enhance the effects of contextual language learning. An experiment with 46 English as foreign language learners showed better retention after two weeks with the proposed system as compared to ordinary flashcards. However, the impact was not universally beneficial to all learners. An analysis of participant attributes that were gathered through surveys and questionnaires shows a link between personality and learning traits and affinity for learning with this system. This result indicates that the proposed system provides a significant advantage in vocabulary retention for some users, while other users should stay with traditional flashcard applications. The implications of this study indicate the need for the development of more personalized learning applications.
]]>Informatics doi: 10.3390/informatics10040090
Authors: Egor Ushakov Anton Naumov Vladislav Fomberg Polina Vishnyakova Aleksandra Asaturova Alina Badlaeva Anna Tregubova Evgeny Karpulevich Gennady Sukhikh Timur Fatkhudinov
H-score is a semi-quantitative method used to assess the presence and distribution of proteins in tissue samples by combining the intensity of staining and the percentage of stained nuclei. It is widely used but time-consuming and can be limited in terms of accuracy and precision. Computer-aided methods may help overcome these limitations and improve the efficiency of pathologists’ workflows. In this work, we developed a model EndoNet for automatic H-score calculation on histological slides. Our proposed method uses neural networks and consists of two main parts. The first is a detection model which predicts the keypoints of centers of nuclei. The second is an H-score module that calculates the value of the H-score using mean pixel values of predicted keypoints. Our model was trained and validated on 1780 annotated tiles with a shape of 100 × 100 µm and we achieved 0.77 mAP on a test dataset. We obtained our best results in H-score calculation; these results proved superior to QuPath predictions. Moreover, the model can be adjusted to a specific specialist or whole laboratory to reproduce the manner of calculating the H-score. Thus, EndoNet is effective and robust in the analysis of histology slides, which can improve and significantly accelerate the work of pathologists.
]]>Informatics doi: 10.3390/informatics10040089
Authors: Jaskaran Gill Madhu Chetty Suryani Lim Jennifer Hallinan
Relation extraction from biological publications plays a pivotal role in accelerating scientific discovery and advancing medical research. While vast amounts of this knowledge is stored within the published literature, extracting it manually from this continually growing volume of documents is becoming increasingly arduous. Recently, attention has been focused towards automatically extracting such knowledge using pre-trained Large Language Models (LLM) and deep-learning algorithms for automated relation extraction. However, the complex syntactic structure of biological sentences, with nested entities and domain-specific terminology, and insufficient annotated training corpora, poses major challenges in accurately capturing entity relationships from the unstructured data. To address these issues, in this paper, we propose a Knowledge-based Intelligent Text Simplification (KITS) approach focused on the accurate extraction of biological relations. KITS is able to precisely and accurately capture the relational context among various binary relations within the sentence, alongside preventing any potential changes in meaning for those sentences being simplified by KITS. The experiments show that the proposed technique, using well-known performance metrics, resulted in a 21% increase in precision, with only 25% of sentences simplified in the Learning Language in Logic (LLL) dataset. Combining the proposed method with BioBERT, the popular pre-trained LLM was able to outperform other state-of-the-art methods.
]]>Informatics doi: 10.3390/informatics10040088
Authors: Paraskevas Koukaras Dimitrios Rousidis Christos Tjortjis
The identification and analysis of sentiment polarity in microblog data has drawn increased attention. Researchers and practitioners attempt to extract knowledge by evaluating public sentiment in response to global events. This study aimed to evaluate public attitudes towards the spread of COVID-19 by performing sentiment analysis on over 2.1 million tweets in English. The implications included the generation of insights for timely disease outbreak prediction and assertions regarding worldwide events, which can help policymakers take suitable actions. We investigated whether there was a correlation between public sentiment and the number of cases and deaths attributed to COVID-19. The research design integrated text preprocessing (regular expression operations, (de)tokenization, stopwords), sentiment polarization analysis via TextBlob, hypothesis formulation (null hypothesis testing), and statistical analysis (Pearson coefficient and p-value) to produce the results. The key findings highlight a correlation between sentiment polarity and deaths, starting at 41 days before and expanding up to 3 days after counting. Twitter users reacted to increased numbers of COVID-19-related deaths after four days by posting tweets with fading sentiment polarization. We also detected a strong correlation between COVID-19 Twitter conversation polarity and reported cases and a weak correlation between polarity and reported deaths.
]]>Informatics doi: 10.3390/informatics10040087
Authors: Joost C. F. de Winter Dimitra Dodou Arno H. A. Stienen
ChatGPT is widely used among students, a situation that challenges educators. The current paper presents two strategies that do not push educators into a defensive role but can empower them. Firstly, we show, based on statistical analysis, that ChatGPT use can be recognized from certain keywords such as ‘delves’ and ‘crucial’. This insight allows educators to detect ChatGPT-assisted work more effectively. Secondly, we illustrate that ChatGPT can be used to assess texts written by students. The latter topic was presented in two interactive workshops provided to educators and educational specialists. The results of the workshops, where prompts were tested live, indicated that ChatGPT, provided a targeted prompt is used, is good at recognizing errors in texts but not consistent in grading. Ethical and copyright concerns were raised as well in the workshops. In conclusion, the methods presented in this paper may help fortify the teaching methods of educators. The computer scripts that we used for live prompting are available and enable educators to give similar workshops.
]]>Informatics doi: 10.3390/informatics10040086
Authors: Brian Rizqi Paradisiaca Darnoto Daniel Siahaan Diana Purwitasari
Persuasive content in online news contains elements that aim to persuade its readers and may not necessarily include factual information. Since a news article only has some sentences that indicate persuasiveness, it would be quite challenging to differentiate news with or without the persuasive content. Recognizing persuasive sentences with a text summarization and classification approach is important to understand persuasive messages effectively. Text summarization identifies arguments and key points, while classification separates persuasive sentences based on the linguistic and semantic features used. Our proposed architecture includes text summarization approaches to shorten sentences without persuasive content and then using classifiers model to detect those with persuasive indication. In this paper, we compare the performance of latent semantic analysis (LSA) and TextRank in text summarization methods, the latter of which has outperformed in all trials, and also two classifiers of convolutional neural network (CNN) and bidirectional long short-term memory (BiLSTM). We have prepared a dataset (±1700 data and manually persuasiveness-labeled) consisting of news articles written in the Indonesian language collected from a nationwide electronic news portal. Comparative studies in our experimental results show that the TextRank–BERT–BiLSTM model achieved the highest accuracy of 95% in detecting persuasive news. The text summarization methods were able to generate detailed and precise summaries of the news articles and the deep learning models were able to effectively differentiate between persuasive news and real news.
]]>Informatics doi: 10.3390/informatics10040085
Authors: Moonkyoung Jang
This study delves into the determinants influencing individuals’ intentions to adopt telemedicine apps during the COVID-19 pandemic. The study aims to offer a comprehensive framework for understanding behavioral intentions by leveraging the Technology Acceptance Model (TAM), supplemented by e-health literacy and social influence variables. The study analyzes survey data from 364 adults using partial least squares structural equation modeling (PLS-SEM) to empirically examine the internal relationships within the model. Results indicated that e-health literacy, attitude, and social influence significantly impacted the intention to use telemedicine apps. Notably, e-health literacy positively influenced both perceived usefulness and ease of use, expanding beyond mere usage intention. The study underscored the substantial role of social influence in predicting the intention to use telemedicine apps, challenging the traditional oversight of social influence in the TAM framework. The findings will help researchers, practitioners, and governments understand how social influence and e-health literacy influence the adoption of telehealth apps and promote the use of telehealth apps through enhancing social influence and e-health literacy.
]]>Informatics doi: 10.3390/informatics10040084
Authors: Evaristus D. Madyatmadja Corinthias P. M. Sianipar Cristofer Wijaya David J. M. Sembiring
Crowdsourcing has gradually become an effective e-government process to gather citizen complaints over the implementation of various public services. In practice, the collected complaints form a massive dataset, making it difficult for government officers to analyze the big data effectively. It is consequently vital to use data mining algorithms to classify the citizen complaint data for efficient follow-up actions. However, different classification algorithms produce varied classification accuracies. Thus, this study aimed to compare the accuracy of several classification algorithms on crowdsourced citizen complaint data. Taking the case of the LAKSA app in Tangerang City, Indonesia, this study included k-Nearest Neighbors, Random Forest, Support Vector Machine, and AdaBoost for the accuracy assessment. The data were taken from crowdsourced citizen complaints submitted to the LAKSA app, including those aggregated from official social media channels, from May 2021 to April 2022. The results showed SVM with a linear kernel as the most accurate among the assessed algorithms (89.2%). In contrast, AdaBoost (base learner: Decision Trees) produced the lowest accuracy. Still, the accuracy levels of all algorithms varied in parallel to the amount of training data available for the actual classification categories. Overall, the assessments on all algorithms indicated that their accuracies were insignificantly different, with an overall variation of 4.3%. The AdaBoost-based classification, in particular, showed its large dependence on the choice of base learners. Looking at the method and results, this study contributes to e-government, data mining, and big data discourses. This research recommends that governments continuously conduct supervised training of classification algorithms over their crowdsourced citizen complaints to seek the highest accuracy possible, paving the way for smart and sustainable governance.
]]>Informatics doi: 10.3390/informatics10040083
Authors: Hendrik Ballhausen Ludwig Christian Hinske
Privacy-preserving computation (PPC) enables encrypted computation of private data. While advantageous in theory, the complex technology has steep barriers to entry in practice. Here, we derive design goals and principles for a middleware that encapsulates the demanding cryptography server side and provides a simple-to-use interface to client-side application developers. The resulting architecture, “Federated Secure Computing”, offloads computing-intensive tasks to the server and separates concerns of cryptography and business logic. It provides microservices through an Open API 3.0 definition and hosts multiple protocols through self-discovered plugins. It requires only minimal DevSecOps capabilities and is straightforward and secure. Finally, it is small enough to work in the internet of things (IoT) and in propaedeutic settings on consumer hardware. We provide benchmarks for calculations with a secure multiparty computation (SMPC) protocol, both for vertically and horizontally partitioned data. Runtimes are in the range of seconds on both dedicated workstations and IoT devices such as Raspberry Pi or smartphones. A reference implementation is available as free and open source software under the MIT license.
]]>Informatics doi: 10.3390/informatics10040082
Authors: Luke Balcombe
Artificial intelligence (AI) chatbots have gained prominence since 2022. Powered by big data, natural language processing (NLP) and machine learning (ML) algorithms, they offer the potential to expand capabilities, improve productivity and provide guidance and support in various domains. Human–Artificial Intelligence (HAI) is proposed to help with the integration of human values, empathy and ethical considerations into AI in order to address the limitations of AI chatbots and enhance their effectiveness. Mental health is a critical global concern, with a substantial impact on individuals, communities and economies. Digital mental health solutions, leveraging AI and ML, have emerged to address the challenges of access, stigma and cost in mental health care. Despite their potential, ethical and legal implications surrounding these technologies remain uncertain. This narrative literature review explores the potential of AI chatbots to revolutionize digital mental health while emphasizing the need for ethical, responsible and trustworthy AI algorithms. The review is guided by three key research questions: the impact of AI chatbots on technology integration, the balance between benefits and harms, and the mitigation of bias and prejudice in AI applications. Methodologically, the review involves extensive database and search engine searches, utilizing keywords related to AI chatbots and digital mental health. Peer-reviewed journal articles and media sources were purposively selected to address the research questions, resulting in a comprehensive analysis of the current state of knowledge on this evolving topic. In conclusion, AI chatbots hold promise in transforming digital mental health but must navigate complex ethical and practical challenges. The integration of HAI principles, responsible regulation and scoping reviews are crucial to maximizing their benefits while minimizing potential risks. Collaborative approaches and modern educational solutions may enhance responsible use and mitigate biases in AI applications, ensuring a more inclusive and effective digital mental health landscape.
]]>Informatics doi: 10.3390/informatics10040081
Authors: Célia Tavares Luciana Oliveira Pedro Duarte Manuel Moreira da Silva
According to a recent study by OpenAI, Open Research, and the University of Pennsylvania, large language models (LLMs) based on artificial intelligence (AI), such as generative pretrained transformers (GPTs), may have potential implications for the job market, specifically regarding occupations that demand writing or programming skills. This research points out that interpreters and translators are one of the main occupations with greater exposure to AI in the US job market (76.5%), in a trend that is expected to affect other regions of the globe. This article, following a mixed-methods survey-based research approach, provides insights into the awareness and knowledge about AI among Portuguese language service providers (LSPs), specifically regarding neural machine translation (NMT) and large language models (LLM), their actual use and usefulness, as well as their potential influence on work performance and the labour market. The results show that most professionals are unable to identify whether AI and/or automation technologies support the tools that are most used in the profession. The usefulness of AI is essentially low to moderate and the professionals who are less familiar with it and less knowledgeable also demonstrate a lack of trust in it. Two thirds of the sample estimate negative or very negative effects of AI in their profession, expressing the devaluation and replacement of experts, the reduction of income, and the reconfiguration of the career of translator to mere post-editors as major concerns.
]]>Informatics doi: 10.3390/informatics10040080
Authors: Patrick Cheong-Iao Pang Megan Munsie Shanton Chang
People are increasingly seeking complex health information online. However, how they access this information and how influential it is on their health choices remains poorly understood. Google Analytics (GA) is a widely used web analytics tool and it has been used in academic research to study health information-seeking behaviors. Nevertheless, it is rarely used to study the navigation flows of health websites. To demonstrate the usefulness of GA data, we adopted both top-down and bottom-up approaches to study how web visitors navigate within a website delivering complex health information about stem cell research using GA’s device, traffic and path data. Custom Treemap and Sankey visualizations were used to illustrate the navigation flows extracted from these data in a more understandable manner. Our methodology reveals that different device and traffic types expose dissimilar search approaches. Through the visualizations, popular web pages and content categories frequently browsed together can be identified. Information on a website that is often overlooked but needed by many users can also be discovered. Our proposed method can identify content requiring improvements, enhance usability and guide a design for better addressing the needs of different audiences. This paper has implications for how web designers can use GA to help them determine users’ priorities and behaviors when navigating complex information. It highlights that even where there is complex health information, users may still want more direct and easy-to-understand navigations to retrieve such information.
]]>Informatics doi: 10.3390/informatics10040079
Authors: Anam Shahil-Feroz Haleema Yasmin Sarah Saleem Zulfiqar Bhutta Emily Seto
This study assessed the usability of the smartphone app, named “Raabta” from the perspective of pregnant women at high risk of preeclampsia to improve the Raabta app for future implementation. Think-aloud and task-completion techniques were used with a purposive sample of 14 pregnant women at high risk of preeclampsia. The sessions were audio-recorded and later professionally transcribed for thematic analysis. The study generated learnings associated with four themes: improving the clarity of instructions, messaging, and terminology; accessibility for non-tech savvy and illiterate Urdu users; enhancing visuals and icons for user engagement; and simplifying navigation and functionality. Overall, user feedback emphasized the importance of enhancing the clarity of instructions, messaging, and terminology within the Raabta app. Voice messages and visuals were valued by users, particularly among the non-tech savvy and illiterate Urdu users, as they enhance accessibility and enable independent monitoring. Suggestions were made to enhance user engagement through visual improvements such as enhanced graphics and culturally aligned color schemes. Lastly, users highlighted the need for improved navigation both between screens and within screens to enhance the overall user experience. The Raabta app prototype will be modified based on the feedback of the users to address the unique needs of diverse groups.
]]>Informatics doi: 10.3390/informatics10040078
Authors: Andreas Dengel Rupert Gehrlein David Fernes Sebastian Görlich Jonas Maurer Hai Hoang Pham Gabriel Großmann Niklas Dietrich genannt Eisermann
In the current era of artificial intelligence, large language models such as ChatGPT and BARD are being increasingly used for various applications, such as language translation, text generation, and human-like conversation. The fact that these models consist of large amounts of data, including many different opinions and perspectives, could introduce the possibility of a new qualitative research approach: Due to the probabilistic character of their answers, “interviewing” these large language models could give insights into public opinions in a way that otherwise only interviews with large groups of subjects could deliver. However, it is not yet clear if qualitative content analysis research methods can be applied to interviews with these models. Evaluating the applicability of qualitative research methods to interviews with large language models could foster our understanding of their abilities and limitations. In this paper, we examine the applicability of qualitative content analysis research methods to interviews with ChatGPT in English, ChatGPT in German, and BARD in English on the relevance of computer science in K-12 education, which was used as an exemplary topic. We found that the answers produced by these models strongly depended on the provided context, and the same model could produce heavily differing results for the same questions. From these results and the insights throughout the process, we formulated guidelines for conducting and analyzing interviews with large language models. Our findings suggest that qualitative content analysis research methods can indeed be applied to interviews with large language models, but with careful consideration of contextual factors that may affect the responses produced by these models. The guidelines we provide can aid researchers and practitioners in conducting more nuanced and insightful interviews with large language models. From an overall view of our results, we generally do not recommend using interviews with large language models for research purposes, due to their highly unpredictable results. However, we suggest using these models as exploration tools for gaining different perspectives on research topics and for testing interview guidelines before conducting real-world interviews.
]]>Informatics doi: 10.3390/informatics10040077
Authors: Fan Zhang Melissa Petersen Leigh Johnson James Hall Raymond F. Palmer Sid E. O’Bryant on behalf of the Health and Aging Brain Study (HABS–HD) Study Team on behalf of the Health and Aging Brain Study (HABS–HD) Study Team
The Health and Aging Brain Study–Health Disparities (HABS–HD) project seeks to understand the biological, social, and environmental factors that impact brain aging among diverse communities. A common issue for HABS–HD is missing data. It is impossible to achieve accurate machine learning (ML) if data contain missing values. Therefore, developing a new imputation methodology has become an urgent task for HABS–HD. The three missing data assumptions, (1) missing completely at random (MCAR), (2) missing at random (MAR), and (3) missing not at random (MNAR), necessitate distinct imputation approaches for each mechanism of missingness. Several popular imputation methods, including listwise deletion, min, mean, predictive mean matching (PMM), classification and regression trees (CART), and missForest, may result in biased outcomes and reduced statistical power when applied to downstream analyses such as testing hypotheses related to clinical variables or utilizing machine learning to predict AD or MCI. Moreover, these commonly used imputation techniques can produce unreliable estimates of missing values if they do not account for the missingness mechanisms or if there is an inconsistency between the imputation method and the missing data mechanism in HABS–HD. Therefore, we proposed a three-step workflow to handle missing data in HABS–HD: (1) missing data evaluation, (2) imputation, and (3) imputation evaluation. First, we explored the missingness in HABS–HD. Then, we developed a machine learning-based multiple imputation method (MLMI) for imputing missing values. We built four ML-based imputation models (support vector machine (SVM), random forest (RF), extreme gradient boosting (XGB), and lasso and elastic-net regularized generalized linear model (GLMNET)) and adapted the four ML-based models to multiple imputations using the simple averaging method. Lastly, we evaluated and compared MLMI with other common methods. Our results showed that the three-step workflow worked well for handling missing values in HABS–HD and the ML-based multiple imputation method outperformed other common methods in terms of prediction performance and change in distribution and correlation. The choice of missing handling methodology has a significant impact on the accompanying statistical analyses of HABS–HD. The conceptual three-step workflow and the ML-based multiple imputation method perform well for our Alzheimer’s disease models. They can also be applied to other disease data analyses.
]]>Informatics doi: 10.3390/informatics10040076
Authors: Samuel R. Schrader Eren Gultepe
The evaluation of similarities between natural languages often relies on prior knowledge of the languages being studied. We describe three methods for building phylogenetic trees and clustering languages without the use of language-specific information. The input to our methods is a set of document vectors trained on a corpus of parallel translations of the Bible into 22 Indo-European languages, representing 4 language families: Indo-Iranian, Slavic, Germanic, and Romance. This text corpus consists of a set of 532,092 Bible verses, with 24,186 identical verses translated into each language. The methods are (A) hierarchical clustering using distance between language vector centroids, (B) hierarchical clustering using a network-derived distance measure, and (C) Deep Embedded Clustering (DEC) of language vectors. We evaluate our methods using a ground-truth tree and language families derived from said tree. All three achieve clustering F-scores above 0.9 on the Indo-Iranian and Slavic families; most confusion is between the Germanic and Romance families. The mean F-scores across all families are 0.864 (centroid clustering), 0.953 (network partitioning), and 0.763 (DEC). This shows that document vectors can be used to capture and compare linguistic features of multilingual texts, and thus could help extend language similarity and other translation studies research.
]]>Informatics doi: 10.3390/informatics10030075
Authors: Nevcihan Toraman Aycan Pekpazar Cigdem Altin Gumussoy
The aim of this study is to conceptualize website usability and develop a survey instrument to measure related concepts from the perspective of end users. We designed a three-stage methodology. First, concepts related to website usability were derived using content analysis technique. A total of 16 constructs measuring website usability were defined with their explanations and corresponding open codes. Second, a survey instrument was developed according to the defined open codes and the literature. The instrument was first validated using face validity, pilot testing (n = 30), and content validity (n = 40). Third, the survey instrument was validated using explanatory and confirmatory analyses. In the explanatory analysis, 785 questionnaires were collected from e-commerce website users to validate the factor structure of website usability. For confirmatory factor analysis, a new sample collected from 1086 users of e-commerce websites was used to confirm the measurement model. In addition, nomological validation was conducted by analyzing the effect of website usability concepts on three key factors: “continued intention to use”, “satisfaction”, and “brand loyalty”.
]]>Informatics doi: 10.3390/informatics10030074
Authors: Bisni Fahad Mon Asma Wasfi Mohammad Hayajneh Ahmad Slim Najah Abu Ali
The utilization of reinforcement learning (RL) within the field of education holds the potential to bring about a significant shift in the way students approach and engage with learning and how teachers evaluate student progress. The use of RL in education allows for personalized and adaptive learning, where the difficulty level can be adjusted based on a student’s performance. As a result, this could result in heightened levels of motivation and engagement among students. The aim of this article is to investigate the applications and techniques of RL in education and determine its potential impact on enhancing educational outcomes. It compares the various policies induced by RL with baselines and identifies four distinct RL techniques: the Markov decision process, partially observable Markov decision process, deep RL network, and Markov chain, as well as their application in education. The main focus of the article is to identify best practices for incorporating RL into educational settings to achieve effective and rewarding outcomes. To accomplish this, the article thoroughly examines the existing literature on using RL in education and its potential to advance educational technology. This work provides a thorough analysis of the various techniques and applications of RL in education to answer questions related to the effectiveness of RL in education and its future prospects. The findings of this study will provide researchers with a benchmark to compare the usefulness and effectiveness of commonly employed RL algorithms and provide direction for future research in education.
]]>Informatics doi: 10.3390/informatics10030073
Authors: Vanessa García-Pineda Alejandro Valencia-Arias Juan Camilo Patiño-Vanegas Juan José Flores Cueto Diana Arango-Botero Angel Marcelo Rojas Coronel Paula Andrea Rodríguez-Correa
This article aims to examine the research trends in the development of mobile networks from machine learning. The methodological approach starts from an analysis of 260 academic documents selected from the Scopus and Web of Science databases and is based on the parameters of the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) statement. Quantity, quality and structure indicators are calculated in order to contextualize the documents’ thematic evolution. The results reveal that, in relation to the publications by country, the United States and China, who are competing for fifth generation (5G) network coverage and are responsible for manufacturing devices for mobile networks, stand out. Most of the research on the subject focuses on the optimization of resources and traffic to guarantee the best management and availability of a network due to the high demand for resources and greater amount of traffic generated by the many Internet of Things (IoT) devices that are being developed for the market. It is concluded that thematic trends focus on generating algorithms for recognizing and learning the data in the network and on trained models that draw from the available data to improve the experience of connecting to mobile networks.
]]>Informatics doi: 10.3390/informatics10030072
Authors: Diego Renza Dora Ballesteros
CNN models can have millions of parameters, which makes them unattractive for some applications that require fast inference times or small memory footprints. To overcome this problem, one alternative is to identify and remove weights that have a small impact on the loss function of the algorithm, which is known as pruning. Typically, pruning methods are compared in terms of performance (e.g., accuracy), model size and inference speed. However, it is unusual to evaluate whether a pruned model preserves regions of importance in an image when performing inference. Consequently, we propose a metric to assess the impact of a pruning method based on images obtained by model interpretation (specifically, class activation maps). These images are spatially and spectrally compared and integrated by the harmonic mean for all samples in the test dataset. The results show that although the accuracy in a pruned model may remain relatively constant, the areas of attention for decision making are not necessarily preserved. Furthermore, the performance of pruning methods can be easily compared as a function of the proposed metric.
]]>Informatics doi: 10.3390/informatics10030071
Authors: Omar Flor-Unda Freddy Simbaña Xavier Larriva-Novo Ángel Acuña Rolando Tipán Patricia Acosta-Vargas
Vulnerabilities in cyber defense in the countries of the Latin American region have favored the activities of cybercriminals from different parts of the world who have carried out a growing number of cyberattacks that affect public and private services and compromise the integrity of users and organizations. This article describes the most representative vulnerabilities related to cyberattacks that have affected different sectors of countries in the Latin American region. A systematic review of repositories and the scientific literature was conducted, considering journal articles, conference proceedings, and reports from official bodies and leading brands of cybersecurity systems. The cybersecurity vulnerabilities identified in the countries of the Latin American region are low cybersecurity awareness, lack of standards and regulations, use of outdated software, security gaps in critical infrastructure, and lack of training and professional specialization.
]]>Informatics doi: 10.3390/informatics10030070
Authors: Ashlin Lee
The digital humanities and social sciences are critical for addressing societal challenges such as climate change and disaster risk reduction. One way in which the digital humanities and social sciences add value, particularly in an increasingly digitised society, is by engaging different communities through digital services and products. Alongside this observation, the field of user experience (UX) has also become popular in industrial settings. UX specifically concerns designing and developing digital products and solutions, and, while it is popular in business and other academic domains, there is disquiet in the digital humanities/social sciences towards UX and a general lack of engagement. This paper shares the reflections and insights of a digital humanities/social science practitioner working on a UX project to build a prototype demonstrator for disaster risk reduction. Insights come from formal developmental and participatory evaluation activities, as well as qualitative self-reflection. The paper identifies lessons learnt, noting challenges experienced—including feelings of uncertainty and platform dependency—and reflects on the hesitancy practitioners may have and potential barriers in participation between UX and the digital humanities/social science. It concludes that digital humanities/social science practitioners have few skill barriers and offer a valued perspective, but unclear opportunities for critical engagement may present a barrier.
]]>Informatics doi: 10.3390/informatics10030069
Authors: Beenish Moalla Chaudhry Shekufeh Shafeie Mona Mohamed
Theoretical models play a vital role in understanding the barriers and facilitators for the acceptance or rejection of emerging technologies. We conducted a narrative review of theoretical models predicting acceptance and adoption of human enhancement embeddable technologies to assess how well those models have studied unique attributes and qualities of embeddables and to identify gaps in the literature. Our broad search across multiple databases and Google Scholar identified 16 relevant articles published since 2016. We discovered that three main theoretical models: the technology acceptance model (TAM), unified theory of acceptance and use of technology (UTAUT), and cognitive–affective–normative (CAN) model have been consistently used and refined to explain the acceptance of human enhancement embeddable technology. Psychological constructs such as self-efficacy, motivation, self-determination, and demographic factors were also explored as mediating and moderating variables. Based on our analysis, we collated the verified determinants into a comprehensive model, modifying the CAN model. We also identified gaps in the literature and recommended a further exploration of design elements and psychological constructs. Additionally, we suggest investigating other models such as the matching person and technology model (MPTM), the hedonic-motivation system adoption model (HMSAM), and the value-based adoption model (VAM) to provide a more nuanced understanding of embeddable technologies’ adoption. Our study not only synthesizes the current state of research but also provides a robust framework for future investigations. By offering insights into the complex interplay of factors influencing the adoption of embeddable technologies, we contribute to the development of more effective strategies for design, implementation, and acceptance, thereby paving the way for the successful integration of these technologies into everyday life.
]]>Informatics doi: 10.3390/informatics10030068
Authors: Konstantinos I. Roumeliotis Nikolaos D. Tselikas
In the present-day digital landscape, websites have increasingly relied on digital marketing practices, notably search engine optimization (SEO), as a vital component in promoting sustainable growth. The traffic a website receives directly determines its development and success. As such, website owners frequently engage the services of SEO experts to enhance their website’s visibility and increase traffic. These specialists employ premium SEO audit tools that crawl the website’s source code to identify structural changes necessary to comply with specific ranking criteria, commonly called SEO factors. Working collaboratively with developers, SEO specialists implement technical changes to the source code and await the results. The cost of purchasing premium SEO audit tools or hiring an SEO specialist typically ranges in the thousands of dollars per year. Against this backdrop, this research endeavors to provide an open-source Python-based Machine Learning SEO software tool to the general public, catering to the needs of both website owners and SEO specialists. The tool analyzes the top-ranking websites for a given search term, assessing their on-page and off-page SEO strategies, and provides recommendations to enhance a website’s performance to surpass its competition. The tool yields remarkable results, boosting average daily organic traffic from 10 to 143 visitors.
]]>Informatics doi: 10.3390/informatics10030067
Authors: Fatma Taher Omar Al Fandi Mousa Al Kfairy Hussam Al Hamadi Saed Alrabaee
There are a variety of reasons why smartphones have grown so pervasive in our daily lives. While their benefits are undeniable, Android users must be vigilant against malicious apps. The goal of this study was to develop a broad framework for detecting Android malware using multiple deep learning classifiers; this framework was given the name DroidMDetection. To provide precise, dynamic, Android malware detection and clustering of different families of malware, the framework makes use of unique methodologies built based on deep learning and natural language processing (NLP) techniques. When compared to other similar works, DroidMDetection (1) uses API calls and intents in addition to the common permissions to accomplish broad malware analysis, (2) uses digests of features in which a deep auto-encoder generates to cluster the detected malware samples into malware family groups, and (3) benefits from both methods of feature extraction and selection. Numerous reference datasets were used to conduct in-depth analyses of the framework. DroidMDetection’s detection rate was high, and the created clusters were relatively consistent, no matter the evaluation parameters. DroidMDetection surpasses state-of-the-art solutions MaMaDroid, DroidMalwareDetector, MalDozer, and DroidAPIMiner across all metrics we used to measure their effectiveness.
]]>Informatics doi: 10.3390/informatics10030066
Authors: Thanapong Champahom Panuwat Wisutwattanasak Chamroeun Se Chinnakrit Banyong Sajjakaj Jomnonkwao Vatanavongs Ratanavaraha
Among several approaches to analyzing crash research, the use of machine learning and econometric analysis has found potential in the analysis. This study aims to empirically examine factors influencing the single-vehicle crash for personal cars and trucks using decision trees (DT) and mixed binary logit with heterogeneity in means and variances (RPBLHMV) and compare model accuracy. The data in this study were obtained from the Department of Highway during 2011–2017, and the results indicated that the RPBLHMV was superior due to its higher overall prediction accuracy, sensitivity, and specificity values when compared to the DT model. According to the RPBLHMV results, car models showed that injury severity was associated with driver gender, seat belt, mount the island, defect equipment, and safety equipment. For the truck model, it was found that crashes located at intersections or medians, mounts on the island, and safety equipment have a significant influence on injury severity. DT results also showed that running off-road and hitting safety equipment can reduce the risk of death for car and truck drivers. This finding can illustrate the difference causing the dependent variable in each model. The RPBLHMV showed the ability to capture random parameters and unobserved heterogeneity. But DT can be easily used to provide variable importance and show which factor has the most significance by sequencing. Each model has advantages and disadvantages. The study findings can give relevant authorities choices for measures and policy improvement based on two analysis methods in accordance with their policy design. Therefore, whether advocating road safety or improving policy measures, the use of appropriate methods can increase operational efficiency.
]]>Informatics doi: 10.3390/informatics10030065
Authors: Aditya Singhal Vijay Mago
The use of Twitter by healthcare organizations is an effective means of disseminating medical information to the public. However, the content of tweets can be influenced by various factors, such as health emergencies and medical breakthroughs. In this study, we conducted a discourse analysis to better understand how public and private healthcare organizations use Twitter and the factors that influence the content of their tweets. Data were collected from the Twitter accounts of five private pharmaceutical companies, two US and two Canadian public health agencies, and the World Health Organization from 1 January 2020, to 31 December 2022. The study applied topic modeling and association rule mining to identify text patterns that influence the content of tweets across different Twitter accounts. The findings revealed that building a reputation on Twitter goes beyond just evaluating the popularity of a tweet in the online sphere. Topic modeling, when applied synchronously with hashtag and tagging analysis can provide an increase in tweet popularity. Additionally, the study showed differences in language use and style across the Twitter accounts’ categories and discussed how the impact of popular association rules could translate to significantly more user engagement. Overall, the results of this study provide insights into natural language processing for health literacy and present a way for organizations to structure their future content to ensure maximum public engagement.
]]>Informatics doi: 10.3390/informatics10030064
Authors: Prathamesh Lahande Parag Kaveri Jatinderkumar Saini
Cloud computing delivers robust computational services by processing tasks on its virtual machines (VMs) using resource-scheduling algorithms. The cloud’s existing algorithms provide limited results due to inappropriate resource scheduling. Additionally, these algorithms cannot process tasks generating faults while being computed. The primary reason for this is that these existing algorithms need an intelligence mechanism to enhance their abilities. To provide an intelligence mechanism to improve the resource-scheduling process and provision the fault-tolerance mechanism, an algorithm named reinforcement learning-shortest job first (RL-SJF) has been implemented by integrating the RL technique with the existing SJF algorithm. An experiment was conducted in a simulation platform to compare the working of RL-SJF with SJF, and challenging tasks were computed in multiple scenarios. The experimental results convey that the RL-SJF algorithm enhances the resource-scheduling process by improving the aggregate cost by 14.88% compared to the SJF algorithm. Additionally, the RL-SJF algorithm provided a fault-tolerance mechanism by computing 55.52% of the total tasks compared to 11.11% of the SJF algorithm. Thus, the RL-SJF algorithm improves the overall cloud performance and provides the ideal quality of service (QoS).
]]>Informatics doi: 10.3390/informatics10030063
Authors: Dimitris C. Gkikas Prokopis K. Theodoridis Theodoros Theodoridis Marios C. Gkikas
This study aims to provide a method that will assist decision makers in managing large datasets, eliminating the decision risk and highlighting significant subsets of data with certain weight. Thus, binary decision tree (BDT) and genetic algorithm (GA) methods are combined using a wrapping technique. The BDT algorithm is used to classify data in a tree structure, while the GA is used to identify the best attribute combinations from a set of possible combinations, referred to as generations. The study seeks to address the problem of overfitting that may occur when classifying large datasets by reducing the number of attributes used in classification. Using the GA, the number of selected attributes is minimized, reducing the risk of overfitting. The algorithm produces many attribute sets that are classified using the BDT algorithm and are assigned a fitness number based on their accuracy. The fittest set of attributes, or chromosomes, as well as the BDTs, are then selected for further analysis. The training process uses the data of a chemical analysis of wines grown in the same region but derived from three different cultivars. The results demonstrate the effectiveness of this innovative approach in defining certain ingredients and weights of wine’s origin.
]]>Informatics doi: 10.3390/informatics10030062
Authors: Arturo Berrones-Santos Franco Bagnoli
The dichotomy in power consumption between digital and biological information processing systems is an intriguing open question related at its core with the necessity for a more thorough understanding of the thermodynamics of the logic of computing. To contribute in this regard, we put forward a model that implements the Boltzmann machine (BM) approach to computation through an electric substrate under thermal fluctuations and dissipation. The resulting network has precisely defined statistical properties, which are consistent with the data that are accessible to the BM. It is shown that by the proposed model, it is possible to design neural-inspired logic gates capable of universal Turing computation under similar thermal conditions to those found in biological neural networks and with information processing and storage electric potentials at comparable scales.
]]>Informatics doi: 10.3390/informatics10030061
Authors: Andrew Vargo Benjamin Tag Chris Blakely Koichi Kise
Online knowledge-based peer-production communities, like question and answer sites (Q&A), often rely on gamification, e.g., through reputation points, to incentivize users to contribute frequently and effectively. These gamification techniques are important for achieving the critical mass that sustains a community and enticing new users to join. However, aging communities tend to build “poverty traps” that act as barriers for new users. In this paper, we present our investigation of 32 domain communities from Stack Exchange and our analysis of how different subjects impact the development of early user advantage. Our results raise important questions about the accessibility of knowledge-based peer-production communities. We consider the analysis results in the context of changing information needs and the relevance of Q&A in the future. Our findings inform policy design for building more equitable knowledge-based peer-production communities and increasing the accessibility to existing ones.
]]>Informatics doi: 10.3390/informatics10030060
Authors: Raza Nowrozy Khandakar Ahmed Hua Wang Timothy Mcintosh
This paper proposed a novel privacy model for Electronic Health Records (EHR) systems utilizing a conceptual privacy ontology and Machine Learning (ML) methodologies. It underscores the challenges currently faced by EHR systems such as balancing privacy and accessibility, user-friendliness, and legal compliance. To address these challenges, the study developed a universal privacy model designed to efficiently manage and share patients’ personal and sensitive data across different platforms, such as MHR and NHS systems. The research employed various BERT techniques to differentiate between legitimate and illegitimate privacy policies. Among them, Distil BERT emerged as the most accurate, demonstrating the potential of our ML-based approach to effectively identify inadequate privacy policies. This paper outlines future research directions, emphasizing the need for comprehensive evaluations, testing in real-world case studies, the investigation of adaptive frameworks, ethical implications, and fostering stakeholder collaboration. This research offers a pioneering approach towards enhancing healthcare information privacy, providing an innovative foundation for future work in this field.
]]>Informatics doi: 10.3390/informatics10030059
Authors: Michail Danousis Christos Goumopoulos
With age, a decline in motor and cognitive functionality is inevitable, and it greatly affects the quality of life of the elderly and their ability to live independently. Early detection of these types of decline can enable timely interventions and support for maintaining functional independence and improving overall well-being. This paper explores the potential of the GAME2AWE platform in assessing the motor and cognitive condition of seniors based on their in-game performance data. The proposed methodology involves developing machine learning models to explore the predictive power of features that are derived from the data collected during gameplay on the GAME2AWE platform. Through a study involving fifteen elderly participants, we demonstrate that utilizing in-game data can achieve a high classification performance when predicting the motor and cognitive states. Various machine learning techniques were used but Random Forest outperformed the other models, achieving a classification accuracy ranging from 93.6% for cognitive screening to 95.6% for motor assessment. These results highlight the potential of using exergames within a technology-rich environment as an effective means of capturing the health status of seniors. This approach opens up new possibilities for objective and non-invasive health assessment, facilitating early detections and interventions to improve the well-being of seniors.
]]>Informatics doi: 10.3390/informatics10030058
Authors: Michael Joseph Dominic Roberts Randy Connolly Joel Conley Janet Miller
Over the past two decades, the internet has become an increasingly important venue for political expression, community building, and social activism. Scholars in a wide range of disciplines have endeavored to understand and measure how these transformations have affected individuals’ civic attitudes and behaviors. The Digital Citizenship Scale (original and revised form) has become one of the most widely used instruments for measuring and evaluating these changes, but to date, no study has investigated how digital citizenship behaviors relate to exogenous variables. Using the classic Big Five Factor model of personality (Openness to experience, Conscientiousness, Extroversion, Agreeableness, and Neuroticism), this study investigated how personality traits relate to the key components of digital citizenship. Survey results were gathered across three countries (n = 1820), and analysis revealed that personality traits map uniquely on to digital citizenship in comparison to traditional forms of civic engagement. The implications of these findings are discussed.
]]>Informatics doi: 10.3390/informatics10030057
Authors: Marina Aivazidi Christos Michalakelis
Innovative learning methods including the increasing use of Information and Communication Technologies (ICT) applications are transforming the contemporary educational process. Teachers’ perceptions of ICT, self-efficacy on computers and demographics are some of the factors that have been found to impact the use of ICT in the educational process. The aim of the present research is to analyze the perceptions of primary school teachers about ICT and how they affect their use in the educational process, through the case of Greece. To do so, primary research was carried out. Data from 285 valid questionnaires were statistically analyzed using descriptive statistics, principal components analysis, correlation and regression analysis. The main results were in accordance with the relevant literature, indicating the impact of teachers’ self-efficacy, perceptions and demographics on ICT use in the educational process. These results provide useful insights for the achievement of a successful implementation of ICT in education.
]]>Informatics doi: 10.3390/informatics10030056
Authors: Celso A. R. L. Brennand Rodolfo Meneguette Geraldo P. Rocha Filho
Congestion in large cities is widely recognized as a problem that impacts various aspects of society, including the economy and public health. To support the urban traffic system and to mitigate traffic congestion and the damage it causes, in this article we propose an assistant Intelligent Transport Systems (ITS) service for traffic management in Vehicular Networks (VANET), which we name FOXS-GSC, for Fast Offset Xpath Service with hexaGonS Communication. FOXS-GSC uses a VANET communication and fog computing paradigm to detect and recommend an alternative vehicle route to avoid traffic jams. Unlike the previous solutions in the literature, the proposed service offers a versatile approach in which traffic road classification and route suggestions can be made by infrastructure or by the vehicle itself without compromising the quality of the route service. To achieve this, the service operates in a decentralized way, and the components of the service (vehicles/infrastructure) exchange messages containing vehicle information and regional traffic information. For communication, the proposed approach uses a new dedicated multi-hop protocol that has been specifically designed based on the characteristics and requirements of a vehicle routing service. Therefore, by adapting to the inherent characteristics of a vehicle routing service, such as the density of regions, the proposed communication protocol both enhances reliability and improves the overall efficiency of the vehicle routing service. Simulation results comparing FOXS-GSC with baseline solutions and other proposals from the literature demonstrate its significant impact, reducing network congestion by up to 95% while maintaining a coverage of 97% across various scenery characteristics. Concerning road traffic efficiency, the traffic quality is increasing by 29%, for a reduction in carbon emissions of 10%.
]]>Informatics doi: 10.3390/informatics10030055
Authors: Jie Xu Xing He Wei Shao Jiang Bian Russell Terry
Up to 20% of renal masses ≤4 cm is found to be benign at the time of surgical excision, raising concern for overtreatment. However, the risk of malignancy is currently unable to be accurately predicted prior to surgery using imaging alone. The objective of this study is to propose a machine learning (ML) framework for pre-operative renal tumor classification using readily available clinical and CT imaging data. We tested both traditional ML methods (i.e., XGBoost, random forest (RF)) and deep learning (DL) methods (i.e., multilayer perceptron (MLP), 3D convolutional neural network (3DCNN)) to build the classification model. We discovered that the combination of clinical and radiomics features produced the best results (i.e., AUC [95% CI] of 0.719 [0.712–0.726], a precision [95% CI] of 0.976 [0.975–0.978], a recall [95% CI] of 0.683 [0.675–0.691], and a specificity [95% CI] of 0.827 [0.817–0.837]). Our analysis revealed that employing ML models with CT scans and clinical data holds promise for classifying the risk of renal malignancy. Future work should focus on externally validating the proposed model and features to better support clinical decision-making in renal cancer diagnosis.
]]>Informatics doi: 10.3390/informatics10020054
Authors: Olegs Cernisevs Yelena Popova Dmitrijs Cernisevs
Risk management is a highly important issue for Fintech companies; moreover, it is very specific and puts forward the serious requirements toward the top management of any financial institution. This study was devoted to specifying the risk factors affecting the finance and capital adequacy of financial institutions. The authors considered the different types of risks in combination, whereas other scholars usually analyze risks in isolation; however, the authors believe that it is necessary to consider their mutual impact. The risks were estimated using the PLS-SEM method in Smart PLS-4 software. The quality of the obtained model is very high according to all indicators. Five hypotheses related to finance and five hypotheses related to capital adequacy were considered. The impact of AML, cyber, and governance risks on capital adequacy was confirmed; the effect of governance and operational risks on finance was also confirmed. Other risks have no impact on finance and capital adequacy. It is interesting that risks associated with staff have no impact on finance and capital adequacy. The findings of this study can be easily applied by any financial institution for risk analysis. Moreover, this study can serve toward a better collaboration of scholars investigating the Fintech activities and practitioners working in this sphere. The authors present a novel approach for enhancing key performance indicators (KPIs) for Fintech companies, proposing utilizing metrics that are derived from the company’s specific risks, thereby introducing an innovative method for selecting KPIs based on the inherent risks associated with the Fintech’s business model. This model aligns the KPIs with the unique risk profile of the company, fostering a fresh perspective on performance measurement within the Fintech industry.
]]>Informatics doi: 10.3390/informatics10020053
Authors: Muhammad Mahreza Maulana Arif Imam Suroso Yani Nurhadryani Kudang Boro Seminar
One way to improve Indonesia’s ranking in terms of ease of conducting business is by taking a closer look at the business licensing process. This study aims to carry out an assessment using a smart governance framework and recommendation capabilities from the Enterprise System (ES). As a result, the recommendations for improvement with the expected priority are generated. The stages of this research are observing the process of making bio-business permits, followed by interviews related to several Enterprise Architecture (EA) capabilities, and providing recommendations based on the results of the maturity level of IT governance. These recommendations are then mapped into an impact—effort matrix for program prioritization. The recommendations for bio-business licenses can also be used to improve the process for other business licenses. Implementation of the EA framework has been proven to align technology, organization, and processes so that it can support continuous improvement processes.
]]>Informatics doi: 10.3390/informatics10020052
Authors: William Villegas-Ch Isabel Urbina-Camacho Joselin García-Ortiz
Using camera-based algorithms to detect abnormal patterns in children’s handwriting has become a promising tool in education and occupational therapy. This study analyzes the performance of a camera- and tablet-based handwriting verification algorithm to detect abnormal patterns in handwriting samples processed from 71 students of different grades. The study results revealed that the algorithm saw abnormal patterns in 20% of the handwriting samples processed, which included practices such as delayed typing speed, excessive pen pressure, irregular slant, and lack of word spacing. In addition, it was observed that the detection accuracy of the algorithm was 95% when comparing the camera data with the abnormal patterns detected, which indicates a high reliability in the results obtained. The highlight of the study was the feedback provided to children and teachers on the camera data and any abnormal patterns detected. This can significantly impact students’ awareness and improvement of writing skills by providing real-time feedback on their writing and allowing them to adjust to correct detected abnormal patterns.
]]>Informatics doi: 10.3390/informatics10020051
Authors: Tjokorda Agung Budi Wirayuda Rinaldi Munir Achmad Imam Kistijantoro
In computer vision, ethnicity classification tasks utilize images containing human faces to extract ethnicity labels. Ethnicity is one of the soft biometric feature categories useful in data analysis for commercial, public, and health sectors. Ethnicity classification begins with face detection as a preprocessing process to determine a human’s presence; then, the feature representation is extracted from the isolated facial image to predict the ethnicity class. This study utilized four handcrafted features (multi-local binary pattern (MLBP), histogram of gradient (HOG), color histogram, and speeded-up-robust-features-based (SURF-based)) as the basis for the generation of a compact-fusion feature. The compact-fusion framework involves optimal feature selection, compact feature extraction, and compact-fusion feature representation. The final feature representation was trained and tested with the SVM One Versus All classifier for ethnicity classification. When it was evaluated in two large datasets, UTKFace and Fair Face, the proposed framework achieved accuracy levels of 89.14%, 82.19%, and 73.87%, respectively, for the UTKFace dataset with four or five classes and the Fair Face dataset with four classes. Furthermore, the compact-fusion feature with a small number of features at 4790, constructed based on conventional handcrafted features, achieved competitive results compared with state-of-the-art methods using a deep-learning-based approach.
]]>Informatics doi: 10.3390/informatics10020050
Authors: Krishna Raj Raghavendran Ahmed Elragal
In the context of developing machine learning models, until and unless we have the required data engineering and machine learning development competencies as well as the time to train and test different machine learning models and tune their hyperparameters, it is worth trying out the automatic machine learning features provided by several cloud-based and cloud-agnostic platforms. This paper explores the possibility of generating automatic machine learning models with low-code experience. We developed criteria to compare different machine learning platforms for generating automatic machine learning models and presenting their results. Thereafter, lessons learned by developing automatic machine learning models from a sample dataset across four different machine learning platforms were elucidated. We also interviewed machine learning experts to conceptualize their domain-specific problems that automatic machine learning platforms can address. Results showed that automatic machine learning platforms can provide a fast track for organizations seeking the digitalization of their businesses. Automatic machine learning platforms help produce results, especially for time-constrained projects where resources are lacking. The contribution of this paper is in the form of a lab experiment in which we demonstrate how low-code platforms can provide a viable option to many business cases and, henceforth, provide a lane that is faster than the usual hiring and training of already scarce data scientists and to analytics projects that suffer from overruns.
]]>Informatics doi: 10.3390/informatics10020049
Authors: Antony Bryant
In November 2022, OpenAI launched ChatGPT, an AI chatbot that gained over 100 million users by February 2023 [...]
]]>Informatics doi: 10.3390/informatics10020048
Authors: Charlee Kaewrat Poonpong Boonbrahm Bukhoree Sahoh
Unsuitable shoe shapes and sizes are a critical reason for unhealthy feet, may severely contribute to chronic injuries such as foot ulcers in susceptible people (e.g., diabetes patients), and thus need accurate measurements in the manner of expert-based procedures. However, the manual measure of such accurate shapes and sizes is labor-intensive, time-consuming, and impractical to apply in a real-time system. This research proposes a foot-detection approach using expert-like measurements to address this concern. It combines the seven-foot dimensions model and the light detection and ranging sensor to encode foot shapes and sizes and detect the dimension surfaces. The graph-based algorithms are developed to present seven-foot dimensions and visualize the shoe’s model based on the augmented reality (AR) technique. The results show that our approach can detect shapes and sizes more effectively than the traditional approach, helps the system imitate expert-like measurements accurately, and can be employed in intelligent applications for susceptible people-based feet measurements.
]]>Informatics doi: 10.3390/informatics10020047
Authors: Mónica Cruz Abílio Oliveira Alessandro Pinheiro
We were promised that the Metaverse would revolutionize our lives, social interactions, work, and business. However, how and when will this happen? We have seen the growth and development of technology, but there is no agreement or prediction about a specific time, and we can only follow the how question. To investigate more leads about this concept, we considered a main research question: How is the Metaverse actually being perceived? This question is connected with three objectives: to verify how the Metaverse is being represented and characterized, identify the main dimensions that facilitate or influence the acceptance of the Metaverse, and identify the leading technologies that suit the Metaverse concept. This study consisted of a documental analysis—or meta-analysis—of fifty of the most relevant scientific papers (taking into account some inclusion criteria) published in the last three years, using the Leximancer software to create concept maps to illustrate the main concepts and themes extracted from the articles to understand their associations or relations with the Metaverse concept. This study provided us with essential findings about how this concept has been perceived and allowed us to answer our objectives, contributing to a scientific discussion on the topic, and provided some valid suggestions for future research, which is already in progress. It also provided new leads on approaching this concept in development.
]]>Informatics doi: 10.3390/informatics10020046
Authors: Zhaoyi Chen Yuchen Yang Dazheng Zhang Jingchuan Guo Yi Guo Xia Hu Yong Chen Jiang Bian
Alzheimer’s disease (AD) and AD-related dementias (AD/ADRD) are a group of progressive neurodegenerative diseases. The progression of AD can be conceptualized as a continuum in which patients progress from normal cognition to preclinical AD (i.e., no symptoms but biological changes in the brain) to mild cognitive impairment (MCI) due to AD (i.e., mild symptoms but not interfere with daily activities), followed by increasing severity of dementia due to AD. Early detection and prediction models for the transition of MCI to AD/ADRD are needed, and efforts have been made to build predictions of MCI conversion to AD/ADRD. However, most existing studies developing such prediction models did not consider the competing risks of death, which may result in biased risk estimates. In this study, we aim to develop a prediction model for AD/ADRD among patients with MCI considering the competing risks of death using a semi-competing risk approach.
]]>Informatics doi: 10.3390/informatics10020045
Authors: Belen Bermejo Carlos Juiz David Cortes Jeroen Oskam Teemu Moilanen Jouko Loijas Praneschen Govender Jennifer Hussey Alexander Lennart Schmidt Ralf Burbach Daniel King Colin O'Connor Davin Dunlea
During the last few years, learning techniques have changed, both in basic education and in higher education. This change has been accompanied by new technologies such as Augmented Reality (AR) and Virtual Reality (AR). The combination of these technologies in education has allowed a greater immersion, positively affecting the learning and teaching processes. In addition, since the COVID-19 pandemic, this trend has been growing due to the diversity of the different fields of application of these technologies, such as heterogeneity in their combination and their different experiences. It is necessary to review the state of the art to determine the effectiveness of the application of these technologies in the field of university higher education. In the present paper, this aim is achieved by performing a systematic literature review from 2012 to 2022. A total of 129 papers were analyzed. Studies in our review concluded that the application of AR/VR improves learning immersion, especially in hospitality, medicine, and science studies. However, there are also negative effects of using these technologies, such as visual exhaustion and mental fatigue.
]]>Informatics doi: 10.3390/informatics10020044
Authors: Anna Christina Situmorang Muhammad Suryanegara Dadang Gunawan Filbert H. Juwono
In Indonesia, there is still a disparity in telecommunications access, with most rural areas experiencing “no signal” or “blank spots.” In contrast, urban areas enjoy modern and societally-beneficial technologies. A comprehensive framework is needed to address the disparity in telecommunications access between “rich” and “poor” groups in urban and rural/remote areas, respectively. This paper proposes a framework, built by the mathematical model, that can be used as a reference for the Indonesian government in constructing the nation’s telecommunications infrastructure. The framework categorizes Indonesian administrative regions into four grids: Grid #1: “fostered” districts; Grid #2: “developing” districts; Grid #3: “developed” districts; and Grid #4: “independent-advanced” districts. To determine where each district falls in these grids, we propose a novel statistical approach using 17 indicators involving a telecommunications network and socioeconomic factors. The proposed framework results in a grid visualization of 7232 districts in Indonesia. Finally, as this paper is replete with academic research approaches and mathematical model perspectives, it is expected that the results may be a valuable input to the development of the country’s telecommunications policy.
]]>Informatics doi: 10.3390/informatics10020043
Authors: Reuben Tamakloe
Studies have explored the factors influencing the safety of PTWs; however, very little has been carried out to comprehensively investigate the factors influencing fatal PTW crashes while considering the fault status of the rider in crash hotspot areas. This study employs spatio-temporal hotspot analysis and association rule mining techniques to discover hidden associations between crash risk factors that lead to fatal PTW crashes considering the fault status of the rider at statistically significant PTW crash hotspots in South Korea from 2012 to 2017. The results indicate the presence of consecutively fatal PTW crash hotspots concentrated within Korea’s densely populated capital, Seoul, and new hotspots near its periphery. According to the results, violations such as over-speeding and red-light running were critical contributory factors influencing PTW crashes at hotspots during summer and at intersections. Interestingly, while reckless riding was the main traffic violation leading to PTW rider at-fault crashes at hotspots, violations such as improper safety distance and red-light running were strongly associated with PTW rider not-at-fault crashes at hotspots. In addition, while PTW rider at-fault crashes are likely to occur during summer, PTW rider not-at-fault crashes mostly occur during spring. The findings could be used for developing targeted policies for improving PTW safety at hotspots.
]]>Informatics doi: 10.3390/informatics10020042
Authors: Angelica Lo Duca Andrea Marchetti Manuela Moretti Francesca Diana Mafalda Toniazzi Andrea D’Errico
The Jewish community archive in Pisa owns a vast collection of documents and manuscripts that date back centuries. These documents contain valuable genealogical information, including birth, marriage, and death records. This paper aims to describe the preliminary results of the Archivio Storico della Comunita Ebraica di Pisa (ASCEPI) project, with a focus on the extraction of data from the Nati, Morti e Ballottati (NMB) Registry document in the archive. The NMB Registry contains about 1900 records of births, deaths, and balloted individuals within the Jewish community in Pisa. The study uses a semiautomatic pipeline of digitization, transcription, and Natural Language Processing (NLP) techniques to extract personal data such as names, surnames, birth and death dates, and parental names from each record. The extracted data are then used to build a knowledge base and a genealogical tree for a representative family, Supino. This study demonstrates the potential of using NLP and rule-based techniques to extract valuable information from historical documents and to construct genealogical trees.
]]>Informatics doi: 10.3390/informatics10020041
Authors: Mohammad Daradkeh
Digital platform business model innovation is a rapidly evolving field, yet the literature on resource, complementary, and ecological boundaries remains limited, leaving a significant gap in our understanding of the factors that shape the success of these platforms. This paper explores the mechanisms by which digital platforms enable business model innovation, a topic of significant theoretical and practical importance that has yet to be fully examined. Through a review of the existing literature and an examination of the connotations of digital platforms, the design of platform boundaries, and the deployment of boundary resources, the study finds that (1) the uncertainty of complementors and complementary products drives business model innovation in digital platforms; (2) the design of resource, complementary, and ecological system boundaries is crucial to digital platform business models and manages complementor and complementary product uncertainty while promoting value co-creation; and (3) boundary resources establish, manage, and sustain cross-border relationships that impact value creation and capture. Based on these findings, four research propositions are proposed to guide future research on digital platform business model innovation and provide insights for effectively innovating business models and influencing value creation and capture.
]]>Informatics doi: 10.3390/informatics10020040
Authors: Andrei Konstantinov Lev Utkin Vladimir Muliukha
This paper provides new models of the attention-based random forests called LARF (leaf attention-based random forest). The first idea behind the models is to introduce a two-level attention, where one of the levels is the “leaf” attention, and the attention mechanism is applied to every leaf of trees. The second level is the tree attention depending on the “leaf” attention. The second idea is to replace the softmax operation in the attention with the weighted sum of the softmax operations with different parameters. It is implemented by applying a mixture of Huber’s contamination models and can be regarded as an analog of the multi-head attention, with “heads” defined by selecting a value of the softmax parameter. Attention parameters are simply trained by solving the quadratic optimization problem. To simplify the tuning process of the models, it is proposed to convert the tuning contamination parameters into trainable parameters and to compute them by solving the quadratic optimization problem. Many numerical experiments with real datasets are performed for studying LARFs. The code of the proposed algorithms is available.
]]>Informatics doi: 10.3390/informatics10020039
Authors: Luke Balcombe Diego De Leo
There are positives and negatives of using YouTube in terms of loneliness and mental health. YouTube’s streaming content is an amazing resource, however, there may be bias or errors in its recommendation algorithms. Parasocial relationships can also complicate the impact of YouTube use. Intervention may be necessary when problematic and risky content is associated with unhealthy behaviors and negative impacts on mental health. Children and adolescents are particularly vulnerable. Although YouTube might assist in connecting with peers, there are privacy, safety, and quality issues to consider. This paper is an integrative review of the positive and negative impacts of YouTube with the aim to inform the design and development of a technology-based intervention to improve mental health. The impact of YouTube use on loneliness and mental health was explored by synthesizing a purposive selection (n = 32) of the empirical and theoretical literature. Next, we explored human–computer interaction issues and proposed a concept whereby an independent-of-YouTube algorithmic recommendation system steers users toward verified positive mental health content or promotions.
]]>Informatics doi: 10.3390/informatics10020038
Authors: Alejandro Valencia-Arias Diana María Arango-Botero Sebastián Cardona-Acevedo Sharon Soledad Paredes Delgado Ada Gallegos
The COVID-19 pandemic and the boom of fake news cluttering the internet have revealed the power of social media today. However, young people are not yet aware of their role in the digital age, even though they are the main users of social media. As a result, the belief that older adults are responsible for information is being re-evaluated. In light of this, the present study was aimed at identifying the factors associated with the spread of fake news among young people in Medellín (Colombia). A total of 404 self-administered questionnaires were processed in a sample of people between the ages of 18 and 34 and analyzed using statistical techniques, such as exploratory factor analysis and structural equation modeling. The results suggest that the instantaneous sharing of fake news is linked to people’s desire to raise awareness among their inner circle, particularly when the messages shared are consistent with their perceptions and beliefs, or to the lack of time to properly verify their accuracy. Finally, passive corrective actions were found to have a less significant impact in the Colombian context than in the context of the original model, which may be explained by cultural factors.
]]>Informatics doi: 10.3390/informatics10020037
Authors: Malinka Ivanova Tsvetelina Petrova
During the pandemic, universities were forced to convert their educational process online. Students had to adapt to new educational conditions and the proposed online environment. Now, we are back to the traditional blended learning environment and wish to understand the students’ attitudes and perceptions of online learning, knowing that they are able to compare blended and online modes. The aim of this paper is to present the performed predictive analysis regarding the students’ online learning performance taking into account their opinion. The predictive models are created through a supervised machine learning algorithm based on Artificial Neural Networks and are characterized with high accuracy. The analysis is based on generated synthetic datasets, ensuring a high level of students’ privacy preservation.
]]>Informatics doi: 10.3390/informatics10020036
Authors: Diogo Santos Elsa Cardoso Isabel Machado Alexandre
Online commerce has been growing rapidly in an increasingly digital world, and gamification, the practice of designing games in a context outside the industry itself, can be an effective strategy to stimulate consumer engagement and conversion rate. This paper describes the design process involved in introducing gamification into an online shop that is supported by two game servers of the same kind, namely one in the United States of America (US) and another in Portugal (PT). Through the various phases of the design thinking process, a gamified system was implemented to meet the needs of various types of users frequently found in the shops. The gamification elements used were intended to increase user engagement with the shops so that they would become more aware of existing products and the introduction of new products, promoting purchase through intangible challenges and rewards. The impacts on server revenues and user satisfaction (N = 138) were evaluated one month after introducing the gamification techniques. The results show that gamification has a positive effect on users, with a significant increase in consumer interaction in both shops. However, from a business point of view, the results show only an increase in revenue for the US shop, while the Portuguese shop shows no significant differences compared to previous months. Of the two user groups analyzed, only those who frequent the US shop show receptivity toward intangible rewards, with tangible rewards (discounts) being a greater motivating factor for both groups.
]]>Informatics doi: 10.3390/informatics10020035
Authors: Shefa M. Tawalbeh Ahmed Al-Omari Lina M. K. Al-Ebbini Hiam Alquran
Jordanian healthcare institutes have launched several programs since 2009 to establish health information systems (HISs). Nowadays, the generic expectation is that the use of HIS resources is performed on daily basis among healthcare staff. However, there can be still a noticeable barrier due to a lack of knowledge if medical doctors do not receive proper training on existing HISs. Moreover, the lack of studies on this area hinders the clarity about the received versus the required training skills among medical doctors. To support this research initiative, survey data have been collected from specialized medical doctors who are currently affiliated with five Jordanian universities to assess their need for HIS training. The results also aim to explore the extent of medical doctors’ use of HIS resources in Jordan. Moreover, they examine whether medical doctors require additional training on using HIS resources or not, as well as the main areas of required training programs. Specifically, this paper highlights the main topics that can be suitable subjects for enhanced training programs. The results show that most respondents use HISs in their daily clinical practices. However, most of them have not taken professional training on such systems. Hence, most of the respondents reported the need for additional training programs on several aspects of HIS resources. Moreover, based on the survey results, the most significant areas that require training are biomedical data analysis, artificial intelligence in medicine, health care management, and recent advances in electronic health records, respectively. Therefore, specialized medical doctors in Jordan need training on extracting useful and potential features of HISs. Education and training professionals in healthcare are recommended to establish training programs in Jordanian healthcare centers, which can further improve the quality of healthcare.
]]>Informatics doi: 10.3390/informatics10020034
Authors: Ridwan Ilyas Masayu Leylia Khodra Rinaldi Munir Rila Mandala Dwi Hendratmo Widyantoro
The paraphrase generator for citation sentences is used to produce several sentence alternatives to avoid plagiarism. Furthermore, the generation results need to pay attention to semantic similarity and lexical divergence standards. This study proposed the StoPGEN model as an algorithm for generating citation paraphrase sentences with stochastic output. The generation process is guided by an objective function using a simulated annealing algorithm to maintain the properties of semantic similarity and lexical divergence. The objective function is created by combining the two factors that maintain these properties. This study combined METEOR and PINC Scores in a linear weighting function that can be adjusted for its value tendency in one of the matrix functions. The dataset of citation sentences that had been labeled with paraphrases was used to test StoPGEN and other models for comparison. The StoPGEN model, with the citation sentences dataset, produced a BLEU score of 55.37, outperforming the bidirectional LSTM method with a value of 28.93. StoPGEN was also tested using Quora data by changing the language source in the architecture section resulting in a BLEU score of 22.37, outperforming UPSA 18.21. In addition, the qualitative evaluation results of the citation sentence generation based on respondents obtained an acceptance value of 50.80.
]]>Informatics doi: 10.3390/informatics10020033
Authors: Amanda L. Luo Akshay Ravi Simone Arvisais-Anhalt Anoop N. Muniyappa Xinran Liu Shan Wang
(1) One in four hospital readmissions is potentially preventable. Machine learning (ML) models have been developed to predict hospital readmissions and risk-stratify patients, but thus far they have been limited in clinical applicability, timeliness, and generalizability. (2) Methods: Using deidentified clinical data from the University of California, San Francisco (UCSF) between January 2016 and November 2021, we developed and compared four supervised ML models (logistic regression, random forest, gradient boosting, and XGBoost) to predict 30-day readmissions for adults admitted to a UCSF hospital. (3) Results: Of 147,358 inpatient encounters, 20,747 (13.9%) patients were readmitted within 30 days of discharge. The final model selected was XGBoost, which had an area under the receiver operating characteristic curve of 0.783 and an area under the precision-recall curve of 0.434. The most important features by Shapley Additive Explanations were days since last admission, discharge department, and inpatient length of stay. (4) Conclusions: We developed and internally validated a supervised ML model to predict 30-day readmissions in a US-based healthcare system. This model has several advantages including state-of-the-art performance metrics, the use of clinical data, the use of features available within 24 h of discharge, and generalizability to multiple disease states.
]]>Informatics doi: 10.3390/informatics10010032
Authors: Ezekiel Bernardo Rosemary Seva
Explainable Artificial Intelligence (XAI) has successfully solved the black box paradox of Artificial Intelligence (AI). By providing human-level insights on AI, it allowed users to understand its inner workings even with limited knowledge of the machine learning algorithms it uses. As a result, the field grew, and development flourished. However, concerns have been expressed that the techniques are limited in terms of to whom they are applicable and how their effect can be leveraged. Currently, most XAI techniques have been designed by developers. Though needed and valuable, XAI is more critical for an end-user, considering transparency cleaves on trust and adoption. This study aims to understand and conceptualize an end-user-centric XAI to fill in the lack of end-user understanding. Considering recent findings of related studies, this study focuses on design conceptualization and affective analysis. Data from 202 participants were collected from an online survey to identify the vital XAI design components and testbed experimentation to explore the affective and trust change per design configuration. The results show that affective is a viable trust calibration route for XAI. In terms of design, explanation form, communication style, and presence of supplementary information are the components users look for in an effective XAI. Lastly, anxiety about AI, incidental emotion, perceived AI reliability, and experience using the system are significant moderators of the trust calibration process for an end-user.
]]>Informatics doi: 10.3390/informatics10010031
Authors: Dipima Buragohain Grisana Punpeng Sureenate Jaratjarungkiat Sushank Chaudhary
Recent technology implementation in learning has inspired language educators to employ various e-learning techniques, strategies, and applications in their pedagogical practices while aiming at improving specific learning efficiencies of students. The current study attempts to blend e-learning activities, including blogging, video making, online exercises, and digital storyboarding, with English language teaching and explores its impact on engineering cohorts at a public university in Malaysia. The longitudinal research study used three digital applications—Voyant Tools, Lumos Text Complexity Analyzer, and Advanced Text Analyzer—to analyze the data collected through a variety of digital assignments and activities from two English language courses during the researched academic semesters. Contributing to the available literature on the significance of integrating technology innovation with language learning, the study found that implementing e-learning activities can provide substantial insights into improving the learners’ different linguistic competencies, including writing competency, reading comprehension, and vocabulary enhancement. Moreover, the implementation of such innovative technology can motivate students to engage in more peer interactivity, learning engagement, and self-directed learning.
]]>Informatics doi: 10.3390/informatics10010030
Authors: Ignacio Redondo Gloria Aznar
Website publishers cannot monetize the ad impressions that are prevented by ad-blockers. Publishers can then employ anti-ad-blockers that force users to choose between either accepting ad impressions by whitelisting the website in the ad-blocker, or leaving the website without accessing the content. This study delineates the mechanisms of how willingness to whitelist/leave the website are affected by the request’s sensitivity to recipients as well as the users’ psychological reactance and evaluation of the website advertising. We tested the proposed relationships using an online panel sample of 500 ad-blocker users, who were asked about their willingness to whitelist/leave their favorite online newspaper after receiving a hypothetical anti-ad-blocker request—four alternative requests with different sensitivity levels were created and randomly assigned to the participants. The results confirmed that (a) the request’s sensitivity can improve the recipient’s compliance, (b) users’ psychological reactance plays an important role in explaining the overall phenomenon, and (c) a favorable evaluation of the website advertising can improve willingness to whitelist. These findings help to better understand user response to anti-ad-blockers and may also help publishers increase their whitelist ratios.
]]>Informatics doi: 10.3390/informatics10010029
Authors: Godwin Kaisara Kelvin Joseph Bwalya
Mobile learning is a global trend, which has become more widespread in the post-COVID-19 pandemic era. However, with the adoption of mobile learning comes new assessment approaches to evaluate the understanding of the acquired information and knowledge. Nevertheless, there is scant knowledge of how to enhance assessment information integrity in mobile learning assessments. Due to the importance of assessments in evaluating knowledge, integrity is the sine qua non of online assessments. This research focuses on the strategies universities could use to improve assessment information integrity. This research adopts a qualitative design, employing interviews with academics as well as teaching and learning support staff for data collection. The findings reveal five strategies that academics and support staff recommend to enhance assessment information integrity in mobile learning. The theoretical and practical implications are discussed, as well as future research directions.
]]>Informatics doi: 10.3390/informatics10010028
Authors: Mohammad Alauthman Ahmad Al-qerem Bilal Sowan Ayoub Alsarhan Mohammed Eshtay Amjad Aldweesh Nauman Aslam
Developing an effective classification model in the medical field is challenging due to limited datasets. To address this issue, this study proposes using a generative adversarial network (GAN) as a data-augmentation technique. The research aims to enhance the classifier’s generalization performance, stability, and precision through the generation of synthetic data that closely resemble real data. We employed feature selection and applied five classification algorithms to thirteen benchmark medical datasets, augmented using the least-square GAN (LS-GAN). Evaluation of the generated samples using different ratios of augmented data showed that the support vector machine model outperforms other methods with larger samples. The proposed data augmentation approach using a GAN presents a promising solution for enhancing the performance of classification models in the healthcare field.
]]>Informatics doi: 10.3390/informatics10010027
Authors: Sara G. Fahmy Khaled M. Abdelgaber Omar H. Karam Doaa S. Elzanfaly
The mechanisms of information diffusion in Online Social Networks (OSNs) have been studied extensively from various perspectives with some focus on identifying and modeling the role of heterogeneous nodes. However, none of these studies have considered the influence of fake accounts on human accounts and how this will affect the rumor diffusion process. This paper aims to present a new information diffusion model that characterizes the role of bots in the rumor diffusion process in OSNs. The proposed SIhIbR model extends the classical SIR model by introducing two types of infected users with different infection rates: the users who are infected by human (Ih) accounts with a normal infection rate and the users who are infected by bot accounts (Ib) with a different diffusion rate that reflects the intent and steadiness of this type of account to spread the rumors. The influence of fake accounts on human accounts diffusion rate has been measured using the social impact theory, as it better reflects the deliberate behavior of bot accounts to spread a rumor to a large portion of the network by considering both the strength and the bias of the source node. The experiment results show that the accuracy of the SIhIbR model outperforms the SIR model when simulating the rumor diffusion process in the existence of fake accounts. It has been concluded that fake accounts accelerate the rumor diffusion process as they impact many people in a short time.
]]>Informatics doi: 10.3390/informatics10010026
Authors: Witthaya Hosap Chaowanan Khundam Patibut Preeyawongsakul Varunyu Vorachart Frédéric Noël
This study aimed to examine the use of light and color in digital paintings and their effect on audiences’ perceptions of environmental issues. Five digital paintings depicting environmental issues have been designed. Digital painting techniques created black-and-white, monochrome, and color images. Each image used utopian and dystopian visualization concepts to communicate hope and despair. In the experiment, 225 volunteers representing students in colleges were separated into three independent groups: the first group was offered black-and-white images, the second group was offered monochromatic images, and the third group was offered color images. After viewing each image, participants were asked to complete questionnaires about their emotions and cognitions regarding environmental issues, including identifying hope and despair and the artist’s perspective at the end. The analysis showed no differences in emotions and cognitions among participants. However, monochromatic images were the most emotionally expressive. The results indicated that the surrounding atmosphere of the images created despair, whereas objects inspired hope. Artists should emphasize the composition of the atmosphere and the objects in the image to convey the concepts of utopia and dystopia to raise awareness of environmental issues.
]]>Informatics doi: 10.3390/informatics10010025
Authors: Pedro Nogueira Leandro Pereira Ana Simões Alvaro Dias Renato Lopes da Costa
Companies try to acquire the finest advantages and techniques in a technologically advanced and end-to-end market to have a stronger foothold there. Although empirical research on this topic links IT to a decline in vertical integration, corporations are increasingly using this corporate strategy. The goal of this study is to show how over the past 22 years, scientific literature has changed with regard to how information technology (IT) affects vertical integration, one of the main types of corporate strategies. The findings demonstrated that vertical integration has been evolving in a balanced manner in a technological environment. Three categories—information technology, innovation, and processes—help explain this association and were discovered through cluster analysis. The direction of operational integration, the degree of industry concentration, demand unpredictability, and innovation should all be considered while making integration decisions.
]]>Informatics doi: 10.3390/informatics10010024
Authors: Giuseppe Ciaburro Sankar Padmanabhan Yassine Maleh Virginia Puyana-Romero
The modern conception of industrial production recognizes the increasingly crucial role of maintenance. Currently, maintenance is thought of as a service that aims to maintain the efficiency of equipment and systems while also taking quality, energy efficiency, and safety requirements into consideration. In this study, a new methodology for automating the fan maintenance procedures was developed. An approach based on the recording of the acoustic emission and the failure diagnosis using deep learning was evaluated for the detection of dust deposits on the blades of an axial fan. Two operating conditions have been foreseen: No-Fault, and Fault. In the No-Fault condition, the fan blades are perfectly clean while in the Fault condition, deposits of material have been artificially created. Utilizing a pre-trained network (SqueezeNet) built on the ImageNet dataset, the acquired data were used to build an algorithm based on convolutional neural networks (CNN). The transfer learning applied to the images of the spectrograms extracted from the recordings of the acoustic emission of the fan, in the two operating conditions, returned excellent results (accuracy = 0.95), confirming the excellent performance of the methodology.
]]>Informatics doi: 10.3390/informatics10010023
Authors: Sofía Ramos-Pulido Neil Hernández-Gress Gabriela Torres-Delgado
This study shows the significant features predicting graduates’ job levels, particularly high-level positions. Moreover, it shows that data science methodologies can accurately predict graduate outcomes. The dataset used to analyze graduate outcomes was derived from a private educational institution survey. The original dataset contains information on 17,898 graduates and approximately 148 features. Three machine learning algorithms, namely, decision trees, random forest, and gradient boosting, were used for data analysis. These three machine learning models were compared with ordinal regression. The results indicate that gradient boosting is the best predictive model, which is 6% higher than the ordinal regression accuracy. The SHapley Additive exPlanations (SHAP), a novel methodology to extract the significant features of different machine learning algorithms, was then used to extract the most important features of the gradient boosting model. Current salary is the most important feature in predicting job levels. Interestingly, graduates who realized the importance of communication skills and teamwork to be good leaders also had higher job positions. Finally, general relevant features to predict job levels include the number of people directly in charge, company size, seniority, and satisfaction with income.
]]>Informatics doi: 10.3390/informatics10010022
Authors: Prianto Budi Saptono Sabina Hodžić Ismail Khozen Gustofan Mahmud Intan Pratiwi Dwi Purwanto Muhamad Akbar Aditama Nisa’ul Haq Siti Khodijah
The effectiveness of the e-tax system in encouraging tax compliance has been largely unexplored. Thus, the current study aims to examine the interrelationship between technological predictors in explaining tax compliance intention among certified tax professionals. Based on the literature on information system success and tax compliance intention, this paper proposed an expanded conceptual framework that incorporates convenience and perception of reduced compliance costs as predictors and satisfaction as a mediator. The data were collected from 650 tax professionals who used e-Filing and 492 who used e-Form through an online survey and analyzed using hierarchical multiple regression. The empirical results suggest that participants’ perceived service quality of e-Filing services and perceptions of reduced compliance costs positively influence users’ willingness to comply with tax regulations. The latter predictor is also, and only, significant among e-Form users. The empirical results also provide statistical evidence for the mediating role of satisfaction in the relationship between all predictors and tax compliance intention. This study encourages tax policymakers and e-tax filing providers to improve their services to increase user satisfaction and tax compliance.
]]>Informatics doi: 10.3390/informatics10010021
Authors: Abhilash Pati Manoranjan Parhi Mohammad Alnabhan Binod Kumar Pattanayak Ahmad Khader Habboush Mohammad K. Al Nawayseh
Recently, it has proven difficult to make an immediate remote diagnosis of any coronary illness, including heart disease, diabetes, etc. The drawbacks of cloud computing infrastructures, such as excessive latency, bandwidth, energy consumption, security, and privacy concerns, have lately been addressed by Fog computing with IoT applications. In this study, an IoT-Fog-Cloud integrated system, called a Fog-empowered framework for real-time analysis in heart patients using ENsemble Deep learning (FRIEND), has been introduced that can instantaneously facilitate remote diagnosis of heart patients. The proposed system was trained on the combined dataset of Long-Beach, Cleveland, Switzerland, and Hungarian heart disease datasets. We first tested the model with eight basic ML approaches, including the decision tree, logistic regression, random forest, naive Bayes, k-nearest neighbors, support vector machine, AdaBoost, and XGBoost approaches, and then applied ensemble methods including bagging classifiers, weighted averaging, and soft and hard voting to achieve enhanced outcomes and a deep neural network, a deep learning approach, with the ensemble methods. These models were validated using 16 performance and 9 network parameters to justify this work. The accuracy, PPV, TPR, TNR, and F1 scores of the experiments reached 94.27%, 97.59%, 96.09%, 75.44%, and 96.83%, respectively, which were comparatively higher when the deep neural network was assembled with bagging and hard-voting classifiers. The user-friendliness and the inclusion of Fog computing principles, instantaneous remote cardiac patient diagnosis, low latency, and low energy consumption, etc., are advantages confirmed according to the achieved experimental results.
]]>Informatics doi: 10.3390/informatics10010020
Authors: Attila Varga László Révész
The authors of the present study explored how ICT devices used in P.E. lessons determine psychomotor performance, perceived motivational climate, and motivation. The students were allowed to use ICT devices (smartphone, webpages, Facebook) during a four-week intervention. In the course of the research project aimed to assess the impact of the application of ICT devices on performance and motivation, the participants were divided into two test groups and one control group. The sample consisted of secondary school students including 21 males and 64 females with the Mage = 16.72 years. The results showed that in groups where ICT devices were used, performance (p = 0.04) and task orientation (p = 0.00) significantly improved. Meanwhile, in the group in which ICT devices were not used, the intervention resulted in improved performance (p = 0.00) and by the end of the project, this trend was coupled with increased Ego orientation (p = 0.00) and higher rate of amotivation (p = 0.04). It can be concluded that the use of ICT tools has a positive impact on performance and motivation.
]]>Informatics doi: 10.3390/informatics10010019
Authors: Mohammad Daradkeh
The importance of business analytics (BA) in driving knowledge generation and business innovation has been widely discussed in both the academic and business communities. However, empirical research on the relationship between knowledge orientation and business analytics capabilities in driving business model innovation remains scarce. Drawing on the knowledge-based view and dynamic capabilities theory, this study develops a model to investigate the interplay between knowledge orientation and BA capabilities in driving business model innovation. It also explores the moderating role of industry type on this relationship. To test the model, data were collected from a cross-sectional sample of 207 firms (high-tech and non-high-tech industries). Descriptive and structural equation modeling (SEM) were used to test the hypotheses. The findings showed that knowledge orientation and BA capabilities are significantly and positively related to business model innovation. Knowledge commitment, shared vision, and open-mindedness are significantly and positively related to BA perception and recognition capabilities and BA integration capabilities. BA capabilities mediated the relationship between knowledge orientation and business model innovation. The path mechanism of knowledge orientation → BA capabilities → business model innovation shows that industry type has a moderating effect on knowledge orientation and BA capabilities, as well as BA capabilities and business model innovation. This study provides empirically proven insights and practical guidance on the dynamics and mechanisms of BA and organizational knowledge capabilities and their impact on business model innovation.
]]>Informatics doi: 10.3390/informatics10010018
Authors: Rocco A. Scollo Antonio G. Spampinato Georgia Fargetta Vincenzo Cutello Mario Pavone
Disease phenotypes are generally caused by the failure of gene modules which often have similar biological roles. Through the study of biological networks, it is possible to identify the intrinsic structure of molecular interactions in order to identify the so-called “disease modules”. Community detection is an interesting and valuable approach to discovering the structure of the community in a complex network, revealing the internal organization of the nodes, and has become a leading research topic in the analysis of complex networks. This work investigates the link between biological modules and network communities in test-case biological networks that are commonly used as a reference point and which include Protein–Protein Interaction Networks, Metabolic Networks and Transcriptional Regulation Networks. In order to identify small and structurally well-defined communities in the biological context, a hybrid immune metaheuristic algorithm Hybrid-IA is proposed and compared with several metaheuristics, hyper-heuristics, and the well-known greedy algorithm Louvain, with respect to modularity maximization. Considering the limitation of modularity optimization, which can fail to identify smaller communities, the reliability of Hybrid-IA was also analyzed with respect to three well-known sensitivity analysis measures (NMI, ARI and NVI) that assess how similar the detected communities are to real ones. By inspecting all outcomes and the performed comparisons, we will see that on one hand Hybrid-IA finds slightly lower modularity values than Louvain, but outperforms all other metaheuristics, while on the other hand, it can detect communities more similar to the real ones when compared to those detected by Louvain.
]]>Informatics doi: 10.3390/informatics10010017
Authors: David Dias José Silvestre Silva Alexandre Bernardino
This work proposes a tool to predict the risk of road accidents. The developed system consists of three steps: data selection and collection, preprocessing, and the use of mining algorithms. The data were imported from the Portuguese National Guard database, and they related to accidents that occurred from 2019 to 2021. The results allowed us to conclude that the highest concentration of accidents occurs during the time interval from 17:00 to 20:00, and that rain is the meteorological factor with the greatest effect on the probability of an accident occurring. Additionally, we concluded that Friday is the day of the week on which more accidents occur than on other days. These results are of importance to the decision makers responsible for planning the most effective allocation of resources for traffic surveillance.
]]>Informatics doi: 10.3390/informatics10010016
Authors: Lisa Ariellah Ward Gulzar H. Shah Jeffery A. Jones Linda Kimsey Hani Samawi
This paper examines the efficacy of telemedicine (TM) technology compared to traditional face-to-face (F2F) visits as an alternative healthcare delivery service for managing diabetes in populations residing in urban medically underserved areas (UMUPAs). Retrospective electronic patient health records (ePHR) with type 2 diabetes mellitus (T2DM) were examined from 1 January 2019 to 30 June 2021. Multiple linear regression models indicated that T2DM patients with uncontrolled diabetes utilizing TM were similar to traditional visits in lowering hemoglobin (HbA1c) levels. The healthcare service type significantly predicted HbA1c % values, as the regression coefficient for TM (vs. F2F) showed a significant negative association (B = −0.339, p < 0.001), suggesting that patients using TM were likely to have 0.34 lower HbA1c % values on average when compared with F2F visits. The regression coefficient for female (vs. male) gender showed a positive association (B = 0.190, p < 0.034), with HbA1c % levels showing that female patients had 0.19 higher HbA1c levels than males. Age (B = −0.026, p < 0.001) was a significant predictor of HbA1c % levels, with 0.026 lower HbA1c % levels for each year’s increase in age. Black adults (B = 0.888, p < 0.001), on average, were more likely to have 0.888 higher HbA1c % levels when compared with White adults.
]]>Informatics doi: 10.3390/informatics10010015
Authors: Francisco Javier Moreno Arboleda Georgia Garani Simon Zea Gallego
In this paper, a measure is proposed that, based on the trajectories of moving objects, computes the speed limit rate in each of the cells in which a region is segmented (the space where the objects move). The time is also segmented into intervals. In this way, the behavior of moving objects can be analyzed with regard to their speed in a cell for a given time interval. An implementation of the corresponding algorithm for this measure and several experiments were conducted with the trajectories of taxis in Porto (Portugal). The results showed that the speed limit rate measure can be helpful for detecting patterns of movement, e.g., in a day (morning hours vs. night hours) or on different days of the week (weekdays vs. weekends). This measure might also serve as a rough estimate for congestion in a (sub)region. This may be useful for traffic analysis, including traffic prediction.
]]>Informatics doi: 10.3390/informatics10010014
Authors: De-Graft Joe Opoku Srinath Perera Robert Osei-Kyei Maria Rashidi Keivan Bamdad Tosin Famakinwa
Digital twin (DT) has gained significant recognition among researchers due to its potential across industries. With the prime goal of solving numerous challenges confronting the construction industry (CI), DT in recent years has witnessed several applications in the CI. Hence, researchers have been advocating for DT adoption to tackle the challenges of the CI. Notwithstanding, a distinguishable set of barriers that oppose the adoption of DT in the CI has not been determined. Therefore, this paper identifies the barriers and incorporates them into a classified framework to enhance the roadmap for adopting DT in the CI. This research conducts an extensive review of the literature and analyses the barriers whilst integrating the science mapping technique. Using Scopus, ScienceDirect, and Web of Science databases, 154 related bibliographic records were identified and analysed using science mapping, while 40 carefully selected relevant publications were systematically reviewed. From the review, the top five barriers identified include low level of knowledge, low level of technology acceptance, lack of clear DT value propositions, project complexities, and static nature of building data. The results show that the UK, China, the USA, and Germany are the countries spearheading the DT adoption in the CI, while only a small number of institutions from Australia, the UK, Algeria, and Greece have established institutional collaborations for DT research. A conceptual framework was developed on the basis of 30 identified barriers to support the DT adoption roadmap. The main categories of the framework comprise stakeholder-oriented, industry-related, construction-enterprise-related, and technology-related barriers. The identified barriers and the framework will guide and broaden the knowledge of DT, which is critical for successful adoption in the construction industry.
]]>Informatics doi: 10.3390/informatics10010013
Authors: Alan E. Stewart Matthew J. Bolton
We review the emergence of digital weather information, the history of human embodied knowing about weather, and two perspectives on cognition, one of which is symbolic (amodal, abstract, and arbitrary) and the other being embodied (embodied, extended, embedded, and enacted) to address the question: Beyond the general weather information they provide, to what extent can digital devices be used in an embodied way to extend a person’s pick-up of weather information? This is an interesting question to examine because human weather information and knowledge has a long past in our evolutionary history. Our human ancestors had to pick-up immediate information from the environment (including the weather) to survive. Digital weather information and knowing has a comparatively short past and a promising future. After reviewing these relevant topics, we concluded that, with the possible exception of weather radar apps, nothing currently exists in the form of digital products than can extend the immediate sensory reach of people to alert them about just-about-to-occur weather—at least not in the embodied forms of information. We believe that people who are weather salient (i.e., have a strong psychological attunement to the weather) may be in the best position going forward to integrate digital weather knowing with that which is embodied.
]]>Informatics doi: 10.3390/informatics10010012
Authors: Ali Azadi Francisco José García-Peñalvo
Nowadays, according to the intention of many hospitals and medical centers to computerize their processes and medical treatments, including data forms and medical images, which are generating a considerable amount of data, IT specialists and data scientists who are oriented to eHealth and related issues know the importance of data integration and its benefits. This study indicates the significance of data integration, especially in medical information systems. It means that the medical subsystems in the HIS (hospital information system) must be integrated, and it is also necessary to unify with the MIS (management information system). In this paper, the accuracy level of the extracted reports from the information system (to evaluate the staff’s performance) will be measured in two ways: (1) At first, the performance of the clinic reception staff will be evaluated. In this way, the personnel attendance system is an independent and separate software, and the mentioned evaluation has been performed by its report. (2) The following year, in the same location, the same evaluation has been performed based on the data extracted from the personnel attendance subsystem, which has been added to the medical information system as an integrated information system. After comparing the accuracy level of both ways, this paper concludes that when the personnel attendance subsystem as a part of the MIS has been unified with the HIS, the reports and, consequently, management decisions will be more accurate; therefore, the managers and decision-makers will perceive the importance of data integration more than in the past.
]]>Informatics doi: 10.3390/informatics10010011
Authors: Informatics Editorial Office Informatics Editorial Office
High-quality academic publishing is built on rigorous peer review [...]
]]>Informatics doi: 10.3390/informatics10010010
Authors: Kokoy Siti Komariah Ariana Tulus Purnomo Ardianto Satriawan Muhammad Ogin Hasanuddin Casi Setianingsih Bong-Kee Sin
To pursue a healthy lifestyle, people are increasingly concerned about their food ingredients. Recently, it has become a common practice to use an online recipe to select the ingredients that match an individual’s meal plan and healthy diet preference. The information from online recipes can be extracted and used to develop various food-related applications. Named entity recognition (NER) is often used to extract such information. However, the problem in building an NER system lies in the massive amount of data needed to train the classifier, especially on a specific domain, such as food. There are food NER datasets available, but they are still quite limited. Thus, we proposed an iterative self-training approach called semi-supervised multi-model prediction technique (SMPT) to construct a food ingredient NER dataset. SMPT is a deep ensemble learning model that employs the concept of self-training and uses multiple pre-trained language models in the iterative data labeling process, with a voting mechanism used as the final decision to determine the entity’s label. Utilizing the SMPT, we have created a new annotated dataset of ingredient entities obtained from the Allrecipes website named FINER. Finally, this study aims to use the FINER dataset as an alternative resource to support food computing research and development.
]]>Informatics doi: 10.3390/informatics10010009
Authors: Vikram Puri Subhra Mondal Subhankar Das Vasiliki G. Vrana
Blockchain and immersive technology are the pioneers in bringing digitalization to tourism, and researchers worldwide are exploring many facets of these techniques. This paper analyzes the various aspects of blockchain technology and its potential use in tourism. We explore high-frequency keywords, perform network analysis of relevant publications to analyze patterns, and introduce machine learning techniques to facilitate systematic reviews. We focused on 94 publications from Web Science that dealt with blockchain implementation in tourism from 2017 to 2022. We used Vosviewer for network analysis and artificial intelligence models with the help of machine learning tools to predict the relevance of the work. Many reviewed articles mainly deal with blockchain in tourism and related terms such as smart tourism and crypto tourism. This study is the first attempt to use text analysis to improve the topic modeling of blockchain in tourism. It comprehensively analyzes the technology’s potential use in the hospitality, accommodation, and booking industry. In this context, the paper provides significant value to researchers by giving an insight into the trends and keyword patterns. Tourism still has many unexplored areas; journal articles should also feature special studies on this topic.
]]>Informatics doi: 10.3390/informatics10010008
Authors: Yifei Wang Shiyang Chen Guobin Chen Ethan Shurberg Hang Liu Pengyu Hong
This work considers the task of representation learning on the attributed relational graph (ARG). Both the nodes and edges in an ARG are associated with attributes/features allowing ARGs to encode rich structural information widely observed in real applications. Existing graph neural networks offer limited ability to capture complex interactions within local structural contexts, which hinders them from taking advantage of the expression power of ARGs. We propose motif convolution module (MCM), a new motif-based graph representation learning technique to better utilize local structural information. The ability to handle continuous edge and node features is one of MCM’s advantages over existing motif-based models. MCM builds a motif vocabulary in an unsupervised way and deploys a novel motif convolution operation to extract the local structural context of individual nodes, which is then used to learn higher level node representations via multilayer perceptron and/or message passing in graph neural networks. When compared with other graph learning approaches to classifying synthetic graphs, our approach is substantially better at capturing structural context. We also demonstrate the performance and explainability advantages of our approach by applying it to several molecular benchmarks.
]]>Informatics doi: 10.3390/informatics10010007
Authors: My Villius Zetterholm Päivi Jokela
The COVID-19 pandemic constitutes a wicked problem that is defined by rapidly evolving and dynamic conditions, where the physical world changes (e.g., pathogens mutate) and, in parallel, our understanding and knowledge rapidly progress. Various preventive measures have been developed or proposed to manage the situation, including digital preventive technologies to support contact tracing or physical distancing. The complexity of the pandemic and the rapidly evolving nature of the situation pose challenges for the design of effective preventive technologies. The aim of this conceptual paper is to apply a systems thinking model, DSRP (distinctions, systems, relations, perspectives) to explain the underlying assumptions, patterns, and connections of the pandemic domain, as well as to identify potential leverage points for design of preventive technologies. Two different design approaches, contact tracing and nudging for distance, are compared, focusing on how their design and preventive logic are related to system complexity. The analysis explains why a contact tracing technology involves more complexity, which can challenge both implementation and user understanding. A system utilizing nudges can operate using a more distinct system boundary, which can benefit understanding and implementation. However, frequent nudges might pose challenges for user experience. This further implies that these technologies have different contextual requirements and are useful at different levels in society. The main contribution of this work is to show how systems thinking can organize our understanding and guide the design of preventive technologies in the context of epidemics and pandemics.
]]>Informatics doi: 10.3390/informatics10010006
Authors: Dario Benvenuti Lauren S. Ferro Andrea Marrella Tiziana Catarci
Video games are an established medium that provides interactive entertainment beyond pure enjoyment in many contexts. Game designers create dedicated tutorials to teach players the game mechanisms and rules, such as the conventions for interaction, control schemes, core game mechanics, etc. While effective tutorial design is considered a crucial aspect to support this learning process, the existing literature approaches focus on designing ad hoc tutorials for specific game genres rather than investigating the impact of different tutorial styles on game learnability and player engagement. In this paper, we tackle this challenge by presenting a general-purpose approach aimed at supporting game designers in the identification of the most suitable tutorial style for a specific genre of video games. The approach is evaluated in the context of a simple first-person shooter (FPS) mainstream video game built by the authors through a controlled comparative user experiment involving 46 players.
]]>Informatics doi: 10.3390/informatics10010005
Authors: Andrea Pozzi Enrico Barbierato Daniele Toti
In the last decade, the techniques of news aggregation and summarization have been increasingly gaining relevance for providing users on the web with condensed and unbiased information. Indeed, the recent development of successful machine learning algorithms, such as those based on the transformers architecture, have made it possible to create effective tools for capturing and elaborating news from the Internet. In this regard, this work proposes, for the first time in the literature to the best of the authors’ knowledge, a methodology for the application of such techniques in news related to cryptocurrencies and the blockchain, whose quick reading can be deemed as extremely useful to operators in the financial sector. Specifically, cutting-edge solutions in the field of natural language processing were employed to cluster news by topic and summarize the corresponding articles published by different newspapers. The results achieved on 22,282 news articles show the effectiveness of the proposed methodology in most of the cases, with 86.8% of the examined summaries being considered as coherent and 95.7% of the corresponding articles correctly aggregated. This methodology was implemented in a freely accessible web application.
]]>Informatics doi: 10.3390/informatics10010004
Authors: Sergey Deryabin Igor Temkin Ulvi Rzazade Egor Kondratev
The article is devoted to methods and models of designing systems for the digital transformation of industrial enterprises within the framework of the Industry 4.0 concept. The purpose of this work is to formalize a new notation for graphical modeling of the architecture of complex large-scale systems with data-centric microservice architectures and to present a variant of the reference model of such an architecture for creating an autonomously functioning industrial enterprise. The paper provides a list and justification for the use of functional components of a data-centric microservice architecture based on the analysis of modern approaches to building systems and the authors’ own results obtained during the implementation of a number of projects. The problems of using traditional graphical modeling notations to represent a data-centric microservice architecture are considered. Examples of designing a model of such an architecture for a mining enterprise are given.
]]>Informatics doi: 10.3390/informatics10010003
Authors: Warodom Werapun Tanakorn Karode Tanwa Arpornthip Jakapan Suaboot Esther Sangiamkul Pawita Boonrat
Decentralized finance (DeFi) has exploded in popularity with a billion-dollar market cap. While uncollateralized lending, known as a flash loan, emerged from DeFi, it has become a primary tool used by attackers to drain investment tokens from DeFi networks. The existing countermeasures seem practical, but no comprehensive quantitative analysis framework was available to test them. This paper proposes the Flash loan Attack Analysis (FAA) framework, which aids security practitioners in understanding the DeFi system’s effects on preventative methods when various factors change. The quantitative predictions can help security professionals in identifying hidden dangers and more efficiently adopting countermeasure strategies. The simulation predicts that the existing strategy, fair reserves, can fully protect the platform in a typical market environment; however, in a highly volatile market where the token price drops by 60% in a single hour, it will be broken, causing more than $8 million in damage.
]]>