<?xml version="1.0" encoding="UTF-8"?>
<rdf:RDF xmlns="http://purl.org/rss/1.0/"
 xmlns:dc="http://purl.org/dc/elements/1.1/"
 xmlns:dcterms="http://purl.org/dc/terms/"
 xmlns:cc="http://web.resource.org/cc/"
 xmlns:prism="http://prismstandard.org/namespaces/basic/2.0/"
 xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#"
 xmlns:admin="http://webns.net/mvcb/"
 xmlns:content="http://purl.org/rss/1.0/modules/content/">
    <channel rdf:about="https://www.mdpi.com/rss/journal/analytics">
		<title>Analytics</title>
		<description>Latest open access articles published in Analytics at https://www.mdpi.com/journal/analytics</description>
		<link>https://www.mdpi.com/journal/analytics</link>
		<admin:generatorAgent rdf:resource="https://www.mdpi.com/journal/analytics"/>
		<admin:errorReportsTo rdf:resource="mailto:support@mdpi.com"/>
		<dc:publisher>MDPI</dc:publisher>
		<dc:language>en</dc:language>
		<dc:rights>Creative Commons Attribution (CC-BY)</dc:rights>
						<prism:copyright>MDPI</prism:copyright>
		<prism:rightsAgent>support@mdpi.com</prism:rightsAgent>
		<image rdf:resource="https://pub.mdpi-res.com/img/design/mdpi-pub-logo.png?13cf3b5bd783e021?1773146959"/>
				<items>
			<rdf:Seq>
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/5/1/14" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/5/1/13" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/5/1/12" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/5/1/11" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/5/1/10" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/5/1/9" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/5/1/8" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/5/1/7" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/5/1/6" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/5/1/5" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/5/1/4" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/5/1/3" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/5/1/2" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/5/1/1" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/4/4/36" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/4/4/35" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/4/4/34" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/4/4/33" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/4/4/32" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/4/4/31" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/4/4/30" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/4/4/29" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/4/4/28" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/4/4/27" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/4/4/26" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/4/3/25" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/4/3/24" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/4/3/23" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/4/3/22" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/4/3/21" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/4/3/20" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/4/3/19" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/4/3/18" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/4/3/17" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/4/2/16" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/4/2/15" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/4/2/14" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/4/2/13" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/4/2/12" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/4/2/11" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/4/1/10" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/4/1/9" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/4/1/8" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/4/1/7" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/4/1/6" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/4/1/5" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/4/1/4" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/4/1/3" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/4/1/2" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/4/1/1" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/3/4/28" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/3/4/27" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/3/4/26" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/3/4/25" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/3/4/24" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/3/4/23" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/3/4/22" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/3/3/21" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/3/3/20" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/3/3/19" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/3/3/18" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/3/3/17" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/3/3/16" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/3/3/15" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/3/2/14" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/3/2/13" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/3/2/12" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/3/2/11" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/3/2/10" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/3/1/9" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/3/1/8" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/3/1/7" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/3/1/6" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/3/1/5" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/3/1/4" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/3/1/3" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/3/1/2" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/3/1/1" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/2/4/46" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/2/4/45" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/2/4/44" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/2/4/43" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/2/4/42" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/2/4/41" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/2/3/40" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/2/3/39" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/2/3/38" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/2/3/37" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/2/3/36" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/2/3/35" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/2/3/34" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/2/3/33" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/2/3/32" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/2/3/31" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/2/3/30" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/2/2/29" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/2/2/28" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/2/2/27" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/2/2/26" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2813-2203/2/2/25" />
                    	</rdf:Seq>
		</items>
				<cc:license rdf:resource="https://creativecommons.org/licenses/by/4.0/" />
	</channel>

        <item rdf:about="https://www.mdpi.com/2813-2203/5/1/14">

	<title>Analytics, Vol. 5, Pages 14: A Decade of Evolution: Evaluating Student Preferences for Degree Selection in the Spanish Public University System Through Directional Community Analysis (2014&amp;ndash;2023)</title>
	<link>https://www.mdpi.com/2813-2203/5/1/14</link>
	<description>The Spanish Public University System (SUPE) assigns student placements through a multi-step application process governed by legal criteria. Analyzing how students move between different degree programs during this process is crucial for universities to optimize and plan their academic offerings. This paper analyzes a decade of student pre-registration data (2014&amp;amp;ndash;2023) to track evolving preferences and mobility between degrees. We model this process as a directed graph, mapping student traffic and studying the formation of directional communities within the degree network. A significant challenge is the weakly connected and poorly conditioned nature of these graphs, which impedes standard community detection algorithms. Extending prior work that relied on manually set thresholds for pruning edges, we propose a novel adaptive pruning algorithm that requires no manual intervention. Applying this method to annual data improves community detection performance and reveals gradual shifts in student preferences and demand, particularly in response to new degrees. These insights provide a valuable decision-making tool for higher education institutions, helping them refine their degree offerings in response to evolving trends.</description>
	<pubDate>2026-03-11</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 5, Pages 14: A Decade of Evolution: Evaluating Student Preferences for Degree Selection in the Spanish Public University System Through Directional Community Analysis (2014&amp;ndash;2023)</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/5/1/14">doi: 10.3390/analytics5010014</a></p>
	<p>Authors:
		José-Miguel Montañana
		Antonio Hervás
		Pedro-Pablo Soriano-Jiménez
		</p>
	<p>The Spanish Public University System (SUPE) assigns student placements through a multi-step application process governed by legal criteria. Analyzing how students move between different degree programs during this process is crucial for universities to optimize and plan their academic offerings. This paper analyzes a decade of student pre-registration data (2014&amp;amp;ndash;2023) to track evolving preferences and mobility between degrees. We model this process as a directed graph, mapping student traffic and studying the formation of directional communities within the degree network. A significant challenge is the weakly connected and poorly conditioned nature of these graphs, which impedes standard community detection algorithms. Extending prior work that relied on manually set thresholds for pruning edges, we propose a novel adaptive pruning algorithm that requires no manual intervention. Applying this method to annual data improves community detection performance and reveals gradual shifts in student preferences and demand, particularly in response to new degrees. These insights provide a valuable decision-making tool for higher education institutions, helping them refine their degree offerings in response to evolving trends.</p>
	]]></content:encoded>

	<dc:title>A Decade of Evolution: Evaluating Student Preferences for Degree Selection in the Spanish Public University System Through Directional Community Analysis (2014&amp;amp;ndash;2023)</dc:title>
			<dc:creator>José-Miguel Montañana</dc:creator>
			<dc:creator>Antonio Hervás</dc:creator>
			<dc:creator>Pedro-Pablo Soriano-Jiménez</dc:creator>
		<dc:identifier>doi: 10.3390/analytics5010014</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2026-03-11</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2026-03-11</prism:publicationDate>
	<prism:volume>5</prism:volume>
	<prism:number>1</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>14</prism:startingPage>
		<prism:doi>10.3390/analytics5010014</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/5/1/14</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/5/1/13">

	<title>Analytics, Vol. 5, Pages 13: Distributed Orders Management in Make-to-Order Supply Chain Networks Using Game-Based Alternating Direction Method of Multipliers</title>
	<link>https://www.mdpi.com/2813-2203/5/1/13</link>
	<description>Operations scheduling of mass customized products is vital in the modern make-to-order (MTO) supply chains. In these systems, order acceptance decisions should be coordinated with available capacity in different sections of the supply chain while considering their potential correlations and interactions. One of the fundamental challenges in optimization of these systems is the computation time of solving models with multiple coupling constraints between supply chain units. This paper addresses this issue by proposing a game-based framework that decomposes the related mixed integer programming mathematical model and it is coordinated and solved using integrated game-based Alternating Direction Method of Multipliers (ADMM). The proposed Stackelberg Leader-Follower game optimizes order acceptance decisions while considering the requirements in supply, production planning, maintenance, inventory, and distribution units. To validate the efficiency of the proposed framework, the model is tested with a simulated four-layer supply chain. The results of experiments proved that decompositions of the model to smaller subsections and solving it in a distributed manner not only optimizes supply chain participating units but also coordinate their movements to achieve the global optimal solution. The proposed framework offers managers a practical decision layer that preserve local autonomy of the supply chain units and reduce their data sharing and computation burdens and concerns.</description>
	<pubDate>2026-03-09</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 5, Pages 13: Distributed Orders Management in Make-to-Order Supply Chain Networks Using Game-Based Alternating Direction Method of Multipliers</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/5/1/13">doi: 10.3390/analytics5010013</a></p>
	<p>Authors:
		Amirhosein Gholami
		Nasim Nezamoddini
		Mohammad T. Khasawneh
		</p>
	<p>Operations scheduling of mass customized products is vital in the modern make-to-order (MTO) supply chains. In these systems, order acceptance decisions should be coordinated with available capacity in different sections of the supply chain while considering their potential correlations and interactions. One of the fundamental challenges in optimization of these systems is the computation time of solving models with multiple coupling constraints between supply chain units. This paper addresses this issue by proposing a game-based framework that decomposes the related mixed integer programming mathematical model and it is coordinated and solved using integrated game-based Alternating Direction Method of Multipliers (ADMM). The proposed Stackelberg Leader-Follower game optimizes order acceptance decisions while considering the requirements in supply, production planning, maintenance, inventory, and distribution units. To validate the efficiency of the proposed framework, the model is tested with a simulated four-layer supply chain. The results of experiments proved that decompositions of the model to smaller subsections and solving it in a distributed manner not only optimizes supply chain participating units but also coordinate their movements to achieve the global optimal solution. The proposed framework offers managers a practical decision layer that preserve local autonomy of the supply chain units and reduce their data sharing and computation burdens and concerns.</p>
	]]></content:encoded>

	<dc:title>Distributed Orders Management in Make-to-Order Supply Chain Networks Using Game-Based Alternating Direction Method of Multipliers</dc:title>
			<dc:creator>Amirhosein Gholami</dc:creator>
			<dc:creator>Nasim Nezamoddini</dc:creator>
			<dc:creator>Mohammad T. Khasawneh</dc:creator>
		<dc:identifier>doi: 10.3390/analytics5010013</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2026-03-09</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2026-03-09</prism:publicationDate>
	<prism:volume>5</prism:volume>
	<prism:number>1</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>13</prism:startingPage>
		<prism:doi>10.3390/analytics5010013</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/5/1/13</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/5/1/12">

	<title>Analytics, Vol. 5, Pages 12: Operationalising CTT and IRT in Spreadsheets: A Methodological Demonstration for Classroom Assessment</title>
	<link>https://www.mdpi.com/2813-2203/5/1/12</link>
	<description>The evaluation of student performance often relies on basic spreadsheet outputs that provide limited insight into item functioning. This study presents a methodological demonstration showing how widely available spreadsheet software can be transformed into a practical environment for psychometric analysis. Using a simulated dataset of 40 students responding to 20 dichotomous items, spreadsheet formulas were developed to compute descriptive statistics and Classical Test Theory (CTT) indices, including item difficulty, discrimination, and corrected item&amp;amp;ndash;total correlations. The demonstration was extended to Item Response Theory (IRT) through the implementation of 1PL, 2PL, and 3PL logistic models using forward-calculated item parameters. A smaller dataset of 10 students and 10 items was used to illustrate the interpretability of the indices and the generation of Item Characteristic Curves (ICCs). Results show that spreadsheets can support teachers in in-terpreting test data beyond total scores, enabling the identification of weak items, refinement of distractors, and construction of small-scale item banks aligned with competence-based curricula. The approach contributes to Sustainable Development Goal 4 (SDG 4) by promoting accessible, equitable, and high-quality assessment practices. Limitations include the instability of IRT parameter estimation in small samples and the need for teacher training. Future research should apply the approach to real classroom data, explore automation within spreadsheet environments, and examine the integration of artificial intelligence for adaptive assessment.</description>
	<pubDate>2026-02-24</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 5, Pages 12: Operationalising CTT and IRT in Spreadsheets: A Methodological Demonstration for Classroom Assessment</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/5/1/12">doi: 10.3390/analytics5010012</a></p>
	<p>Authors:
		António Faria
		Guilhermina Lobato Miranda
		</p>
	<p>The evaluation of student performance often relies on basic spreadsheet outputs that provide limited insight into item functioning. This study presents a methodological demonstration showing how widely available spreadsheet software can be transformed into a practical environment for psychometric analysis. Using a simulated dataset of 40 students responding to 20 dichotomous items, spreadsheet formulas were developed to compute descriptive statistics and Classical Test Theory (CTT) indices, including item difficulty, discrimination, and corrected item&amp;amp;ndash;total correlations. The demonstration was extended to Item Response Theory (IRT) through the implementation of 1PL, 2PL, and 3PL logistic models using forward-calculated item parameters. A smaller dataset of 10 students and 10 items was used to illustrate the interpretability of the indices and the generation of Item Characteristic Curves (ICCs). Results show that spreadsheets can support teachers in in-terpreting test data beyond total scores, enabling the identification of weak items, refinement of distractors, and construction of small-scale item banks aligned with competence-based curricula. The approach contributes to Sustainable Development Goal 4 (SDG 4) by promoting accessible, equitable, and high-quality assessment practices. Limitations include the instability of IRT parameter estimation in small samples and the need for teacher training. Future research should apply the approach to real classroom data, explore automation within spreadsheet environments, and examine the integration of artificial intelligence for adaptive assessment.</p>
	]]></content:encoded>

	<dc:title>Operationalising CTT and IRT in Spreadsheets: A Methodological Demonstration for Classroom Assessment</dc:title>
			<dc:creator>António Faria</dc:creator>
			<dc:creator>Guilhermina Lobato Miranda</dc:creator>
		<dc:identifier>doi: 10.3390/analytics5010012</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2026-02-24</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2026-02-24</prism:publicationDate>
	<prism:volume>5</prism:volume>
	<prism:number>1</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>12</prism:startingPage>
		<prism:doi>10.3390/analytics5010012</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/5/1/12</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/5/1/11">

	<title>Analytics, Vol. 5, Pages 11: Integrating Deep Learning Nodes into an Augmented Decision Tree for Automated Medical Coding</title>
	<link>https://www.mdpi.com/2813-2203/5/1/11</link>
	<description>Accurate assignment of International Classification of Diseases (ICD) codes is essential for healthcare analytics, billing, and clinical research. However, manual coding remains time-consuming and error-prone due to the scale and complexity of the ICD taxonomy. While hierarchical deep learning approaches have improved automated coding, their deployment across large taxonomies raises scalability and efficiency concerns. To address these limitations, we introduce the Augmented Decision Tree (ADT) framework, which integrates deep learning with symbolic rule-based logic for automated medical coding. ADT employs an automated lexical screening mechanism to dynamically select the most appropriate modeling strategy for each decision node, thereby minimizing manual configuration. Nodes with high keyword distinctiveness are handled by symbolic rules, while semantically ambiguous nodes are assigned to deep contextual models fine-tuned from PubMedBERT. This selective design eliminates the need to train a deep learning model at every node, significantly reducing computational cost. A case study demonstrates that this hybrid and adaptive ADT approach supports scalable and efficient ICD coding. Experimental results show that ADT outperforms a pure decision tree baseline and achieves accuracy comparable to that of a full deep learning-based decision tree, while requiring substantially less training time and computational resources.</description>
	<pubDate>2026-02-12</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 5, Pages 11: Integrating Deep Learning Nodes into an Augmented Decision Tree for Automated Medical Coding</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/5/1/11">doi: 10.3390/analytics5010011</a></p>
	<p>Authors:
		Spoorthi Bhat
		Veda Sahaja Bandi
		Haiping Xu
		Joshua Carberry
		</p>
	<p>Accurate assignment of International Classification of Diseases (ICD) codes is essential for healthcare analytics, billing, and clinical research. However, manual coding remains time-consuming and error-prone due to the scale and complexity of the ICD taxonomy. While hierarchical deep learning approaches have improved automated coding, their deployment across large taxonomies raises scalability and efficiency concerns. To address these limitations, we introduce the Augmented Decision Tree (ADT) framework, which integrates deep learning with symbolic rule-based logic for automated medical coding. ADT employs an automated lexical screening mechanism to dynamically select the most appropriate modeling strategy for each decision node, thereby minimizing manual configuration. Nodes with high keyword distinctiveness are handled by symbolic rules, while semantically ambiguous nodes are assigned to deep contextual models fine-tuned from PubMedBERT. This selective design eliminates the need to train a deep learning model at every node, significantly reducing computational cost. A case study demonstrates that this hybrid and adaptive ADT approach supports scalable and efficient ICD coding. Experimental results show that ADT outperforms a pure decision tree baseline and achieves accuracy comparable to that of a full deep learning-based decision tree, while requiring substantially less training time and computational resources.</p>
	]]></content:encoded>

	<dc:title>Integrating Deep Learning Nodes into an Augmented Decision Tree for Automated Medical Coding</dc:title>
			<dc:creator>Spoorthi Bhat</dc:creator>
			<dc:creator>Veda Sahaja Bandi</dc:creator>
			<dc:creator>Haiping Xu</dc:creator>
			<dc:creator>Joshua Carberry</dc:creator>
		<dc:identifier>doi: 10.3390/analytics5010011</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2026-02-12</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2026-02-12</prism:publicationDate>
	<prism:volume>5</prism:volume>
	<prism:number>1</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>11</prism:startingPage>
		<prism:doi>10.3390/analytics5010011</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/5/1/11</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/5/1/10">

	<title>Analytics, Vol. 5, Pages 10: Site Selection for Solar Photovoltaic Power Plant Using MCDM Method with New De-i-Fuzzification Technique</title>
	<link>https://www.mdpi.com/2813-2203/5/1/10</link>
	<description>Choosing sites for solar photovoltaic (PV) power plants in developing countries like India is a crucial task while considering multiple conflicting factors and sub-factors simultaneously. Multi-criteria decision-making (MCDM) is an optimisation method that provides a framework for handling such situations in an intuitionistic fuzzy environment. The complexity and uncertainty associated with the site selection model are dealt with professionally. The Criteria Importance Through Intercriteria Correlation (CRITIC) method is applied to determine the relative importance of the criteria, identifying airflow speed as the most influential factor, followed by humidity ratio, level of dust haze, availability of labour and resources, and ecological effects. This shows that airflow speed plays an important role in the power plant&amp;amp;rsquo;s efficiency and performance. The Vlse Kriterijumska Optimizacija I Kompromisno Re&amp;amp;scaron;enje (VIKOR) method is then used to prioritise the alternatives as potential locations for setting up a solar PV power plant in India. A new de-i-fuzzification method based on the relative difference between two real numbers is also proposed. Sensitivity analyses and comparative studies are conducted to assess the robustness and effectiveness of the framework. Overall, the results demonstrate that the proposed framework is useful and effective for optimising site selection for solar power plants in India.</description>
	<pubDate>2026-02-09</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 5, Pages 10: Site Selection for Solar Photovoltaic Power Plant Using MCDM Method with New De-i-Fuzzification Technique</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/5/1/10">doi: 10.3390/analytics5010010</a></p>
	<p>Authors:
		Kamal Hossain Gazi
		Asesh Kumar Mukherjee
		Shashi Bajaj Mukherjee
		Sankar Prasad Mondal
		Soheil Salahshour
		Arijit Ghosh
		</p>
	<p>Choosing sites for solar photovoltaic (PV) power plants in developing countries like India is a crucial task while considering multiple conflicting factors and sub-factors simultaneously. Multi-criteria decision-making (MCDM) is an optimisation method that provides a framework for handling such situations in an intuitionistic fuzzy environment. The complexity and uncertainty associated with the site selection model are dealt with professionally. The Criteria Importance Through Intercriteria Correlation (CRITIC) method is applied to determine the relative importance of the criteria, identifying airflow speed as the most influential factor, followed by humidity ratio, level of dust haze, availability of labour and resources, and ecological effects. This shows that airflow speed plays an important role in the power plant&amp;amp;rsquo;s efficiency and performance. The Vlse Kriterijumska Optimizacija I Kompromisno Re&amp;amp;scaron;enje (VIKOR) method is then used to prioritise the alternatives as potential locations for setting up a solar PV power plant in India. A new de-i-fuzzification method based on the relative difference between two real numbers is also proposed. Sensitivity analyses and comparative studies are conducted to assess the robustness and effectiveness of the framework. Overall, the results demonstrate that the proposed framework is useful and effective for optimising site selection for solar power plants in India.</p>
	]]></content:encoded>

	<dc:title>Site Selection for Solar Photovoltaic Power Plant Using MCDM Method with New De-i-Fuzzification Technique</dc:title>
			<dc:creator>Kamal Hossain Gazi</dc:creator>
			<dc:creator>Asesh Kumar Mukherjee</dc:creator>
			<dc:creator>Shashi Bajaj Mukherjee</dc:creator>
			<dc:creator>Sankar Prasad Mondal</dc:creator>
			<dc:creator>Soheil Salahshour</dc:creator>
			<dc:creator>Arijit Ghosh</dc:creator>
		<dc:identifier>doi: 10.3390/analytics5010010</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2026-02-09</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2026-02-09</prism:publicationDate>
	<prism:volume>5</prism:volume>
	<prism:number>1</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>10</prism:startingPage>
		<prism:doi>10.3390/analytics5010010</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/5/1/10</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/5/1/9">

	<title>Analytics, Vol. 5, Pages 9: Denoising Stock Price Time Series with Singular Spectrum Analysis for Enhanced Deep Learning Forecasting</title>
	<link>https://www.mdpi.com/2813-2203/5/1/9</link>
	<description>Aim: Stock price prediction remains a highly challenging task due to the complex and nonlinear nature of financial time series data. While deep learning (DL) has shown promise in capturing these nonlinear patterns, its effectiveness is often hindered by the low signal-to-noise ratio inherent in market data. This study aims to enhance the stock predictive performance and trading outcomes by integrating Singular Spectrum Analysis (SSA) with deep learning models for stock price forecasting and strategy development on the Australian Securities Exchange (ASX)50 index. Method: The proposed framework begins by applying SSA to decompose raw stock price time series into interpretable components, effectively isolating meaningful trends and eliminating noise. The denoised sequences are then used to train a suite of deep learning architectures, including Convolutional Neural Networks (CNN), Long Short-Term Memory (LSTM), and hybrid CNN-LSTM models. These models are evaluated based on their forecasting accuracy and the profitability of the trading strategies derived from their predictions. Results: Experimental results demonstrated that the SSA-DL framework significantly improved the prediction accuracy and trading performance compared to baseline DL models trained on raw data. The best-performing model, SSA-CNN-LSTM, achieved a Sharpe Ratio of 1.88 and a return on investment (ROI) of 67%, indicating robust risk-adjusted returns and effective exploitation of the underlying market conditions. Conclusions: The integration of Singular Spectrum Analysis with deep learning offers a powerful approach to stock price prediction in noisy financial environments. By denoising input data prior to model training, the SSA-DL framework enhanced signal clarity, improved forecast reliability, and enabled the construction of profitable trading strategies. These findings suggested a strong potential for SSA-based preprocessing in financial time series modeling.</description>
	<pubDate>2026-01-27</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 5, Pages 9: Denoising Stock Price Time Series with Singular Spectrum Analysis for Enhanced Deep Learning Forecasting</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/5/1/9">doi: 10.3390/analytics5010009</a></p>
	<p>Authors:
		Carol Anne Hargreaves
		Zixian Fan
		</p>
	<p>Aim: Stock price prediction remains a highly challenging task due to the complex and nonlinear nature of financial time series data. While deep learning (DL) has shown promise in capturing these nonlinear patterns, its effectiveness is often hindered by the low signal-to-noise ratio inherent in market data. This study aims to enhance the stock predictive performance and trading outcomes by integrating Singular Spectrum Analysis (SSA) with deep learning models for stock price forecasting and strategy development on the Australian Securities Exchange (ASX)50 index. Method: The proposed framework begins by applying SSA to decompose raw stock price time series into interpretable components, effectively isolating meaningful trends and eliminating noise. The denoised sequences are then used to train a suite of deep learning architectures, including Convolutional Neural Networks (CNN), Long Short-Term Memory (LSTM), and hybrid CNN-LSTM models. These models are evaluated based on their forecasting accuracy and the profitability of the trading strategies derived from their predictions. Results: Experimental results demonstrated that the SSA-DL framework significantly improved the prediction accuracy and trading performance compared to baseline DL models trained on raw data. The best-performing model, SSA-CNN-LSTM, achieved a Sharpe Ratio of 1.88 and a return on investment (ROI) of 67%, indicating robust risk-adjusted returns and effective exploitation of the underlying market conditions. Conclusions: The integration of Singular Spectrum Analysis with deep learning offers a powerful approach to stock price prediction in noisy financial environments. By denoising input data prior to model training, the SSA-DL framework enhanced signal clarity, improved forecast reliability, and enabled the construction of profitable trading strategies. These findings suggested a strong potential for SSA-based preprocessing in financial time series modeling.</p>
	]]></content:encoded>

	<dc:title>Denoising Stock Price Time Series with Singular Spectrum Analysis for Enhanced Deep Learning Forecasting</dc:title>
			<dc:creator>Carol Anne Hargreaves</dc:creator>
			<dc:creator>Zixian Fan</dc:creator>
		<dc:identifier>doi: 10.3390/analytics5010009</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2026-01-27</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2026-01-27</prism:publicationDate>
	<prism:volume>5</prism:volume>
	<prism:number>1</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>9</prism:startingPage>
		<prism:doi>10.3390/analytics5010009</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/5/1/9</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/5/1/8">

	<title>Analytics, Vol. 5, Pages 8: From Models to Metrics: A Governance Framework for Large Language Models in Enterprise AI and Analytics</title>
	<link>https://www.mdpi.com/2813-2203/5/1/8</link>
	<description>Large language models (LLMs) and other foundation models are rapidly being woven into enterprise analytics workflows, where they assist with data exploration, forecasting, decision support, and automation. These systems can feel like powerful new teammates: creative, scalable, and tireless. Yet they also introduce distinctive risks related to opacity, brittleness, bias, and misalignment with organizational goals. Existing work on AI ethics, alignment, and governance provides valuable principles and technical safeguards, but enterprises still lack practical frameworks that connect these ideas to the specific metrics, controls, and workflows by which analytics teams design, deploy, and monitor LLM-powered systems. This paper proposes a conceptual governance framework for enterprise AI and analytics that is explicitly centered on LLMs embedded in analytics pipelines. The framework adopts a three-layered perspective&amp;amp;mdash;model and data alignment, system and workflow alignment, and ecosystem and governance alignment&amp;amp;mdash;that links technical properties of models to enterprise analytics practices, performance indicators, and oversight mechanisms. In practical terms, the framework shows how model and workflow choices translate into concrete metrics and inform real deployment, monitoring, and scaling decisions for LLM-powered analytics. We also illustrate how this framework can guide the design of controls for metrics, monitoring, human-in-the-loop structures, and incident response in LLM-driven analytics. The paper concludes with implications for analytics leaders and governance teams seeking to operationalize responsible, scalable use of LLMs in enterprise settings.</description>
	<pubDate>2026-01-11</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 5, Pages 8: From Models to Metrics: A Governance Framework for Large Language Models in Enterprise AI and Analytics</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/5/1/8">doi: 10.3390/analytics5010008</a></p>
	<p>Authors:
		Darshan Desai
		Ashish Desai
		</p>
	<p>Large language models (LLMs) and other foundation models are rapidly being woven into enterprise analytics workflows, where they assist with data exploration, forecasting, decision support, and automation. These systems can feel like powerful new teammates: creative, scalable, and tireless. Yet they also introduce distinctive risks related to opacity, brittleness, bias, and misalignment with organizational goals. Existing work on AI ethics, alignment, and governance provides valuable principles and technical safeguards, but enterprises still lack practical frameworks that connect these ideas to the specific metrics, controls, and workflows by which analytics teams design, deploy, and monitor LLM-powered systems. This paper proposes a conceptual governance framework for enterprise AI and analytics that is explicitly centered on LLMs embedded in analytics pipelines. The framework adopts a three-layered perspective&amp;amp;mdash;model and data alignment, system and workflow alignment, and ecosystem and governance alignment&amp;amp;mdash;that links technical properties of models to enterprise analytics practices, performance indicators, and oversight mechanisms. In practical terms, the framework shows how model and workflow choices translate into concrete metrics and inform real deployment, monitoring, and scaling decisions for LLM-powered analytics. We also illustrate how this framework can guide the design of controls for metrics, monitoring, human-in-the-loop structures, and incident response in LLM-driven analytics. The paper concludes with implications for analytics leaders and governance teams seeking to operationalize responsible, scalable use of LLMs in enterprise settings.</p>
	]]></content:encoded>

	<dc:title>From Models to Metrics: A Governance Framework for Large Language Models in Enterprise AI and Analytics</dc:title>
			<dc:creator>Darshan Desai</dc:creator>
			<dc:creator>Ashish Desai</dc:creator>
		<dc:identifier>doi: 10.3390/analytics5010008</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2026-01-11</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2026-01-11</prism:publicationDate>
	<prism:volume>5</prism:volume>
	<prism:number>1</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>8</prism:startingPage>
		<prism:doi>10.3390/analytics5010008</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/5/1/8</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/5/1/7">

	<title>Analytics, Vol. 5, Pages 7: Predicting ESG Scores Using Machine Learning for Data-Driven Sustainable Investment</title>
	<link>https://www.mdpi.com/2813-2203/5/1/7</link>
	<description>Environmental, social and governance (ESG) metrics increasingly inform sustainable investment yet suffer from inter-rater heterogeneity and incomplete reporting, limiting their utility for forward-looking allocation. In this study, we developed and validated a two-level stacked-ensemble machine-learning framework to predict total ESG risk scores for S&amp;amp;amp;P 500 firms using a comprehensive feature set comprising pillar sub-scores, controversy measures, firm financials, categorical descriptors and geospatial environmental indicators. Data pre-processing combined median/mean imputation, one-hot encoding, normalization and rigorous feature engineering; models were trained with an 80:20 train&amp;amp;ndash;test split and hyperparameters tuned by k-fold cross-validation. The stacked ensemble substantially outperformed single-model baselines (RMSE = 1.006, MAE = 0.664, MAPE = 3.13%, R2 = 0.979, CV_RMSE_Mean = 1.383, CV_R2_Mean = 0.957), with LightGBM and gradient boosting as competitive comparators. Permutation importance and correlation analysis identified environmental and social components as primary drivers (environmental importance = 0.41; social = 0.32), with potential multicollinearity between component and aggregate scores. This study concludes that ensemble-based predictive analytics can produce reliable, actionable ESG estimates to enhance screening and prioritization in sustainable investment, while recommending human review for extreme predictions and further work to harmonize cross-provider score divergence.</description>
	<pubDate>2026-01-09</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 5, Pages 7: Predicting ESG Scores Using Machine Learning for Data-Driven Sustainable Investment</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/5/1/7">doi: 10.3390/analytics5010007</a></p>
	<p>Authors:
		Sanskruti Patel
		Abhay Nath
		Pranav Desai
		</p>
	<p>Environmental, social and governance (ESG) metrics increasingly inform sustainable investment yet suffer from inter-rater heterogeneity and incomplete reporting, limiting their utility for forward-looking allocation. In this study, we developed and validated a two-level stacked-ensemble machine-learning framework to predict total ESG risk scores for S&amp;amp;amp;P 500 firms using a comprehensive feature set comprising pillar sub-scores, controversy measures, firm financials, categorical descriptors and geospatial environmental indicators. Data pre-processing combined median/mean imputation, one-hot encoding, normalization and rigorous feature engineering; models were trained with an 80:20 train&amp;amp;ndash;test split and hyperparameters tuned by k-fold cross-validation. The stacked ensemble substantially outperformed single-model baselines (RMSE = 1.006, MAE = 0.664, MAPE = 3.13%, R2 = 0.979, CV_RMSE_Mean = 1.383, CV_R2_Mean = 0.957), with LightGBM and gradient boosting as competitive comparators. Permutation importance and correlation analysis identified environmental and social components as primary drivers (environmental importance = 0.41; social = 0.32), with potential multicollinearity between component and aggregate scores. This study concludes that ensemble-based predictive analytics can produce reliable, actionable ESG estimates to enhance screening and prioritization in sustainable investment, while recommending human review for extreme predictions and further work to harmonize cross-provider score divergence.</p>
	]]></content:encoded>

	<dc:title>Predicting ESG Scores Using Machine Learning for Data-Driven Sustainable Investment</dc:title>
			<dc:creator>Sanskruti Patel</dc:creator>
			<dc:creator>Abhay Nath</dc:creator>
			<dc:creator>Pranav Desai</dc:creator>
		<dc:identifier>doi: 10.3390/analytics5010007</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2026-01-09</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2026-01-09</prism:publicationDate>
	<prism:volume>5</prism:volume>
	<prism:number>1</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>7</prism:startingPage>
		<prism:doi>10.3390/analytics5010007</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/5/1/7</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/5/1/6">

	<title>Analytics, Vol. 5, Pages 6: Interference-Driven Scaling Variability in Burst-Based Loopless Invasion Percolation Models of Induced Seismicity</title>
	<link>https://www.mdpi.com/2813-2203/5/1/6</link>
	<description>Many fluid-injection sequences display burst-like seismicity with approximate power-law event-size distributions whose exponents drift between catalogs. Classical percolation models instead predict fixed, dimension-dependent exponents and do not specify which geometric mechanisms could underlie such b-value variability. We address this gap using two loopless invasion percolation variants&amp;amp;mdash;the constrained Leath invasion percolation (CLIP) and avalanche invasion percolation (AIP) models&amp;amp;mdash;to generate synthetic burst catalogs and quantify how burst geometry modifies size&amp;amp;ndash;frequency statistics. For each model we measure burst-size distributions and an interference fraction, defined as the proportion of attempted growth steps that terminate on previously activated bonds. Single-burst clusters recover the Fisher exponent of classical percolation, whereas multi-burst sequences show systematic, dimension-dependent drift of the effective exponent with a burst number that is strongly correlated with the interference fraction. CLIP and AIP are indistinguishable under these diagnostics, indicating that interference-driven exponent drift is a generic feature of burst growth rather than a model-specific artifact. Mapping the size-distribution exponent to an equivalent Gutenberg&amp;amp;ndash;Richter b-value shows that increasing interference suppresses large bursts and produces b value ranges comparable to those reported for injection-induced seismicity, supporting the interpretation of interference as a geometric proxy for mechanical inhibition that limits the growth of large events in real fracture networks.</description>
	<pubDate>2026-01-06</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 5, Pages 6: Interference-Driven Scaling Variability in Burst-Based Loopless Invasion Percolation Models of Induced Seismicity</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/5/1/6">doi: 10.3390/analytics5010006</a></p>
	<p>Authors:
		Ian Baughman
		John B. Rundle
		</p>
	<p>Many fluid-injection sequences display burst-like seismicity with approximate power-law event-size distributions whose exponents drift between catalogs. Classical percolation models instead predict fixed, dimension-dependent exponents and do not specify which geometric mechanisms could underlie such b-value variability. We address this gap using two loopless invasion percolation variants&amp;amp;mdash;the constrained Leath invasion percolation (CLIP) and avalanche invasion percolation (AIP) models&amp;amp;mdash;to generate synthetic burst catalogs and quantify how burst geometry modifies size&amp;amp;ndash;frequency statistics. For each model we measure burst-size distributions and an interference fraction, defined as the proportion of attempted growth steps that terminate on previously activated bonds. Single-burst clusters recover the Fisher exponent of classical percolation, whereas multi-burst sequences show systematic, dimension-dependent drift of the effective exponent with a burst number that is strongly correlated with the interference fraction. CLIP and AIP are indistinguishable under these diagnostics, indicating that interference-driven exponent drift is a generic feature of burst growth rather than a model-specific artifact. Mapping the size-distribution exponent to an equivalent Gutenberg&amp;amp;ndash;Richter b-value shows that increasing interference suppresses large bursts and produces b value ranges comparable to those reported for injection-induced seismicity, supporting the interpretation of interference as a geometric proxy for mechanical inhibition that limits the growth of large events in real fracture networks.</p>
	]]></content:encoded>

	<dc:title>Interference-Driven Scaling Variability in Burst-Based Loopless Invasion Percolation Models of Induced Seismicity</dc:title>
			<dc:creator>Ian Baughman</dc:creator>
			<dc:creator>John B. Rundle</dc:creator>
		<dc:identifier>doi: 10.3390/analytics5010006</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2026-01-06</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2026-01-06</prism:publicationDate>
	<prism:volume>5</prism:volume>
	<prism:number>1</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>6</prism:startingPage>
		<prism:doi>10.3390/analytics5010006</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/5/1/6</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/5/1/5">

	<title>Analytics, Vol. 5, Pages 5: PSYCH&amp;mdash;Psychometric Assessment of Large Language Model Characters: An Exploration of the German Language</title>
	<link>https://www.mdpi.com/2813-2203/5/1/5</link>
	<description>Background: Existing evaluations of large language models (LLMs) largely emphasize linguistic and factual performance, while their psychometric characteristics and behavioral biases remain insufficiently examined, particularly beyond English-language contexts. This study presents a systematic psychometric screening of LLMs in German using the validated Big Five Inventory-2 (BFI-2). Methods: Thirty-two contemporary commercial and open-source LLMs completed all 60 BFI-2 items 60 times each (once with and once without having to justify their answers), yielding over 330,000 responses. Models answered independently, under male and female impersonation, and with and without required justifications. Responses were compared to German human reference data using Welch&amp;amp;rsquo;s t-tests (p&amp;amp;lt;0.01) to assess deviations, response stability, justification effects, and gender differences. Results: At the domain level, LLM personality profiles broadly align with human means. Facet-level analyses, however, reveal systematic deviations, including inflated agreement&amp;amp;mdash;especially in Agreeableness and Aesthetic Sensitivity&amp;amp;mdash;and reduced Negative Emotionality. Only a few models show minimal deviations. Justification prompts significantly altered responses in 56% of models, often increasing variability. Commercial models exhibited substantially higher response stability than open-source models. Gender impersonation affected up to 25% of BFI-2 items, reflecting and occasionally amplifying human gender differences. Conclusions: This study introduces a reproducible psychometric framework for benchmarking LLM behavior against validated human norms and shows that LLMs produce stable yet systematically biased personality-like response patterns. Psychometric screening could therefore complement traditional LLM evaluation in sensitive applications.</description>
	<pubDate>2026-01-06</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 5, Pages 5: PSYCH&amp;mdash;Psychometric Assessment of Large Language Model Characters: An Exploration of the German Language</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/5/1/5">doi: 10.3390/analytics5010005</a></p>
	<p>Authors:
		Nane Kratzke
		Niklas Beuter
		André Drews
		Monique Janneck
		</p>
	<p>Background: Existing evaluations of large language models (LLMs) largely emphasize linguistic and factual performance, while their psychometric characteristics and behavioral biases remain insufficiently examined, particularly beyond English-language contexts. This study presents a systematic psychometric screening of LLMs in German using the validated Big Five Inventory-2 (BFI-2). Methods: Thirty-two contemporary commercial and open-source LLMs completed all 60 BFI-2 items 60 times each (once with and once without having to justify their answers), yielding over 330,000 responses. Models answered independently, under male and female impersonation, and with and without required justifications. Responses were compared to German human reference data using Welch&amp;amp;rsquo;s t-tests (p&amp;amp;lt;0.01) to assess deviations, response stability, justification effects, and gender differences. Results: At the domain level, LLM personality profiles broadly align with human means. Facet-level analyses, however, reveal systematic deviations, including inflated agreement&amp;amp;mdash;especially in Agreeableness and Aesthetic Sensitivity&amp;amp;mdash;and reduced Negative Emotionality. Only a few models show minimal deviations. Justification prompts significantly altered responses in 56% of models, often increasing variability. Commercial models exhibited substantially higher response stability than open-source models. Gender impersonation affected up to 25% of BFI-2 items, reflecting and occasionally amplifying human gender differences. Conclusions: This study introduces a reproducible psychometric framework for benchmarking LLM behavior against validated human norms and shows that LLMs produce stable yet systematically biased personality-like response patterns. Psychometric screening could therefore complement traditional LLM evaluation in sensitive applications.</p>
	]]></content:encoded>

	<dc:title>PSYCH&amp;amp;mdash;Psychometric Assessment of Large Language Model Characters: An Exploration of the German Language</dc:title>
			<dc:creator>Nane Kratzke</dc:creator>
			<dc:creator>Niklas Beuter</dc:creator>
			<dc:creator>André Drews</dc:creator>
			<dc:creator>Monique Janneck</dc:creator>
		<dc:identifier>doi: 10.3390/analytics5010005</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2026-01-06</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2026-01-06</prism:publicationDate>
	<prism:volume>5</prism:volume>
	<prism:number>1</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>5</prism:startingPage>
		<prism:doi>10.3390/analytics5010005</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/5/1/5</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/5/1/4">

	<title>Analytics, Vol. 5, Pages 4: GSM: An Integrated GAM&amp;ndash;SHAP&amp;ndash;MCDA Framework for Stroke Risk Assessment</title>
	<link>https://www.mdpi.com/2813-2203/5/1/4</link>
	<description>This study proposes GSM, an interpretable and operational GAM-SHAP-MCDA framework for stroke risk stratification by integrating generalized additive models (GAMs), a point-based clinical scoring system, SHAP-based explainability, and multi-criteria decision analysis (MCDA). Using a publicly available dataset of n=5110 individuals (4.87% stroke prevalence), a GAM was fitted to capture nonlinear effects of key physiological predictors, including age, average blood glucose level, and body mass index (BMI), together with linear effects for hypertension, heart disease, and categorical covariates. The estimated smooth functions revealed strong age-related risk acceleration beyond 60 years, threshold behavior for glucose levels above approximately 180mg/dL, and a non-monotonic BMI association with peak risk at moderate BMI ranges. In a comparative evaluation, the GAM achieved superior discrimination and calibration relative to classical logistic regression, with a mean AUC of 0.846 versus 0.812 and a lower Brier score (0.045 vs. 0.051). A calibration analysis yielded an intercept of &amp;amp;minus;0.04 and a slope of 1.03, indicating near-ideal agreement between the predicted and observed risks. While high-capacity ensemble models such as XGBoost achieved slightly higher AUC values (0.862), the GAM attained near-upper-bound performance while retaining full interpretability. To enhance clinical usability, the GAM smooth effects were discretized into clinically interpretable bands and converted into an additive point-based risk score ranging from 0 to 42, which was subsequently calibrated to absolute stroke probability. The calibrated probabilities were incorporated into the TOPSIS and VIKOR MCDA frameworks, producing transparent and robust patient prioritization rankings. A SHAP analysis confirmed age, glucose, and cardiometabolic factors as dominant global contributors, aligning with the learned GAM structure. Overall, the proposed GAM&amp;amp;ndash;SHAP&amp;amp;ndash;MCDA framework demonstrates that near-state-of-the-art predictive performance can be achieved alongside transparency, calibration, and decision-oriented interpretability, supporting ethical and practical deployment of medical artificial intelligence for stroke risk assessment.</description>
	<pubDate>2025-12-29</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 5, Pages 4: GSM: An Integrated GAM&amp;ndash;SHAP&amp;ndash;MCDA Framework for Stroke Risk Assessment</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/5/1/4">doi: 10.3390/analytics5010004</a></p>
	<p>Authors:
		Rilwan Mustapha
		Ashiribo Wusu
		Olusola Olabanjo
		Bamidele Adetunji
		</p>
	<p>This study proposes GSM, an interpretable and operational GAM-SHAP-MCDA framework for stroke risk stratification by integrating generalized additive models (GAMs), a point-based clinical scoring system, SHAP-based explainability, and multi-criteria decision analysis (MCDA). Using a publicly available dataset of n=5110 individuals (4.87% stroke prevalence), a GAM was fitted to capture nonlinear effects of key physiological predictors, including age, average blood glucose level, and body mass index (BMI), together with linear effects for hypertension, heart disease, and categorical covariates. The estimated smooth functions revealed strong age-related risk acceleration beyond 60 years, threshold behavior for glucose levels above approximately 180mg/dL, and a non-monotonic BMI association with peak risk at moderate BMI ranges. In a comparative evaluation, the GAM achieved superior discrimination and calibration relative to classical logistic regression, with a mean AUC of 0.846 versus 0.812 and a lower Brier score (0.045 vs. 0.051). A calibration analysis yielded an intercept of &amp;amp;minus;0.04 and a slope of 1.03, indicating near-ideal agreement between the predicted and observed risks. While high-capacity ensemble models such as XGBoost achieved slightly higher AUC values (0.862), the GAM attained near-upper-bound performance while retaining full interpretability. To enhance clinical usability, the GAM smooth effects were discretized into clinically interpretable bands and converted into an additive point-based risk score ranging from 0 to 42, which was subsequently calibrated to absolute stroke probability. The calibrated probabilities were incorporated into the TOPSIS and VIKOR MCDA frameworks, producing transparent and robust patient prioritization rankings. A SHAP analysis confirmed age, glucose, and cardiometabolic factors as dominant global contributors, aligning with the learned GAM structure. Overall, the proposed GAM&amp;amp;ndash;SHAP&amp;amp;ndash;MCDA framework demonstrates that near-state-of-the-art predictive performance can be achieved alongside transparency, calibration, and decision-oriented interpretability, supporting ethical and practical deployment of medical artificial intelligence for stroke risk assessment.</p>
	]]></content:encoded>

	<dc:title>GSM: An Integrated GAM&amp;amp;ndash;SHAP&amp;amp;ndash;MCDA Framework for Stroke Risk Assessment</dc:title>
			<dc:creator>Rilwan Mustapha</dc:creator>
			<dc:creator>Ashiribo Wusu</dc:creator>
			<dc:creator>Olusola Olabanjo</dc:creator>
			<dc:creator>Bamidele Adetunji</dc:creator>
		<dc:identifier>doi: 10.3390/analytics5010004</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2025-12-29</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2025-12-29</prism:publicationDate>
	<prism:volume>5</prism:volume>
	<prism:number>1</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>4</prism:startingPage>
		<prism:doi>10.3390/analytics5010004</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/5/1/4</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/5/1/3">

	<title>Analytics, Vol. 5, Pages 3: Can Length Limit for App Titles Benefit Consumers?</title>
	<link>https://www.mdpi.com/2813-2203/5/1/3</link>
	<description>The App Store introduced a title-length limit for mobile apps in 2016, and similar policies were later adopted across the industry. This issue drew considerable attention from industry practitioners in the 2010s. Using both empirical and theoretical approaches, this paper examines the effectiveness of this policy and its welfare implications. Title length became an issue because some sellers assemble meaningful keywords in the app title to convey information to consumers, while others combine irrelevant yet popular keywords in an attempt to increase their app&amp;amp;rsquo;s downloads. We hypothesize that when titles are short, title length is positively associated with an app&amp;amp;rsquo;s performance because both honest and opportunistic sellers coexist in the market. However, due to the presence of opportunistic sellers, once titles become too long, this positive relationship disappears. We examine this hypothesis using a random sample of 1998 apps from the App Store in 2015. Our results show that for apps with titles longer than 30 characters, title length remains positively associated with app performance. However, for titles exceeding 50 characters, we do not have sufficient evidence to conclude that further increases in length continue to generate additional downloads. To interpret our empirical findings, we construct communication games between an app seller and a consumer, in which the equilibrium is characterized by a threshold. Based on our model and empirical observations, the 30-character limit might hurt consumers.</description>
	<pubDate>2025-12-29</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 5, Pages 3: Can Length Limit for App Titles Benefit Consumers?</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/5/1/3">doi: 10.3390/analytics5010003</a></p>
	<p>Authors:
		Saori Chiba
		Yu-Hsi Liu
		Chien-Yuan Sher
		Min-Hsueh Tsai
		</p>
	<p>The App Store introduced a title-length limit for mobile apps in 2016, and similar policies were later adopted across the industry. This issue drew considerable attention from industry practitioners in the 2010s. Using both empirical and theoretical approaches, this paper examines the effectiveness of this policy and its welfare implications. Title length became an issue because some sellers assemble meaningful keywords in the app title to convey information to consumers, while others combine irrelevant yet popular keywords in an attempt to increase their app&amp;amp;rsquo;s downloads. We hypothesize that when titles are short, title length is positively associated with an app&amp;amp;rsquo;s performance because both honest and opportunistic sellers coexist in the market. However, due to the presence of opportunistic sellers, once titles become too long, this positive relationship disappears. We examine this hypothesis using a random sample of 1998 apps from the App Store in 2015. Our results show that for apps with titles longer than 30 characters, title length remains positively associated with app performance. However, for titles exceeding 50 characters, we do not have sufficient evidence to conclude that further increases in length continue to generate additional downloads. To interpret our empirical findings, we construct communication games between an app seller and a consumer, in which the equilibrium is characterized by a threshold. Based on our model and empirical observations, the 30-character limit might hurt consumers.</p>
	]]></content:encoded>

	<dc:title>Can Length Limit for App Titles Benefit Consumers?</dc:title>
			<dc:creator>Saori Chiba</dc:creator>
			<dc:creator>Yu-Hsi Liu</dc:creator>
			<dc:creator>Chien-Yuan Sher</dc:creator>
			<dc:creator>Min-Hsueh Tsai</dc:creator>
		<dc:identifier>doi: 10.3390/analytics5010003</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2025-12-29</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2025-12-29</prism:publicationDate>
	<prism:volume>5</prism:volume>
	<prism:number>1</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>3</prism:startingPage>
		<prism:doi>10.3390/analytics5010003</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/5/1/3</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/5/1/2">

	<title>Analytics, Vol. 5, Pages 2: A Threshold Selection Method in Code Plagiarism Checking Function for Code Writing Problem in Java Programming Learning Assistant System Considering AI-Generated Codes</title>
	<link>https://www.mdpi.com/2813-2203/5/1/2</link>
	<description>To support novice learners, the Java programming learning assistant system (JPLAS) has been developed with various features. Among them, code writing problem (CWP) assigns writing an answer code that passes a given test code. The correctness of an answer code is validated by running it on JUnit. In previous works, we implemented a code plagiarism checking function that calculates the similarity score for each pair of answer codes based on the Levenshtein distance. When the score is higher than a given threshold, this pair is regarded as plagiarism. However, a method for finding the proper threshold has not been studied. In addition, AI-generated codes have become threats in plagiarism, as AI has grown in popularity, which should be investigated. In this paper, we propose a threshold selection method based on Tukey&amp;amp;rsquo;s IQR fences. It uses a custom upper threshold derived from the statistical distribution of similarity scores for each assignment. To better accommodate skewed similarity distributions, the method introduces a simple percentile-based adjustment for determining the upper threshold. We also design prompts to generate answer codes using generative AI and apply them to four AI models. For evaluation, we used a total of 745 source codes of two datasets. The first dataset consists of 420 answer codes across 12 CWP instances from 35 first-year undergraduate students in the State Polytechnic of Malang, Indonesia (POLINEMA). The second dataset includes 325 answer codes across five CWP assignments from 65 third-year undergraduate students at Okayama University, Japan. The applications of our proposals found the following: (1) any pair of student codes whose score is higher than the selected threshold has some evidence of plagiarism, (2) some student codes have a higher similarity than the threshold with AI-generated codes, indicating the use of generative AI, and (3) multiple AI models can generate code that resembles student-written code, despite adopting different implementations. The validity of our proposal is confirmed.</description>
	<pubDate>2025-12-26</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 5, Pages 2: A Threshold Selection Method in Code Plagiarism Checking Function for Code Writing Problem in Java Programming Learning Assistant System Considering AI-Generated Codes</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/5/1/2">doi: 10.3390/analytics5010002</a></p>
	<p>Authors:
		Perwira Annissa Dyah Permatasari
		Mustika Mentari
		Safira Adine Kinari
		Soe Thandar Aung
		Nobuo Funabiki
		Htoo Htoo Sandi Kyaw
		Khaing Hsu Wai
		</p>
	<p>To support novice learners, the Java programming learning assistant system (JPLAS) has been developed with various features. Among them, code writing problem (CWP) assigns writing an answer code that passes a given test code. The correctness of an answer code is validated by running it on JUnit. In previous works, we implemented a code plagiarism checking function that calculates the similarity score for each pair of answer codes based on the Levenshtein distance. When the score is higher than a given threshold, this pair is regarded as plagiarism. However, a method for finding the proper threshold has not been studied. In addition, AI-generated codes have become threats in plagiarism, as AI has grown in popularity, which should be investigated. In this paper, we propose a threshold selection method based on Tukey&amp;amp;rsquo;s IQR fences. It uses a custom upper threshold derived from the statistical distribution of similarity scores for each assignment. To better accommodate skewed similarity distributions, the method introduces a simple percentile-based adjustment for determining the upper threshold. We also design prompts to generate answer codes using generative AI and apply them to four AI models. For evaluation, we used a total of 745 source codes of two datasets. The first dataset consists of 420 answer codes across 12 CWP instances from 35 first-year undergraduate students in the State Polytechnic of Malang, Indonesia (POLINEMA). The second dataset includes 325 answer codes across five CWP assignments from 65 third-year undergraduate students at Okayama University, Japan. The applications of our proposals found the following: (1) any pair of student codes whose score is higher than the selected threshold has some evidence of plagiarism, (2) some student codes have a higher similarity than the threshold with AI-generated codes, indicating the use of generative AI, and (3) multiple AI models can generate code that resembles student-written code, despite adopting different implementations. The validity of our proposal is confirmed.</p>
	]]></content:encoded>

	<dc:title>A Threshold Selection Method in Code Plagiarism Checking Function for Code Writing Problem in Java Programming Learning Assistant System Considering AI-Generated Codes</dc:title>
			<dc:creator>Perwira Annissa Dyah Permatasari</dc:creator>
			<dc:creator>Mustika Mentari</dc:creator>
			<dc:creator>Safira Adine Kinari</dc:creator>
			<dc:creator>Soe Thandar Aung</dc:creator>
			<dc:creator>Nobuo Funabiki</dc:creator>
			<dc:creator>Htoo Htoo Sandi Kyaw</dc:creator>
			<dc:creator>Khaing Hsu Wai</dc:creator>
		<dc:identifier>doi: 10.3390/analytics5010002</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2025-12-26</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2025-12-26</prism:publicationDate>
	<prism:volume>5</prism:volume>
	<prism:number>1</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>2</prism:startingPage>
		<prism:doi>10.3390/analytics5010002</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/5/1/2</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/5/1/1">

	<title>Analytics, Vol. 5, Pages 1: A Novel Magnificent Frigatebird Optimization Algorithm with Proposed Movement Strategies for Enhanced Global Search</title>
	<link>https://www.mdpi.com/2813-2203/5/1/1</link>
	<description>Global optimization is a fundamental tool for addressing complex and nonlinear problems across scientific and technological domains. The primary objective of this work is to enhance the efficiency, stability, and convergence speed of the Magnificent Frigatebird Optimization (MFO) algorithm by introducing new strategies that strengthen both global exploration and local exploitation. To this end, we propose an improved version of MFO that incorporates three novel movement strategies (aggressive, conservative, and mixed), a BFGS-based local search procedure for more accurate solution refinement, and a dynamic termination criterion capable of detecting stagnation and reducing unnecessary function evaluations. The algorithm is extensively evaluated on a diverse set of benchmark functions, demonstrating substantially lower computational cost and higher reliability compared to classical evolutionary and swarm-based methods. The results confirm the effectiveness of the proposed modifications and highlight the potential of the enhanced MFO for application to demanding real-world optimization problems.</description>
	<pubDate>2025-12-23</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 5, Pages 1: A Novel Magnificent Frigatebird Optimization Algorithm with Proposed Movement Strategies for Enhanced Global Search</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/5/1/1">doi: 10.3390/analytics5010001</a></p>
	<p>Authors:
		Glykeria Kyrou
		Vasileios Charilogis
		Ioannis G. Tsoulos
		</p>
	<p>Global optimization is a fundamental tool for addressing complex and nonlinear problems across scientific and technological domains. The primary objective of this work is to enhance the efficiency, stability, and convergence speed of the Magnificent Frigatebird Optimization (MFO) algorithm by introducing new strategies that strengthen both global exploration and local exploitation. To this end, we propose an improved version of MFO that incorporates three novel movement strategies (aggressive, conservative, and mixed), a BFGS-based local search procedure for more accurate solution refinement, and a dynamic termination criterion capable of detecting stagnation and reducing unnecessary function evaluations. The algorithm is extensively evaluated on a diverse set of benchmark functions, demonstrating substantially lower computational cost and higher reliability compared to classical evolutionary and swarm-based methods. The results confirm the effectiveness of the proposed modifications and highlight the potential of the enhanced MFO for application to demanding real-world optimization problems.</p>
	]]></content:encoded>

	<dc:title>A Novel Magnificent Frigatebird Optimization Algorithm with Proposed Movement Strategies for Enhanced Global Search</dc:title>
			<dc:creator>Glykeria Kyrou</dc:creator>
			<dc:creator>Vasileios Charilogis</dc:creator>
			<dc:creator>Ioannis G. Tsoulos</dc:creator>
		<dc:identifier>doi: 10.3390/analytics5010001</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2025-12-23</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2025-12-23</prism:publicationDate>
	<prism:volume>5</prism:volume>
	<prism:number>1</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>1</prism:startingPage>
		<prism:doi>10.3390/analytics5010001</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/5/1/1</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/4/4/36">

	<title>Analytics, Vol. 4, Pages 36: Assessing the Impact of Capital Expenditure on Corporate Profitability in South Korea&amp;rsquo;s Electronics Industry: A Regression Analysis Approach</title>
	<link>https://www.mdpi.com/2813-2203/4/4/36</link>
	<description>This study investigates the relationship between capital expenditure (CAPEX) and long-term corporate profitability in South Korea&amp;amp;rsquo;s electronics industry. Using panel data from 126 listed electronics firms covering 2005&amp;amp;ndash;2019, the research applies fixed-effects regression analysis to examine how CAPEX influences profitability, measured by EBITDA/total assets. The results confirm that CAPEX exerts a positive and statistically significant long-term effect on profitability, with stronger but not significantly different impacts for large firms compared to SMEs. The findings contribute to empirical evidence on capital investment efficiency and the implications of economies and diseconomies of scale in capital-intensive industries.</description>
	<pubDate>2025-12-10</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 4, Pages 36: Assessing the Impact of Capital Expenditure on Corporate Profitability in South Korea&amp;rsquo;s Electronics Industry: A Regression Analysis Approach</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/4/4/36">doi: 10.3390/analytics4040036</a></p>
	<p>Authors:
		Bomee Park
		Tetiana Paientko
		</p>
	<p>This study investigates the relationship between capital expenditure (CAPEX) and long-term corporate profitability in South Korea&amp;amp;rsquo;s electronics industry. Using panel data from 126 listed electronics firms covering 2005&amp;amp;ndash;2019, the research applies fixed-effects regression analysis to examine how CAPEX influences profitability, measured by EBITDA/total assets. The results confirm that CAPEX exerts a positive and statistically significant long-term effect on profitability, with stronger but not significantly different impacts for large firms compared to SMEs. The findings contribute to empirical evidence on capital investment efficiency and the implications of economies and diseconomies of scale in capital-intensive industries.</p>
	]]></content:encoded>

	<dc:title>Assessing the Impact of Capital Expenditure on Corporate Profitability in South Korea&amp;amp;rsquo;s Electronics Industry: A Regression Analysis Approach</dc:title>
			<dc:creator>Bomee Park</dc:creator>
			<dc:creator>Tetiana Paientko</dc:creator>
		<dc:identifier>doi: 10.3390/analytics4040036</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2025-12-10</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2025-12-10</prism:publicationDate>
	<prism:volume>4</prism:volume>
	<prism:number>4</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>36</prism:startingPage>
		<prism:doi>10.3390/analytics4040036</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/4/4/36</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/4/4/35">

	<title>Analytics, Vol. 4, Pages 35: Option Pricing in the Approach of Integrating Market Risk Premium: Application to OTM Options</title>
	<link>https://www.mdpi.com/2813-2203/4/4/35</link>
	<description>In this research, we summarize the results of implementing the market risk premium into the option valuation formulas of the Black&amp;amp;ndash;Scholes&amp;amp;ndash;Merton model for out-of-the-money (OTM) options. We show that derivative prices can partly depend on systematic market risk, which the BSM model ignores by construction. Specifically, empirical studies are conducted using 50ETF options obtained from the Shanghai Stock Exchange, covering the periods from January 2018 to September 2022 and from December 2023 to October 2025. The pricing of the OTM options shows that the adjusted BSM formulas exhibit better pricing performance compared with the market prices of the OTM options tested. Furthermore, a framework for the empirical analysis of option prices based on the Capital Asset Pricing Model (CAPM) or factor models is discussed, which may lead to option formulas using non-homogeneous heat equations. The later proposal requires further statistical testing using real market data but offers an alternative to the existing risk-neutral valuation of options.</description>
	<pubDate>2025-11-21</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 4, Pages 35: Option Pricing in the Approach of Integrating Market Risk Premium: Application to OTM Options</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/4/4/35">doi: 10.3390/analytics4040035</a></p>
	<p>Authors:
		David Liu
		</p>
	<p>In this research, we summarize the results of implementing the market risk premium into the option valuation formulas of the Black&amp;amp;ndash;Scholes&amp;amp;ndash;Merton model for out-of-the-money (OTM) options. We show that derivative prices can partly depend on systematic market risk, which the BSM model ignores by construction. Specifically, empirical studies are conducted using 50ETF options obtained from the Shanghai Stock Exchange, covering the periods from January 2018 to September 2022 and from December 2023 to October 2025. The pricing of the OTM options shows that the adjusted BSM formulas exhibit better pricing performance compared with the market prices of the OTM options tested. Furthermore, a framework for the empirical analysis of option prices based on the Capital Asset Pricing Model (CAPM) or factor models is discussed, which may lead to option formulas using non-homogeneous heat equations. The later proposal requires further statistical testing using real market data but offers an alternative to the existing risk-neutral valuation of options.</p>
	]]></content:encoded>

	<dc:title>Option Pricing in the Approach of Integrating Market Risk Premium: Application to OTM Options</dc:title>
			<dc:creator>David Liu</dc:creator>
		<dc:identifier>doi: 10.3390/analytics4040035</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2025-11-21</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2025-11-21</prism:publicationDate>
	<prism:volume>4</prism:volume>
	<prism:number>4</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>35</prism:startingPage>
		<prism:doi>10.3390/analytics4040035</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/4/4/35</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/4/4/34">

	<title>Analytics, Vol. 4, Pages 34: Fan Loyalty and Price Elasticity in Sport: Insights from Major League Baseball&amp;rsquo;s Post-Pandemic Recovery</title>
	<link>https://www.mdpi.com/2813-2203/4/4/34</link>
	<description>The COVID-19 pandemic disrupted traditional patterns of sport consumption, raising questions about whether fans would return to stadiums and how sensitive they would be to ticket prices in the recovery period. This study reconceptualizes ticket price elasticity as a market-based indicator of fan loyalty and applies it to Major League Baseball (MLB) during 2021&amp;amp;ndash;2023. Using team&amp;amp;ndash;season attendance data from Baseball-Reference, primary-market ticket prices from the Team Marketing Report Fan Cost Index, and secondary-market prices from TicketIQ, we estimate log&amp;amp;ndash;log fixed-effects panel models to separate causal price responses from popularity-driven correlations. The results show a strongly negative elasticity of attendance with respect to primary-market prices (&amp;amp;beta; &amp;amp;asymp; &amp;amp;minus;7.93, p &amp;amp;lt; 0.001), indicating that higher ticket prices substantially reduce attendance, while secondary-market prices are positively associated with attendance, reflecting demand shocks rather than causal effects. Heterogeneity analyses reveal that brand strength, team performance, and game salience significantly moderate elasticity, supporting the interpretation of inelastic demand as revealed loyalty. These findings highlight the potential of elasticity as a Fan Loyalty Index, providing a replicable framework for measuring consumer resilience. The study offers practical insights for pricing strategy, fan segmentation, and engagement, while emphasizing the broader social role of sport in restoring community identity during post-pandemic recovery.</description>
	<pubDate>2025-11-21</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 4, Pages 34: Fan Loyalty and Price Elasticity in Sport: Insights from Major League Baseball&amp;rsquo;s Post-Pandemic Recovery</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/4/4/34">doi: 10.3390/analytics4040034</a></p>
	<p>Authors:
		Soojin Choi
		Fang Zheng
		Seung-Man Lee
		</p>
	<p>The COVID-19 pandemic disrupted traditional patterns of sport consumption, raising questions about whether fans would return to stadiums and how sensitive they would be to ticket prices in the recovery period. This study reconceptualizes ticket price elasticity as a market-based indicator of fan loyalty and applies it to Major League Baseball (MLB) during 2021&amp;amp;ndash;2023. Using team&amp;amp;ndash;season attendance data from Baseball-Reference, primary-market ticket prices from the Team Marketing Report Fan Cost Index, and secondary-market prices from TicketIQ, we estimate log&amp;amp;ndash;log fixed-effects panel models to separate causal price responses from popularity-driven correlations. The results show a strongly negative elasticity of attendance with respect to primary-market prices (&amp;amp;beta; &amp;amp;asymp; &amp;amp;minus;7.93, p &amp;amp;lt; 0.001), indicating that higher ticket prices substantially reduce attendance, while secondary-market prices are positively associated with attendance, reflecting demand shocks rather than causal effects. Heterogeneity analyses reveal that brand strength, team performance, and game salience significantly moderate elasticity, supporting the interpretation of inelastic demand as revealed loyalty. These findings highlight the potential of elasticity as a Fan Loyalty Index, providing a replicable framework for measuring consumer resilience. The study offers practical insights for pricing strategy, fan segmentation, and engagement, while emphasizing the broader social role of sport in restoring community identity during post-pandemic recovery.</p>
	]]></content:encoded>

	<dc:title>Fan Loyalty and Price Elasticity in Sport: Insights from Major League Baseball&amp;amp;rsquo;s Post-Pandemic Recovery</dc:title>
			<dc:creator>Soojin Choi</dc:creator>
			<dc:creator>Fang Zheng</dc:creator>
			<dc:creator>Seung-Man Lee</dc:creator>
		<dc:identifier>doi: 10.3390/analytics4040034</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2025-11-21</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2025-11-21</prism:publicationDate>
	<prism:volume>4</prism:volume>
	<prism:number>4</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>34</prism:startingPage>
		<prism:doi>10.3390/analytics4040034</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/4/4/34</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/4/4/33">

	<title>Analytics, Vol. 4, Pages 33: AI-Powered Chatbot for FDA Drug Labeling Information Retrieval: OpenAI GPT for Grounded Question Answering</title>
	<link>https://www.mdpi.com/2813-2203/4/4/33</link>
	<description>This study presents the development of an AI-powered chatbot designed to facilitate accurate and efficient retrieval of information from the FDA drug labeling documents. Leveraging OpenAI&amp;amp;rsquo;s GPT-3.5-turbo model within a controlled, document-grounded question&amp;amp;ndash;answering framework, Chatbot was created, which can provide users with answers that are strictly limited to the content of the uploaded drug label, thereby minimizing hallucinations and enhancing traceability. A user-friendly interface built with Streamlit allows users to upload FDA labeling PDFs and pose natural language queries. The chatbot extracts relevant sections using PyMuPDF and regex-based segmentation and generates responses constrained to those sections. To evaluate performance, semantic similarity scores were computed between generated answers and ground truth text using Sentence Transformers. Results across 10 breast cancer drug labels demonstrate high semantic alignment, with most scores ranging from 0.7 to 0.9, indicating reliable summarization and contextual fidelity. The chatbot achieved high semantic similarity scores (&amp;amp;ge;0.95 for concise sections) and ROUGE scores, confirming strong semantic and textual alignment. Comparative analysis with GPT-5-chat and NotebookLM demonstrated that our approach maintains accuracy and section-specific fidelity across models. The current work is limited to a small dataset, focused on breast cancer drugs. Future work will expand to diverse therapeutic areas and incorporate BERTScore and expert-based validation.</description>
	<pubDate>2025-11-17</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 4, Pages 33: AI-Powered Chatbot for FDA Drug Labeling Information Retrieval: OpenAI GPT for Grounded Question Answering</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/4/4/33">doi: 10.3390/analytics4040033</a></p>
	<p>Authors:
		Manasa Koppula
		Fnu Madhulika
		Navya Sreeramoju
		Praveen Kolimi
		</p>
	<p>This study presents the development of an AI-powered chatbot designed to facilitate accurate and efficient retrieval of information from the FDA drug labeling documents. Leveraging OpenAI&amp;amp;rsquo;s GPT-3.5-turbo model within a controlled, document-grounded question&amp;amp;ndash;answering framework, Chatbot was created, which can provide users with answers that are strictly limited to the content of the uploaded drug label, thereby minimizing hallucinations and enhancing traceability. A user-friendly interface built with Streamlit allows users to upload FDA labeling PDFs and pose natural language queries. The chatbot extracts relevant sections using PyMuPDF and regex-based segmentation and generates responses constrained to those sections. To evaluate performance, semantic similarity scores were computed between generated answers and ground truth text using Sentence Transformers. Results across 10 breast cancer drug labels demonstrate high semantic alignment, with most scores ranging from 0.7 to 0.9, indicating reliable summarization and contextual fidelity. The chatbot achieved high semantic similarity scores (&amp;amp;ge;0.95 for concise sections) and ROUGE scores, confirming strong semantic and textual alignment. Comparative analysis with GPT-5-chat and NotebookLM demonstrated that our approach maintains accuracy and section-specific fidelity across models. The current work is limited to a small dataset, focused on breast cancer drugs. Future work will expand to diverse therapeutic areas and incorporate BERTScore and expert-based validation.</p>
	]]></content:encoded>

	<dc:title>AI-Powered Chatbot for FDA Drug Labeling Information Retrieval: OpenAI GPT for Grounded Question Answering</dc:title>
			<dc:creator>Manasa Koppula</dc:creator>
			<dc:creator>Fnu Madhulika</dc:creator>
			<dc:creator>Navya Sreeramoju</dc:creator>
			<dc:creator>Praveen Kolimi</dc:creator>
		<dc:identifier>doi: 10.3390/analytics4040033</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2025-11-17</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2025-11-17</prism:publicationDate>
	<prism:volume>4</prism:volume>
	<prism:number>4</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>33</prism:startingPage>
		<prism:doi>10.3390/analytics4040033</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/4/4/33</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/4/4/32">

	<title>Analytics, Vol. 4, Pages 32: Scale-Invariant Correspondence Analysis of Compositional Data</title>
	<link>https://www.mdpi.com/2813-2203/4/4/32</link>
	<description>Correspondence analysis is a dimension reduction technique for visualizing a non-negative matrix N=(nij) of size I&amp;amp;times;J, particularly contingency tables or compositional datasets, but it depends on the row and column marginals of N. Three complementary transformations of the data T(N)=(T(ainijbj)) render CA of T(N) invariant for any ai&amp;amp;gt;0 and bj&amp;amp;gt;0: first, Greenacre&amp;amp;rsquo;s scale-invariant approach, valid for positive data; second, Goodman&amp;amp;rsquo;s marginal-free correspondence analysis, valid for positive or moderately sparse data; third, correspondence analysis of the sign-transformed matrix, sign(N)=(sign(nij)), valid for sparse or extremely sparse data. We demonstrate these three methods on four real-world datasets with varying levels of sparsity to compare their exploratory performance.</description>
	<pubDate>2025-11-12</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 4, Pages 32: Scale-Invariant Correspondence Analysis of Compositional Data</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/4/4/32">doi: 10.3390/analytics4040032</a></p>
	<p>Authors:
		Vartan Choulakian
		Jacques Allard
		</p>
	<p>Correspondence analysis is a dimension reduction technique for visualizing a non-negative matrix N=(nij) of size I&amp;amp;times;J, particularly contingency tables or compositional datasets, but it depends on the row and column marginals of N. Three complementary transformations of the data T(N)=(T(ainijbj)) render CA of T(N) invariant for any ai&amp;amp;gt;0 and bj&amp;amp;gt;0: first, Greenacre&amp;amp;rsquo;s scale-invariant approach, valid for positive data; second, Goodman&amp;amp;rsquo;s marginal-free correspondence analysis, valid for positive or moderately sparse data; third, correspondence analysis of the sign-transformed matrix, sign(N)=(sign(nij)), valid for sparse or extremely sparse data. We demonstrate these three methods on four real-world datasets with varying levels of sparsity to compare their exploratory performance.</p>
	]]></content:encoded>

	<dc:title>Scale-Invariant Correspondence Analysis of Compositional Data</dc:title>
			<dc:creator>Vartan Choulakian</dc:creator>
			<dc:creator>Jacques Allard</dc:creator>
		<dc:identifier>doi: 10.3390/analytics4040032</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2025-11-12</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2025-11-12</prism:publicationDate>
	<prism:volume>4</prism:volume>
	<prism:number>4</prism:number>
	<prism:section>Review</prism:section>
	<prism:startingPage>32</prism:startingPage>
		<prism:doi>10.3390/analytics4040032</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/4/4/32</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/4/4/31">

	<title>Analytics, Vol. 4, Pages 31: PlayMyData: A Statistical Analysis of a Video Game Dataset on Review Scores and Gaming Platforms</title>
	<link>https://www.mdpi.com/2813-2203/4/4/31</link>
	<description>In recent years, video games have become an increasingly popular form of entertainment and enjoyment for consumers of all ages. Given their rapid rise in production, projects such as PlayMyData aim to organize the immense amounts of data that accompany these games into sets of data for public use in research, primarily games bound specifically to modern platforms that are still being actively developed or further improved. This study aims to examine the particular differences in video game review scores using this set of data across the four listed platforms&amp;amp;mdash;Nintendo, Xbox, PlayStation, and PC&amp;amp;mdash;for different gaming titles relating to each platform. Through analysis of variance (ANOVA) testing and several other statistical analyses, significant differences between the platforms were observed, with PC games receiving the highest amount of positive scores and consistently outperforming the other three platforms, Xbox and PlayStation trailing behind PC, and Nintendo receiving the lowest review scores overall. These results illustrate the influence of platforms and their differences on player ratings and provide insight for developers and market analysts seeking to develop and invest in console platform video games.</description>
	<pubDate>2025-11-11</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 4, Pages 31: PlayMyData: A Statistical Analysis of a Video Game Dataset on Review Scores and Gaming Platforms</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/4/4/31">doi: 10.3390/analytics4040031</a></p>
	<p>Authors:
		Christian Ellington
		Paramahansa Pramanik
		Haley K. Robinson
		</p>
	<p>In recent years, video games have become an increasingly popular form of entertainment and enjoyment for consumers of all ages. Given their rapid rise in production, projects such as PlayMyData aim to organize the immense amounts of data that accompany these games into sets of data for public use in research, primarily games bound specifically to modern platforms that are still being actively developed or further improved. This study aims to examine the particular differences in video game review scores using this set of data across the four listed platforms&amp;amp;mdash;Nintendo, Xbox, PlayStation, and PC&amp;amp;mdash;for different gaming titles relating to each platform. Through analysis of variance (ANOVA) testing and several other statistical analyses, significant differences between the platforms were observed, with PC games receiving the highest amount of positive scores and consistently outperforming the other three platforms, Xbox and PlayStation trailing behind PC, and Nintendo receiving the lowest review scores overall. These results illustrate the influence of platforms and their differences on player ratings and provide insight for developers and market analysts seeking to develop and invest in console platform video games.</p>
	]]></content:encoded>

	<dc:title>PlayMyData: A Statistical Analysis of a Video Game Dataset on Review Scores and Gaming Platforms</dc:title>
			<dc:creator>Christian Ellington</dc:creator>
			<dc:creator>Paramahansa Pramanik</dc:creator>
			<dc:creator>Haley K. Robinson</dc:creator>
		<dc:identifier>doi: 10.3390/analytics4040031</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2025-11-11</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2025-11-11</prism:publicationDate>
	<prism:volume>4</prism:volume>
	<prism:number>4</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>31</prism:startingPage>
		<prism:doi>10.3390/analytics4040031</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/4/4/31</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/4/4/30">

	<title>Analytics, Vol. 4, Pages 30: System Inertia Cost Forecasting Using Machine Learning: A Data-Driven Approach for Grid Energy Trading in Great Britain</title>
	<link>https://www.mdpi.com/2813-2203/4/4/30</link>
	<description>As modern power systems integrate more renewable and decentralised generation, maintaining grid stability has become increasingly challenging. This study proposes a data-driven machine learning framework for forecasting system inertia service costs&amp;amp;mdash;a key yet underexplored variable influencing energy trading and frequency stability in Great Britain. Using eight years (2017&amp;amp;ndash;2024) of National Energy System Operator (NESO) data, four models&amp;amp;mdash;Long Short-Term Memory (LSTM), Residual LSTM, eXtreme Gradient Boosting (XGBoost), and Light Gradient-Boosting Machine (LightGBM)&amp;amp;mdash;are comparatively analysed. LSTM-based models capture temporal dependencies, while ensemble methods effectively handle nonlinear feature relationships. Results demonstrate that LightGBM achieves the highest predictive accuracy, offering a robust method for inertia cost estimation and market intelligence. The framework contributes to strategic procurement planning and supports market design for a more resilient, cost-effective grid.</description>
	<pubDate>2025-10-23</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 4, Pages 30: System Inertia Cost Forecasting Using Machine Learning: A Data-Driven Approach for Grid Energy Trading in Great Britain</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/4/4/30">doi: 10.3390/analytics4040030</a></p>
	<p>Authors:
		Maitreyee Dey
		Soumya Prakash Rana
		Preeti Patel
		</p>
	<p>As modern power systems integrate more renewable and decentralised generation, maintaining grid stability has become increasingly challenging. This study proposes a data-driven machine learning framework for forecasting system inertia service costs&amp;amp;mdash;a key yet underexplored variable influencing energy trading and frequency stability in Great Britain. Using eight years (2017&amp;amp;ndash;2024) of National Energy System Operator (NESO) data, four models&amp;amp;mdash;Long Short-Term Memory (LSTM), Residual LSTM, eXtreme Gradient Boosting (XGBoost), and Light Gradient-Boosting Machine (LightGBM)&amp;amp;mdash;are comparatively analysed. LSTM-based models capture temporal dependencies, while ensemble methods effectively handle nonlinear feature relationships. Results demonstrate that LightGBM achieves the highest predictive accuracy, offering a robust method for inertia cost estimation and market intelligence. The framework contributes to strategic procurement planning and supports market design for a more resilient, cost-effective grid.</p>
	]]></content:encoded>

	<dc:title>System Inertia Cost Forecasting Using Machine Learning: A Data-Driven Approach for Grid Energy Trading in Great Britain</dc:title>
			<dc:creator>Maitreyee Dey</dc:creator>
			<dc:creator>Soumya Prakash Rana</dc:creator>
			<dc:creator>Preeti Patel</dc:creator>
		<dc:identifier>doi: 10.3390/analytics4040030</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2025-10-23</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2025-10-23</prism:publicationDate>
	<prism:volume>4</prism:volume>
	<prism:number>4</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>30</prism:startingPage>
		<prism:doi>10.3390/analytics4040030</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/4/4/30</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/4/4/29">

	<title>Analytics, Vol. 4, Pages 29: Distributional CNN-LSTM, KDE, and Copula Approaches for Multimodal Multivariate Data: Assessing Conditional Treatment Effects</title>
	<link>https://www.mdpi.com/2813-2203/4/4/29</link>
	<description>We introduce a distributional CNN-LSTM framework for probabilistic multivariate modeling and heterogeneous treatment effect (HTE) estimation. The model jointly captures complex dependencies among multiple outcomes and enables precise estimation of individual-level conditional average treatment effects (CATEs). In simulation studies with multivariate Gaussian mixtures, the CNN-LSTM demonstrates robust density estimation and strong CATE recovery, particularly as mixture complexity increases, while classical methods such as Kernel Density Estimation (KDE) and Gaussian Copulas may achieve higher log-likelihood or coverage in simpler scenarios. On real-world datasets, including Iris and Criteo Uplift, the CNN-LSTM achieves the lowest CATE RMSE, confirming its practical utility for individualized prediction, although KDE and Gaussian Copula approaches may perform better on global likelihood or coverage metrics. These results indicate that the CNN-LSTM can be trained efficiently on moderate-sized datasets while maintaining stable predictive performance. Overall, the framework is particularly valuable in applications requiring accurate individual-level effect estimation and handling of multimodal heterogeneity&amp;amp;mdash;such as personalized medicine, economic policy evaluation, and environmental risk assessment&amp;amp;mdash;with its primary strength being superior CATE recovery under complex outcome distributions, even when likelihood-based metrics favor simpler baselines.</description>
	<pubDate>2025-10-21</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 4, Pages 29: Distributional CNN-LSTM, KDE, and Copula Approaches for Multimodal Multivariate Data: Assessing Conditional Treatment Effects</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/4/4/29">doi: 10.3390/analytics4040029</a></p>
	<p>Authors:
		Jong-Min Kim
		</p>
	<p>We introduce a distributional CNN-LSTM framework for probabilistic multivariate modeling and heterogeneous treatment effect (HTE) estimation. The model jointly captures complex dependencies among multiple outcomes and enables precise estimation of individual-level conditional average treatment effects (CATEs). In simulation studies with multivariate Gaussian mixtures, the CNN-LSTM demonstrates robust density estimation and strong CATE recovery, particularly as mixture complexity increases, while classical methods such as Kernel Density Estimation (KDE) and Gaussian Copulas may achieve higher log-likelihood or coverage in simpler scenarios. On real-world datasets, including Iris and Criteo Uplift, the CNN-LSTM achieves the lowest CATE RMSE, confirming its practical utility for individualized prediction, although KDE and Gaussian Copula approaches may perform better on global likelihood or coverage metrics. These results indicate that the CNN-LSTM can be trained efficiently on moderate-sized datasets while maintaining stable predictive performance. Overall, the framework is particularly valuable in applications requiring accurate individual-level effect estimation and handling of multimodal heterogeneity&amp;amp;mdash;such as personalized medicine, economic policy evaluation, and environmental risk assessment&amp;amp;mdash;with its primary strength being superior CATE recovery under complex outcome distributions, even when likelihood-based metrics favor simpler baselines.</p>
	]]></content:encoded>

	<dc:title>Distributional CNN-LSTM, KDE, and Copula Approaches for Multimodal Multivariate Data: Assessing Conditional Treatment Effects</dc:title>
			<dc:creator>Jong-Min Kim</dc:creator>
		<dc:identifier>doi: 10.3390/analytics4040029</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2025-10-21</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2025-10-21</prism:publicationDate>
	<prism:volume>4</prism:volume>
	<prism:number>4</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>29</prism:startingPage>
		<prism:doi>10.3390/analytics4040029</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/4/4/29</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/4/4/28">

	<title>Analytics, Vol. 4, Pages 28: Reservoir Computation with Networks of Differentiating Neuron Ring Oscillators</title>
	<link>https://www.mdpi.com/2813-2203/4/4/28</link>
	<description>Reservoir computing is an approach to machine learning that leverages the dynamics of a complex system alongside a simple, often linear, machine learning model for a designated task. While many efforts have previously focused their attention on integrating neurons, which produce an output in response to large, sustained inputs, we focus on using differentiating neurons, which produce an output in response to large changes in input. Here, we introduce a small-world graph built from rings of differentiating neurons as a Reservoir Computing substrate. We find the coupling strength and network topology that enable these small-world networks to function as an effective reservoir. The dynamics of differentiating neurons naturally give rise to oscillatory dynamics when arranged in rings, where we study their computational use in the Reservoir Computing setting. We demonstrate the efficacy of these networks in the MNIST digit recognition task, achieving comparable performance of 90.65% to existing Reservoir Computing approaches. Beyond accuracy, we conduct systematic analysis of our reservoir&amp;amp;rsquo;s internal dynamics using three complementary complexity measures that quantify neuronal activity balance, input dependence, and effective dimensionality. Our analysis reveals that optimal performance emerges when the reservoir operates with intermediate levels of neural entropy and input sensitivity, consistent with the edge-of-chaos hypothesis, where the system balances stability and responsiveness. The findings suggest that differentiating neurons can be a potential alternative to integrating neurons and can provide a sustainable future alternative for power-hungry AI applications.</description>
	<pubDate>2025-10-20</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 4, Pages 28: Reservoir Computation with Networks of Differentiating Neuron Ring Oscillators</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/4/4/28">doi: 10.3390/analytics4040028</a></p>
	<p>Authors:
		Alexander Yeung
		Peter DelMastro
		Arjun Karuvally
		Hava Siegelmann
		Edward Rietman
		Hananel Hazan
		</p>
	<p>Reservoir computing is an approach to machine learning that leverages the dynamics of a complex system alongside a simple, often linear, machine learning model for a designated task. While many efforts have previously focused their attention on integrating neurons, which produce an output in response to large, sustained inputs, we focus on using differentiating neurons, which produce an output in response to large changes in input. Here, we introduce a small-world graph built from rings of differentiating neurons as a Reservoir Computing substrate. We find the coupling strength and network topology that enable these small-world networks to function as an effective reservoir. The dynamics of differentiating neurons naturally give rise to oscillatory dynamics when arranged in rings, where we study their computational use in the Reservoir Computing setting. We demonstrate the efficacy of these networks in the MNIST digit recognition task, achieving comparable performance of 90.65% to existing Reservoir Computing approaches. Beyond accuracy, we conduct systematic analysis of our reservoir&amp;amp;rsquo;s internal dynamics using three complementary complexity measures that quantify neuronal activity balance, input dependence, and effective dimensionality. Our analysis reveals that optimal performance emerges when the reservoir operates with intermediate levels of neural entropy and input sensitivity, consistent with the edge-of-chaos hypothesis, where the system balances stability and responsiveness. The findings suggest that differentiating neurons can be a potential alternative to integrating neurons and can provide a sustainable future alternative for power-hungry AI applications.</p>
	]]></content:encoded>

	<dc:title>Reservoir Computation with Networks of Differentiating Neuron Ring Oscillators</dc:title>
			<dc:creator>Alexander Yeung</dc:creator>
			<dc:creator>Peter DelMastro</dc:creator>
			<dc:creator>Arjun Karuvally</dc:creator>
			<dc:creator>Hava Siegelmann</dc:creator>
			<dc:creator>Edward Rietman</dc:creator>
			<dc:creator>Hananel Hazan</dc:creator>
		<dc:identifier>doi: 10.3390/analytics4040028</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2025-10-20</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2025-10-20</prism:publicationDate>
	<prism:volume>4</prism:volume>
	<prism:number>4</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>28</prism:startingPage>
		<prism:doi>10.3390/analytics4040028</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/4/4/28</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/4/4/27">

	<title>Analytics, Vol. 4, Pages 27: Multiplicative Decomposition Model to Predict UK&amp;rsquo;s Long-Term Electricity Demand with Monthly and Hourly Resolution</title>
	<link>https://www.mdpi.com/2813-2203/4/4/27</link>
	<description>The UK electricity market is changing to adapt to Net Zero targets and respond to disruptions like the Russia&amp;amp;ndash;Ukraine war. This requires strategic planning to decide on the construction of new electricity generation plants for a resilient UK electricity grid. Such planning is based on forecasting the UK electricity demand long-term (from 1 year and beyond). In this paper, we propose a long-term predictive model by identifying the main components of the UK electricity demand, modelling each of these components, and combining them in a multiplicative manner to deliver a single long-term prediction. To the best of our knowledge, this study is the first to apply a multiplicative decomposition model for long-term predictions at both monthly and hourly resolutions, combining neural networks with Fourier analysis. This approach is extremely flexible and accurate, with a mean absolute percentage error of 4.16% and 8.62% in predicting the monthly and hourly electricity demand, respectively, from 2019 to 2021.</description>
	<pubDate>2025-10-06</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 4, Pages 27: Multiplicative Decomposition Model to Predict UK&amp;rsquo;s Long-Term Electricity Demand with Monthly and Hourly Resolution</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/4/4/27">doi: 10.3390/analytics4040027</a></p>
	<p>Authors:
		Marie Baillon
		María Carmen Romano
		Ekkehard Ullner
		</p>
	<p>The UK electricity market is changing to adapt to Net Zero targets and respond to disruptions like the Russia&amp;amp;ndash;Ukraine war. This requires strategic planning to decide on the construction of new electricity generation plants for a resilient UK electricity grid. Such planning is based on forecasting the UK electricity demand long-term (from 1 year and beyond). In this paper, we propose a long-term predictive model by identifying the main components of the UK electricity demand, modelling each of these components, and combining them in a multiplicative manner to deliver a single long-term prediction. To the best of our knowledge, this study is the first to apply a multiplicative decomposition model for long-term predictions at both monthly and hourly resolutions, combining neural networks with Fourier analysis. This approach is extremely flexible and accurate, with a mean absolute percentage error of 4.16% and 8.62% in predicting the monthly and hourly electricity demand, respectively, from 2019 to 2021.</p>
	]]></content:encoded>

	<dc:title>Multiplicative Decomposition Model to Predict UK&amp;amp;rsquo;s Long-Term Electricity Demand with Monthly and Hourly Resolution</dc:title>
			<dc:creator>Marie Baillon</dc:creator>
			<dc:creator>María Carmen Romano</dc:creator>
			<dc:creator>Ekkehard Ullner</dc:creator>
		<dc:identifier>doi: 10.3390/analytics4040027</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2025-10-06</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2025-10-06</prism:publicationDate>
	<prism:volume>4</prism:volume>
	<prism:number>4</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>27</prism:startingPage>
		<prism:doi>10.3390/analytics4040027</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/4/4/27</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/4/4/26">

	<title>Analytics, Vol. 4, Pages 26: Fairness in Predictive Marketing: Auditing and Mitigating Demographic Bias in Machine Learning for Customer Targeting</title>
	<link>https://www.mdpi.com/2813-2203/4/4/26</link>
	<description>As organizations increasingly turn to machine learning for customer segmentation and targeted marketing, concerns about fairness and algorithmic bias have become more urgent. This study presents a comprehensive fairness audit and mitigation framework for predictive marketing models using the Bank Marketing dataset. We train logistic regression and random forest classifiers to predict customer subscription behavior and evaluate their performance across key demographic groups, including age, education, and job type. Using model explainability techniques such as SHAP and fairness metrics including disparate impact and true positive rate parity, we uncover notable disparities in model behavior that could result in discriminatory targeting. We implement three mitigation strategies&amp;amp;mdash;reweighing, threshold adjustment, and feature exclusion&amp;amp;mdash;and assess their effectiveness in improving fairness while preserving business-relevant performance metrics. Among these, reweighing produced the most balanced outcome, raising the Disparate Impact Ratio for older individuals from 0.65 to 0.82 and reducing the true positive rate parity gap by over 40%, with only a modest decline in precision (from 0.78 to 0.76). We propose a replicable workflow for embedding fairness auditing into enterprise BI systems and highlight the strategic importance of ethical AI practices in building accountable and inclusive marketing technologies. technologies.</description>
	<pubDate>2025-10-01</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 4, Pages 26: Fairness in Predictive Marketing: Auditing and Mitigating Demographic Bias in Machine Learning for Customer Targeting</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/4/4/26">doi: 10.3390/analytics4040026</a></p>
	<p>Authors:
		Sayee Phaneendhar Pasupuleti
		Jagadeesh Kola
		Sai Phaneendra Manikantesh Kodete
		Sree Harsha Palli
		</p>
	<p>As organizations increasingly turn to machine learning for customer segmentation and targeted marketing, concerns about fairness and algorithmic bias have become more urgent. This study presents a comprehensive fairness audit and mitigation framework for predictive marketing models using the Bank Marketing dataset. We train logistic regression and random forest classifiers to predict customer subscription behavior and evaluate their performance across key demographic groups, including age, education, and job type. Using model explainability techniques such as SHAP and fairness metrics including disparate impact and true positive rate parity, we uncover notable disparities in model behavior that could result in discriminatory targeting. We implement three mitigation strategies&amp;amp;mdash;reweighing, threshold adjustment, and feature exclusion&amp;amp;mdash;and assess their effectiveness in improving fairness while preserving business-relevant performance metrics. Among these, reweighing produced the most balanced outcome, raising the Disparate Impact Ratio for older individuals from 0.65 to 0.82 and reducing the true positive rate parity gap by over 40%, with only a modest decline in precision (from 0.78 to 0.76). We propose a replicable workflow for embedding fairness auditing into enterprise BI systems and highlight the strategic importance of ethical AI practices in building accountable and inclusive marketing technologies. technologies.</p>
	]]></content:encoded>

	<dc:title>Fairness in Predictive Marketing: Auditing and Mitigating Demographic Bias in Machine Learning for Customer Targeting</dc:title>
			<dc:creator>Sayee Phaneendhar Pasupuleti</dc:creator>
			<dc:creator>Jagadeesh Kola</dc:creator>
			<dc:creator>Sai Phaneendra Manikantesh Kodete</dc:creator>
			<dc:creator>Sree Harsha Palli</dc:creator>
		<dc:identifier>doi: 10.3390/analytics4040026</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2025-10-01</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2025-10-01</prism:publicationDate>
	<prism:volume>4</prism:volume>
	<prism:number>4</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>26</prism:startingPage>
		<prism:doi>10.3390/analytics4040026</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/4/4/26</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/4/3/25">

	<title>Analytics, Vol. 4, Pages 25: Evolution Cybercrime&amp;mdash;Key Trends, Cybersecurity Threats, and Mitigation Strategies from Historical Data</title>
	<link>https://www.mdpi.com/2813-2203/4/3/25</link>
	<description>The landscape of cybercrime has undergone significant transformations over the past decade. Present-day threats include AI-generated attacks, deep fakes, 5G network vulnerabilities, cryptojacking, and supply chain attacks, among others. To remain resilient against contemporary threats, it is essential to examine historical data to gain insights that can inform cybersecurity strategies, policy decisions, and public awareness campaigns. This paper presents a comprehensive analysis of the evolution of cyber trends in state-sponsored attacks over the past 20 years, based on the council on foreign relations state-sponsored cyber operations (2005&amp;amp;ndash;present). The study explores the key trends, patterns, and demographic shifts in cybercrime victims, the evolution of complaints and losses, and the most prevalent cyber threats over the years. It also investigates the geographical distribution, the gender disparity in victimization, the temporal peaks of specific scams, and the most frequently reported internet crimes. The findings reveal a traditional cyber landscape, with cyber threats becoming more sophisticated and monetized. Finally, the article proposes areas for further exploration through a comprehensive analysis. It provides a detailed chronicle of the trajectory of cybercrimes, offering insights into its past, present, and future.</description>
	<pubDate>2025-09-18</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 4, Pages 25: Evolution Cybercrime&amp;mdash;Key Trends, Cybersecurity Threats, and Mitigation Strategies from Historical Data</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/4/3/25">doi: 10.3390/analytics4030025</a></p>
	<p>Authors:
		Muhammad Abdullah
		Muhammad Munib Nawaz
		Bilal Saleem
		Maila Zahra
		Effa binte Ashfaq
		Zia Muhammad
		</p>
	<p>The landscape of cybercrime has undergone significant transformations over the past decade. Present-day threats include AI-generated attacks, deep fakes, 5G network vulnerabilities, cryptojacking, and supply chain attacks, among others. To remain resilient against contemporary threats, it is essential to examine historical data to gain insights that can inform cybersecurity strategies, policy decisions, and public awareness campaigns. This paper presents a comprehensive analysis of the evolution of cyber trends in state-sponsored attacks over the past 20 years, based on the council on foreign relations state-sponsored cyber operations (2005&amp;amp;ndash;present). The study explores the key trends, patterns, and demographic shifts in cybercrime victims, the evolution of complaints and losses, and the most prevalent cyber threats over the years. It also investigates the geographical distribution, the gender disparity in victimization, the temporal peaks of specific scams, and the most frequently reported internet crimes. The findings reveal a traditional cyber landscape, with cyber threats becoming more sophisticated and monetized. Finally, the article proposes areas for further exploration through a comprehensive analysis. It provides a detailed chronicle of the trajectory of cybercrimes, offering insights into its past, present, and future.</p>
	]]></content:encoded>

	<dc:title>Evolution Cybercrime&amp;amp;mdash;Key Trends, Cybersecurity Threats, and Mitigation Strategies from Historical Data</dc:title>
			<dc:creator>Muhammad Abdullah</dc:creator>
			<dc:creator>Muhammad Munib Nawaz</dc:creator>
			<dc:creator>Bilal Saleem</dc:creator>
			<dc:creator>Maila Zahra</dc:creator>
			<dc:creator>Effa binte Ashfaq</dc:creator>
			<dc:creator>Zia Muhammad</dc:creator>
		<dc:identifier>doi: 10.3390/analytics4030025</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2025-09-18</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2025-09-18</prism:publicationDate>
	<prism:volume>4</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Review</prism:section>
	<prism:startingPage>25</prism:startingPage>
		<prism:doi>10.3390/analytics4030025</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/4/3/25</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/4/3/24">

	<title>Analytics, Vol. 4, Pages 24: Meta-Analysis of Artificial Intelligence&amp;rsquo;s Influence on Competitive Dynamics for Small- and Medium-Sized Financial Institutions</title>
	<link>https://www.mdpi.com/2813-2203/4/3/24</link>
	<description>Artificial intelligence adoption in financial services presents uncertain implications for competitive dynamics, particularly for smaller institutions. The literature on AI in finance is growing, but there remains a notable absence regarding the impacts on small- and medium-sized financial services firms. We conduct a meta-analysis combining a systematic literature review, sentiment bibliometrics, and network analysis to examine how AI is transforming competition across different firm sizes in the financial sector. Our analysis of 160 publications reveals predominantly positive academic sentiment toward AI in finance (mean positive sentiment 0.725 versus negative 0.586, Cohen&amp;amp;rsquo;s d = 0.790, p &amp;amp;lt; 0.0001), with anticipatory sentiment increasing significantly over time (&amp;amp;beta;=2.10&amp;amp;times;10&amp;amp;minus;2,p=0.007). However, network analysis reveals substantial conceptual fragmentation in the research discourse, with a low connectivity coefficient (&amp;amp;#981;=0.125) indicating that the field lacks unified terminology. These findings expose a critical knowledge gap: while scholars increasingly view AI as competitively advantageous, research has not coalesced around coherent models for understanding differential impacts across firm sizes. The absence of size-specific research leaves practitioners and policymakers without clear guidance on how AI adoption affects competitive positioning, particularly for smaller institutions that may face resource constraints or technological barriers. The research fragmentation identified here has direct implications for strategic planning, regulatory approaches, and employment dynamics in financial services.</description>
	<pubDate>2025-09-18</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 4, Pages 24: Meta-Analysis of Artificial Intelligence&amp;rsquo;s Influence on Competitive Dynamics for Small- and Medium-Sized Financial Institutions</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/4/3/24">doi: 10.3390/analytics4030024</a></p>
	<p>Authors:
		Macy Cudmore
		David Mattie
		</p>
	<p>Artificial intelligence adoption in financial services presents uncertain implications for competitive dynamics, particularly for smaller institutions. The literature on AI in finance is growing, but there remains a notable absence regarding the impacts on small- and medium-sized financial services firms. We conduct a meta-analysis combining a systematic literature review, sentiment bibliometrics, and network analysis to examine how AI is transforming competition across different firm sizes in the financial sector. Our analysis of 160 publications reveals predominantly positive academic sentiment toward AI in finance (mean positive sentiment 0.725 versus negative 0.586, Cohen&amp;amp;rsquo;s d = 0.790, p &amp;amp;lt; 0.0001), with anticipatory sentiment increasing significantly over time (&amp;amp;beta;=2.10&amp;amp;times;10&amp;amp;minus;2,p=0.007). However, network analysis reveals substantial conceptual fragmentation in the research discourse, with a low connectivity coefficient (&amp;amp;#981;=0.125) indicating that the field lacks unified terminology. These findings expose a critical knowledge gap: while scholars increasingly view AI as competitively advantageous, research has not coalesced around coherent models for understanding differential impacts across firm sizes. The absence of size-specific research leaves practitioners and policymakers without clear guidance on how AI adoption affects competitive positioning, particularly for smaller institutions that may face resource constraints or technological barriers. The research fragmentation identified here has direct implications for strategic planning, regulatory approaches, and employment dynamics in financial services.</p>
	]]></content:encoded>

	<dc:title>Meta-Analysis of Artificial Intelligence&amp;amp;rsquo;s Influence on Competitive Dynamics for Small- and Medium-Sized Financial Institutions</dc:title>
			<dc:creator>Macy Cudmore</dc:creator>
			<dc:creator>David Mattie</dc:creator>
		<dc:identifier>doi: 10.3390/analytics4030024</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2025-09-18</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2025-09-18</prism:publicationDate>
	<prism:volume>4</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>24</prism:startingPage>
		<prism:doi>10.3390/analytics4030024</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/4/3/24</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/4/3/23">

	<title>Analytics, Vol. 4, Pages 23: Game-Theoretic Analysis of MEV Attacks and Mitigation Strategies in Decentralized Finance</title>
	<link>https://www.mdpi.com/2813-2203/4/3/23</link>
	<description>Maximal Extractable Value (MEV) presents a significant challenge to the fairness and efficiency of decentralized finance (DeFi). This paper provides a game-theoretic analysis of the strategic interactions within the MEV supply chain, involving searchers, builders, and validators. A three-stage game of incomplete information is developed to model these interactions. The analysis derives the Perfect Bayesian Nash Equilibria for primary MEV attack vectors, such as sandwich attacks, and formally characterizes attacker behavior. The research demonstrates that the competitive dynamics of the current MEV market are best described as Bertrand-style competition, which compels rational actors to engage in aggressive extraction that reduces overall system welfare in a prisoner&amp;amp;rsquo;s dilemma-like outcome. To address these issues, the paper proposes and evaluates mechanism design solutions, including commit&amp;amp;ndash;reveal schemes and threshold encryption. The potential of these solutions to mitigate harmful MEV is quantified. Theoretical models are validated against on-chain data from the Ethereum blockchain, showing a close alignment between theoretical predictions and empirically observed market behavior.</description>
	<pubDate>2025-09-15</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 4, Pages 23: Game-Theoretic Analysis of MEV Attacks and Mitigation Strategies in Decentralized Finance</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/4/3/23">doi: 10.3390/analytics4030023</a></p>
	<p>Authors:
		Benjamin Appiah
		Daniel Commey
		Winful Bagyl-Bac
		Laurene Adjei
		Ebenezer Owusu
		</p>
	<p>Maximal Extractable Value (MEV) presents a significant challenge to the fairness and efficiency of decentralized finance (DeFi). This paper provides a game-theoretic analysis of the strategic interactions within the MEV supply chain, involving searchers, builders, and validators. A three-stage game of incomplete information is developed to model these interactions. The analysis derives the Perfect Bayesian Nash Equilibria for primary MEV attack vectors, such as sandwich attacks, and formally characterizes attacker behavior. The research demonstrates that the competitive dynamics of the current MEV market are best described as Bertrand-style competition, which compels rational actors to engage in aggressive extraction that reduces overall system welfare in a prisoner&amp;amp;rsquo;s dilemma-like outcome. To address these issues, the paper proposes and evaluates mechanism design solutions, including commit&amp;amp;ndash;reveal schemes and threshold encryption. The potential of these solutions to mitigate harmful MEV is quantified. Theoretical models are validated against on-chain data from the Ethereum blockchain, showing a close alignment between theoretical predictions and empirically observed market behavior.</p>
	]]></content:encoded>

	<dc:title>Game-Theoretic Analysis of MEV Attacks and Mitigation Strategies in Decentralized Finance</dc:title>
			<dc:creator>Benjamin Appiah</dc:creator>
			<dc:creator>Daniel Commey</dc:creator>
			<dc:creator>Winful Bagyl-Bac</dc:creator>
			<dc:creator>Laurene Adjei</dc:creator>
			<dc:creator>Ebenezer Owusu</dc:creator>
		<dc:identifier>doi: 10.3390/analytics4030023</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2025-09-15</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2025-09-15</prism:publicationDate>
	<prism:volume>4</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>23</prism:startingPage>
		<prism:doi>10.3390/analytics4030023</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/4/3/23</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/4/3/22">

	<title>Analytics, Vol. 4, Pages 22: Bankruptcy Prediction Using Machine Learning and Data Preprocessing Techniques</title>
	<link>https://www.mdpi.com/2813-2203/4/3/22</link>
	<description>Bankruptcy prediction is critical for financial risk management. This study demonstrates that machine learning models, particularly Random Forest, can substantially improve prediction accuracy compared to traditional approaches. Using data from 8262 U.S. firms (1999&amp;amp;ndash;2018), we evaluate Logistic Regression, SVM, Random Forest, ANN, and RNN in combination with robust data preprocessing steps. Random Forest achieved the highest prediction accuracy (~95%), far surpassing Logistic Regression (~57%). Key preprocessing steps included feature engineering of financial ratios, feature selection, class balancing using SMOTE, and scaling. The findings highlight that ensemble and deep learning models&amp;amp;mdash;particularly Random Forest and ANN&amp;amp;mdash;offer strong predictive performance, suggesting their suitability for early-warning financial distress systems.</description>
	<pubDate>2025-09-10</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 4, Pages 22: Bankruptcy Prediction Using Machine Learning and Data Preprocessing Techniques</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/4/3/22">doi: 10.3390/analytics4030022</a></p>
	<p>Authors:
		Kamil Samara
		Apurva Shinde
		</p>
	<p>Bankruptcy prediction is critical for financial risk management. This study demonstrates that machine learning models, particularly Random Forest, can substantially improve prediction accuracy compared to traditional approaches. Using data from 8262 U.S. firms (1999&amp;amp;ndash;2018), we evaluate Logistic Regression, SVM, Random Forest, ANN, and RNN in combination with robust data preprocessing steps. Random Forest achieved the highest prediction accuracy (~95%), far surpassing Logistic Regression (~57%). Key preprocessing steps included feature engineering of financial ratios, feature selection, class balancing using SMOTE, and scaling. The findings highlight that ensemble and deep learning models&amp;amp;mdash;particularly Random Forest and ANN&amp;amp;mdash;offer strong predictive performance, suggesting their suitability for early-warning financial distress systems.</p>
	]]></content:encoded>

	<dc:title>Bankruptcy Prediction Using Machine Learning and Data Preprocessing Techniques</dc:title>
			<dc:creator>Kamil Samara</dc:creator>
			<dc:creator>Apurva Shinde</dc:creator>
		<dc:identifier>doi: 10.3390/analytics4030022</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2025-09-10</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2025-09-10</prism:publicationDate>
	<prism:volume>4</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>22</prism:startingPage>
		<prism:doi>10.3390/analytics4030022</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/4/3/22</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/4/3/21">

	<title>Analytics, Vol. 4, Pages 21: Accurate Analytical Forms of Heaviside and Ramp Function</title>
	<link>https://www.mdpi.com/2813-2203/4/3/21</link>
	<description>In this paper, explicit exact representations of the Unit Step Function and Ramp Function are obtained. These important functions constitute fundamental concepts of operational calculus together with digital signal processing theory and are also involved in many other areas of applied sciences and engineering practices. In particular, according to a rigorous process from the viewpoint of Mathematical Analysis, the Unit Step Function and the Ramp Function are equivalently performed as bi-parametric single-valued functions with only one constraint imposed on each parameter. The novelty of this work, when compared with other investigations concerning accurate and/or approximate forms of Unit Step Function and/or Ramp Function, is that the proposed exact formulae are not exhibited in terms of miscellaneous special functions, e.g., Gamma Function, Biexponential Function, or any other special functions, such as Error Function, Complementary Error Function, Hyperbolic Function, or Orthogonal Polynomials. In this framework, one may deduce that these formulae may be much more practical, flexible, and useful in the computational procedures that are inserted into operational calculus and digital signal processing techniques as well as other engineering practices.</description>
	<pubDate>2025-08-26</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 4, Pages 21: Accurate Analytical Forms of Heaviside and Ramp Function</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/4/3/21">doi: 10.3390/analytics4030021</a></p>
	<p>Authors:
		John Constantine Venetis
		</p>
	<p>In this paper, explicit exact representations of the Unit Step Function and Ramp Function are obtained. These important functions constitute fundamental concepts of operational calculus together with digital signal processing theory and are also involved in many other areas of applied sciences and engineering practices. In particular, according to a rigorous process from the viewpoint of Mathematical Analysis, the Unit Step Function and the Ramp Function are equivalently performed as bi-parametric single-valued functions with only one constraint imposed on each parameter. The novelty of this work, when compared with other investigations concerning accurate and/or approximate forms of Unit Step Function and/or Ramp Function, is that the proposed exact formulae are not exhibited in terms of miscellaneous special functions, e.g., Gamma Function, Biexponential Function, or any other special functions, such as Error Function, Complementary Error Function, Hyperbolic Function, or Orthogonal Polynomials. In this framework, one may deduce that these formulae may be much more practical, flexible, and useful in the computational procedures that are inserted into operational calculus and digital signal processing techniques as well as other engineering practices.</p>
	]]></content:encoded>

	<dc:title>Accurate Analytical Forms of Heaviside and Ramp Function</dc:title>
			<dc:creator>John Constantine Venetis</dc:creator>
		<dc:identifier>doi: 10.3390/analytics4030021</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2025-08-26</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2025-08-26</prism:publicationDate>
	<prism:volume>4</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>21</prism:startingPage>
		<prism:doi>10.3390/analytics4030021</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/4/3/21</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/4/3/20">

	<title>Analytics, Vol. 4, Pages 20: LINEX Loss-Based Estimation of Expected Arrival Time of Next Event from HPP and NHPP Processes Past Truncated Time</title>
	<link>https://www.mdpi.com/2813-2203/4/3/20</link>
	<description>This article introduces a computational tool for Bayesian estimation of the expected time until the next event occurs in both homogeneous Poisson processes (HPPs) and non-homogeneous Poisson processes (NHPPs), following a truncated time. The estimation utilizes the linear exponential (LINEX) asymmetric loss function and incorporates both gamma and non-informative priors. Furthermore, it presents a minimax-type criterion to ascertain the optimal sample size required to achieve a specified percentage reduction in posterior risk. Simulation studies indicate that estimators employing gamma priors for both HPP and NHPP demonstrate greater accuracy compared to those based on non-informative priors and maximum likelihood estimates (MLE), provided that the proposed data-driven method for selecting hyperparameters is applied.</description>
	<pubDate>2025-08-26</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 4, Pages 20: LINEX Loss-Based Estimation of Expected Arrival Time of Next Event from HPP and NHPP Processes Past Truncated Time</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/4/3/20">doi: 10.3390/analytics4030020</a></p>
	<p>Authors:
		M. S. Aminzadeh
		</p>
	<p>This article introduces a computational tool for Bayesian estimation of the expected time until the next event occurs in both homogeneous Poisson processes (HPPs) and non-homogeneous Poisson processes (NHPPs), following a truncated time. The estimation utilizes the linear exponential (LINEX) asymmetric loss function and incorporates both gamma and non-informative priors. Furthermore, it presents a minimax-type criterion to ascertain the optimal sample size required to achieve a specified percentage reduction in posterior risk. Simulation studies indicate that estimators employing gamma priors for both HPP and NHPP demonstrate greater accuracy compared to those based on non-informative priors and maximum likelihood estimates (MLE), provided that the proposed data-driven method for selecting hyperparameters is applied.</p>
	]]></content:encoded>

	<dc:title>LINEX Loss-Based Estimation of Expected Arrival Time of Next Event from HPP and NHPP Processes Past Truncated Time</dc:title>
			<dc:creator>M. S. Aminzadeh</dc:creator>
		<dc:identifier>doi: 10.3390/analytics4030020</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2025-08-26</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2025-08-26</prism:publicationDate>
	<prism:volume>4</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>20</prism:startingPage>
		<prism:doi>10.3390/analytics4030020</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/4/3/20</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/4/3/19">

	<title>Analytics, Vol. 4, Pages 19: A Bounded Sine Skewed Model for Hydrological Data Analysis</title>
	<link>https://www.mdpi.com/2813-2203/4/3/19</link>
	<description>Hydrological time series frequently exhibit periodic trends with variables such as rainfall, runoff, and evaporation rates often following annual cycles. Seasonal variations further contribute to the complexity of these data sets. A critical aspect of analyzing such phenomena is estimating realistic return intervals, making the precise determination of these values essential. Given this importance, selecting an appropriate probability distribution is paramount. To address this need, we introduce a flexible probability model specifically designed to capture periodicity in hydrological data. We thoroughly examine its fundamental mathematical and statistical properties, including the asymptotic behavior of the probability density function (PDF) and hazard rate function (HRF), to enhance predictive accuracy. Our analysis reveals that the PDF exhibits polynomial decay as x&amp;amp;rarr;&amp;amp;infin;, ensuring heavy-tailed behavior suitable for extreme events. The HRF demonstrates decreasing or non-monotonic trends, reflecting variable failure risks over time. Additionally, we conduct a simulation study to evaluate the performance of the estimation method. Based on these results, we refine return period estimates, providing more reliable and robust hydrological assessments. This approach ensures that the model not only fits observed data but also captures the underlying dynamics of hydrological extremes.</description>
	<pubDate>2025-08-13</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 4, Pages 19: A Bounded Sine Skewed Model for Hydrological Data Analysis</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/4/3/19">doi: 10.3390/analytics4030019</a></p>
	<p>Authors:
		Tassaddaq Hussain
		Mohammad Shakil
		Mohammad Ahsanullah
		Bhuiyan Mohammad Golam Kibria
		</p>
	<p>Hydrological time series frequently exhibit periodic trends with variables such as rainfall, runoff, and evaporation rates often following annual cycles. Seasonal variations further contribute to the complexity of these data sets. A critical aspect of analyzing such phenomena is estimating realistic return intervals, making the precise determination of these values essential. Given this importance, selecting an appropriate probability distribution is paramount. To address this need, we introduce a flexible probability model specifically designed to capture periodicity in hydrological data. We thoroughly examine its fundamental mathematical and statistical properties, including the asymptotic behavior of the probability density function (PDF) and hazard rate function (HRF), to enhance predictive accuracy. Our analysis reveals that the PDF exhibits polynomial decay as x&amp;amp;rarr;&amp;amp;infin;, ensuring heavy-tailed behavior suitable for extreme events. The HRF demonstrates decreasing or non-monotonic trends, reflecting variable failure risks over time. Additionally, we conduct a simulation study to evaluate the performance of the estimation method. Based on these results, we refine return period estimates, providing more reliable and robust hydrological assessments. This approach ensures that the model not only fits observed data but also captures the underlying dynamics of hydrological extremes.</p>
	]]></content:encoded>

	<dc:title>A Bounded Sine Skewed Model for Hydrological Data Analysis</dc:title>
			<dc:creator>Tassaddaq Hussain</dc:creator>
			<dc:creator>Mohammad Shakil</dc:creator>
			<dc:creator>Mohammad Ahsanullah</dc:creator>
			<dc:creator>Bhuiyan Mohammad Golam Kibria</dc:creator>
		<dc:identifier>doi: 10.3390/analytics4030019</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2025-08-13</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2025-08-13</prism:publicationDate>
	<prism:volume>4</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>19</prism:startingPage>
		<prism:doi>10.3390/analytics4030019</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/4/3/19</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/4/3/18">

	<title>Analytics, Vol. 4, Pages 18: Predictive Framework for Regional Patent Output Using Digital Economic Indicators: A Stacked Machine Learning and Geospatial Ensemble to Address R&amp;amp;D Disparities</title>
	<link>https://www.mdpi.com/2813-2203/4/3/18</link>
	<description>As digital transformation becomes an increasingly central focus of national and regional policy agendas, parallel efforts are intensifying to stimulate innovation as a critical driver of firm competitiveness and high-quality economic growth. However, regional disparities in innovation capacity persist. This study proposes an integrated framework in which regionally tracked digital economy indicators are leveraged to predict firm-level innovation performance, measured through patent activity, across China. Drawing on a comprehensive dataset covering 13 digital economic indicators from 2013 to 2022, this study spans core, broad, and narrow dimensions of digital development. Spatial dependencies among these indicators are assessed using global and local spatial autocorrelation measures, including Moran&amp;amp;rsquo;s I and Geary&amp;amp;rsquo;s C, to provide actionable insights for constructing innovation-conducive environments. To model the predictive relationship between digital metrics and innovation output, this study employs a suite of supervised machine learning techniques&amp;amp;mdash;Random Forest, Extreme Learning Machine (ELM), Support Vector Machine (SVM), XGBoost, and stacked ensemble approaches. Our findings demonstrate the potential of digital infrastructure metrics to serve as early indicators of regional innovation capacity, offering a data-driven foundation for targeted policymaking, strategic resource allocation, and the design of adaptive digital innovation ecosystems.</description>
	<pubDate>2025-07-08</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 4, Pages 18: Predictive Framework for Regional Patent Output Using Digital Economic Indicators: A Stacked Machine Learning and Geospatial Ensemble to Address R&amp;amp;D Disparities</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/4/3/18">doi: 10.3390/analytics4030018</a></p>
	<p>Authors:
		Amelia Zhao
		Peng Wang
		</p>
	<p>As digital transformation becomes an increasingly central focus of national and regional policy agendas, parallel efforts are intensifying to stimulate innovation as a critical driver of firm competitiveness and high-quality economic growth. However, regional disparities in innovation capacity persist. This study proposes an integrated framework in which regionally tracked digital economy indicators are leveraged to predict firm-level innovation performance, measured through patent activity, across China. Drawing on a comprehensive dataset covering 13 digital economic indicators from 2013 to 2022, this study spans core, broad, and narrow dimensions of digital development. Spatial dependencies among these indicators are assessed using global and local spatial autocorrelation measures, including Moran&amp;amp;rsquo;s I and Geary&amp;amp;rsquo;s C, to provide actionable insights for constructing innovation-conducive environments. To model the predictive relationship between digital metrics and innovation output, this study employs a suite of supervised machine learning techniques&amp;amp;mdash;Random Forest, Extreme Learning Machine (ELM), Support Vector Machine (SVM), XGBoost, and stacked ensemble approaches. Our findings demonstrate the potential of digital infrastructure metrics to serve as early indicators of regional innovation capacity, offering a data-driven foundation for targeted policymaking, strategic resource allocation, and the design of adaptive digital innovation ecosystems.</p>
	]]></content:encoded>

	<dc:title>Predictive Framework for Regional Patent Output Using Digital Economic Indicators: A Stacked Machine Learning and Geospatial Ensemble to Address R&amp;amp;amp;D Disparities</dc:title>
			<dc:creator>Amelia Zhao</dc:creator>
			<dc:creator>Peng Wang</dc:creator>
		<dc:identifier>doi: 10.3390/analytics4030018</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2025-07-08</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2025-07-08</prism:publicationDate>
	<prism:volume>4</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>18</prism:startingPage>
		<prism:doi>10.3390/analytics4030018</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/4/3/18</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/4/3/17">

	<title>Analytics, Vol. 4, Pages 17: Domestication of Source Text in Literary Translation Prevails over Foreignization</title>
	<link>https://www.mdpi.com/2813-2203/4/3/17</link>
	<description>Domestication is a translation theory in which the source text (to be translated) is matched to the foreign reader by erasing its original linguistic and cultural difference. This match aims at making the target text (translated text) more fluent. On the contrary, foreignization is a translation theory in which the foreign reader is matched to the source text. This paper mathematically explores the degree of domestication/foreignization in current translation practice of texts written in alphabetical languages. A geometrical representation of texts, based on linear combinations of deep&amp;amp;ndash;language parameters, allows us (a) to calculate a domestication index which measures how much domestication is applied to the source text and (b) to distinguish language families. An expansion index measures the relative spread around mean values. This paper reports statistics and results on translations of (a) Greek New Testament books in Latin and in 35 modern languages, belonging to diverse language families; and (b) English novels in Western languages. English and French, although attributed to different language families, mathematically almost coincide. The requirement of making the target text more fluent makes domestication, with varying degrees, universally adopted, so that a blind comparison of the same linguistic parameters of a text and its translation hardly indicates that they refer to each other.</description>
	<pubDate>2025-06-20</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 4, Pages 17: Domestication of Source Text in Literary Translation Prevails over Foreignization</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/4/3/17">doi: 10.3390/analytics4030017</a></p>
	<p>Authors:
		Emilio Matricciani
		</p>
	<p>Domestication is a translation theory in which the source text (to be translated) is matched to the foreign reader by erasing its original linguistic and cultural difference. This match aims at making the target text (translated text) more fluent. On the contrary, foreignization is a translation theory in which the foreign reader is matched to the source text. This paper mathematically explores the degree of domestication/foreignization in current translation practice of texts written in alphabetical languages. A geometrical representation of texts, based on linear combinations of deep&amp;amp;ndash;language parameters, allows us (a) to calculate a domestication index which measures how much domestication is applied to the source text and (b) to distinguish language families. An expansion index measures the relative spread around mean values. This paper reports statistics and results on translations of (a) Greek New Testament books in Latin and in 35 modern languages, belonging to diverse language families; and (b) English novels in Western languages. English and French, although attributed to different language families, mathematically almost coincide. The requirement of making the target text more fluent makes domestication, with varying degrees, universally adopted, so that a blind comparison of the same linguistic parameters of a text and its translation hardly indicates that they refer to each other.</p>
	]]></content:encoded>

	<dc:title>Domestication of Source Text in Literary Translation Prevails over Foreignization</dc:title>
			<dc:creator>Emilio Matricciani</dc:creator>
		<dc:identifier>doi: 10.3390/analytics4030017</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2025-06-20</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2025-06-20</prism:publicationDate>
	<prism:volume>4</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>17</prism:startingPage>
		<prism:doi>10.3390/analytics4030017</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/4/3/17</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/4/2/16">

	<title>Analytics, Vol. 4, Pages 16: The Classical Model of Type-Token Systems Compared with Items from the Standardized Project Gutenberg Corpus</title>
	<link>https://www.mdpi.com/2813-2203/4/2/16</link>
	<description>We compare the &amp;amp;ldquo;classical&amp;amp;rdquo; equations of type-token systems, namely Zipf&amp;amp;rsquo;s laws, Heaps&amp;amp;rsquo; law and the relationships between their indices, with data selected from the Standardized Project Gutenberg Corpus (SPGC). Selected items all exceed 100,000 word-tokens and are trimmed to 100,000 word-tokens each. With the most egregious anomalies removed, a dataset of 8432 items is examined in terms of the relationships between the Zipf and Heaps&amp;amp;rsquo; indices computed using the Maximum Likelihood algorithm. Zipf&amp;amp;rsquo;s second (size) law indices suggest that the types vs. frequency distribution is log&amp;amp;ndash;log convex, with the high and low frequency indices showing weak but significant negative correlation. Under certain circumstances, the classical equations work tolerably well, though the level of agreement depends heavily on the type of literature and the language (Finnish being notably anomalous). The frequency vs. rank characteristics exhibit log&amp;amp;ndash;log linearity in the &amp;amp;ldquo;middle range&amp;amp;rdquo; (ranks 100&amp;amp;ndash;1000), as characterised by the Kolmogorov&amp;amp;ndash;Smirnov significance. For most items, the Heaps&amp;amp;rsquo; index correlates strongly with the low frequency Zipf index in a manner consistent with classical theory, while the high frequency indices are largely uncorrelated. This is consistent with a simple simulation.</description>
	<pubDate>2025-06-05</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 4, Pages 16: The Classical Model of Type-Token Systems Compared with Items from the Standardized Project Gutenberg Corpus</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/4/2/16">doi: 10.3390/analytics4020016</a></p>
	<p>Authors:
		Martin Tunnicliffe
		Gordon Hunter
		</p>
	<p>We compare the &amp;amp;ldquo;classical&amp;amp;rdquo; equations of type-token systems, namely Zipf&amp;amp;rsquo;s laws, Heaps&amp;amp;rsquo; law and the relationships between their indices, with data selected from the Standardized Project Gutenberg Corpus (SPGC). Selected items all exceed 100,000 word-tokens and are trimmed to 100,000 word-tokens each. With the most egregious anomalies removed, a dataset of 8432 items is examined in terms of the relationships between the Zipf and Heaps&amp;amp;rsquo; indices computed using the Maximum Likelihood algorithm. Zipf&amp;amp;rsquo;s second (size) law indices suggest that the types vs. frequency distribution is log&amp;amp;ndash;log convex, with the high and low frequency indices showing weak but significant negative correlation. Under certain circumstances, the classical equations work tolerably well, though the level of agreement depends heavily on the type of literature and the language (Finnish being notably anomalous). The frequency vs. rank characteristics exhibit log&amp;amp;ndash;log linearity in the &amp;amp;ldquo;middle range&amp;amp;rdquo; (ranks 100&amp;amp;ndash;1000), as characterised by the Kolmogorov&amp;amp;ndash;Smirnov significance. For most items, the Heaps&amp;amp;rsquo; index correlates strongly with the low frequency Zipf index in a manner consistent with classical theory, while the high frequency indices are largely uncorrelated. This is consistent with a simple simulation.</p>
	]]></content:encoded>

	<dc:title>The Classical Model of Type-Token Systems Compared with Items from the Standardized Project Gutenberg Corpus</dc:title>
			<dc:creator>Martin Tunnicliffe</dc:creator>
			<dc:creator>Gordon Hunter</dc:creator>
		<dc:identifier>doi: 10.3390/analytics4020016</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2025-06-05</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2025-06-05</prism:publicationDate>
	<prism:volume>4</prism:volume>
	<prism:number>2</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>16</prism:startingPage>
		<prism:doi>10.3390/analytics4020016</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/4/2/16</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/4/2/15">

	<title>Analytics, Vol. 4, Pages 15: Multiplicity Adjustments for Differences in Proportion Parameters in Multiple-Sample Misclassified Binary Data</title>
	<link>https://www.mdpi.com/2813-2203/4/2/15</link>
	<description>Generally, following an omnibus (overall equality) test, multiple pairwise comparison (MPC) tests are typically conducted as the second step in a sequential testing procedure to identify which specific pairs (e.g., proportions) exhibit significant differences. In this manuscript, we develop maximum likelihood estimation (MLE) methods to construct three different types of confidence intervals (CIs) for multiple pairwise differences in proportions, specifically in contexts where both types of misclassifications (i.e., over-reporting and under-reporting) exist in multiple-sample binomial data. Our closed-form algorithm is straightforward to implement. Consequently, when dealing with multiple sample proportions, we can readily apply MPC adjustment procedures&amp;amp;mdash;such as Bonferroni, &amp;amp;Scaron;id&amp;amp;aacute;k, and Dunn&amp;amp;mdash;to address the issue of multiplicity. This manuscript advances the existing literature by extending from scenarios with only one type of misclassification to those involving both. Furthermore, we demonstrate our methods using a real-world data example.</description>
	<pubDate>2025-05-28</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 4, Pages 15: Multiplicity Adjustments for Differences in Proportion Parameters in Multiple-Sample Misclassified Binary Data</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/4/2/15">doi: 10.3390/analytics4020015</a></p>
	<p>Authors:
		Dewi Rahardja
		</p>
	<p>Generally, following an omnibus (overall equality) test, multiple pairwise comparison (MPC) tests are typically conducted as the second step in a sequential testing procedure to identify which specific pairs (e.g., proportions) exhibit significant differences. In this manuscript, we develop maximum likelihood estimation (MLE) methods to construct three different types of confidence intervals (CIs) for multiple pairwise differences in proportions, specifically in contexts where both types of misclassifications (i.e., over-reporting and under-reporting) exist in multiple-sample binomial data. Our closed-form algorithm is straightforward to implement. Consequently, when dealing with multiple sample proportions, we can readily apply MPC adjustment procedures&amp;amp;mdash;such as Bonferroni, &amp;amp;Scaron;id&amp;amp;aacute;k, and Dunn&amp;amp;mdash;to address the issue of multiplicity. This manuscript advances the existing literature by extending from scenarios with only one type of misclassification to those involving both. Furthermore, we demonstrate our methods using a real-world data example.</p>
	]]></content:encoded>

	<dc:title>Multiplicity Adjustments for Differences in Proportion Parameters in Multiple-Sample Misclassified Binary Data</dc:title>
			<dc:creator>Dewi Rahardja</dc:creator>
		<dc:identifier>doi: 10.3390/analytics4020015</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2025-05-28</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2025-05-28</prism:publicationDate>
	<prism:volume>4</prism:volume>
	<prism:number>2</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>15</prism:startingPage>
		<prism:doi>10.3390/analytics4020015</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/4/2/15</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/4/2/14">

	<title>Analytics, Vol. 4, Pages 14: Analytical Modeling of Ancillary Items</title>
	<link>https://www.mdpi.com/2813-2203/4/2/14</link>
	<description>Airlines profitability increasingly depends on the sale of ancillary items such as seat selection, baggage fees, etc. The modeling of ancillary items is becoming more important in the analytics literature. Much of the modeling is stylized and not immediately applicable. This paper contains a review of the approaches and modeling assumptions made in the literature. The focus is on the assumptions made so that models may be evaluated for how effective they are for applications and to highlight gaps in the literature.</description>
	<pubDate>2025-05-19</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 4, Pages 14: Analytical Modeling of Ancillary Items</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/4/2/14">doi: 10.3390/analytics4020014</a></p>
	<p>Authors:
		John Wilson
		</p>
	<p>Airlines profitability increasingly depends on the sale of ancillary items such as seat selection, baggage fees, etc. The modeling of ancillary items is becoming more important in the analytics literature. Much of the modeling is stylized and not immediately applicable. This paper contains a review of the approaches and modeling assumptions made in the literature. The focus is on the assumptions made so that models may be evaluated for how effective they are for applications and to highlight gaps in the literature.</p>
	]]></content:encoded>

	<dc:title>Analytical Modeling of Ancillary Items</dc:title>
			<dc:creator>John Wilson</dc:creator>
		<dc:identifier>doi: 10.3390/analytics4020014</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2025-05-19</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2025-05-19</prism:publicationDate>
	<prism:volume>4</prism:volume>
	<prism:number>2</prism:number>
	<prism:section>Review</prism:section>
	<prism:startingPage>14</prism:startingPage>
		<prism:doi>10.3390/analytics4020014</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/4/2/14</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/4/2/13">

	<title>Analytics, Vol. 4, Pages 13: Artificial Intelligence Applied to the Analysis of Biblical Scriptures: A Systematic Review</title>
	<link>https://www.mdpi.com/2813-2203/4/2/13</link>
	<description>The Holy Bible is the most read book in the world, originally written in Aramaic, Hebrew, and Greek over a time span in the order of centuries by many people, and formed by a combination of various literary styles, such as stories, prophecies, poetry, instructions, and others. As such, the Bible is a complex text to be analyzed by humans and machines. This paper provides a systematic survey of the application of Artificial Intelligence (AI) and some of its subareas to the analysis of the Biblical scriptures. Emphasis is given to what types of tasks are being solved, what are the main AI algorithms used, and their limitations. The findings deliver a general perspective on how this field is being developed, along with its limitations and gaps. This research follows a procedure based on three steps: planning (defining the review protocol), conducting (performing the survey), and reporting (formatting the report). The results obtained show there are seven main tasks solved by AI in the Bible analysis: machine translation, authorship identification, part of speech tagging (PoS tagging), semantic annotation, clustering, categorization, and Biblical interpretation. Also, the classes of AI techniques with better performance when applied to Biblical text research are machine learning, neural networks, and deep learning. The main challenges in the field involve the nature and style of the language used in the Bible, among others.</description>
	<pubDate>2025-04-11</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 4, Pages 13: Artificial Intelligence Applied to the Analysis of Biblical Scriptures: A Systematic Review</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/4/2/13">doi: 10.3390/analytics4020013</a></p>
	<p>Authors:
		Bruno Cesar Lima
		Nizam Omar
		Israel Avansi
		Leandro Nunes de Castro
		</p>
	<p>The Holy Bible is the most read book in the world, originally written in Aramaic, Hebrew, and Greek over a time span in the order of centuries by many people, and formed by a combination of various literary styles, such as stories, prophecies, poetry, instructions, and others. As such, the Bible is a complex text to be analyzed by humans and machines. This paper provides a systematic survey of the application of Artificial Intelligence (AI) and some of its subareas to the analysis of the Biblical scriptures. Emphasis is given to what types of tasks are being solved, what are the main AI algorithms used, and their limitations. The findings deliver a general perspective on how this field is being developed, along with its limitations and gaps. This research follows a procedure based on three steps: planning (defining the review protocol), conducting (performing the survey), and reporting (formatting the report). The results obtained show there are seven main tasks solved by AI in the Bible analysis: machine translation, authorship identification, part of speech tagging (PoS tagging), semantic annotation, clustering, categorization, and Biblical interpretation. Also, the classes of AI techniques with better performance when applied to Biblical text research are machine learning, neural networks, and deep learning. The main challenges in the field involve the nature and style of the language used in the Bible, among others.</p>
	]]></content:encoded>

	<dc:title>Artificial Intelligence Applied to the Analysis of Biblical Scriptures: A Systematic Review</dc:title>
			<dc:creator>Bruno Cesar Lima</dc:creator>
			<dc:creator>Nizam Omar</dc:creator>
			<dc:creator>Israel Avansi</dc:creator>
			<dc:creator>Leandro Nunes de Castro</dc:creator>
		<dc:identifier>doi: 10.3390/analytics4020013</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2025-04-11</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2025-04-11</prism:publicationDate>
	<prism:volume>4</prism:volume>
	<prism:number>2</prism:number>
	<prism:section>Systematic Review</prism:section>
	<prism:startingPage>13</prism:startingPage>
		<prism:doi>10.3390/analytics4020013</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/4/2/13</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/4/2/12">

	<title>Analytics, Vol. 4, Pages 12: Traffic Prediction with Data Fusion and Machine Learning</title>
	<link>https://www.mdpi.com/2813-2203/4/2/12</link>
	<description>Traffic prediction, as a core task to alleviate urban congestion and optimize the transport system, has limitations in the integration of multimodal data, making it difficult to comprehensively capture the complex spatio-temporal characteristics of the transport system. Although some studies have attempted to introduce multimodal data, they mostly rely on resource-intensive deep neural network architectures, which have difficultly meeting the demands of practical applications. To this end, we propose a traffic prediction framework based on simple machine learning techniques that effectively integrates property features, amenity features, and emotion features (PAE features). Validated with large-scale real datasets, the method demonstrates excellent prediction performance while significantly reducing computational complexity and deployment costs. This study demonstrates the great potential of simple machine learning techniques in multimodal data fusion, provides an efficient and practical solution for traffic prediction, and offers an effective alternative to resource-intensive deep learning methods, opening up new paths for building scalable traffic prediction systems.</description>
	<pubDate>2025-04-09</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 4, Pages 12: Traffic Prediction with Data Fusion and Machine Learning</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/4/2/12">doi: 10.3390/analytics4020012</a></p>
	<p>Authors:
		Juntao Qiu
		Yaping Zhao
		</p>
	<p>Traffic prediction, as a core task to alleviate urban congestion and optimize the transport system, has limitations in the integration of multimodal data, making it difficult to comprehensively capture the complex spatio-temporal characteristics of the transport system. Although some studies have attempted to introduce multimodal data, they mostly rely on resource-intensive deep neural network architectures, which have difficultly meeting the demands of practical applications. To this end, we propose a traffic prediction framework based on simple machine learning techniques that effectively integrates property features, amenity features, and emotion features (PAE features). Validated with large-scale real datasets, the method demonstrates excellent prediction performance while significantly reducing computational complexity and deployment costs. This study demonstrates the great potential of simple machine learning techniques in multimodal data fusion, provides an efficient and practical solution for traffic prediction, and offers an effective alternative to resource-intensive deep learning methods, opening up new paths for building scalable traffic prediction systems.</p>
	]]></content:encoded>

	<dc:title>Traffic Prediction with Data Fusion and Machine Learning</dc:title>
			<dc:creator>Juntao Qiu</dc:creator>
			<dc:creator>Yaping Zhao</dc:creator>
		<dc:identifier>doi: 10.3390/analytics4020012</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2025-04-09</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2025-04-09</prism:publicationDate>
	<prism:volume>4</prism:volume>
	<prism:number>2</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>12</prism:startingPage>
		<prism:doi>10.3390/analytics4020012</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/4/2/12</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/4/2/11">

	<title>Analytics, Vol. 4, Pages 11: Copula-Based Bayesian Model for Detecting Differential Gene Expression</title>
	<link>https://www.mdpi.com/2813-2203/4/2/11</link>
	<description>Deoxyribonucleic acid, more commonly known as DNA, is a fundamental genetic material in all living organisms, containing thousands of genes, but only a subset exhibit differential expression and play a crucial role in diseases. Microarray technology has revolutionized the study of gene expression, with two primary types available for expression analysis: spotted cDNA arrays and oligonucleotide arrays. This research focuses on the statistical analysis of data from spotted cDNA microarrays. Numerous models have been developed to identify differentially expressed genes based on the red and green fluorescence intensities measured using these arrays. We propose a novel approach using a Gaussian copula model to characterize the joint distribution of red and green intensities, effectively capturing their dependence structure. Given the right-skewed nature of the intensity distributions, we model the marginal distributions using gamma distributions. Differentially expressed genes are identified using the Bayes estimate under our proposed copula framework. To evaluate the performance of our model, we conduct simulation studies to assess parameter estimation accuracy. Our results demonstrate that the proposed approach outperforms existing methods reported in the literature. Finally, we apply our model to Escherichia coli microarray data, illustrating its practical utility in gene expression analysis.</description>
	<pubDate>2025-04-03</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 4, Pages 11: Copula-Based Bayesian Model for Detecting Differential Gene Expression</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/4/2/11">doi: 10.3390/analytics4020011</a></p>
	<p>Authors:
		Prasansha Liyanaarachchi
		N. Rao Chaganty
		</p>
	<p>Deoxyribonucleic acid, more commonly known as DNA, is a fundamental genetic material in all living organisms, containing thousands of genes, but only a subset exhibit differential expression and play a crucial role in diseases. Microarray technology has revolutionized the study of gene expression, with two primary types available for expression analysis: spotted cDNA arrays and oligonucleotide arrays. This research focuses on the statistical analysis of data from spotted cDNA microarrays. Numerous models have been developed to identify differentially expressed genes based on the red and green fluorescence intensities measured using these arrays. We propose a novel approach using a Gaussian copula model to characterize the joint distribution of red and green intensities, effectively capturing their dependence structure. Given the right-skewed nature of the intensity distributions, we model the marginal distributions using gamma distributions. Differentially expressed genes are identified using the Bayes estimate under our proposed copula framework. To evaluate the performance of our model, we conduct simulation studies to assess parameter estimation accuracy. Our results demonstrate that the proposed approach outperforms existing methods reported in the literature. Finally, we apply our model to Escherichia coli microarray data, illustrating its practical utility in gene expression analysis.</p>
	]]></content:encoded>

	<dc:title>Copula-Based Bayesian Model for Detecting Differential Gene Expression</dc:title>
			<dc:creator>Prasansha Liyanaarachchi</dc:creator>
			<dc:creator>N. Rao Chaganty</dc:creator>
		<dc:identifier>doi: 10.3390/analytics4020011</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2025-04-03</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2025-04-03</prism:publicationDate>
	<prism:volume>4</prism:volume>
	<prism:number>2</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>11</prism:startingPage>
		<prism:doi>10.3390/analytics4020011</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/4/2/11</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/4/1/10">

	<title>Analytics, Vol. 4, Pages 10: Unveiling the Impact of Socioeconomic and Demographic Factors on Graduate Salaries: A Machine Learning Explanatory Analytical Approach Using Higher Education Statistical Agency Data</title>
	<link>https://www.mdpi.com/2813-2203/4/1/10</link>
	<description>Graduate salaries are a significant concern for graduates, employers, and policymakers, as various factors influence them. This study investigates determinants of graduate salaries in the UK, utilising survey data from HESA (Higher Education Statistical Agency) and integrating advanced machine learning (ML) explanatory techniques with statistical analytical methodologies. By employing multi-stage analyses alongside machine learning models such as decision trees, random forests and the explainability with SHAP stands for (Shapley Additive exPanations), this study investigates the influence of 21 socioeconomic and demographic variables on graduate salary outcomes. Key variables, including institutional reputation, age at graduation, socioeconomic classification, job qualification requirements, and domicile, emerged as critical determinants, with institutional reputation proving the most significant. Among ML methods, the decision tree achieved a standout with the highest accuracy through rigorous optimisation techniques, including oversampling and undersampling. SHAP highlighted the top 12 influential variables, providing actionable insights into the interplay between individual and systemic factors. Furthermore, the statistical analysis using ANOVA (Analysis of Variance) validated the significance of these variables, revealing intricate interactions that shape graduate salary dynamics. Additionally, domain experts&amp;amp;rsquo; opinions are also analysed to authenticate the findings. This research makes a unique contribution by combining qualitative contextual analysis with quantitative methodologies, machine learning explainability and domain experts&amp;amp;rsquo; views on addressing gaps in the existing identification of graduate salary predicting components. Additionally, the findings inform policy and educational interventions to reduce wage inequalities and promote equitable career opportunities. Despite limitations, such as the UK-specific dataset and the focus on socioeconomic and demographic variables, this study lays a robust foundation for future research in predictive modelling and graduate outcomes.</description>
	<pubDate>2025-03-11</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 4, Pages 10: Unveiling the Impact of Socioeconomic and Demographic Factors on Graduate Salaries: A Machine Learning Explanatory Analytical Approach Using Higher Education Statistical Agency Data</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/4/1/10">doi: 10.3390/analytics4010010</a></p>
	<p>Authors:
		Bassey Henshaw
		Bhupesh Kumar Mishra
		William Sayers
		Zeeshan Pervez
		</p>
	<p>Graduate salaries are a significant concern for graduates, employers, and policymakers, as various factors influence them. This study investigates determinants of graduate salaries in the UK, utilising survey data from HESA (Higher Education Statistical Agency) and integrating advanced machine learning (ML) explanatory techniques with statistical analytical methodologies. By employing multi-stage analyses alongside machine learning models such as decision trees, random forests and the explainability with SHAP stands for (Shapley Additive exPanations), this study investigates the influence of 21 socioeconomic and demographic variables on graduate salary outcomes. Key variables, including institutional reputation, age at graduation, socioeconomic classification, job qualification requirements, and domicile, emerged as critical determinants, with institutional reputation proving the most significant. Among ML methods, the decision tree achieved a standout with the highest accuracy through rigorous optimisation techniques, including oversampling and undersampling. SHAP highlighted the top 12 influential variables, providing actionable insights into the interplay between individual and systemic factors. Furthermore, the statistical analysis using ANOVA (Analysis of Variance) validated the significance of these variables, revealing intricate interactions that shape graduate salary dynamics. Additionally, domain experts&amp;amp;rsquo; opinions are also analysed to authenticate the findings. This research makes a unique contribution by combining qualitative contextual analysis with quantitative methodologies, machine learning explainability and domain experts&amp;amp;rsquo; views on addressing gaps in the existing identification of graduate salary predicting components. Additionally, the findings inform policy and educational interventions to reduce wage inequalities and promote equitable career opportunities. Despite limitations, such as the UK-specific dataset and the focus on socioeconomic and demographic variables, this study lays a robust foundation for future research in predictive modelling and graduate outcomes.</p>
	]]></content:encoded>

	<dc:title>Unveiling the Impact of Socioeconomic and Demographic Factors on Graduate Salaries: A Machine Learning Explanatory Analytical Approach Using Higher Education Statistical Agency Data</dc:title>
			<dc:creator>Bassey Henshaw</dc:creator>
			<dc:creator>Bhupesh Kumar Mishra</dc:creator>
			<dc:creator>William Sayers</dc:creator>
			<dc:creator>Zeeshan Pervez</dc:creator>
		<dc:identifier>doi: 10.3390/analytics4010010</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2025-03-11</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2025-03-11</prism:publicationDate>
	<prism:volume>4</prism:volume>
	<prism:number>1</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>10</prism:startingPage>
		<prism:doi>10.3390/analytics4010010</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/4/1/10</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/4/1/9">

	<title>Analytics, Vol. 4, Pages 9: Updated Aims and Scope of Analytics</title>
	<link>https://www.mdpi.com/2813-2203/4/1/9</link>
	<description>Analytics [...]</description>
	<pubDate>2025-03-06</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 4, Pages 9: Updated Aims and Scope of Analytics</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/4/1/9">doi: 10.3390/analytics4010009</a></p>
	<p>Authors:
		Carson K. Leung
		</p>
	<p>Analytics [...]</p>
	]]></content:encoded>

	<dc:title>Updated Aims and Scope of Analytics</dc:title>
			<dc:creator>Carson K. Leung</dc:creator>
		<dc:identifier>doi: 10.3390/analytics4010009</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2025-03-06</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2025-03-06</prism:publicationDate>
	<prism:volume>4</prism:volume>
	<prism:number>1</prism:number>
	<prism:section>Editorial</prism:section>
	<prism:startingPage>9</prism:startingPage>
		<prism:doi>10.3390/analytics4010009</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/4/1/9</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/4/1/8">

	<title>Analytics, Vol. 4, Pages 8: The Role of Cognitive Performance in Older Europeans&amp;rsquo; General Health: Insights from Relative Importance Analysis</title>
	<link>https://www.mdpi.com/2813-2203/4/1/8</link>
	<description>This study explores the role of cognitive performance in the general health of older Europeans aged 50 and over, focusing on gender differences, using data from 336,500 respondents in the sixth wave of the Survey of Health, Aging, and Retirement in Europe (SHARE). Cognitive functioning was assessed through self-rated reading and writing skills, orientation in time, numeracy, memory, verbal fluency, and word-list learning. General health status was estimated by constructing a composite index of physical and mental health-related measures, including chronic diseases, mobility limitations, depressive symptoms, self-perceived health, and the Global Activity Limitation Indicator. Participants were classified into good or poor health status, and logistic regression models assessed the predictive significance of cognitive variables on general health, supplemented by a relative importance analysis to estimate relative effect sizes. The results indicated that males had a 51.1% lower risk of reporting poor health than females, and older age was associated with a 4.0% increase in the odds of reporting worse health for both genders. Memory was the strongest predictor of health status (26% of the model R2), with a greater relative contribution than the other cognitive variables. No significant gender differences were found. While this study estimates the odds of reporting poorer health in relation to gender and various cognitive characteristics, adopting a lifespan approach could provide valuable insights into the longitudinal associations between cognitive functioning and health outcomes.</description>
	<pubDate>2025-03-04</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 4, Pages 8: The Role of Cognitive Performance in Older Europeans&amp;rsquo; General Health: Insights from Relative Importance Analysis</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/4/1/8">doi: 10.3390/analytics4010008</a></p>
	<p>Authors:
		Eleni Serafetinidou
		Christina Parpoula
		</p>
	<p>This study explores the role of cognitive performance in the general health of older Europeans aged 50 and over, focusing on gender differences, using data from 336,500 respondents in the sixth wave of the Survey of Health, Aging, and Retirement in Europe (SHARE). Cognitive functioning was assessed through self-rated reading and writing skills, orientation in time, numeracy, memory, verbal fluency, and word-list learning. General health status was estimated by constructing a composite index of physical and mental health-related measures, including chronic diseases, mobility limitations, depressive symptoms, self-perceived health, and the Global Activity Limitation Indicator. Participants were classified into good or poor health status, and logistic regression models assessed the predictive significance of cognitive variables on general health, supplemented by a relative importance analysis to estimate relative effect sizes. The results indicated that males had a 51.1% lower risk of reporting poor health than females, and older age was associated with a 4.0% increase in the odds of reporting worse health for both genders. Memory was the strongest predictor of health status (26% of the model R2), with a greater relative contribution than the other cognitive variables. No significant gender differences were found. While this study estimates the odds of reporting poorer health in relation to gender and various cognitive characteristics, adopting a lifespan approach could provide valuable insights into the longitudinal associations between cognitive functioning and health outcomes.</p>
	]]></content:encoded>

	<dc:title>The Role of Cognitive Performance in Older Europeans&amp;amp;rsquo; General Health: Insights from Relative Importance Analysis</dc:title>
			<dc:creator>Eleni Serafetinidou</dc:creator>
			<dc:creator>Christina Parpoula</dc:creator>
		<dc:identifier>doi: 10.3390/analytics4010008</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2025-03-04</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2025-03-04</prism:publicationDate>
	<prism:volume>4</prism:volume>
	<prism:number>1</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>8</prism:startingPage>
		<prism:doi>10.3390/analytics4010008</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/4/1/8</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/4/1/7">

	<title>Analytics, Vol. 4, Pages 7: Towards Visual Analytics for Explainable AI in Industrial Applications</title>
	<link>https://www.mdpi.com/2813-2203/4/1/7</link>
	<description>As the levels of automation and reliance on modern artificial intelligence (AI) approaches increase across multiple industries, the importance of the human-centered perspective becomes more evident. Various actors in such industrial applications, including equipment operators and decision makers, have their needs and preferences that often do not align with the decisions produced by black-box models, potentially leading to mistrust and wasted productivity gain opportunities. In this paper, we examine these issues through the lenses of visual analytics and, more broadly, interactive visualization, and we argue that the methods and techniques from these fields can lead to advances in both academic research and industrial innovations concerning the explainability of AI models. To address the existing gap within and across the research and application fields, we propose a conceptual framework for visual analytics design and evaluation for such scenarios, followed by a preliminary roadmap and call to action for the respective communities.</description>
	<pubDate>2025-02-12</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 4, Pages 7: Towards Visual Analytics for Explainable AI in Industrial Applications</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/4/1/7">doi: 10.3390/analytics4010007</a></p>
	<p>Authors:
		Kostiantyn Kucher
		Elmira Zohrevandi
		Carl A. L. Westin
		</p>
	<p>As the levels of automation and reliance on modern artificial intelligence (AI) approaches increase across multiple industries, the importance of the human-centered perspective becomes more evident. Various actors in such industrial applications, including equipment operators and decision makers, have their needs and preferences that often do not align with the decisions produced by black-box models, potentially leading to mistrust and wasted productivity gain opportunities. In this paper, we examine these issues through the lenses of visual analytics and, more broadly, interactive visualization, and we argue that the methods and techniques from these fields can lead to advances in both academic research and industrial innovations concerning the explainability of AI models. To address the existing gap within and across the research and application fields, we propose a conceptual framework for visual analytics design and evaluation for such scenarios, followed by a preliminary roadmap and call to action for the respective communities.</p>
	]]></content:encoded>

	<dc:title>Towards Visual Analytics for Explainable AI in Industrial Applications</dc:title>
			<dc:creator>Kostiantyn Kucher</dc:creator>
			<dc:creator>Elmira Zohrevandi</dc:creator>
			<dc:creator>Carl A. L. Westin</dc:creator>
		<dc:identifier>doi: 10.3390/analytics4010007</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2025-02-12</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2025-02-12</prism:publicationDate>
	<prism:volume>4</prism:volume>
	<prism:number>1</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>7</prism:startingPage>
		<prism:doi>10.3390/analytics4010007</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/4/1/7</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/4/1/6">

	<title>Analytics, Vol. 4, Pages 6: Monetary Policy Sentiment and Its Influence on Healthcare and Technology Markets: A Transformer Model Approach</title>
	<link>https://www.mdpi.com/2813-2203/4/1/6</link>
	<description>This study investigates how the Federal Open Market Committee&amp;amp;rsquo;s (FOMC) statements impact healthcare spending, mental health trends, and stock performance in healthcare and tech sectors By analyzing FOMC&amp;amp;rsquo;s sentiment from 2018 to 2024, we found that higher sentiment correlates with increased depressive disorders (2019&amp;amp;ndash;2021) and tech stock returns, especially for the &amp;amp;ldquo;Magnificent Seven&amp;amp;rdquo; (like Apple and Amazon). Although healthcare stocks showed weaker ties to sentiment, Granger causality tests suggest some influence, hinting at ways to adjust stock strategies based on FOMC trends. These results highlight how central bank communication can shape both mental health dynamics and investment decisions in healthcare and technology.</description>
	<pubDate>2025-02-11</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 4, Pages 6: Monetary Policy Sentiment and Its Influence on Healthcare and Technology Markets: A Transformer Model Approach</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/4/1/6">doi: 10.3390/analytics4010006</a></p>
	<p>Authors:
		Dongnan Liu
		Jong-Min Kim
		</p>
	<p>This study investigates how the Federal Open Market Committee&amp;amp;rsquo;s (FOMC) statements impact healthcare spending, mental health trends, and stock performance in healthcare and tech sectors By analyzing FOMC&amp;amp;rsquo;s sentiment from 2018 to 2024, we found that higher sentiment correlates with increased depressive disorders (2019&amp;amp;ndash;2021) and tech stock returns, especially for the &amp;amp;ldquo;Magnificent Seven&amp;amp;rdquo; (like Apple and Amazon). Although healthcare stocks showed weaker ties to sentiment, Granger causality tests suggest some influence, hinting at ways to adjust stock strategies based on FOMC trends. These results highlight how central bank communication can shape both mental health dynamics and investment decisions in healthcare and technology.</p>
	]]></content:encoded>

	<dc:title>Monetary Policy Sentiment and Its Influence on Healthcare and Technology Markets: A Transformer Model Approach</dc:title>
			<dc:creator>Dongnan Liu</dc:creator>
			<dc:creator>Jong-Min Kim</dc:creator>
		<dc:identifier>doi: 10.3390/analytics4010006</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2025-02-11</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2025-02-11</prism:publicationDate>
	<prism:volume>4</prism:volume>
	<prism:number>1</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>6</prism:startingPage>
		<prism:doi>10.3390/analytics4010006</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/4/1/6</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/4/1/5">

	<title>Analytics, Vol. 4, Pages 5: A Comparative Analysis of Machine Learning and Deep Learning Techniques for Accurate Market Price Forecasting</title>
	<link>https://www.mdpi.com/2813-2203/4/1/5</link>
	<description>This study compares three machine learning and deep learning models&amp;amp;mdash;Support Vector Regression (SVR), Recurrent Neural Networks (RNN), and Long Short-Term Memory (LSTM)&amp;amp;mdash;for predicting market prices using the NGX All-Share Index dataset. The models were evaluated using multiple error metrics, including Mean Absolute Error (MAE), Mean Squared Error (MSE), Root Mean Square Error (RMSE), Mean Percentage Error (MPE), and R-squared. RNN and LSTM were tested with both 30 and 60-day windows, with performance compared to SVR. LSTM delivered better R-squared values, with a 60-day LSTM achieving the best accuracy (R-squared = 0.993) when using a combination of endogenous market data and technical indicators. SVR showed reliable results in certain scenarios but struggled in fold 2 with a sudden spike that shows a high probability of not capturing the entire underlying NGX pattern in the dataset correctly, as witnessed by the high validation loss during the period. Additionally, RNN faced the vanishing gradient problem that limits its long-term performance. Despite challenges, LSTM&amp;amp;rsquo;s ability to handle temporal dependencies, especially with the inclusion of On-Balance Volume, led to significant improvements in prediction accuracy. The use of the Optuna optimisation framework further enhanced model training and hyperparameter tuning, contributing to the performance of the LSTM model.</description>
	<pubDate>2025-02-11</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 4, Pages 5: A Comparative Analysis of Machine Learning and Deep Learning Techniques for Accurate Market Price Forecasting</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/4/1/5">doi: 10.3390/analytics4010005</a></p>
	<p>Authors:
		Olamilekan Shobayo
		Sidikat Adeyemi-Longe
		Olusogo Popoola
		Obinna Okoyeigbo
		</p>
	<p>This study compares three machine learning and deep learning models&amp;amp;mdash;Support Vector Regression (SVR), Recurrent Neural Networks (RNN), and Long Short-Term Memory (LSTM)&amp;amp;mdash;for predicting market prices using the NGX All-Share Index dataset. The models were evaluated using multiple error metrics, including Mean Absolute Error (MAE), Mean Squared Error (MSE), Root Mean Square Error (RMSE), Mean Percentage Error (MPE), and R-squared. RNN and LSTM were tested with both 30 and 60-day windows, with performance compared to SVR. LSTM delivered better R-squared values, with a 60-day LSTM achieving the best accuracy (R-squared = 0.993) when using a combination of endogenous market data and technical indicators. SVR showed reliable results in certain scenarios but struggled in fold 2 with a sudden spike that shows a high probability of not capturing the entire underlying NGX pattern in the dataset correctly, as witnessed by the high validation loss during the period. Additionally, RNN faced the vanishing gradient problem that limits its long-term performance. Despite challenges, LSTM&amp;amp;rsquo;s ability to handle temporal dependencies, especially with the inclusion of On-Balance Volume, led to significant improvements in prediction accuracy. The use of the Optuna optimisation framework further enhanced model training and hyperparameter tuning, contributing to the performance of the LSTM model.</p>
	]]></content:encoded>

	<dc:title>A Comparative Analysis of Machine Learning and Deep Learning Techniques for Accurate Market Price Forecasting</dc:title>
			<dc:creator>Olamilekan Shobayo</dc:creator>
			<dc:creator>Sidikat Adeyemi-Longe</dc:creator>
			<dc:creator>Olusogo Popoola</dc:creator>
			<dc:creator>Obinna Okoyeigbo</dc:creator>
		<dc:identifier>doi: 10.3390/analytics4010005</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2025-02-11</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2025-02-11</prism:publicationDate>
	<prism:volume>4</prism:volume>
	<prism:number>1</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>5</prism:startingPage>
		<prism:doi>10.3390/analytics4010005</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/4/1/5</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/4/1/4">

	<title>Analytics, Vol. 4, Pages 4: Personalizing Multimedia Content Recommendations for Intelligent Vehicles Through Text&amp;ndash;Image Embedding Approaches</title>
	<link>https://www.mdpi.com/2813-2203/4/1/4</link>
	<description>The ability to automate and personalize the recommendation of multimedia contents to consumers has been gaining significant attention recently. The burgeoning demand for digitization and automation of formerly analog communication processes has caught the attention of researchers and professionals alike. In light of the recent interest and anticipated transition to fully autonomous vehicles, this study proposes a text&amp;amp;ndash;image embedding method recommender system for the optimization of personalized multimedia content for in-vehicle infotainment. This study leverages existing pre-trained text embedding models and pre-trained image feature extraction methods. Previous research to date has focused mainly on textual-only or image-only analyses. By employing similarity measurements, this study demonstrates how recommendation of the most relevant multimedia content to consumers is enhanced through text&amp;amp;ndash;image embedding.</description>
	<pubDate>2025-02-05</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 4, Pages 4: Personalizing Multimedia Content Recommendations for Intelligent Vehicles Through Text&amp;ndash;Image Embedding Approaches</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/4/1/4">doi: 10.3390/analytics4010004</a></p>
	<p>Authors:
		Jin-A Choi
		Taekeun Hong
		Kiho Lim
		</p>
	<p>The ability to automate and personalize the recommendation of multimedia contents to consumers has been gaining significant attention recently. The burgeoning demand for digitization and automation of formerly analog communication processes has caught the attention of researchers and professionals alike. In light of the recent interest and anticipated transition to fully autonomous vehicles, this study proposes a text&amp;amp;ndash;image embedding method recommender system for the optimization of personalized multimedia content for in-vehicle infotainment. This study leverages existing pre-trained text embedding models and pre-trained image feature extraction methods. Previous research to date has focused mainly on textual-only or image-only analyses. By employing similarity measurements, this study demonstrates how recommendation of the most relevant multimedia content to consumers is enhanced through text&amp;amp;ndash;image embedding.</p>
	]]></content:encoded>

	<dc:title>Personalizing Multimedia Content Recommendations for Intelligent Vehicles Through Text&amp;amp;ndash;Image Embedding Approaches</dc:title>
			<dc:creator>Jin-A Choi</dc:creator>
			<dc:creator>Taekeun Hong</dc:creator>
			<dc:creator>Kiho Lim</dc:creator>
		<dc:identifier>doi: 10.3390/analytics4010004</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2025-02-05</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2025-02-05</prism:publicationDate>
	<prism:volume>4</prism:volume>
	<prism:number>1</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>4</prism:startingPage>
		<prism:doi>10.3390/analytics4010004</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/4/1/4</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/4/1/3">

	<title>Analytics, Vol. 4, Pages 3: A Fuzzy Analytical Network Process Framework for Prioritizing Competitive Intelligence in Startups</title>
	<link>https://www.mdpi.com/2813-2203/4/1/3</link>
	<description>Competitive intelligence (CI) is a critical tool for startups, enabling informed decision making through the systematic gathering and analysis of relevant information. This study aims to identify and prioritize the key factors influencing CI in startups, providing actionable insights for entrepreneurs, educators, and support organizations. Through a systematic literature review, key variables and components impacting competitive intelligence were identified. Two surveys were conducted to refine these components. The first employed a five-point Likert scale to evaluate the significance of each component, while the second used a pairwise comparison approach involving ten experts in CI and startup mentorship. Utilizing the fuzzy Analytical Network Process (ANP), this study ranked Technology Intelligence as the most critical factor, followed by market and Strategic Intelligence. Competitor Intelligence and Internet intelligence were deemed moderately important, while Organizational Intelligence ranked lowest. These findings emphasize the importance of technology-driven insights and market awareness in fostering startups&amp;amp;rsquo; competitive advantage and informed decision making. This study provides a structured framework to guide startups in prioritizing CI efforts, offering practical strategies for navigating dynamic market conditions and achieving long-term success.</description>
	<pubDate>2025-01-14</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 4, Pages 3: A Fuzzy Analytical Network Process Framework for Prioritizing Competitive Intelligence in Startups</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/4/1/3">doi: 10.3390/analytics4010003</a></p>
	<p>Authors:
		Arman Golshan
		Soheila Sardar
		Seyed Faraz Mahdavi Ardestani
		Paria Sadeghian
		</p>
	<p>Competitive intelligence (CI) is a critical tool for startups, enabling informed decision making through the systematic gathering and analysis of relevant information. This study aims to identify and prioritize the key factors influencing CI in startups, providing actionable insights for entrepreneurs, educators, and support organizations. Through a systematic literature review, key variables and components impacting competitive intelligence were identified. Two surveys were conducted to refine these components. The first employed a five-point Likert scale to evaluate the significance of each component, while the second used a pairwise comparison approach involving ten experts in CI and startup mentorship. Utilizing the fuzzy Analytical Network Process (ANP), this study ranked Technology Intelligence as the most critical factor, followed by market and Strategic Intelligence. Competitor Intelligence and Internet intelligence were deemed moderately important, while Organizational Intelligence ranked lowest. These findings emphasize the importance of technology-driven insights and market awareness in fostering startups&amp;amp;rsquo; competitive advantage and informed decision making. This study provides a structured framework to guide startups in prioritizing CI efforts, offering practical strategies for navigating dynamic market conditions and achieving long-term success.</p>
	]]></content:encoded>

	<dc:title>A Fuzzy Analytical Network Process Framework for Prioritizing Competitive Intelligence in Startups</dc:title>
			<dc:creator>Arman Golshan</dc:creator>
			<dc:creator>Soheila Sardar</dc:creator>
			<dc:creator>Seyed Faraz Mahdavi Ardestani</dc:creator>
			<dc:creator>Paria Sadeghian</dc:creator>
		<dc:identifier>doi: 10.3390/analytics4010003</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2025-01-14</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2025-01-14</prism:publicationDate>
	<prism:volume>4</prism:volume>
	<prism:number>1</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>3</prism:startingPage>
		<prism:doi>10.3390/analytics4010003</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/4/1/3</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/4/1/2">

	<title>Analytics, Vol. 4, Pages 2: Use of Hazard Functions for Determining Power-Law Behaviour in Data</title>
	<link>https://www.mdpi.com/2813-2203/4/1/2</link>
	<description>Determining the &amp;amp;lsquo;best-fitting&amp;amp;rsquo; distribution for data is an important problem in data analysis. Specifically, observing how the distribution of data changes as values below (or above) a threshold are omitted from analyses can be of use in various applications, from animal movement to the modelling of natural phenomena. Such truncated distributions, known as hazard functions, are widely studied and well understood in survival analysis, although rarely widely used in data analysis. Here, by considering the hazard and reverse-hazard functions, we demonstrate a qualitative assessment of the &amp;amp;lsquo;best-fit&amp;amp;rsquo; distribution of data. Specifically, we highlight the potential advantages of this method when determining whether power-law behaviour may or may not be present in data. Finally, we demonstrate this approach using some real-world datasets.</description>
	<pubDate>2025-01-09</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 4, Pages 2: Use of Hazard Functions for Determining Power-Law Behaviour in Data</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/4/1/2">doi: 10.3390/analytics4010002</a></p>
	<p>Authors:
		Joseph D. Bailey
		</p>
	<p>Determining the &amp;amp;lsquo;best-fitting&amp;amp;rsquo; distribution for data is an important problem in data analysis. Specifically, observing how the distribution of data changes as values below (or above) a threshold are omitted from analyses can be of use in various applications, from animal movement to the modelling of natural phenomena. Such truncated distributions, known as hazard functions, are widely studied and well understood in survival analysis, although rarely widely used in data analysis. Here, by considering the hazard and reverse-hazard functions, we demonstrate a qualitative assessment of the &amp;amp;lsquo;best-fit&amp;amp;rsquo; distribution of data. Specifically, we highlight the potential advantages of this method when determining whether power-law behaviour may or may not be present in data. Finally, we demonstrate this approach using some real-world datasets.</p>
	]]></content:encoded>

	<dc:title>Use of Hazard Functions for Determining Power-Law Behaviour in Data</dc:title>
			<dc:creator>Joseph D. Bailey</dc:creator>
		<dc:identifier>doi: 10.3390/analytics4010002</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2025-01-09</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2025-01-09</prism:publicationDate>
	<prism:volume>4</prism:volume>
	<prism:number>1</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>2</prism:startingPage>
		<prism:doi>10.3390/analytics4010002</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/4/1/2</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/4/1/1">

	<title>Analytics, Vol. 4, Pages 1: Uncovering Patterns and Trends in Big Data-Driven Research Through Text Mining of NSF Award Synopses</title>
	<link>https://www.mdpi.com/2813-2203/4/1/1</link>
	<description>The rapid expansion of big data has transformed research practices across disciplines, yet disparities exist in its adoption among U.S. institutions of higher education. This study examines trends in NSF-funded big data-driven research across research domains, institutional classifications, and directorates. Using a quantitative approach and natural language processing (NLP) techniques, we analyzed NSF awards from 2006 to 2022, focusing on seven NSF research areas: Biological Sciences, Computer and Information Science and Engineering, Engineering, Geosciences, Mathematical and Physical Sciences, Social, Behavioral and Economic Sciences, and STEM Education (formally known as Education and Human Resources). Findings indicate a significant increase in big data-related awards over time, with CISE (Computer and Information Science and Engineering) leading in funding. Machine learning and artificial intelligence are dominant themes across all institutions&amp;amp;rsquo; classifications. Results show that R1 and non-minority-serving institutions receive the majority of big data-driven research funding, though HBCUs have seen recent growth due to national diversity initiatives. Topic modeling reveals key subdomains such as cybersecurity and bioinformatics benefiting from big data, while areas like Biological Sciences and Social Sciences engage less with these methods. These findings suggest the need for broader support and funding to foster equitable adoption of big data methods across institutions and disciplines.</description>
	<pubDate>2025-01-06</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 4, Pages 1: Uncovering Patterns and Trends in Big Data-Driven Research Through Text Mining of NSF Award Synopses</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/4/1/1">doi: 10.3390/analytics4010001</a></p>
	<p>Authors:
		Arielle King
		Sayed A. Mostafa
		</p>
	<p>The rapid expansion of big data has transformed research practices across disciplines, yet disparities exist in its adoption among U.S. institutions of higher education. This study examines trends in NSF-funded big data-driven research across research domains, institutional classifications, and directorates. Using a quantitative approach and natural language processing (NLP) techniques, we analyzed NSF awards from 2006 to 2022, focusing on seven NSF research areas: Biological Sciences, Computer and Information Science and Engineering, Engineering, Geosciences, Mathematical and Physical Sciences, Social, Behavioral and Economic Sciences, and STEM Education (formally known as Education and Human Resources). Findings indicate a significant increase in big data-related awards over time, with CISE (Computer and Information Science and Engineering) leading in funding. Machine learning and artificial intelligence are dominant themes across all institutions&amp;amp;rsquo; classifications. Results show that R1 and non-minority-serving institutions receive the majority of big data-driven research funding, though HBCUs have seen recent growth due to national diversity initiatives. Topic modeling reveals key subdomains such as cybersecurity and bioinformatics benefiting from big data, while areas like Biological Sciences and Social Sciences engage less with these methods. These findings suggest the need for broader support and funding to foster equitable adoption of big data methods across institutions and disciplines.</p>
	]]></content:encoded>

	<dc:title>Uncovering Patterns and Trends in Big Data-Driven Research Through Text Mining of NSF Award Synopses</dc:title>
			<dc:creator>Arielle King</dc:creator>
			<dc:creator>Sayed A. Mostafa</dc:creator>
		<dc:identifier>doi: 10.3390/analytics4010001</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2025-01-06</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2025-01-06</prism:publicationDate>
	<prism:volume>4</prism:volume>
	<prism:number>1</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>1</prism:startingPage>
		<prism:doi>10.3390/analytics4010001</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/4/1/1</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/3/4/28">

	<title>Analytics, Vol. 3, Pages 493-507: Advancements in Predictive Maintenance: A Bibliometric Review of Diagnostic Models Using Machine Learning Techniques</title>
	<link>https://www.mdpi.com/2813-2203/3/4/28</link>
	<description>This bibliometric review investigates the advancements in machine learning techniques for predictive maintenance, focusing on the use of Artificial Neural Networks (ANNs) and Support Vector Machines (SVMs) for fault detection in wheelset axle bearings. Using data from Scopus and Web of Science, the review analyses key trends, influential publications, and significant contributions to the field from 2000 to 2024. The findings highlight the performance of ANNs in handling large datasets and modelling complex, non-linear relationships, as well as the high accuracy of SVMs in fault classification tasks, particularly with small-to-medium-sized datasets. However, the study also identifies several limitations, including the dependency on high-quality data, significant computational resource requirements, limited model adaptability, interpretability challenges, and practical implementation complexities. This review provides valuable insights for researchers and engineers, guiding the selection of appropriate diagnostic models and highlighting opportunities for future research. Addressing the identified limitations is crucial for the broader adoption and effectiveness of machine learning-based predictive maintenance strategies across various industrial contexts.</description>
	<pubDate>2024-12-10</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 3, Pages 493-507: Advancements in Predictive Maintenance: A Bibliometric Review of Diagnostic Models Using Machine Learning Techniques</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/3/4/28">doi: 10.3390/analytics3040028</a></p>
	<p>Authors:
		Nontuthuzelo Lindokuhle Vithi
		Colin Chibaya
		</p>
	<p>This bibliometric review investigates the advancements in machine learning techniques for predictive maintenance, focusing on the use of Artificial Neural Networks (ANNs) and Support Vector Machines (SVMs) for fault detection in wheelset axle bearings. Using data from Scopus and Web of Science, the review analyses key trends, influential publications, and significant contributions to the field from 2000 to 2024. The findings highlight the performance of ANNs in handling large datasets and modelling complex, non-linear relationships, as well as the high accuracy of SVMs in fault classification tasks, particularly with small-to-medium-sized datasets. However, the study also identifies several limitations, including the dependency on high-quality data, significant computational resource requirements, limited model adaptability, interpretability challenges, and practical implementation complexities. This review provides valuable insights for researchers and engineers, guiding the selection of appropriate diagnostic models and highlighting opportunities for future research. Addressing the identified limitations is crucial for the broader adoption and effectiveness of machine learning-based predictive maintenance strategies across various industrial contexts.</p>
	]]></content:encoded>

	<dc:title>Advancements in Predictive Maintenance: A Bibliometric Review of Diagnostic Models Using Machine Learning Techniques</dc:title>
			<dc:creator>Nontuthuzelo Lindokuhle Vithi</dc:creator>
			<dc:creator>Colin Chibaya</dc:creator>
		<dc:identifier>doi: 10.3390/analytics3040028</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2024-12-10</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2024-12-10</prism:publicationDate>
	<prism:volume>3</prism:volume>
	<prism:number>4</prism:number>
	<prism:section>Review</prism:section>
	<prism:startingPage>493</prism:startingPage>
		<prism:doi>10.3390/analytics3040028</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/3/4/28</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/3/4/27">

	<title>Analytics, Vol. 3, Pages 476-492: NPI-WGNN: A Weighted Graph Neural Network Leveraging Centrality Measures and High-Order Common Neighbor Similarity for Accurate ncRNA&amp;ndash;Protein Interaction Prediction</title>
	<link>https://www.mdpi.com/2813-2203/3/4/27</link>
	<description>Predicting ncRNA&amp;amp;ndash;protein interactions (NPIs) is essential for understanding regulatory roles in cellular processes and disease mechanisms, yet experimental methods are costly and time-consuming. In this study, we propose NPI-WGNN, a novel weighted graph neural network model designed to enhance NPI prediction by incorporating topological insights from graph structures. Our approach introduces a bipartite version of the high-order common neighbor (HOCN) similarity metric to assign edge weights in an ncRNA&amp;amp;ndash;protein network, refining node embeddings via weighted node2vec. We further enrich these embeddings with centrality measures, such as degree and Katz centralities, to capture network hierarchy and connectivity. To optimize prediction accuracy, we employ a hybrid GNN architecture that combines graph convolutional network (GCN), graph attention network (GAT), and GraphSAGE layers, each contributing unique advantages: GraphSAGE offers scalability, GCN provides a global structural perspective, and GAT applies dynamic neighbor weighting. An ablation study confirms the complementary strengths of these layers, showing that their integration improves predictive accuracy and robustness across varied graph complexities. Experimental results on three benchmark datasets demonstrate that NPI-WGNN outperforms state-of-the-art methods, achieving up to 96.1% accuracy, 97.5% sensitivity, and an F1-score of 0.96, positioning it as a robust and accurate framework for ncRNA&amp;amp;ndash;protein interaction prediction.</description>
	<pubDate>2024-12-02</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 3, Pages 476-492: NPI-WGNN: A Weighted Graph Neural Network Leveraging Centrality Measures and High-Order Common Neighbor Similarity for Accurate ncRNA&amp;ndash;Protein Interaction Prediction</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/3/4/27">doi: 10.3390/analytics3040027</a></p>
	<p>Authors:
		Fatemeh Khoushehgir
		Zahra Noshad
		Morteza Noshad
		Sadegh Sulaimany
		</p>
	<p>Predicting ncRNA&amp;amp;ndash;protein interactions (NPIs) is essential for understanding regulatory roles in cellular processes and disease mechanisms, yet experimental methods are costly and time-consuming. In this study, we propose NPI-WGNN, a novel weighted graph neural network model designed to enhance NPI prediction by incorporating topological insights from graph structures. Our approach introduces a bipartite version of the high-order common neighbor (HOCN) similarity metric to assign edge weights in an ncRNA&amp;amp;ndash;protein network, refining node embeddings via weighted node2vec. We further enrich these embeddings with centrality measures, such as degree and Katz centralities, to capture network hierarchy and connectivity. To optimize prediction accuracy, we employ a hybrid GNN architecture that combines graph convolutional network (GCN), graph attention network (GAT), and GraphSAGE layers, each contributing unique advantages: GraphSAGE offers scalability, GCN provides a global structural perspective, and GAT applies dynamic neighbor weighting. An ablation study confirms the complementary strengths of these layers, showing that their integration improves predictive accuracy and robustness across varied graph complexities. Experimental results on three benchmark datasets demonstrate that NPI-WGNN outperforms state-of-the-art methods, achieving up to 96.1% accuracy, 97.5% sensitivity, and an F1-score of 0.96, positioning it as a robust and accurate framework for ncRNA&amp;amp;ndash;protein interaction prediction.</p>
	]]></content:encoded>

	<dc:title>NPI-WGNN: A Weighted Graph Neural Network Leveraging Centrality Measures and High-Order Common Neighbor Similarity for Accurate ncRNA&amp;amp;ndash;Protein Interaction Prediction</dc:title>
			<dc:creator>Fatemeh Khoushehgir</dc:creator>
			<dc:creator>Zahra Noshad</dc:creator>
			<dc:creator>Morteza Noshad</dc:creator>
			<dc:creator>Sadegh Sulaimany</dc:creator>
		<dc:identifier>doi: 10.3390/analytics3040027</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2024-12-02</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2024-12-02</prism:publicationDate>
	<prism:volume>3</prism:volume>
	<prism:number>4</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>476</prism:startingPage>
		<prism:doi>10.3390/analytics3040027</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/3/4/27</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/3/4/26">

	<title>Analytics, Vol. 3, Pages 461-475: Breast Cancer Classification Using Fine-Tuned SWIN Transformer Model on Mammographic Images</title>
	<link>https://www.mdpi.com/2813-2203/3/4/26</link>
	<description>Breast cancer is the most prevalent type of disease among women. It has become one of the foremost causes of death among women globally. Early detection plays a significant role in administering personalized treatment and improving patient outcomes. Mammography procedures are often used to detect early-stage cancer cells. This traditional method of mammography while valuable has limitations in its potential for false positives and negatives, patient discomfort, and radiation exposure. Therefore, there is a probe for more accurate techniques required in detecting breast cancer, leading to exploring the potential of machine learning in the classification of diagnostic images due to its efficiency and accuracy. This study conducted a comparative analysis of pre-trained CNNs (ResNet50 and VGG16) and vision transformers (ViT-base and SWIN transformer) with the inclusion of ViT-base trained from scratch model architectures to effectively classify mammographic breast cancer images into benign and malignant cases. The SWIN transformer exhibits superior performance with 99.9% accuracy and a precision of 99.8%. These findings demonstrate the efficiency of deep learning to accurately classify mammographic breast cancer images for the diagnosis of breast cancer, leading to improvements in patient outcomes.</description>
	<pubDate>2024-11-11</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 3, Pages 461-475: Breast Cancer Classification Using Fine-Tuned SWIN Transformer Model on Mammographic Images</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/3/4/26">doi: 10.3390/analytics3040026</a></p>
	<p>Authors:
		Oluwatosin Tanimola
		Olamilekan Shobayo
		Olusogo Popoola
		Obinna Okoyeigbo
		</p>
	<p>Breast cancer is the most prevalent type of disease among women. It has become one of the foremost causes of death among women globally. Early detection plays a significant role in administering personalized treatment and improving patient outcomes. Mammography procedures are often used to detect early-stage cancer cells. This traditional method of mammography while valuable has limitations in its potential for false positives and negatives, patient discomfort, and radiation exposure. Therefore, there is a probe for more accurate techniques required in detecting breast cancer, leading to exploring the potential of machine learning in the classification of diagnostic images due to its efficiency and accuracy. This study conducted a comparative analysis of pre-trained CNNs (ResNet50 and VGG16) and vision transformers (ViT-base and SWIN transformer) with the inclusion of ViT-base trained from scratch model architectures to effectively classify mammographic breast cancer images into benign and malignant cases. The SWIN transformer exhibits superior performance with 99.9% accuracy and a precision of 99.8%. These findings demonstrate the efficiency of deep learning to accurately classify mammographic breast cancer images for the diagnosis of breast cancer, leading to improvements in patient outcomes.</p>
	]]></content:encoded>

	<dc:title>Breast Cancer Classification Using Fine-Tuned SWIN Transformer Model on Mammographic Images</dc:title>
			<dc:creator>Oluwatosin Tanimola</dc:creator>
			<dc:creator>Olamilekan Shobayo</dc:creator>
			<dc:creator>Olusogo Popoola</dc:creator>
			<dc:creator>Obinna Okoyeigbo</dc:creator>
		<dc:identifier>doi: 10.3390/analytics3040026</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2024-11-11</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2024-11-11</prism:publicationDate>
	<prism:volume>3</prism:volume>
	<prism:number>4</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>461</prism:startingPage>
		<prism:doi>10.3390/analytics3040026</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/3/4/26</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/3/4/25">

	<title>Analytics, Vol. 3, Pages 449-460: Modified Bayesian Information Criterion for Item Response Models in Planned Missingness Test Designs</title>
	<link>https://www.mdpi.com/2813-2203/3/4/25</link>
	<description>The Bayesian information criterion (BIC) is a widely used statistical tool originally derived for fully observed data. The BIC formula includes the sample size and the number of estimated parameters in the penalty term. However, not all variables are available for every subject in planned missingness designs. This article demonstrates that a modified BIC, tailored for planned missingness designs, outperforms the original BIC. The modification adjusts the penalty term by using the average number of estimable parameters per subject rather than the total number of model parameters. This new criterion was successfully applied to item response theory models in two simulation studies. We recommend that future studies utilizing planned missingness designs adopt the modified BIC formula proposed here.</description>
	<pubDate>2024-11-08</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 3, Pages 449-460: Modified Bayesian Information Criterion for Item Response Models in Planned Missingness Test Designs</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/3/4/25">doi: 10.3390/analytics3040025</a></p>
	<p>Authors:
		Alexander Robitzsch
		</p>
	<p>The Bayesian information criterion (BIC) is a widely used statistical tool originally derived for fully observed data. The BIC formula includes the sample size and the number of estimated parameters in the penalty term. However, not all variables are available for every subject in planned missingness designs. This article demonstrates that a modified BIC, tailored for planned missingness designs, outperforms the original BIC. The modification adjusts the penalty term by using the average number of estimable parameters per subject rather than the total number of model parameters. This new criterion was successfully applied to item response theory models in two simulation studies. We recommend that future studies utilizing planned missingness designs adopt the modified BIC formula proposed here.</p>
	]]></content:encoded>

	<dc:title>Modified Bayesian Information Criterion for Item Response Models in Planned Missingness Test Designs</dc:title>
			<dc:creator>Alexander Robitzsch</dc:creator>
		<dc:identifier>doi: 10.3390/analytics3040025</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2024-11-08</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2024-11-08</prism:publicationDate>
	<prism:volume>3</prism:volume>
	<prism:number>4</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>449</prism:startingPage>
		<prism:doi>10.3390/analytics3040025</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/3/4/25</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/3/4/24">

	<title>Analytics, Vol. 3, Pages 439-448: Adaptive Weighted Multiview Kernel Matrix Factorization and Its Application in Alzheimer&amp;rsquo;s Disease Analysis</title>
	<link>https://www.mdpi.com/2813-2203/3/4/24</link>
	<description>Recent technology and equipment advancements have provided us with opportunities to better analyze Alzheimer&amp;amp;rsquo;s disease (AD), where we could collect and employ the data from different image and genetic modalities that may potentially enhance the predictive performance. To perform better clustering in AD analysis, in this paper, we propose a novel model to leverage data from all different modalities/views, which can learn the weights of each view adaptively. Different from previous vanilla Non-negative matrix factorization which assumes data is linearly separable, we propose a simple yet efficient method based on kernel matrix factorization, which is not only able to deal with non-linear data structure but also can achieve better prediction accuracy. Experimental results on the ADNI dataset demonstrate the effectiveness of our proposed method, which indicates promising prospects for kernel application in AD analysis.</description>
	<pubDate>2024-11-04</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 3, Pages 439-448: Adaptive Weighted Multiview Kernel Matrix Factorization and Its Application in Alzheimer&amp;rsquo;s Disease Analysis</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/3/4/24">doi: 10.3390/analytics3040024</a></p>
	<p>Authors:
		Yarui Cao
		Kai Liu
		</p>
	<p>Recent technology and equipment advancements have provided us with opportunities to better analyze Alzheimer&amp;amp;rsquo;s disease (AD), where we could collect and employ the data from different image and genetic modalities that may potentially enhance the predictive performance. To perform better clustering in AD analysis, in this paper, we propose a novel model to leverage data from all different modalities/views, which can learn the weights of each view adaptively. Different from previous vanilla Non-negative matrix factorization which assumes data is linearly separable, we propose a simple yet efficient method based on kernel matrix factorization, which is not only able to deal with non-linear data structure but also can achieve better prediction accuracy. Experimental results on the ADNI dataset demonstrate the effectiveness of our proposed method, which indicates promising prospects for kernel application in AD analysis.</p>
	]]></content:encoded>

	<dc:title>Adaptive Weighted Multiview Kernel Matrix Factorization and Its Application in Alzheimer&amp;amp;rsquo;s Disease Analysis</dc:title>
			<dc:creator>Yarui Cao</dc:creator>
			<dc:creator>Kai Liu</dc:creator>
		<dc:identifier>doi: 10.3390/analytics3040024</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2024-11-04</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2024-11-04</prism:publicationDate>
	<prism:volume>3</prism:volume>
	<prism:number>4</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>439</prism:startingPage>
		<prism:doi>10.3390/analytics3040024</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/3/4/24</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/3/4/23">

	<title>Analytics, Vol. 3, Pages 425-438: Electric Vehicle Sentiment Analysis Using Large Language Models</title>
	<link>https://www.mdpi.com/2813-2203/3/4/23</link>
	<description>Sentiment analysis is a technique used to understand the public&amp;amp;rsquo;s opinion towards an event, product, or organization. For example, sentiment analysis can be used to understand positive or negative opinions or attitudes towards electric vehicle (EV) brands. This provides companies with valuable insight into the public&amp;amp;rsquo;s opinion of their products and brands. In the field of natural language processing (NLP), transformer models have shown great performance compared to traditional machine learning algorithms. However, these models have not been explored extensively in the EV domain. EV companies are becoming significant competitors in the automotive industry and are projected to cover up to 30% of the United States light vehicle market by 2030 In this study, we present a comparative study of large language models (LLMs) including bidirectional encoder representations from transformers (BERT), robustly optimised BERT (RoBERTa), and a generalised autoregressive pre-training method (XLNet) using Lucid Motors and Tesla Motors YouTube datasets. Results evidenced that LLMs like BERT and her variants are off-the-shelf algorithms for sentiment analysis, specifically when fine-tuned. Furthermore, our findings present the need for domain adaptation whilst utilizing LLMs. Finally, the experimental results showed that RoBERTa achieved consistent performance across the EV datasets with an F1 score of at least 92%.</description>
	<pubDate>2024-11-01</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 3, Pages 425-438: Electric Vehicle Sentiment Analysis Using Large Language Models</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/3/4/23">doi: 10.3390/analytics3040023</a></p>
	<p>Authors:
		Hemlata Sharma
		Faiz Ud Din
		Bayode Ogunleye
		</p>
	<p>Sentiment analysis is a technique used to understand the public&amp;amp;rsquo;s opinion towards an event, product, or organization. For example, sentiment analysis can be used to understand positive or negative opinions or attitudes towards electric vehicle (EV) brands. This provides companies with valuable insight into the public&amp;amp;rsquo;s opinion of their products and brands. In the field of natural language processing (NLP), transformer models have shown great performance compared to traditional machine learning algorithms. However, these models have not been explored extensively in the EV domain. EV companies are becoming significant competitors in the automotive industry and are projected to cover up to 30% of the United States light vehicle market by 2030 In this study, we present a comparative study of large language models (LLMs) including bidirectional encoder representations from transformers (BERT), robustly optimised BERT (RoBERTa), and a generalised autoregressive pre-training method (XLNet) using Lucid Motors and Tesla Motors YouTube datasets. Results evidenced that LLMs like BERT and her variants are off-the-shelf algorithms for sentiment analysis, specifically when fine-tuned. Furthermore, our findings present the need for domain adaptation whilst utilizing LLMs. Finally, the experimental results showed that RoBERTa achieved consistent performance across the EV datasets with an F1 score of at least 92%.</p>
	]]></content:encoded>

	<dc:title>Electric Vehicle Sentiment Analysis Using Large Language Models</dc:title>
			<dc:creator>Hemlata Sharma</dc:creator>
			<dc:creator>Faiz Ud Din</dc:creator>
			<dc:creator>Bayode Ogunleye</dc:creator>
		<dc:identifier>doi: 10.3390/analytics3040023</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2024-11-01</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2024-11-01</prism:publicationDate>
	<prism:volume>3</prism:volume>
	<prism:number>4</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>425</prism:startingPage>
		<prism:doi>10.3390/analytics3040023</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/3/4/23</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/3/4/22">

	<title>Analytics, Vol. 3, Pages 406-424: The Analyst&amp;rsquo;s Hierarchy of Needs: Grounded Design Principles for Tailored Intelligence Analysis Tools</title>
	<link>https://www.mdpi.com/2813-2203/3/4/22</link>
	<description>Intelligence analysis involves gathering, analyzing, and interpreting vast amounts of information from diverse sources to generate accurate and timely insights. Tailored tools hold great promise in providing individualized support, enhancing efficiency, and facilitating the identification of crucial intelligence gaps and trends where traditional tools fail. The effectiveness of tailored tools depends on an analyst&amp;amp;rsquo;s unique needs and motivations, as well as the broader context in which they operate. This paper describes a series of focus discovery exercises that revealed a distinct hierarchy of needs for intelligence analysts. This reflection on the balance between competing needs is of particular value in the context of intelligence analysis, where the compartmentalization required for security can make it difficult to group design patterns in stakeholder values. We hope that this study will enable the development of more effective tools, supporting the well-being and performance of intelligence analysts as well as the organizations they serve.</description>
	<pubDate>2024-10-29</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 3, Pages 406-424: The Analyst&amp;rsquo;s Hierarchy of Needs: Grounded Design Principles for Tailored Intelligence Analysis Tools</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/3/4/22">doi: 10.3390/analytics3040022</a></p>
	<p>Authors:
		Antonio E. Girona
		James C. Peters
		Wenyuan Wang
		R. Jordan Crouser
		</p>
	<p>Intelligence analysis involves gathering, analyzing, and interpreting vast amounts of information from diverse sources to generate accurate and timely insights. Tailored tools hold great promise in providing individualized support, enhancing efficiency, and facilitating the identification of crucial intelligence gaps and trends where traditional tools fail. The effectiveness of tailored tools depends on an analyst&amp;amp;rsquo;s unique needs and motivations, as well as the broader context in which they operate. This paper describes a series of focus discovery exercises that revealed a distinct hierarchy of needs for intelligence analysts. This reflection on the balance between competing needs is of particular value in the context of intelligence analysis, where the compartmentalization required for security can make it difficult to group design patterns in stakeholder values. We hope that this study will enable the development of more effective tools, supporting the well-being and performance of intelligence analysts as well as the organizations they serve.</p>
	]]></content:encoded>

	<dc:title>The Analyst&amp;amp;rsquo;s Hierarchy of Needs: Grounded Design Principles for Tailored Intelligence Analysis Tools</dc:title>
			<dc:creator>Antonio E. Girona</dc:creator>
			<dc:creator>James C. Peters</dc:creator>
			<dc:creator>Wenyuan Wang</dc:creator>
			<dc:creator>R. Jordan Crouser</dc:creator>
		<dc:identifier>doi: 10.3390/analytics3040022</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2024-10-29</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2024-10-29</prism:publicationDate>
	<prism:volume>3</prism:volume>
	<prism:number>4</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>406</prism:startingPage>
		<prism:doi>10.3390/analytics3040022</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/3/4/22</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/3/3/21">

	<title>Analytics, Vol. 3, Pages 389-405: Directed Topic Extraction with Side Information for Sustainability Analysis</title>
	<link>https://www.mdpi.com/2813-2203/3/3/21</link>
	<description>Topic analysis represents each document in a text corpus in a low-dimensional latent topic space. In some cases, the desired topic representation is subject to specific requirements or guidelines constituting side information. For instance, sustainability-aware investors might be interested in automatically assessing aspects of firm sustainability based on the textual content of its corporate reports, focusing on the established 17 UN sustainability goals. The main corpus consists of the corporate report texts, while the texts containing the definitions of the 17 UN sustainability goals represent the side information. Under the assumption that both text corpora share a common low-dimensional subspace, we propose representing them in such a space via directed topic extraction using matrix co-factorization. Both the main and the side text corpora are first represented as term&amp;amp;ndash;context matrices, which are then jointly decomposed into word&amp;amp;ndash;topic and topic&amp;amp;ndash;context matrices. The word&amp;amp;ndash;topic matrix is common to both text corpora, whereas the topic&amp;amp;ndash;context matrices contain specific representations in the shared topic space. A nuisance parameter, which allows us to shift the focus between the error minimization of individual factorization terms, controls the extent to which the side information is taken into account. With our approach, documents from the main and the side corpora can be related to each other in the resulting latent topic space. That is, the corporate reports are represented in the same latent topic space as the descriptions of the 17 UN sustainability goals, enabling a structured automatic sustainability assessment of the textual report&amp;amp;rsquo;s content. We provide an algorithm for such directed topic extraction and propose techniques for visualizing and interpreting the results.</description>
	<pubDate>2024-09-11</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 3, Pages 389-405: Directed Topic Extraction with Side Information for Sustainability Analysis</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/3/3/21">doi: 10.3390/analytics3030021</a></p>
	<p>Authors:
		Maria Osipenko
		</p>
	<p>Topic analysis represents each document in a text corpus in a low-dimensional latent topic space. In some cases, the desired topic representation is subject to specific requirements or guidelines constituting side information. For instance, sustainability-aware investors might be interested in automatically assessing aspects of firm sustainability based on the textual content of its corporate reports, focusing on the established 17 UN sustainability goals. The main corpus consists of the corporate report texts, while the texts containing the definitions of the 17 UN sustainability goals represent the side information. Under the assumption that both text corpora share a common low-dimensional subspace, we propose representing them in such a space via directed topic extraction using matrix co-factorization. Both the main and the side text corpora are first represented as term&amp;amp;ndash;context matrices, which are then jointly decomposed into word&amp;amp;ndash;topic and topic&amp;amp;ndash;context matrices. The word&amp;amp;ndash;topic matrix is common to both text corpora, whereas the topic&amp;amp;ndash;context matrices contain specific representations in the shared topic space. A nuisance parameter, which allows us to shift the focus between the error minimization of individual factorization terms, controls the extent to which the side information is taken into account. With our approach, documents from the main and the side corpora can be related to each other in the resulting latent topic space. That is, the corporate reports are represented in the same latent topic space as the descriptions of the 17 UN sustainability goals, enabling a structured automatic sustainability assessment of the textual report&amp;amp;rsquo;s content. We provide an algorithm for such directed topic extraction and propose techniques for visualizing and interpreting the results.</p>
	]]></content:encoded>

	<dc:title>Directed Topic Extraction with Side Information for Sustainability Analysis</dc:title>
			<dc:creator>Maria Osipenko</dc:creator>
		<dc:identifier>doi: 10.3390/analytics3030021</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2024-09-11</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2024-09-11</prism:publicationDate>
	<prism:volume>3</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>389</prism:startingPage>
		<prism:doi>10.3390/analytics3030021</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/3/3/21</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/3/3/20">

	<title>Analytics, Vol. 3, Pages 368-388: SIMEX-Based and Analytical Bias Corrections in Stocking&amp;ndash;Lord Linking</title>
	<link>https://www.mdpi.com/2813-2203/3/3/20</link>
	<description>Stocking&amp;amp;ndash;Lord (SL) linking is a popular linking method for group comparisons based on dichotomous item responses. This article proposes a bias correction technique based on the simulation extrapolation (SIMEX) method for SL linking in the 2PL model in the presence of uniform differential item functioning (DIF). The SIMEX-based method is compared to the analytical bias correction methods of SL linking. It turned out in a simulation study that SIMEX-based SL linking performed best, is easy to implement, and can be adapted to other linking methods straightforwardly.</description>
	<pubDate>2024-08-06</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 3, Pages 368-388: SIMEX-Based and Analytical Bias Corrections in Stocking&amp;ndash;Lord Linking</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/3/3/20">doi: 10.3390/analytics3030020</a></p>
	<p>Authors:
		Alexander Robitzsch
		</p>
	<p>Stocking&amp;amp;ndash;Lord (SL) linking is a popular linking method for group comparisons based on dichotomous item responses. This article proposes a bias correction technique based on the simulation extrapolation (SIMEX) method for SL linking in the 2PL model in the presence of uniform differential item functioning (DIF). The SIMEX-based method is compared to the analytical bias correction methods of SL linking. It turned out in a simulation study that SIMEX-based SL linking performed best, is easy to implement, and can be adapted to other linking methods straightforwardly.</p>
	]]></content:encoded>

	<dc:title>SIMEX-Based and Analytical Bias Corrections in Stocking&amp;amp;ndash;Lord Linking</dc:title>
			<dc:creator>Alexander Robitzsch</dc:creator>
		<dc:identifier>doi: 10.3390/analytics3030020</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2024-08-06</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2024-08-06</prism:publicationDate>
	<prism:volume>3</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>368</prism:startingPage>
		<prism:doi>10.3390/analytics3030020</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/3/3/20</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/3/3/19">

	<title>Analytics, Vol. 3, Pages 344-367: Comparative Analysis of Nature-Inspired Metaheuristic Techniques for Optimizing Phishing Website Detection</title>
	<link>https://www.mdpi.com/2813-2203/3/3/19</link>
	<description>The increasing number, frequency, and sophistication of phishing website-based attacks necessitate the development of robust solutions for detecting phishing websites to enhance the overall security of cyberspace. Drawing inspiration from natural processes, nature-inspired metaheuristic techniques have been proven to be efficient in solving complex optimization problems in diverse domains. Following these successes, this research paper aims to investigate the effectiveness of metaheuristic techniques, particularly Genetic Algorithms (GAs), Differential Evolution (DE), and Particle Swarm Optimization (PSO), in optimizing the hyperparameters of machine learning (ML) algorithms for detecting phishing websites. Using multiple datasets, six ensemble classifiers were trained on each dataset and their hyperparameters were optimized using each metaheuristic technique. As a baseline for assessing performance improvement, the classifiers were also trained with the default hyperparameters. To validate the genuine impact of the techniques over the use of default hyperparameters, we conducted statistical tests on the accuracy scores of all the optimized classifiers. The results show that the GA is the most effective technique, by improving the accuracy scores of all the classifiers, followed by DE, which improved four of the six classifiers. PSO was the least effective, improving only one classifier. It was also found that GA-optimized Gradient Boosting, LGBM and XGBoost were the best classifiers across all the metrics in predicting phishing websites, achieving peak accuracy scores of 98.98%, 99.24%, and 99.47%, respectively.</description>
	<pubDate>2024-08-06</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 3, Pages 344-367: Comparative Analysis of Nature-Inspired Metaheuristic Techniques for Optimizing Phishing Website Detection</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/3/3/19">doi: 10.3390/analytics3030019</a></p>
	<p>Authors:
		Thomas Nagunwa
		</p>
	<p>The increasing number, frequency, and sophistication of phishing website-based attacks necessitate the development of robust solutions for detecting phishing websites to enhance the overall security of cyberspace. Drawing inspiration from natural processes, nature-inspired metaheuristic techniques have been proven to be efficient in solving complex optimization problems in diverse domains. Following these successes, this research paper aims to investigate the effectiveness of metaheuristic techniques, particularly Genetic Algorithms (GAs), Differential Evolution (DE), and Particle Swarm Optimization (PSO), in optimizing the hyperparameters of machine learning (ML) algorithms for detecting phishing websites. Using multiple datasets, six ensemble classifiers were trained on each dataset and their hyperparameters were optimized using each metaheuristic technique. As a baseline for assessing performance improvement, the classifiers were also trained with the default hyperparameters. To validate the genuine impact of the techniques over the use of default hyperparameters, we conducted statistical tests on the accuracy scores of all the optimized classifiers. The results show that the GA is the most effective technique, by improving the accuracy scores of all the classifiers, followed by DE, which improved four of the six classifiers. PSO was the least effective, improving only one classifier. It was also found that GA-optimized Gradient Boosting, LGBM and XGBoost were the best classifiers across all the metrics in predicting phishing websites, achieving peak accuracy scores of 98.98%, 99.24%, and 99.47%, respectively.</p>
	]]></content:encoded>

	<dc:title>Comparative Analysis of Nature-Inspired Metaheuristic Techniques for Optimizing Phishing Website Detection</dc:title>
			<dc:creator>Thomas Nagunwa</dc:creator>
		<dc:identifier>doi: 10.3390/analytics3030019</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2024-08-06</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2024-08-06</prism:publicationDate>
	<prism:volume>3</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>344</prism:startingPage>
		<prism:doi>10.3390/analytics3030019</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/3/3/19</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/3/3/18">

	<title>Analytics, Vol. 3, Pages 318-343: A Longitudinal Tree-Based Framework for Lapse Management in Life Insurance</title>
	<link>https://www.mdpi.com/2813-2203/3/3/18</link>
	<description>Developing an informed lapse management strategy (LMS) is critical for life insurers to improve profitability and gain insight into the risk of their global portfolio. Prior research in actuarial science has shown that targeting policyholders by maximising their individual customer lifetime value is more advantageous than targeting all those likely to lapse. However, most existing lapse analyses do not leverage the variability of features and targets over time. We propose a longitudinal LMS framework, utilising tree-based models for longitudinal data, such as left-truncated and right-censored (LTRC) trees and forests, as well as mixed-effect tree-based models. Our methodology provides time-informed insights, leading to increased precision in targeting. Our findings indicate that the use of longitudinally structured data significantly enhances the precision of models in predicting lapse behaviour, estimating customer lifetime value, and evaluating individual retention gains. The implementation of mixed-effect random forests enables the production of time-varying predictions that are highly relevant for decision-making. This paper contributes to the field of lapse analysis for life insurers by demonstrating the importance of exploiting the complete past trajectory of policyholders, which is often available in insurers&amp;amp;rsquo; information systems but has yet to be fully utilised.</description>
	<pubDate>2024-08-05</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 3, Pages 318-343: A Longitudinal Tree-Based Framework for Lapse Management in Life Insurance</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/3/3/18">doi: 10.3390/analytics3030018</a></p>
	<p>Authors:
		Mathias Valla
		</p>
	<p>Developing an informed lapse management strategy (LMS) is critical for life insurers to improve profitability and gain insight into the risk of their global portfolio. Prior research in actuarial science has shown that targeting policyholders by maximising their individual customer lifetime value is more advantageous than targeting all those likely to lapse. However, most existing lapse analyses do not leverage the variability of features and targets over time. We propose a longitudinal LMS framework, utilising tree-based models for longitudinal data, such as left-truncated and right-censored (LTRC) trees and forests, as well as mixed-effect tree-based models. Our methodology provides time-informed insights, leading to increased precision in targeting. Our findings indicate that the use of longitudinally structured data significantly enhances the precision of models in predicting lapse behaviour, estimating customer lifetime value, and evaluating individual retention gains. The implementation of mixed-effect random forests enables the production of time-varying predictions that are highly relevant for decision-making. This paper contributes to the field of lapse analysis for life insurers by demonstrating the importance of exploiting the complete past trajectory of policyholders, which is often available in insurers&amp;amp;rsquo; information systems but has yet to be fully utilised.</p>
	]]></content:encoded>

	<dc:title>A Longitudinal Tree-Based Framework for Lapse Management in Life Insurance</dc:title>
			<dc:creator>Mathias Valla</dc:creator>
		<dc:identifier>doi: 10.3390/analytics3030018</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2024-08-05</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2024-08-05</prism:publicationDate>
	<prism:volume>3</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>318</prism:startingPage>
		<prism:doi>10.3390/analytics3030018</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/3/3/18</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/3/3/17">

	<title>Analytics, Vol. 3, Pages 297-317: Enhancing Talent Recruitment in Business Intelligence Systems: A Comparative Analysis of Machine Learning Models</title>
	<link>https://www.mdpi.com/2813-2203/3/3/17</link>
	<description>In the competitive field of business intelligence, optimizing talent recruitment through data-driven methodologies is crucial for better decision-making. This study compares the effectiveness of various machine learning models to improve recruitment accuracy and efficiency. Using the recruitment data from a major Yemeni organization (2019&amp;amp;ndash;2022), we evaluated models including K-Nearest Neighbors, Logistic Regression, Support Vector Machine, Naive Bayes, Decision Trees, Random Forest, Gradient Boosting Classifier, AdaBoost Classifier, and Neural Networks. Hyperparameter tuning and cross-validation were used for optimization. The Random Forest model achieved the highest accuracy (92.8%), followed by Neural Networks (92.6%) and Gradient Boosting Classifier (92.5%). These results suggest that advanced machine learning models, particularly Random Forest and Neural Networks, can significantly enhance the recruitment processes in business intelligence systems. This study provides valuable insights for recruiters, advocating for the integration of sophisticated machine learning techniques in talent acquisition strategies.</description>
	<pubDate>2024-07-15</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 3, Pages 297-317: Enhancing Talent Recruitment in Business Intelligence Systems: A Comparative Analysis of Machine Learning Models</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/3/3/17">doi: 10.3390/analytics3030017</a></p>
	<p>Authors:
		Hikmat Al-Quhfa
		Ali Mothana
		Abdussalam Aljbri
		Jie Song
		</p>
	<p>In the competitive field of business intelligence, optimizing talent recruitment through data-driven methodologies is crucial for better decision-making. This study compares the effectiveness of various machine learning models to improve recruitment accuracy and efficiency. Using the recruitment data from a major Yemeni organization (2019&amp;amp;ndash;2022), we evaluated models including K-Nearest Neighbors, Logistic Regression, Support Vector Machine, Naive Bayes, Decision Trees, Random Forest, Gradient Boosting Classifier, AdaBoost Classifier, and Neural Networks. Hyperparameter tuning and cross-validation were used for optimization. The Random Forest model achieved the highest accuracy (92.8%), followed by Neural Networks (92.6%) and Gradient Boosting Classifier (92.5%). These results suggest that advanced machine learning models, particularly Random Forest and Neural Networks, can significantly enhance the recruitment processes in business intelligence systems. This study provides valuable insights for recruiters, advocating for the integration of sophisticated machine learning techniques in talent acquisition strategies.</p>
	]]></content:encoded>

	<dc:title>Enhancing Talent Recruitment in Business Intelligence Systems: A Comparative Analysis of Machine Learning Models</dc:title>
			<dc:creator>Hikmat Al-Quhfa</dc:creator>
			<dc:creator>Ali Mothana</dc:creator>
			<dc:creator>Abdussalam Aljbri</dc:creator>
			<dc:creator>Jie Song</dc:creator>
		<dc:identifier>doi: 10.3390/analytics3030017</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2024-07-15</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2024-07-15</prism:publicationDate>
	<prism:volume>3</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>297</prism:startingPage>
		<prism:doi>10.3390/analytics3030017</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/3/3/17</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/3/3/16">

	<title>Analytics, Vol. 3, Pages 276-296: Modeling Sea Level Rise Using Ensemble Techniques: Impacts on Coastal Adaptation, Freshwater Ecosystems, Agriculture and Infrastructure</title>
	<link>https://www.mdpi.com/2813-2203/3/3/16</link>
	<description>Sea level rise (SLR) is a crucial indicator of climate change, primarily driven by greenhouse gas emissions and the subsequent increase in global temperatures. The impact of SLR, however, varies regionally due to factors such as ocean bathymetry, resulting in distinct shifts across different areas compared to the global average. Understanding the complex factors influencing SLR across diverse spatial scales, along with the associated uncertainties, is essential. This study focuses on the East Coast of the United States and Gulf of Mexico, utilizing historical SLR data from 1993 to 2023. To forecast SLR trends from 2024 to 2103, a weighted ensemble model comprising SARIMAX, LSTM, and exponential smoothing models was employed. Additionally, using historical greenhouse gas data, an ensemble of LSTM models was used to predict real-time SLR values, achieving a testing loss of 0.005. Furthermore, conductance and dissolved oxygen (DO) values were assessed for the entire forecasting period, leveraging forecasted SLR trends to evaluate the impacts on marine life, agriculture, and infrastructure.</description>
	<pubDate>2024-07-05</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 3, Pages 276-296: Modeling Sea Level Rise Using Ensemble Techniques: Impacts on Coastal Adaptation, Freshwater Ecosystems, Agriculture and Infrastructure</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/3/3/16">doi: 10.3390/analytics3030016</a></p>
	<p>Authors:
		Sambandh Bhusan Dhal
		Rishabh Singh
		Tushar Pandey
		Sheelabhadra Dey
		Stavros Kalafatis
		Vivekvardhan Kesireddy
		</p>
	<p>Sea level rise (SLR) is a crucial indicator of climate change, primarily driven by greenhouse gas emissions and the subsequent increase in global temperatures. The impact of SLR, however, varies regionally due to factors such as ocean bathymetry, resulting in distinct shifts across different areas compared to the global average. Understanding the complex factors influencing SLR across diverse spatial scales, along with the associated uncertainties, is essential. This study focuses on the East Coast of the United States and Gulf of Mexico, utilizing historical SLR data from 1993 to 2023. To forecast SLR trends from 2024 to 2103, a weighted ensemble model comprising SARIMAX, LSTM, and exponential smoothing models was employed. Additionally, using historical greenhouse gas data, an ensemble of LSTM models was used to predict real-time SLR values, achieving a testing loss of 0.005. Furthermore, conductance and dissolved oxygen (DO) values were assessed for the entire forecasting period, leveraging forecasted SLR trends to evaluate the impacts on marine life, agriculture, and infrastructure.</p>
	]]></content:encoded>

	<dc:title>Modeling Sea Level Rise Using Ensemble Techniques: Impacts on Coastal Adaptation, Freshwater Ecosystems, Agriculture and Infrastructure</dc:title>
			<dc:creator>Sambandh Bhusan Dhal</dc:creator>
			<dc:creator>Rishabh Singh</dc:creator>
			<dc:creator>Tushar Pandey</dc:creator>
			<dc:creator>Sheelabhadra Dey</dc:creator>
			<dc:creator>Stavros Kalafatis</dc:creator>
			<dc:creator>Vivekvardhan Kesireddy</dc:creator>
		<dc:identifier>doi: 10.3390/analytics3030016</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2024-07-05</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2024-07-05</prism:publicationDate>
	<prism:volume>3</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Communication</prism:section>
	<prism:startingPage>276</prism:startingPage>
		<prism:doi>10.3390/analytics3030016</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/3/3/16</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/3/3/15">

	<title>Analytics, Vol. 3, Pages 255-275: TaskFinder: A Semantics-Based Methodology for Visualization Task Recommendation</title>
	<link>https://www.mdpi.com/2813-2203/3/3/15</link>
	<description>Data visualization has entered the mainstream, and numerous visualization recommender systems have been proposed to assist visualization novices, as well as busy professionals, in selecting the most appropriate type of chart for their data. Given a dataset and a set of user-defined analytical tasks, these systems can make recommendations based on expert coded visualization design principles or empirical models. However, the need to identify the pertinent analytical tasks beforehand still exists and often requires domain expertise. In this work, we aim to automate this step with TaskFinder, a prototype system that leverages the information available in textual documents to understand domain-specific relations between attributes and tasks. TaskFinder employs word vectors as well as a custom dependency parser along with an expert-defined list of task keywords to extract and rank associations between tasks and attributes. It pairs these associations with a statistical analysis of the dataset to filter out tasks irrelevant given the data. TaskFinder ultimately produces a ranked list of attribute&amp;amp;ndash;task pairs. We show that the number of domain articles needed to converge to a recommendation consensus is bounded for our approach. We demonstrate our TaskFinder over multiple domains with varying article types and quantities.</description>
	<pubDate>2024-07-04</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 3, Pages 255-275: TaskFinder: A Semantics-Based Methodology for Visualization Task Recommendation</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/3/3/15">doi: 10.3390/analytics3030015</a></p>
	<p>Authors:
		Darius Coelho
		Bhavya Ghai
		Arjun Krishna
		Maria Velez-Rojas
		Steve Greenspan
		Serge Mankovski
		Klaus Mueller
		</p>
	<p>Data visualization has entered the mainstream, and numerous visualization recommender systems have been proposed to assist visualization novices, as well as busy professionals, in selecting the most appropriate type of chart for their data. Given a dataset and a set of user-defined analytical tasks, these systems can make recommendations based on expert coded visualization design principles or empirical models. However, the need to identify the pertinent analytical tasks beforehand still exists and often requires domain expertise. In this work, we aim to automate this step with TaskFinder, a prototype system that leverages the information available in textual documents to understand domain-specific relations between attributes and tasks. TaskFinder employs word vectors as well as a custom dependency parser along with an expert-defined list of task keywords to extract and rank associations between tasks and attributes. It pairs these associations with a statistical analysis of the dataset to filter out tasks irrelevant given the data. TaskFinder ultimately produces a ranked list of attribute&amp;amp;ndash;task pairs. We show that the number of domain articles needed to converge to a recommendation consensus is bounded for our approach. We demonstrate our TaskFinder over multiple domains with varying article types and quantities.</p>
	]]></content:encoded>

	<dc:title>TaskFinder: A Semantics-Based Methodology for Visualization Task Recommendation</dc:title>
			<dc:creator>Darius Coelho</dc:creator>
			<dc:creator>Bhavya Ghai</dc:creator>
			<dc:creator>Arjun Krishna</dc:creator>
			<dc:creator>Maria Velez-Rojas</dc:creator>
			<dc:creator>Steve Greenspan</dc:creator>
			<dc:creator>Serge Mankovski</dc:creator>
			<dc:creator>Klaus Mueller</dc:creator>
		<dc:identifier>doi: 10.3390/analytics3030015</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2024-07-04</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2024-07-04</prism:publicationDate>
	<prism:volume>3</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>255</prism:startingPage>
		<prism:doi>10.3390/analytics3030015</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/3/3/15</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/3/2/14">

	<title>Analytics, Vol. 3, Pages 241-254: Customer Sentiments in Product Reviews: A Comparative Study with GooglePaLM</title>
	<link>https://www.mdpi.com/2813-2203/3/2/14</link>
	<description>In this work, we evaluated the efficacy of Google&amp;amp;rsquo;s Pathways Language Model (GooglePaLM) in analyzing sentiments expressed in product reviews. Although conventional Natural Language Processing (NLP) techniques such as the rule-based Valence Aware Dictionary for Sentiment Reasoning (VADER) and the long sequence Bidirectional Encoder Representations from Transformers (BERT) model are effective, they frequently encounter difficulties when dealing with intricate linguistic features like sarcasm and contextual nuances commonly found in customer feedback. We performed a sentiment analysis on Amazon&amp;amp;rsquo;s fashion review datasets using the VADER, BERT, and GooglePaLM models, respectively, and compared the results based on evaluation metrics such as precision, recall, accuracy correct positive prediction, and correct negative prediction. We used the default values of the VADER and BERT models and slightly finetuned GooglePaLM with a Temperature of 0.0 and an N-value of 1. We observed that GooglePaLM performed better with correct positive and negative prediction values of 0.91 and 0.93, respectively, followed by BERT and VADER. We concluded that large language models surpass traditional rule-based systems for natural language processing tasks.</description>
	<pubDate>2024-06-18</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 3, Pages 241-254: Customer Sentiments in Product Reviews: A Comparative Study with GooglePaLM</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/3/2/14">doi: 10.3390/analytics3020014</a></p>
	<p>Authors:
		Olamilekan Shobayo
		Swethika Sasikumar
		Sandhya Makkar
		Obinna Okoyeigbo
		</p>
	<p>In this work, we evaluated the efficacy of Google&amp;amp;rsquo;s Pathways Language Model (GooglePaLM) in analyzing sentiments expressed in product reviews. Although conventional Natural Language Processing (NLP) techniques such as the rule-based Valence Aware Dictionary for Sentiment Reasoning (VADER) and the long sequence Bidirectional Encoder Representations from Transformers (BERT) model are effective, they frequently encounter difficulties when dealing with intricate linguistic features like sarcasm and contextual nuances commonly found in customer feedback. We performed a sentiment analysis on Amazon&amp;amp;rsquo;s fashion review datasets using the VADER, BERT, and GooglePaLM models, respectively, and compared the results based on evaluation metrics such as precision, recall, accuracy correct positive prediction, and correct negative prediction. We used the default values of the VADER and BERT models and slightly finetuned GooglePaLM with a Temperature of 0.0 and an N-value of 1. We observed that GooglePaLM performed better with correct positive and negative prediction values of 0.91 and 0.93, respectively, followed by BERT and VADER. We concluded that large language models surpass traditional rule-based systems for natural language processing tasks.</p>
	]]></content:encoded>

	<dc:title>Customer Sentiments in Product Reviews: A Comparative Study with GooglePaLM</dc:title>
			<dc:creator>Olamilekan Shobayo</dc:creator>
			<dc:creator>Swethika Sasikumar</dc:creator>
			<dc:creator>Sandhya Makkar</dc:creator>
			<dc:creator>Obinna Okoyeigbo</dc:creator>
		<dc:identifier>doi: 10.3390/analytics3020014</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2024-06-18</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2024-06-18</prism:publicationDate>
	<prism:volume>3</prism:volume>
	<prism:number>2</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>241</prism:startingPage>
		<prism:doi>10.3390/analytics3020014</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/3/2/14</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/3/2/13">

	<title>Analytics, Vol. 3, Pages 225-240: Improving the Giant-Armadillo Optimization Method</title>
	<link>https://www.mdpi.com/2813-2203/3/2/13</link>
	<description>Global optimization is widely adopted presently in a variety of practical and scientific problems. In this context, a group of widely used techniques are evolutionary techniques. A relatively new evolutionary technique in this direction is that of Giant-Armadillo Optimization, which is based on the hunting strategy of giant armadillos. In this paper, modifications to this technique are proposed, such as the periodic application of a local minimization method as well as the use of modern termination techniques based on statistical observations. The proposed modifications have been tested on a wide series of test functions available from the relevant literature and compared against other evolutionary methods.</description>
	<pubDate>2024-06-10</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 3, Pages 225-240: Improving the Giant-Armadillo Optimization Method</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/3/2/13">doi: 10.3390/analytics3020013</a></p>
	<p>Authors:
		Glykeria Kyrou
		Vasileios Charilogis
		Ioannis G. Tsoulos
		</p>
	<p>Global optimization is widely adopted presently in a variety of practical and scientific problems. In this context, a group of widely used techniques are evolutionary techniques. A relatively new evolutionary technique in this direction is that of Giant-Armadillo Optimization, which is based on the hunting strategy of giant armadillos. In this paper, modifications to this technique are proposed, such as the periodic application of a local minimization method as well as the use of modern termination techniques based on statistical observations. The proposed modifications have been tested on a wide series of test functions available from the relevant literature and compared against other evolutionary methods.</p>
	]]></content:encoded>

	<dc:title>Improving the Giant-Armadillo Optimization Method</dc:title>
			<dc:creator>Glykeria Kyrou</dc:creator>
			<dc:creator>Vasileios Charilogis</dc:creator>
			<dc:creator>Ioannis G. Tsoulos</dc:creator>
		<dc:identifier>doi: 10.3390/analytics3020013</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2024-06-10</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2024-06-10</prism:publicationDate>
	<prism:volume>3</prism:volume>
	<prism:number>2</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>225</prism:startingPage>
		<prism:doi>10.3390/analytics3020013</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/3/2/13</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/3/2/12">

	<title>Analytics, Vol. 3, Pages 221-224: Beyond the ROC Curve: The IMCP Curve</title>
	<link>https://www.mdpi.com/2813-2203/3/2/12</link>
	<description>The ROC curve [...]</description>
	<pubDate>2024-05-27</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 3, Pages 221-224: Beyond the ROC Curve: The IMCP Curve</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/3/2/12">doi: 10.3390/analytics3020012</a></p>
	<p>Authors:
		Jesus S. Aguilar-Ruiz
		</p>
	<p>The ROC curve [...]</p>
	]]></content:encoded>

	<dc:title>Beyond the ROC Curve: The IMCP Curve</dc:title>
			<dc:creator>Jesus S. Aguilar-Ruiz</dc:creator>
		<dc:identifier>doi: 10.3390/analytics3020012</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2024-05-27</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2024-05-27</prism:publicationDate>
	<prism:volume>3</prism:volume>
	<prism:number>2</prism:number>
	<prism:section>Editorial</prism:section>
	<prism:startingPage>221</prism:startingPage>
		<prism:doi>10.3390/analytics3020012</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/3/2/12</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/3/2/11">

	<title>Analytics, Vol. 3, Pages 194-220: Interconnected Markets: Unveiling Volatility Spillovers in Commodities and Energy Markets through BEKK-GARCH Modelling</title>
	<link>https://www.mdpi.com/2813-2203/3/2/11</link>
	<description>Food commodities and energy bills have experienced rapid undulating movements and hikes globally in recent times. This spurred this study to examine the possibility that the shocks that arise from fluctuations of one market spill over to the other and to determine how time-varying the spillovers were across a time. Data were daily frequency (prices of grains and energy products) from 1 July 2019 to 31 December 2022, as quoted in markets. The choice of the period was to capture the COVID pandemic and the Russian&amp;amp;ndash;Ukrainian war as events that could impact volatility. The returns were duly calculated using spreadsheets and subjected to ADF stationarity, co-integration, and the full BEKK-GARCH estimation. The results revealed a prolonged association between returns in the energy markets and food commodity market returns. Both markets were found to have volatility persistence individually, and time-varying bidirectional transmission of volatility across the markets was found. No lagged-effects spillover was found from one market to the other. The findings confirm that shocks that emanate from fluctuations in energy markets are impactful on the volatility of prices in food commodity markets and vice versa, but this impact occurs immediately after the shocks arise or on the same day such variation occurs.</description>
	<pubDate>2024-04-16</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 3, Pages 194-220: Interconnected Markets: Unveiling Volatility Spillovers in Commodities and Energy Markets through BEKK-GARCH Modelling</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/3/2/11">doi: 10.3390/analytics3020011</a></p>
	<p>Authors:
		Tetiana Paientko
		Stanley Amakude
		</p>
	<p>Food commodities and energy bills have experienced rapid undulating movements and hikes globally in recent times. This spurred this study to examine the possibility that the shocks that arise from fluctuations of one market spill over to the other and to determine how time-varying the spillovers were across a time. Data were daily frequency (prices of grains and energy products) from 1 July 2019 to 31 December 2022, as quoted in markets. The choice of the period was to capture the COVID pandemic and the Russian&amp;amp;ndash;Ukrainian war as events that could impact volatility. The returns were duly calculated using spreadsheets and subjected to ADF stationarity, co-integration, and the full BEKK-GARCH estimation. The results revealed a prolonged association between returns in the energy markets and food commodity market returns. Both markets were found to have volatility persistence individually, and time-varying bidirectional transmission of volatility across the markets was found. No lagged-effects spillover was found from one market to the other. The findings confirm that shocks that emanate from fluctuations in energy markets are impactful on the volatility of prices in food commodity markets and vice versa, but this impact occurs immediately after the shocks arise or on the same day such variation occurs.</p>
	]]></content:encoded>

	<dc:title>Interconnected Markets: Unveiling Volatility Spillovers in Commodities and Energy Markets through BEKK-GARCH Modelling</dc:title>
			<dc:creator>Tetiana Paientko</dc:creator>
			<dc:creator>Stanley Amakude</dc:creator>
		<dc:identifier>doi: 10.3390/analytics3020011</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2024-04-16</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2024-04-16</prism:publicationDate>
	<prism:volume>3</prism:volume>
	<prism:number>2</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>194</prism:startingPage>
		<prism:doi>10.3390/analytics3020011</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/3/2/11</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/3/2/10">

	<title>Analytics, Vol. 3, Pages 178-193: Learner Engagement and Demographic Influences in Brazilian Massive Open Online Courses: Aprenda Mais Platform Case Study</title>
	<link>https://www.mdpi.com/2813-2203/3/2/10</link>
	<description>This paper explores the dynamics of student engagement and demographic influences in Massive Open Online Courses (MOOCs). The study analyzes multiple facets of Brazilian MOOC participation, including re-enrollment patterns, course completion rates, and the impact of demographic characteristics on learning outcomes. Using survey data and statistical analyses from the public Aprenda Mais Platform, this study reveals that MOOC learners exhibit a strong tendency toward continuous learning, with a majority re-enrolling in subsequent courses within a short timeframe. The average completion rate across courses is around 42.14%, with learners maintaining consistent academic performance. Demographic factors, notably, race/color and disability, are found to influence enrollment and completion rates, underscoring the importance of inclusive educational practices. Geographical location impacts students&amp;amp;rsquo; decision to enroll in and complete courses, highlighting the necessity for region-specific educational strategies. The research concludes that a diverse array of factors, including content interest, personal motivation, and demographic attributes, shape student engagement in MOOCs. These insights are vital for educators and course designers in creating effective, inclusive, and engaging online learning experiences.</description>
	<pubDate>2024-04-03</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 3, Pages 178-193: Learner Engagement and Demographic Influences in Brazilian Massive Open Online Courses: Aprenda Mais Platform Case Study</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/3/2/10">doi: 10.3390/analytics3020010</a></p>
	<p>Authors:
		Júlia Marques Carvalho da Silva
		Gabriela Hahn Pedroso
		Augusto Basso Veber
		Úrsula Gomes Rosa Maruyama
		</p>
	<p>This paper explores the dynamics of student engagement and demographic influences in Massive Open Online Courses (MOOCs). The study analyzes multiple facets of Brazilian MOOC participation, including re-enrollment patterns, course completion rates, and the impact of demographic characteristics on learning outcomes. Using survey data and statistical analyses from the public Aprenda Mais Platform, this study reveals that MOOC learners exhibit a strong tendency toward continuous learning, with a majority re-enrolling in subsequent courses within a short timeframe. The average completion rate across courses is around 42.14%, with learners maintaining consistent academic performance. Demographic factors, notably, race/color and disability, are found to influence enrollment and completion rates, underscoring the importance of inclusive educational practices. Geographical location impacts students&amp;amp;rsquo; decision to enroll in and complete courses, highlighting the necessity for region-specific educational strategies. The research concludes that a diverse array of factors, including content interest, personal motivation, and demographic attributes, shape student engagement in MOOCs. These insights are vital for educators and course designers in creating effective, inclusive, and engaging online learning experiences.</p>
	]]></content:encoded>

	<dc:title>Learner Engagement and Demographic Influences in Brazilian Massive Open Online Courses: Aprenda Mais Platform Case Study</dc:title>
			<dc:creator>Júlia Marques Carvalho da Silva</dc:creator>
			<dc:creator>Gabriela Hahn Pedroso</dc:creator>
			<dc:creator>Augusto Basso Veber</dc:creator>
			<dc:creator>Úrsula Gomes Rosa Maruyama</dc:creator>
		<dc:identifier>doi: 10.3390/analytics3020010</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2024-04-03</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2024-04-03</prism:publicationDate>
	<prism:volume>3</prism:volume>
	<prism:number>2</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>178</prism:startingPage>
		<prism:doi>10.3390/analytics3020010</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/3/2/10</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/3/1/9">

	<title>Analytics, Vol. 3, Pages 165-177: Optimal Matching with Matching Priority</title>
	<link>https://www.mdpi.com/2813-2203/3/1/9</link>
	<description>Matching algorithms are commonly used to build comparable subsets (matchings) in observational studies. When a complete matching is not possible, some units must necessarily be excluded from the final matching. This may bias the final estimates comparing the two populations, and thus it is important to reduce the number of drops to avoid unsatisfactory results. Greedy matching algorithms may not reach the maximum matching size, thus dropping more units than necessary. Optimal matching algorithms do ensure a maximum matching size, but they implicitly assume that all units have the same matching priority. In this paper, we propose a matching strategy which is order optimal in the sense that it finds a maximum matching size which is consistent with a given matching priority. The strategy is based on an order-optimal matching algorithm originally proposed in connection with assignment problems by D. Gale. When a matching priority is given, the algorithm ensures that the discarded units have the lowest possible matching priority. We discuss the algorithm&amp;amp;rsquo;s complexity and its relation with classic optimal matching. We illustrate its use with a problem in a case study concerning a comparison of female and male executives and a simulation.</description>
	<pubDate>2024-03-19</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 3, Pages 165-177: Optimal Matching with Matching Priority</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/3/1/9">doi: 10.3390/analytics3010009</a></p>
	<p>Authors:
		Massimo Cannas
		Emiliano Sironi
		</p>
	<p>Matching algorithms are commonly used to build comparable subsets (matchings) in observational studies. When a complete matching is not possible, some units must necessarily be excluded from the final matching. This may bias the final estimates comparing the two populations, and thus it is important to reduce the number of drops to avoid unsatisfactory results. Greedy matching algorithms may not reach the maximum matching size, thus dropping more units than necessary. Optimal matching algorithms do ensure a maximum matching size, but they implicitly assume that all units have the same matching priority. In this paper, we propose a matching strategy which is order optimal in the sense that it finds a maximum matching size which is consistent with a given matching priority. The strategy is based on an order-optimal matching algorithm originally proposed in connection with assignment problems by D. Gale. When a matching priority is given, the algorithm ensures that the discarded units have the lowest possible matching priority. We discuss the algorithm&amp;amp;rsquo;s complexity and its relation with classic optimal matching. We illustrate its use with a problem in a case study concerning a comparison of female and male executives and a simulation.</p>
	]]></content:encoded>

	<dc:title>Optimal Matching with Matching Priority</dc:title>
			<dc:creator>Massimo Cannas</dc:creator>
			<dc:creator>Emiliano Sironi</dc:creator>
		<dc:identifier>doi: 10.3390/analytics3010009</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2024-03-19</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2024-03-19</prism:publicationDate>
	<prism:volume>3</prism:volume>
	<prism:number>1</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>165</prism:startingPage>
		<prism:doi>10.3390/analytics3010009</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/3/1/9</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/3/1/8">

	<title>Analytics, Vol. 3, Pages 140-164: Artificial Intelligence and Sustainability&amp;mdash;A Review</title>
	<link>https://www.mdpi.com/2813-2203/3/1/8</link>
	<description>In recent decades, artificial intelligence has undergone transformative advancements, reshaping diverse sectors such as healthcare, transport, agriculture, energy, and the media. Despite the enthusiasm surrounding AI&amp;amp;rsquo;s potential, concerns persist about its potential negative impacts, including substantial energy consumption and ethical challenges. This paper critically reviews the evolving landscape of AI sustainability, addressing economic, social, and environmental dimensions. The literature is systematically categorized into &amp;amp;ldquo;Sustainability of AI&amp;amp;rdquo; and &amp;amp;ldquo;AI for Sustainability&amp;amp;rdquo;, revealing a balanced perspective between the two. The study also identifies a notable trend towards holistic approaches, with a surge in publications and empirical studies since 2019, signaling the field&amp;amp;rsquo;s maturity. Future research directions emphasize delving into the relatively under-explored economic dimension, aligning with the United Nations&amp;amp;rsquo; Sustainable Development Goals (SDGs), and addressing stakeholders&amp;amp;rsquo; influence.</description>
	<pubDate>2024-03-01</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 3, Pages 140-164: Artificial Intelligence and Sustainability&amp;mdash;A Review</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/3/1/8">doi: 10.3390/analytics3010008</a></p>
	<p>Authors:
		Rachit Dhiman
		Sofia Miteff
		Yuancheng Wang
		Shih-Chi Ma
		Ramila Amirikas
		Benjamin Fabian
		</p>
	<p>In recent decades, artificial intelligence has undergone transformative advancements, reshaping diverse sectors such as healthcare, transport, agriculture, energy, and the media. Despite the enthusiasm surrounding AI&amp;amp;rsquo;s potential, concerns persist about its potential negative impacts, including substantial energy consumption and ethical challenges. This paper critically reviews the evolving landscape of AI sustainability, addressing economic, social, and environmental dimensions. The literature is systematically categorized into &amp;amp;ldquo;Sustainability of AI&amp;amp;rdquo; and &amp;amp;ldquo;AI for Sustainability&amp;amp;rdquo;, revealing a balanced perspective between the two. The study also identifies a notable trend towards holistic approaches, with a surge in publications and empirical studies since 2019, signaling the field&amp;amp;rsquo;s maturity. Future research directions emphasize delving into the relatively under-explored economic dimension, aligning with the United Nations&amp;amp;rsquo; Sustainable Development Goals (SDGs), and addressing stakeholders&amp;amp;rsquo; influence.</p>
	]]></content:encoded>

	<dc:title>Artificial Intelligence and Sustainability&amp;amp;mdash;A Review</dc:title>
			<dc:creator>Rachit Dhiman</dc:creator>
			<dc:creator>Sofia Miteff</dc:creator>
			<dc:creator>Yuancheng Wang</dc:creator>
			<dc:creator>Shih-Chi Ma</dc:creator>
			<dc:creator>Ramila Amirikas</dc:creator>
			<dc:creator>Benjamin Fabian</dc:creator>
		<dc:identifier>doi: 10.3390/analytics3010008</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2024-03-01</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2024-03-01</prism:publicationDate>
	<prism:volume>3</prism:volume>
	<prism:number>1</prism:number>
	<prism:section>Review</prism:section>
	<prism:startingPage>140</prism:startingPage>
		<prism:doi>10.3390/analytics3010008</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/3/1/8</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/3/1/7">

	<title>Analytics, Vol. 3, Pages 116-139: Visual Analytics for Robust Investigations of Placental Aquaporin Gene Expression in Response to Maternal SARS-CoV-2 Infection</title>
	<link>https://www.mdpi.com/2813-2203/3/1/7</link>
	<description>The human placenta is a multifunctional, disc-shaped temporary fetal organ that develops in the uterus during pregnancy, connecting the mother and the fetus. The availability of large-scale datasets on the gene expression of placental cell types and scholarly articles documenting adverse pregnancy outcomes from maternal infection warrants the use of computational resources to aid in knowledge generation from disparate data sources. Using maternal Severe Acute Respiratory Syndrome Coronavirus 2 (SARS-CoV-2) infection as a case study in microbial infection, we constructed integrated datasets and implemented visual analytics resources to facilitate robust investigations of placental gene expression data in the dimensions of flow, curation, and analytics. The visual analytics resources and associated datasets can support a greater understanding of SARS-CoV-2-induced changes to the human placental expression levels of 18,882 protein-coding genes and at least 1233 human gene groups/families. We focus this report on the human aquaporin gene family that encodes small integral membrane proteins initially studied for their roles in water transport across cell membranes. Aquaporin-9 (AQP9) was the only aquaporin downregulated in term placental villi from SARS-CoV-2-positive mothers. Previous studies have found that (1) oxygen signaling modulates placental development; (2) oxygen tension could modulate AQP9 expression in the human placenta; and (3) SARS-CoV-2 can disrupt the formation of oxygen-carrying red blood cells in the placenta. Thus, future research could be performed on microbial infection-induced changes to (1) the placental hematopoietic stem and progenitor cells; and (2) placental expression of human aquaporin genes, especially AQP9.</description>
	<pubDate>2024-02-05</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 3, Pages 116-139: Visual Analytics for Robust Investigations of Placental Aquaporin Gene Expression in Response to Maternal SARS-CoV-2 Infection</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/3/1/7">doi: 10.3390/analytics3010007</a></p>
	<p>Authors:
		Raphael D. Isokpehi
		Amos O. Abioye
		Rickeisha S. Hamilton
		Jasmin C. Fryer
		Antoinesha L. Hollman
		Antoinette M. Destefano
		Kehinde B. Ezekiel
		Tyrese L. Taylor
		Shawna F. Brooks
		Matilda O. Johnson
		Olubukola Smile
		Shirma Ramroop-Butts
		Angela U. Makolo
		Albert G. Hayward
		</p>
	<p>The human placenta is a multifunctional, disc-shaped temporary fetal organ that develops in the uterus during pregnancy, connecting the mother and the fetus. The availability of large-scale datasets on the gene expression of placental cell types and scholarly articles documenting adverse pregnancy outcomes from maternal infection warrants the use of computational resources to aid in knowledge generation from disparate data sources. Using maternal Severe Acute Respiratory Syndrome Coronavirus 2 (SARS-CoV-2) infection as a case study in microbial infection, we constructed integrated datasets and implemented visual analytics resources to facilitate robust investigations of placental gene expression data in the dimensions of flow, curation, and analytics. The visual analytics resources and associated datasets can support a greater understanding of SARS-CoV-2-induced changes to the human placental expression levels of 18,882 protein-coding genes and at least 1233 human gene groups/families. We focus this report on the human aquaporin gene family that encodes small integral membrane proteins initially studied for their roles in water transport across cell membranes. Aquaporin-9 (AQP9) was the only aquaporin downregulated in term placental villi from SARS-CoV-2-positive mothers. Previous studies have found that (1) oxygen signaling modulates placental development; (2) oxygen tension could modulate AQP9 expression in the human placenta; and (3) SARS-CoV-2 can disrupt the formation of oxygen-carrying red blood cells in the placenta. Thus, future research could be performed on microbial infection-induced changes to (1) the placental hematopoietic stem and progenitor cells; and (2) placental expression of human aquaporin genes, especially AQP9.</p>
	]]></content:encoded>

	<dc:title>Visual Analytics for Robust Investigations of Placental Aquaporin Gene Expression in Response to Maternal SARS-CoV-2 Infection</dc:title>
			<dc:creator>Raphael D. Isokpehi</dc:creator>
			<dc:creator>Amos O. Abioye</dc:creator>
			<dc:creator>Rickeisha S. Hamilton</dc:creator>
			<dc:creator>Jasmin C. Fryer</dc:creator>
			<dc:creator>Antoinesha L. Hollman</dc:creator>
			<dc:creator>Antoinette M. Destefano</dc:creator>
			<dc:creator>Kehinde B. Ezekiel</dc:creator>
			<dc:creator>Tyrese L. Taylor</dc:creator>
			<dc:creator>Shawna F. Brooks</dc:creator>
			<dc:creator>Matilda O. Johnson</dc:creator>
			<dc:creator>Olubukola Smile</dc:creator>
			<dc:creator>Shirma Ramroop-Butts</dc:creator>
			<dc:creator>Angela U. Makolo</dc:creator>
			<dc:creator>Albert G. Hayward</dc:creator>
		<dc:identifier>doi: 10.3390/analytics3010007</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2024-02-05</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2024-02-05</prism:publicationDate>
	<prism:volume>3</prism:volume>
	<prism:number>1</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>116</prism:startingPage>
		<prism:doi>10.3390/analytics3010007</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/3/1/7</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/3/1/6">

	<title>Analytics, Vol. 3, Pages 84-115: Interoperable Information Flow as Enabler for Efficient Predictive Maintenance</title>
	<link>https://www.mdpi.com/2813-2203/3/1/6</link>
	<description>Industry 4.0 enables the modernisation of machines and opens up the digitalisation of processes in the manufacturing industry. As a result, these machines are ready for predictive maintenance as part of Industry 4.0 services. The benefit of predictive maintenance is that it can significantly extend the life of machines. The integration of predictive maintenance into existing production environments faces challenges in terms of data understanding and data preparation for machines and legacy systems. Current AI frameworks lack adequate support for the ongoing task of data integration. In this context, adequate support means that the data analyst does not need to know the technical background of the pilot&amp;amp;rsquo;s data sources in terms of data formats and schemas. It should be possible to perform data analyses without knowing the characteristics of the pilot&amp;amp;rsquo;s specific data sources. The aim is to achieve a seamless integration of data as information for predictive maintenance. For this purpose, the developed data-sharing infrastructure enables automatic data acquisition and data integration for AI frameworks using interoperability methods. The evaluation, based on two pilot projects, shows that the step of data understanding and data preparation for predictive maintenance is simplified and that the solution is applicable for new pilot projects.</description>
	<pubDate>2024-02-01</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 3, Pages 84-115: Interoperable Information Flow as Enabler for Efficient Predictive Maintenance</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/3/1/6">doi: 10.3390/analytics3010006</a></p>
	<p>Authors:
		Marco Franke
		Quan Deng
		Zisis Kyroudis
		Maria Psarodimou
		Jovana Milenkovic
		Ioannis Meintanis
		Dimitris Lokas
		Stefano Borgia
		Klaus-Dieter Thoben
		</p>
	<p>Industry 4.0 enables the modernisation of machines and opens up the digitalisation of processes in the manufacturing industry. As a result, these machines are ready for predictive maintenance as part of Industry 4.0 services. The benefit of predictive maintenance is that it can significantly extend the life of machines. The integration of predictive maintenance into existing production environments faces challenges in terms of data understanding and data preparation for machines and legacy systems. Current AI frameworks lack adequate support for the ongoing task of data integration. In this context, adequate support means that the data analyst does not need to know the technical background of the pilot&amp;amp;rsquo;s data sources in terms of data formats and schemas. It should be possible to perform data analyses without knowing the characteristics of the pilot&amp;amp;rsquo;s specific data sources. The aim is to achieve a seamless integration of data as information for predictive maintenance. For this purpose, the developed data-sharing infrastructure enables automatic data acquisition and data integration for AI frameworks using interoperability methods. The evaluation, based on two pilot projects, shows that the step of data understanding and data preparation for predictive maintenance is simplified and that the solution is applicable for new pilot projects.</p>
	]]></content:encoded>

	<dc:title>Interoperable Information Flow as Enabler for Efficient Predictive Maintenance</dc:title>
			<dc:creator>Marco Franke</dc:creator>
			<dc:creator>Quan Deng</dc:creator>
			<dc:creator>Zisis Kyroudis</dc:creator>
			<dc:creator>Maria Psarodimou</dc:creator>
			<dc:creator>Jovana Milenkovic</dc:creator>
			<dc:creator>Ioannis Meintanis</dc:creator>
			<dc:creator>Dimitris Lokas</dc:creator>
			<dc:creator>Stefano Borgia</dc:creator>
			<dc:creator>Klaus-Dieter Thoben</dc:creator>
		<dc:identifier>doi: 10.3390/analytics3010006</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2024-02-01</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2024-02-01</prism:publicationDate>
	<prism:volume>3</prism:volume>
	<prism:number>1</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>84</prism:startingPage>
		<prism:doi>10.3390/analytics3010006</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/3/1/6</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/3/1/5">

	<title>Analytics, Vol. 3, Pages 63-83: Analysing the Influence of Macroeconomic Factors on Credit Risk in the UK Banking Sector</title>
	<link>https://www.mdpi.com/2813-2203/3/1/5</link>
	<description>Macroeconomic factors have a critical impact on banking credit risk, which cannot be directly controlled by banks, and therefore, there is a need for an early credit risk warning system based on the macroeconomy. By comparing different predictive models (traditional statistical and machine learning algorithms), this study aims to examine the macroeconomic determinants&amp;amp;rsquo; impact on the UK banking credit risk and assess the most accurate credit risk estimate using predictive analytics. This study found that the variance-based multi-split decision tree algorithm is the most precise predictive model with interpretable, reliable, and robust results. Our model performance achieved 95% accuracy and evidenced that unemployment and inflation rate are significant credit risk predictors in the UK banking context. Our findings provided valuable insights such as a positive association between credit risk and inflation, the unemployment rate, and national savings, as well as a negative relationship between credit risk and national debt, total trade deficit, and national income. In addition, we empirically showed the relationship between national savings and non-performing loans, thus proving the &amp;amp;ldquo;paradox of thrift&amp;amp;rdquo;. These findings benefit the credit risk management team in monitoring the macroeconomic factors&amp;amp;rsquo; thresholds and implementing critical reforms to mitigate credit risk.</description>
	<pubDate>2024-01-26</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 3, Pages 63-83: Analysing the Influence of Macroeconomic Factors on Credit Risk in the UK Banking Sector</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/3/1/5">doi: 10.3390/analytics3010005</a></p>
	<p>Authors:
		Hemlata Sharma
		Aparna Andhalkar
		Oluwaseun Ajao
		Bayode Ogunleye
		</p>
	<p>Macroeconomic factors have a critical impact on banking credit risk, which cannot be directly controlled by banks, and therefore, there is a need for an early credit risk warning system based on the macroeconomy. By comparing different predictive models (traditional statistical and machine learning algorithms), this study aims to examine the macroeconomic determinants&amp;amp;rsquo; impact on the UK banking credit risk and assess the most accurate credit risk estimate using predictive analytics. This study found that the variance-based multi-split decision tree algorithm is the most precise predictive model with interpretable, reliable, and robust results. Our model performance achieved 95% accuracy and evidenced that unemployment and inflation rate are significant credit risk predictors in the UK banking context. Our findings provided valuable insights such as a positive association between credit risk and inflation, the unemployment rate, and national savings, as well as a negative relationship between credit risk and national debt, total trade deficit, and national income. In addition, we empirically showed the relationship between national savings and non-performing loans, thus proving the &amp;amp;ldquo;paradox of thrift&amp;amp;rdquo;. These findings benefit the credit risk management team in monitoring the macroeconomic factors&amp;amp;rsquo; thresholds and implementing critical reforms to mitigate credit risk.</p>
	]]></content:encoded>

	<dc:title>Analysing the Influence of Macroeconomic Factors on Credit Risk in the UK Banking Sector</dc:title>
			<dc:creator>Hemlata Sharma</dc:creator>
			<dc:creator>Aparna Andhalkar</dc:creator>
			<dc:creator>Oluwaseun Ajao</dc:creator>
			<dc:creator>Bayode Ogunleye</dc:creator>
		<dc:identifier>doi: 10.3390/analytics3010005</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2024-01-26</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2024-01-26</prism:publicationDate>
	<prism:volume>3</prism:volume>
	<prism:number>1</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>63</prism:startingPage>
		<prism:doi>10.3390/analytics3010005</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/3/1/5</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/3/1/4">

	<title>Analytics, Vol. 3, Pages 46-62: Code Plagiarism Checking Function and Its Application for Code Writing Problem in Java Programming Learning Assistant System</title>
	<link>https://www.mdpi.com/2813-2203/3/1/4</link>
	<description>A web-based Java programming learning assistant system (JPLAS) has been developed for novice students to study Java programming by themselves while enhancing code reading and code writing skills. One type of the implemented exercise problem is code writing problem (CWP), which asks students to create a source code that can pass the given test code. The correctness of this answer code is validated by running them on JUnit. In previous works, a Python-based answer code validation program was implemented to assist teachers. It automatically verifies the source codes from all the students for one test code, and reports the number of passed test cases by each code in the CSV file. While this program plays a crucial role in checking the correctness of code behaviors, it cannot detect code plagiarism that can often happen in programming courses. In this paper, we implement a code plagiarism checking function in the answer code validation program, and present its application results to a Java programming course at Okayama University, Japan. This function first removes the whitespace characters and the comments using the regular expressions. Next, it calculates the Levenshtein distance and similarity score for each pair of source codes from different students in the class. If the score is larger than a given threshold, they are regarded as plagiarism. Finally, it outputs the scores as a CSV file with the student IDs. For evaluations, we applied the proposed function to a total of 877 source codes for 45 CWP assignments submitted from 9 to 39 students and analyzed the results. It was found that (1) CWP assignments asking for shorter source codes generate higher scores than those for longer codes due to the use of test codes, (2) proper thresholds are different by assignments, and (3) some students often copied source codes from certain students.</description>
	<pubDate>2024-01-17</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 3, Pages 46-62: Code Plagiarism Checking Function and Its Application for Code Writing Problem in Java Programming Learning Assistant System</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/3/1/4">doi: 10.3390/analytics3010004</a></p>
	<p>Authors:
		Ei Ei Htet
		Khaing Hsu Wai
		Soe Thandar Aung
		Nobuo Funabiki
		Xiqin Lu
		Htoo Htoo Sandi Kyaw
		Wen-Chung Kao
		</p>
	<p>A web-based Java programming learning assistant system (JPLAS) has been developed for novice students to study Java programming by themselves while enhancing code reading and code writing skills. One type of the implemented exercise problem is code writing problem (CWP), which asks students to create a source code that can pass the given test code. The correctness of this answer code is validated by running them on JUnit. In previous works, a Python-based answer code validation program was implemented to assist teachers. It automatically verifies the source codes from all the students for one test code, and reports the number of passed test cases by each code in the CSV file. While this program plays a crucial role in checking the correctness of code behaviors, it cannot detect code plagiarism that can often happen in programming courses. In this paper, we implement a code plagiarism checking function in the answer code validation program, and present its application results to a Java programming course at Okayama University, Japan. This function first removes the whitespace characters and the comments using the regular expressions. Next, it calculates the Levenshtein distance and similarity score for each pair of source codes from different students in the class. If the score is larger than a given threshold, they are regarded as plagiarism. Finally, it outputs the scores as a CSV file with the student IDs. For evaluations, we applied the proposed function to a total of 877 source codes for 45 CWP assignments submitted from 9 to 39 students and analyzed the results. It was found that (1) CWP assignments asking for shorter source codes generate higher scores than those for longer codes due to the use of test codes, (2) proper thresholds are different by assignments, and (3) some students often copied source codes from certain students.</p>
	]]></content:encoded>

	<dc:title>Code Plagiarism Checking Function and Its Application for Code Writing Problem in Java Programming Learning Assistant System</dc:title>
			<dc:creator>Ei Ei Htet</dc:creator>
			<dc:creator>Khaing Hsu Wai</dc:creator>
			<dc:creator>Soe Thandar Aung</dc:creator>
			<dc:creator>Nobuo Funabiki</dc:creator>
			<dc:creator>Xiqin Lu</dc:creator>
			<dc:creator>Htoo Htoo Sandi Kyaw</dc:creator>
			<dc:creator>Wen-Chung Kao</dc:creator>
		<dc:identifier>doi: 10.3390/analytics3010004</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2024-01-17</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2024-01-17</prism:publicationDate>
	<prism:volume>3</prism:volume>
	<prism:number>1</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>46</prism:startingPage>
		<prism:doi>10.3390/analytics3010004</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/3/1/4</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/3/1/3">

	<title>Analytics, Vol. 3, Pages 30-45: An Optimal House Price Prediction Algorithm: XGBoost</title>
	<link>https://www.mdpi.com/2813-2203/3/1/3</link>
	<description>An accurate prediction of house prices is a fundamental requirement for various sectors, including real estate and mortgage lending. It is widely recognized that a property&amp;amp;rsquo;s value is not solely determined by its physical attributes but is significantly influenced by its surrounding neighborhood. Meeting the diverse housing needs of individuals while balancing budget constraints is a primary concern for real estate developers. To this end, we addressed the house price prediction problem as a regression task and thus employed various machine learning (ML) techniques capable of expressing the significance of independent variables. We made use of the housing dataset of Ames City in Iowa, USA to compare XGBoost, support vector regressor, random forest regressor, multilayer perceptron, and multiple linear regression algorithms for house price prediction. Afterwards, we identified the key factors that influence housing costs. Our results show that XGBoost is the best performing model for house price prediction. Our findings present valuable insights and tools for stakeholders, facilitating more accurate property price estimates and, in turn, enabling more informed decision making to meet the housing needs of diverse populations while considering budget constraints.</description>
	<pubDate>2024-01-02</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 3, Pages 30-45: An Optimal House Price Prediction Algorithm: XGBoost</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/3/1/3">doi: 10.3390/analytics3010003</a></p>
	<p>Authors:
		Hemlata Sharma
		Hitesh Harsora
		Bayode Ogunleye
		</p>
	<p>An accurate prediction of house prices is a fundamental requirement for various sectors, including real estate and mortgage lending. It is widely recognized that a property&amp;amp;rsquo;s value is not solely determined by its physical attributes but is significantly influenced by its surrounding neighborhood. Meeting the diverse housing needs of individuals while balancing budget constraints is a primary concern for real estate developers. To this end, we addressed the house price prediction problem as a regression task and thus employed various machine learning (ML) techniques capable of expressing the significance of independent variables. We made use of the housing dataset of Ames City in Iowa, USA to compare XGBoost, support vector regressor, random forest regressor, multilayer perceptron, and multiple linear regression algorithms for house price prediction. Afterwards, we identified the key factors that influence housing costs. Our results show that XGBoost is the best performing model for house price prediction. Our findings present valuable insights and tools for stakeholders, facilitating more accurate property price estimates and, in turn, enabling more informed decision making to meet the housing needs of diverse populations while considering budget constraints.</p>
	]]></content:encoded>

	<dc:title>An Optimal House Price Prediction Algorithm: XGBoost</dc:title>
			<dc:creator>Hemlata Sharma</dc:creator>
			<dc:creator>Hitesh Harsora</dc:creator>
			<dc:creator>Bayode Ogunleye</dc:creator>
		<dc:identifier>doi: 10.3390/analytics3010003</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2024-01-02</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2024-01-02</prism:publicationDate>
	<prism:volume>3</prism:volume>
	<prism:number>1</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>30</prism:startingPage>
		<prism:doi>10.3390/analytics3010003</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/3/1/3</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/3/1/2">

	<title>Analytics, Vol. 3, Pages 14-29: Exploring Infant Physical Activity Using a Population-Based Network Analysis Approach</title>
	<link>https://www.mdpi.com/2813-2203/3/1/2</link>
	<description>Background: Physical activity (PA) is an important aspect of infant development and has been shown to have long-term effects on health and well-being. Accurate analysis of infant PA is crucial for understanding their physical development, monitoring health and wellness, as well as identifying areas for improvement. However, individual analysis of infant PA can be challenging and often leads to biased results due to an infant&amp;amp;rsquo;s inability to self-report and constantly changing posture and movement. This manuscript explores a population-based network analysis approach to study infants&amp;amp;rsquo; PA. The network analysis approach allows us to draw conclusions that are generalizable to the entire population and to identify trends and patterns in PA levels. Methods: This study aims to analyze the PA of infants aged 6&amp;amp;ndash;15 months using accelerometer data. A total of 20 infants from different types of childcare settings were recruited, including home-based and center-based care. Each infant wore an accelerometer for four days (2 weekdays, 2 weekend days). Data were analyzed using a network analysis approach, exploring the relationship between PA and various demographic and social factors. Results: The results showed that infants in center-based care have significantly higher levels of PA than those in home-based care. Moreover, the ankle acceleration was much higher than the waist acceleration, and activity patterns differed on weekdays and weekends. Conclusions: This study highlights the need for further research to explore the factors contributing to disparities in PA levels among infants in different childcare settings. Additionally, there is a need to develop effective strategies to promote PA among infants, considering the findings from the network analysis approach. Such efforts can contribute to enhancing infant health and well-being through targeted interventions aimed at increasing PA levels.</description>
	<pubDate>2023-12-31</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 3, Pages 14-29: Exploring Infant Physical Activity Using a Population-Based Network Analysis Approach</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/3/1/2">doi: 10.3390/analytics3010002</a></p>
	<p>Authors:
		Rama Krishna Thelagathoti
		Priyanka Chaudhary
		Brian Knarr
		Michaela Schenkelberg
		Hesham H. Ali
		Danae Dinkel
		</p>
	<p>Background: Physical activity (PA) is an important aspect of infant development and has been shown to have long-term effects on health and well-being. Accurate analysis of infant PA is crucial for understanding their physical development, monitoring health and wellness, as well as identifying areas for improvement. However, individual analysis of infant PA can be challenging and often leads to biased results due to an infant&amp;amp;rsquo;s inability to self-report and constantly changing posture and movement. This manuscript explores a population-based network analysis approach to study infants&amp;amp;rsquo; PA. The network analysis approach allows us to draw conclusions that are generalizable to the entire population and to identify trends and patterns in PA levels. Methods: This study aims to analyze the PA of infants aged 6&amp;amp;ndash;15 months using accelerometer data. A total of 20 infants from different types of childcare settings were recruited, including home-based and center-based care. Each infant wore an accelerometer for four days (2 weekdays, 2 weekend days). Data were analyzed using a network analysis approach, exploring the relationship between PA and various demographic and social factors. Results: The results showed that infants in center-based care have significantly higher levels of PA than those in home-based care. Moreover, the ankle acceleration was much higher than the waist acceleration, and activity patterns differed on weekdays and weekends. Conclusions: This study highlights the need for further research to explore the factors contributing to disparities in PA levels among infants in different childcare settings. Additionally, there is a need to develop effective strategies to promote PA among infants, considering the findings from the network analysis approach. Such efforts can contribute to enhancing infant health and well-being through targeted interventions aimed at increasing PA levels.</p>
	]]></content:encoded>

	<dc:title>Exploring Infant Physical Activity Using a Population-Based Network Analysis Approach</dc:title>
			<dc:creator>Rama Krishna Thelagathoti</dc:creator>
			<dc:creator>Priyanka Chaudhary</dc:creator>
			<dc:creator>Brian Knarr</dc:creator>
			<dc:creator>Michaela Schenkelberg</dc:creator>
			<dc:creator>Hesham H. Ali</dc:creator>
			<dc:creator>Danae Dinkel</dc:creator>
		<dc:identifier>doi: 10.3390/analytics3010002</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2023-12-31</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2023-12-31</prism:publicationDate>
	<prism:volume>3</prism:volume>
	<prism:number>1</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>14</prism:startingPage>
		<prism:doi>10.3390/analytics3010002</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/3/1/2</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/3/1/1">

	<title>Analytics, Vol. 3, Pages 1-13: Does Part of Speech Have an Influence on Cyberbullying Detection?</title>
	<link>https://www.mdpi.com/2813-2203/3/1/1</link>
	<description>With the development of the Internet, the issue of cyberbullying on social media has gained significant attention. Cyberbullying is often expressed in text. Methods of identifying such text via machine learning have been growing, most of which rely on the extraction of part-of-speech (POS) tags to improve their performance. However, the current study only arbitrarily used part-of-speech labels that it considered reasonable, without investigating whether the chosen part-of-speech labels can better enhance the effectiveness of the cyberbullying detection task. In other words, the effectiveness of different part-of-speech labels in the automatic cyberbullying detection task was not proven. This study aimed to investigate the part of speech in statements related to cyberbullying and explore how three classification models (random forest, na&amp;amp;iuml;ve Bayes, and support vector machine) are sensitive to parts of speech in detecting cyberbullying. We also examined which part-of-speech combinations are most appropriate for the models mentioned above. The results of our experiments showed that the predictive performance of different models differs when using different part-of-speech tags as inputs. Random forest showed the best predictive performance, and naive Bayes and support vector machine followed, respectively. Meanwhile, across the different models, the sensitivity to different part-of-speech tags was consistent, with greater sensitivity shown towards nouns, verbs, and measure words, and lower sensitivity shown towards adjectives and pronouns. We also found that the combination of different parts of speech as inputs had an influence on the predictive performance of the models. This study will help researchers to determine which combination of part-of-speech categories is appropriate to improve the accuracy of cyberbullying detection.</description>
	<pubDate>2023-12-21</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 3, Pages 1-13: Does Part of Speech Have an Influence on Cyberbullying Detection?</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/3/1/1">doi: 10.3390/analytics3010001</a></p>
	<p>Authors:
		Jingxiu Huang
		Ruofei Ding
		Yunxiang Zheng
		Xiaomin Wu
		Shumin Chen
		Xiunan Jin
		</p>
	<p>With the development of the Internet, the issue of cyberbullying on social media has gained significant attention. Cyberbullying is often expressed in text. Methods of identifying such text via machine learning have been growing, most of which rely on the extraction of part-of-speech (POS) tags to improve their performance. However, the current study only arbitrarily used part-of-speech labels that it considered reasonable, without investigating whether the chosen part-of-speech labels can better enhance the effectiveness of the cyberbullying detection task. In other words, the effectiveness of different part-of-speech labels in the automatic cyberbullying detection task was not proven. This study aimed to investigate the part of speech in statements related to cyberbullying and explore how three classification models (random forest, na&amp;amp;iuml;ve Bayes, and support vector machine) are sensitive to parts of speech in detecting cyberbullying. We also examined which part-of-speech combinations are most appropriate for the models mentioned above. The results of our experiments showed that the predictive performance of different models differs when using different part-of-speech tags as inputs. Random forest showed the best predictive performance, and naive Bayes and support vector machine followed, respectively. Meanwhile, across the different models, the sensitivity to different part-of-speech tags was consistent, with greater sensitivity shown towards nouns, verbs, and measure words, and lower sensitivity shown towards adjectives and pronouns. We also found that the combination of different parts of speech as inputs had an influence on the predictive performance of the models. This study will help researchers to determine which combination of part-of-speech categories is appropriate to improve the accuracy of cyberbullying detection.</p>
	]]></content:encoded>

	<dc:title>Does Part of Speech Have an Influence on Cyberbullying Detection?</dc:title>
			<dc:creator>Jingxiu Huang</dc:creator>
			<dc:creator>Ruofei Ding</dc:creator>
			<dc:creator>Yunxiang Zheng</dc:creator>
			<dc:creator>Xiaomin Wu</dc:creator>
			<dc:creator>Shumin Chen</dc:creator>
			<dc:creator>Xiunan Jin</dc:creator>
		<dc:identifier>doi: 10.3390/analytics3010001</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2023-12-21</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2023-12-21</prism:publicationDate>
	<prism:volume>3</prism:volume>
	<prism:number>1</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>1</prism:startingPage>
		<prism:doi>10.3390/analytics3010001</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/3/1/1</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/2/4/46">

	<title>Analytics, Vol. 2, Pages 877-898: Learning Analytics in the Era of Large Language Models</title>
	<link>https://www.mdpi.com/2813-2203/2/4/46</link>
	<description>Learning analytics (LA) has the potential to significantly improve teaching and learning, but there are still many areas for improvement in LA research and practice. The literature highlights limitations in every stage of the LA life cycle, including scarce pedagogical grounding and poor design choices in the development of LA, challenges in the implementation of LA with respect to the interpretability of insights, prediction, and actionability of feedback, and lack of generalizability and strong practices in LA evaluation. In this position paper, we advocate for empowering teachers in developing LA solutions. We argue that this would enhance the theoretical basis of LA tools and make them more understandable and practical. We present some instances where process data can be utilized to comprehend learning processes and generate more interpretable LA insights. Additionally, we investigate the potential implementation of large language models (LLMs) in LA to produce comprehensible insights, provide timely and actionable feedback, enhance personalization, and support teachers&amp;amp;rsquo; tasks more extensively.</description>
	<pubDate>2023-11-16</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 2, Pages 877-898: Learning Analytics in the Era of Large Language Models</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/2/4/46">doi: 10.3390/analytics2040046</a></p>
	<p>Authors:
		Elisabetta Mazzullo
		Okan Bulut
		Tarid Wongvorachan
		Bin Tan
		</p>
	<p>Learning analytics (LA) has the potential to significantly improve teaching and learning, but there are still many areas for improvement in LA research and practice. The literature highlights limitations in every stage of the LA life cycle, including scarce pedagogical grounding and poor design choices in the development of LA, challenges in the implementation of LA with respect to the interpretability of insights, prediction, and actionability of feedback, and lack of generalizability and strong practices in LA evaluation. In this position paper, we advocate for empowering teachers in developing LA solutions. We argue that this would enhance the theoretical basis of LA tools and make them more understandable and practical. We present some instances where process data can be utilized to comprehend learning processes and generate more interpretable LA insights. Additionally, we investigate the potential implementation of large language models (LLMs) in LA to produce comprehensible insights, provide timely and actionable feedback, enhance personalization, and support teachers&amp;amp;rsquo; tasks more extensively.</p>
	]]></content:encoded>

	<dc:title>Learning Analytics in the Era of Large Language Models</dc:title>
			<dc:creator>Elisabetta Mazzullo</dc:creator>
			<dc:creator>Okan Bulut</dc:creator>
			<dc:creator>Tarid Wongvorachan</dc:creator>
			<dc:creator>Bin Tan</dc:creator>
		<dc:identifier>doi: 10.3390/analytics2040046</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2023-11-16</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2023-11-16</prism:publicationDate>
	<prism:volume>2</prism:volume>
	<prism:number>4</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>877</prism:startingPage>
		<prism:doi>10.3390/analytics2040046</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/2/4/46</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/2/4/45">

	<title>Analytics, Vol. 2, Pages 853-876: A Comparative Analysis of VirLock and Bacteriophage &amp;#981;6 through the Lens of Game Theory</title>
	<link>https://www.mdpi.com/2813-2203/2/4/45</link>
	<description>The novelty of this paper lies in its perspective, which underscores the fruitful correlation between biological and computer viruses. In the realm of computer science, the study of theoretical concepts often intersects with practical applications. Computer viruses have many common traits with their biological counterparts. Studying their correlation may enhance our perspective and, ultimately, augment our ability to successfully protect our computer systems and data against viruses. Game theory may be an appropriate tool for establishing the link between biological and computer viruses. In this work, we establish correlations between a well-known computer virus, VirLock, with an equally well-studied biological virus, the bacteriophage &amp;amp;#981;6. VirLock is a formidable ransomware that encrypts user files and demands a ransom for data restoration. Drawing a parallel with the biological virus bacteriophage &amp;amp;#981;6, we uncover conceptual links like shared attributes and behaviors, as well as useful insights. Following this line of thought, we suggest efficient strategies based on a game theory perspective, which have the potential to address the infections caused by VirLock, and other viruses with analogous behavior. Moreover, we propose mathematical formulations that integrate real-world variables, providing a means to gauge virus severity and design robust defensive strategies and analytics. This interdisciplinary inquiry, fusing game theory, biology, and computer science, advances our understanding of virus behavior, paving the way for the development of effective countermeasures while presenting an alternative viewpoint. Throughout this theoretical exploration, we contribute to the ongoing discourse on computer virus behavior and stimulate new avenues for addressing digital threats. In particular, the formulas and framework developed in this work can facilitate better risk analysis and assessment, and become useful tools in penetration testing analysis, helping companies and organizations enhance their security.</description>
	<pubDate>2023-11-06</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 2, Pages 853-876: A Comparative Analysis of VirLock and Bacteriophage &amp;#981;6 through the Lens of Game Theory</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/2/4/45">doi: 10.3390/analytics2040045</a></p>
	<p>Authors:
		Dimitris Kostadimas
		Kalliopi Kastampolidou
		Theodore Andronikos
		</p>
	<p>The novelty of this paper lies in its perspective, which underscores the fruitful correlation between biological and computer viruses. In the realm of computer science, the study of theoretical concepts often intersects with practical applications. Computer viruses have many common traits with their biological counterparts. Studying their correlation may enhance our perspective and, ultimately, augment our ability to successfully protect our computer systems and data against viruses. Game theory may be an appropriate tool for establishing the link between biological and computer viruses. In this work, we establish correlations between a well-known computer virus, VirLock, with an equally well-studied biological virus, the bacteriophage &amp;amp;#981;6. VirLock is a formidable ransomware that encrypts user files and demands a ransom for data restoration. Drawing a parallel with the biological virus bacteriophage &amp;amp;#981;6, we uncover conceptual links like shared attributes and behaviors, as well as useful insights. Following this line of thought, we suggest efficient strategies based on a game theory perspective, which have the potential to address the infections caused by VirLock, and other viruses with analogous behavior. Moreover, we propose mathematical formulations that integrate real-world variables, providing a means to gauge virus severity and design robust defensive strategies and analytics. This interdisciplinary inquiry, fusing game theory, biology, and computer science, advances our understanding of virus behavior, paving the way for the development of effective countermeasures while presenting an alternative viewpoint. Throughout this theoretical exploration, we contribute to the ongoing discourse on computer virus behavior and stimulate new avenues for addressing digital threats. In particular, the formulas and framework developed in this work can facilitate better risk analysis and assessment, and become useful tools in penetration testing analysis, helping companies and organizations enhance their security.</p>
	]]></content:encoded>

	<dc:title>A Comparative Analysis of VirLock and Bacteriophage &amp;amp;#981;6 through the Lens of Game Theory</dc:title>
			<dc:creator>Dimitris Kostadimas</dc:creator>
			<dc:creator>Kalliopi Kastampolidou</dc:creator>
			<dc:creator>Theodore Andronikos</dc:creator>
		<dc:identifier>doi: 10.3390/analytics2040045</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2023-11-06</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2023-11-06</prism:publicationDate>
	<prism:volume>2</prism:volume>
	<prism:number>4</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>853</prism:startingPage>
		<prism:doi>10.3390/analytics2040045</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/2/4/45</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/2/4/44">

	<title>Analytics, Vol. 2, Pages 836-852: Can Oral Grades Predict Final Examination Scores? Case Study in a Higher Education Military Academy</title>
	<link>https://www.mdpi.com/2813-2203/2/4/44</link>
	<description>This paper investigates the correlation between oral grades and final written examination grades in a higher education military academy. A quantitative, correlational methodology utilizing linear regression analysis is employed. The data consist of undergraduate telecommunications and electronics engineering students&amp;amp;rsquo; grades in two courses offered during the fourth year of studies, and spans six academic years. Course One covers period 2017&amp;amp;ndash;2022, while Course Two, period 1 spans 2014&amp;amp;ndash;2018 and period 2 spans 2019&amp;amp;ndash;2022. In Course One oral grades are obtained by means of a midterm exam. In Course Two period 1, 30% of the oral grade comes from homework assignments and lab exercises, while the remaining 70% comes from a midterm exam. In Course Two period 2, oral grades are the result of various alternative assessment activities. In all cases, the final grade results from a traditional written examination given at the end of the semester. Correlation and predictive models between oral and final grades were examined. The results of the analysis demonstrated that, (a) under certain conditions, oral grades based more or less on midterm exams can be good predictors of final examination scores; (b) oral grades obtained through alternative assessment activities cannot predict final examination scores.</description>
	<pubDate>2023-11-02</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 2, Pages 836-852: Can Oral Grades Predict Final Examination Scores? Case Study in a Higher Education Military Academy</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/2/4/44">doi: 10.3390/analytics2040044</a></p>
	<p>Authors:
		Antonios Andreatos
		Apostolos Leros
		</p>
	<p>This paper investigates the correlation between oral grades and final written examination grades in a higher education military academy. A quantitative, correlational methodology utilizing linear regression analysis is employed. The data consist of undergraduate telecommunications and electronics engineering students&amp;amp;rsquo; grades in two courses offered during the fourth year of studies, and spans six academic years. Course One covers period 2017&amp;amp;ndash;2022, while Course Two, period 1 spans 2014&amp;amp;ndash;2018 and period 2 spans 2019&amp;amp;ndash;2022. In Course One oral grades are obtained by means of a midterm exam. In Course Two period 1, 30% of the oral grade comes from homework assignments and lab exercises, while the remaining 70% comes from a midterm exam. In Course Two period 2, oral grades are the result of various alternative assessment activities. In all cases, the final grade results from a traditional written examination given at the end of the semester. Correlation and predictive models between oral and final grades were examined. The results of the analysis demonstrated that, (a) under certain conditions, oral grades based more or less on midterm exams can be good predictors of final examination scores; (b) oral grades obtained through alternative assessment activities cannot predict final examination scores.</p>
	]]></content:encoded>

	<dc:title>Can Oral Grades Predict Final Examination Scores? Case Study in a Higher Education Military Academy</dc:title>
			<dc:creator>Antonios Andreatos</dc:creator>
			<dc:creator>Apostolos Leros</dc:creator>
		<dc:identifier>doi: 10.3390/analytics2040044</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2023-11-02</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2023-11-02</prism:publicationDate>
	<prism:volume>2</prism:volume>
	<prism:number>4</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>836</prism:startingPage>
		<prism:doi>10.3390/analytics2040044</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/2/4/44</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/2/4/43">

	<title>Analytics, Vol. 2, Pages 824-835: Relating the Ramsay Quotient Model to the Classical D-Scoring Rule</title>
	<link>https://www.mdpi.com/2813-2203/2/4/43</link>
	<description>In a series of papers, Dimitrov suggested the classical D-scoring rule for scoring items that give difficult items a higher weight while easier items receive a lower weight. The latent D-scoring model has been proposed to serve as a latent mirror of the classical D-scoring model. However, the item weights implied by this latent D-scoring model are typically only weakly related to the weights in the classical D-scoring model. To this end, this article proposes an alternative item response model, the modified Ramsay quotient model, that is better-suited as a latent mirror of the classical D-scoring model. The reasoning is based on analytical arguments and numerical illustrations.</description>
	<pubDate>2023-10-17</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 2, Pages 824-835: Relating the Ramsay Quotient Model to the Classical D-Scoring Rule</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/2/4/43">doi: 10.3390/analytics2040043</a></p>
	<p>Authors:
		Alexander Robitzsch
		</p>
	<p>In a series of papers, Dimitrov suggested the classical D-scoring rule for scoring items that give difficult items a higher weight while easier items receive a lower weight. The latent D-scoring model has been proposed to serve as a latent mirror of the classical D-scoring model. However, the item weights implied by this latent D-scoring model are typically only weakly related to the weights in the classical D-scoring model. To this end, this article proposes an alternative item response model, the modified Ramsay quotient model, that is better-suited as a latent mirror of the classical D-scoring model. The reasoning is based on analytical arguments and numerical illustrations.</p>
	]]></content:encoded>

	<dc:title>Relating the Ramsay Quotient Model to the Classical D-Scoring Rule</dc:title>
			<dc:creator>Alexander Robitzsch</dc:creator>
		<dc:identifier>doi: 10.3390/analytics2040043</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2023-10-17</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2023-10-17</prism:publicationDate>
	<prism:volume>2</prism:volume>
	<prism:number>4</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>824</prism:startingPage>
		<prism:doi>10.3390/analytics2040043</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/2/4/43</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/2/4/42">

	<title>Analytics, Vol. 2, Pages 809-823: An Exploration of Clustering Algorithms for Customer Segmentation in the UK Retail Market</title>
	<link>https://www.mdpi.com/2813-2203/2/4/42</link>
	<description>Recently, peoples&amp;amp;rsquo; awareness of online purchases has significantly risen. This has given rise to online retail platforms and the need for a better understanding of customer purchasing behaviour. Retail companies are pressed with the need to deal with a high volume of customer purchases, which requires sophisticated approaches to perform more accurate and efficient customer segmentation. Customer segmentation is a marketing analytical tool that aids customer-centric service and thus enhances profitability. In this paper, we aim to develop a customer segmentation model to improve decision-making processes in the retail market industry. To achieve this, we employed a UK-based online retail dataset obtained from the UCI machine learning repository. The retail dataset consists of 541,909 customer records and eight features. Our study adopted the RFM (recency, frequency, and monetary) framework to quantify customer values. Thereafter, we compared several state-of-the-art (SOTA) clustering algorithms, namely, K-means clustering, the Gaussian mixture model (GMM), density-based spatial clustering of applications with noise (DBSCAN), agglomerative clustering, and balanced iterative reducing and clustering using hierarchies (BIRCH). The results showed the GMM outperformed other approaches, with a Silhouette Score of 0.80.</description>
	<pubDate>2023-10-12</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 2, Pages 809-823: An Exploration of Clustering Algorithms for Customer Segmentation in the UK Retail Market</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/2/4/42">doi: 10.3390/analytics2040042</a></p>
	<p>Authors:
		Jeen Mary John
		Olamilekan Shobayo
		Bayode Ogunleye
		</p>
	<p>Recently, peoples&amp;amp;rsquo; awareness of online purchases has significantly risen. This has given rise to online retail platforms and the need for a better understanding of customer purchasing behaviour. Retail companies are pressed with the need to deal with a high volume of customer purchases, which requires sophisticated approaches to perform more accurate and efficient customer segmentation. Customer segmentation is a marketing analytical tool that aids customer-centric service and thus enhances profitability. In this paper, we aim to develop a customer segmentation model to improve decision-making processes in the retail market industry. To achieve this, we employed a UK-based online retail dataset obtained from the UCI machine learning repository. The retail dataset consists of 541,909 customer records and eight features. Our study adopted the RFM (recency, frequency, and monetary) framework to quantify customer values. Thereafter, we compared several state-of-the-art (SOTA) clustering algorithms, namely, K-means clustering, the Gaussian mixture model (GMM), density-based spatial clustering of applications with noise (DBSCAN), agglomerative clustering, and balanced iterative reducing and clustering using hierarchies (BIRCH). The results showed the GMM outperformed other approaches, with a Silhouette Score of 0.80.</p>
	]]></content:encoded>

	<dc:title>An Exploration of Clustering Algorithms for Customer Segmentation in the UK Retail Market</dc:title>
			<dc:creator>Jeen Mary John</dc:creator>
			<dc:creator>Olamilekan Shobayo</dc:creator>
			<dc:creator>Bayode Ogunleye</dc:creator>
		<dc:identifier>doi: 10.3390/analytics2040042</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2023-10-12</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2023-10-12</prism:publicationDate>
	<prism:volume>2</prism:volume>
	<prism:number>4</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>809</prism:startingPage>
		<prism:doi>10.3390/analytics2040042</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/2/4/42</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/2/4/41">

	<title>Analytics, Vol. 2, Pages 781-808: A Novel Curve Clustering Method for Functional Data: Applications to COVID-19 and Financial Data</title>
	<link>https://www.mdpi.com/2813-2203/2/4/41</link>
	<description>Functional data analysis has significantly enriched the landscape of existing data analysis methodologies, providing a new framework for comprehending data structures and extracting valuable insights. This paper is dedicated to addressing functional data clustering&amp;amp;mdash;a pivotal challenge within functional data analysis. Our contribution to this field manifests through the introduction of innovative clustering methodologies tailored specifically to functional curves. Initially, we present a proximity measure algorithm designed for functional curve clustering. This innovative clustering approach offers the flexibility to redefine measurement points on continuous functions, adapting to either equidistant or nonuniform arrangements, as dictated by the demands of the proximity measure. Central to this method is the &amp;amp;ldquo;proximity threshold&amp;amp;rdquo;, a critical parameter that governs the cluster count, and its selection is thoroughly explored. Subsequently, we propose a time-shift clustering algorithm designed for time-series data. This approach identifies historical data segments that share patterns similar to those observed in the present. To evaluate the effectiveness of our methodologies, we conduct comparisons with the classic K-means clustering method and apply them to simulated data, yielding encouraging simulation results. Moving beyond simulation, we apply the proposed proximity measure algorithm to COVID-19 data, yielding notable clustering accuracy. Additionally, the time-shift clustering algorithm is employed to analyse NASDAQ Composite data, successfully revealing underlying economic cycles.</description>
	<pubDate>2023-10-08</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 2, Pages 781-808: A Novel Curve Clustering Method for Functional Data: Applications to COVID-19 and Financial Data</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/2/4/41">doi: 10.3390/analytics2040041</a></p>
	<p>Authors:
		Ting Wei
		Bo Wang
		</p>
	<p>Functional data analysis has significantly enriched the landscape of existing data analysis methodologies, providing a new framework for comprehending data structures and extracting valuable insights. This paper is dedicated to addressing functional data clustering&amp;amp;mdash;a pivotal challenge within functional data analysis. Our contribution to this field manifests through the introduction of innovative clustering methodologies tailored specifically to functional curves. Initially, we present a proximity measure algorithm designed for functional curve clustering. This innovative clustering approach offers the flexibility to redefine measurement points on continuous functions, adapting to either equidistant or nonuniform arrangements, as dictated by the demands of the proximity measure. Central to this method is the &amp;amp;ldquo;proximity threshold&amp;amp;rdquo;, a critical parameter that governs the cluster count, and its selection is thoroughly explored. Subsequently, we propose a time-shift clustering algorithm designed for time-series data. This approach identifies historical data segments that share patterns similar to those observed in the present. To evaluate the effectiveness of our methodologies, we conduct comparisons with the classic K-means clustering method and apply them to simulated data, yielding encouraging simulation results. Moving beyond simulation, we apply the proposed proximity measure algorithm to COVID-19 data, yielding notable clustering accuracy. Additionally, the time-shift clustering algorithm is employed to analyse NASDAQ Composite data, successfully revealing underlying economic cycles.</p>
	]]></content:encoded>

	<dc:title>A Novel Curve Clustering Method for Functional Data: Applications to COVID-19 and Financial Data</dc:title>
			<dc:creator>Ting Wei</dc:creator>
			<dc:creator>Bo Wang</dc:creator>
		<dc:identifier>doi: 10.3390/analytics2040041</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2023-10-08</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2023-10-08</prism:publicationDate>
	<prism:volume>2</prism:volume>
	<prism:number>4</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>781</prism:startingPage>
		<prism:doi>10.3390/analytics2040041</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/2/4/41</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/2/3/40">

	<title>Analytics, Vol. 2, Pages 745-780: Image Segmentation of the Sudd Wetlands in South Sudan for Environmental Analytics by GRASS GIS Scripts</title>
	<link>https://www.mdpi.com/2813-2203/2/3/40</link>
	<description>This paper presents the object detection algorithms GRASS GIS applied for Landsat 8-9 OLI/TIRS data. The study area includes the Sudd wetlands located in South Sudan. This study describes a programming method for the automated processing of satellite images for environmental analytics, applying the scripting algorithms of GRASS GIS. This study documents how the land cover changed and developed over time in South Sudan with varying climate and environmental settings, indicating the variations in landscape patterns. A set of modules was used to process satellite images by scripting language. It streamlines the geospatial processing tasks. The functionality of the modules of GRASS GIS to image processing is called within scripts as subprocesses which automate operations. The cutting-edge tools of GRASS GIS present a cost-effective solution to remote sensing data modelling and analysis. This is based on the discrimination of the spectral reflectance of pixels on the raster scenes. Scripting algorithms of remote sensing data processing based on the GRASS GIS syntax are run from the terminal, enabling to pass commands to the module. This ensures the automation and high speed of image processing. The algorithm challenge is that landscape patterns differ substantially, and there are nonlinear dynamics in land cover types due to environmental factors and climate effects. Time series analysis of several multispectral images demonstrated changes in land cover types over the study area of the Sudd, South Sudan affected by environmental degradation of landscapes. The map is generated for each Landsat image from 2015 to 2023 using 481 maximum-likelihood discriminant analysis approaches of classification. The methodology includes image segmentation by &amp;amp;lsquo;i.segment&amp;amp;rsquo; module, image clustering and classification by &amp;amp;lsquo;i.cluster&amp;amp;rsquo; and &amp;amp;lsquo;i.maxlike&amp;amp;rsquo; modules, accuracy assessment by &amp;amp;lsquo;r.kappa&amp;amp;rsquo; module, and computing NDVI and cartographic mapping implemented using GRASS GIS. The benefits of object detection techniques for image analysis are demonstrated with the reported effects of various threshold levels of segmentation. The segmentation was performed 371 times with 90% of the threshold and minsize = 5; the process was converged in 37 to 41 iterations. The following segments are defined for images: 4515 for 2015, 4813 for 2016, 4114 for 2017, 5090 for 2018, 6021 for 2019, 3187 for 2020, 2445 for 2022, and 5181 for 2023. The percent convergence is 98% for the processed images. Detecting variations in land cover patterns is possible using spaceborne datasets and advanced applications of scripting algorithms. The implications of cartographic approach for environmental landscape analysis are discussed. The algorithm for image processing is based on a set of GRASS GIS wrapper functions for automated image classification.</description>
	<pubDate>2023-09-21</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 2, Pages 745-780: Image Segmentation of the Sudd Wetlands in South Sudan for Environmental Analytics by GRASS GIS Scripts</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/2/3/40">doi: 10.3390/analytics2030040</a></p>
	<p>Authors:
		Polina Lemenkova
		</p>
	<p>This paper presents the object detection algorithms GRASS GIS applied for Landsat 8-9 OLI/TIRS data. The study area includes the Sudd wetlands located in South Sudan. This study describes a programming method for the automated processing of satellite images for environmental analytics, applying the scripting algorithms of GRASS GIS. This study documents how the land cover changed and developed over time in South Sudan with varying climate and environmental settings, indicating the variations in landscape patterns. A set of modules was used to process satellite images by scripting language. It streamlines the geospatial processing tasks. The functionality of the modules of GRASS GIS to image processing is called within scripts as subprocesses which automate operations. The cutting-edge tools of GRASS GIS present a cost-effective solution to remote sensing data modelling and analysis. This is based on the discrimination of the spectral reflectance of pixels on the raster scenes. Scripting algorithms of remote sensing data processing based on the GRASS GIS syntax are run from the terminal, enabling to pass commands to the module. This ensures the automation and high speed of image processing. The algorithm challenge is that landscape patterns differ substantially, and there are nonlinear dynamics in land cover types due to environmental factors and climate effects. Time series analysis of several multispectral images demonstrated changes in land cover types over the study area of the Sudd, South Sudan affected by environmental degradation of landscapes. The map is generated for each Landsat image from 2015 to 2023 using 481 maximum-likelihood discriminant analysis approaches of classification. The methodology includes image segmentation by &amp;amp;lsquo;i.segment&amp;amp;rsquo; module, image clustering and classification by &amp;amp;lsquo;i.cluster&amp;amp;rsquo; and &amp;amp;lsquo;i.maxlike&amp;amp;rsquo; modules, accuracy assessment by &amp;amp;lsquo;r.kappa&amp;amp;rsquo; module, and computing NDVI and cartographic mapping implemented using GRASS GIS. The benefits of object detection techniques for image analysis are demonstrated with the reported effects of various threshold levels of segmentation. The segmentation was performed 371 times with 90% of the threshold and minsize = 5; the process was converged in 37 to 41 iterations. The following segments are defined for images: 4515 for 2015, 4813 for 2016, 4114 for 2017, 5090 for 2018, 6021 for 2019, 3187 for 2020, 2445 for 2022, and 5181 for 2023. The percent convergence is 98% for the processed images. Detecting variations in land cover patterns is possible using spaceborne datasets and advanced applications of scripting algorithms. The implications of cartographic approach for environmental landscape analysis are discussed. The algorithm for image processing is based on a set of GRASS GIS wrapper functions for automated image classification.</p>
	]]></content:encoded>

	<dc:title>Image Segmentation of the Sudd Wetlands in South Sudan for Environmental Analytics by GRASS GIS Scripts</dc:title>
			<dc:creator>Polina Lemenkova</dc:creator>
		<dc:identifier>doi: 10.3390/analytics2030040</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2023-09-21</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2023-09-21</prism:publicationDate>
	<prism:volume>2</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>745</prism:startingPage>
		<prism:doi>10.3390/analytics2030040</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/2/3/40</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/2/3/39">

	<title>Analytics, Vol. 2, Pages 708-744: Application of Machine Learning and Deep Learning Models in Prostate Cancer Diagnosis Using Medical Images: A Systematic Review</title>
	<link>https://www.mdpi.com/2813-2203/2/3/39</link>
	<description>Introduction: Prostate cancer (PCa) is one of the deadliest and most common causes of malignancy and death in men worldwide, with a higher prevalence and mortality in developing countries specifically. Factors such as age, family history, race and certain genetic mutations are some of the factors contributing to the occurrence of PCa in men. Recent advances in technology and algorithms gave rise to the computer-aided diagnosis (CAD) of PCa. With the availability of medical image datasets and emerging trends in state-of-the-art machine and deep learning techniques, there has been a growth in recent related publications. Materials and Methods: In this study, we present a systematic review of PCa diagnosis with medical images using machine learning and deep learning techniques. We conducted a thorough review of the relevant studies indexed in four databases (IEEE, PubMed, Springer and ScienceDirect) using the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) guidelines. With well-defined search terms, a total of 608 articles were identified, and 77 met the final inclusion criteria. The key elements in the included papers are presented and conclusions are drawn from them. Results: The findings show that the United States has the most research in PCa diagnosis with machine learning, Magnetic Resonance Images are the most used datasets and transfer learning is the most used method of diagnosing PCa in recent times. In addition, some available PCa datasets and some key considerations for the choice of loss function in the deep learning models are presented. The limitations and lessons learnt are discussed, and some key recommendations are made. Conclusion: The discoveries and the conclusions of this work are organized so as to enable researchers in the same domain to use this work and make crucial implementation decisions.</description>
	<pubDate>2023-09-19</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 2, Pages 708-744: Application of Machine Learning and Deep Learning Models in Prostate Cancer Diagnosis Using Medical Images: A Systematic Review</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/2/3/39">doi: 10.3390/analytics2030039</a></p>
	<p>Authors:
		Olusola Olabanjo
		Ashiribo Wusu
		Mauton Asokere
		Oseni Afisi
		Basheerat Okugbesan
		Olufemi Olabanjo
		Olusegun Folorunso
		Manuel Mazzara
		</p>
	<p>Introduction: Prostate cancer (PCa) is one of the deadliest and most common causes of malignancy and death in men worldwide, with a higher prevalence and mortality in developing countries specifically. Factors such as age, family history, race and certain genetic mutations are some of the factors contributing to the occurrence of PCa in men. Recent advances in technology and algorithms gave rise to the computer-aided diagnosis (CAD) of PCa. With the availability of medical image datasets and emerging trends in state-of-the-art machine and deep learning techniques, there has been a growth in recent related publications. Materials and Methods: In this study, we present a systematic review of PCa diagnosis with medical images using machine learning and deep learning techniques. We conducted a thorough review of the relevant studies indexed in four databases (IEEE, PubMed, Springer and ScienceDirect) using the Preferred Reporting Items for Systematic Reviews and Meta-Analyses (PRISMA) guidelines. With well-defined search terms, a total of 608 articles were identified, and 77 met the final inclusion criteria. The key elements in the included papers are presented and conclusions are drawn from them. Results: The findings show that the United States has the most research in PCa diagnosis with machine learning, Magnetic Resonance Images are the most used datasets and transfer learning is the most used method of diagnosing PCa in recent times. In addition, some available PCa datasets and some key considerations for the choice of loss function in the deep learning models are presented. The limitations and lessons learnt are discussed, and some key recommendations are made. Conclusion: The discoveries and the conclusions of this work are organized so as to enable researchers in the same domain to use this work and make crucial implementation decisions.</p>
	]]></content:encoded>

	<dc:title>Application of Machine Learning and Deep Learning Models in Prostate Cancer Diagnosis Using Medical Images: A Systematic Review</dc:title>
			<dc:creator>Olusola Olabanjo</dc:creator>
			<dc:creator>Ashiribo Wusu</dc:creator>
			<dc:creator>Mauton Asokere</dc:creator>
			<dc:creator>Oseni Afisi</dc:creator>
			<dc:creator>Basheerat Okugbesan</dc:creator>
			<dc:creator>Olufemi Olabanjo</dc:creator>
			<dc:creator>Olusegun Folorunso</dc:creator>
			<dc:creator>Manuel Mazzara</dc:creator>
		<dc:identifier>doi: 10.3390/analytics2030039</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2023-09-19</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2023-09-19</prism:publicationDate>
	<prism:volume>2</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Review</prism:section>
	<prism:startingPage>708</prism:startingPage>
		<prism:doi>10.3390/analytics2030039</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/2/3/39</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/2/3/38">

	<title>Analytics, Vol. 2, Pages 694-707: The Use of a Large Language Model for Cyberbullying Detection</title>
	<link>https://www.mdpi.com/2813-2203/2/3/38</link>
	<description>The dominance of social media has added to the channels of bullying for perpetrators. Unfortunately, cyberbullying (CB) is the most prevalent phenomenon in today&amp;amp;rsquo;s cyber world, and is a severe threat to the mental and physical health of citizens. This opens the need to develop a robust system to prevent bullying content from online forums, blogs, and social media platforms to manage the impact in our society. Several machine learning (ML) algorithms have been proposed for this purpose. However, their performances are not consistent due to high class imbalance and generalisation issues. In recent years, large language models (LLMs) like BERT and RoBERTa have achieved state-of-the-art (SOTA) results in several natural language processing (NLP) tasks. Unfortunately, the LLMs have not been applied extensively for CB detection. In our paper, we explored the use of these models for cyberbullying (CB) detection. We have prepared a new dataset (D2) from existing studies (Formspring and Twitter). Our experimental results for dataset D1 and D2 showed that RoBERTa outperformed other models.</description>
	<pubDate>2023-09-06</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 2, Pages 694-707: The Use of a Large Language Model for Cyberbullying Detection</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/2/3/38">doi: 10.3390/analytics2030038</a></p>
	<p>Authors:
		Bayode Ogunleye
		Babitha Dharmaraj
		</p>
	<p>The dominance of social media has added to the channels of bullying for perpetrators. Unfortunately, cyberbullying (CB) is the most prevalent phenomenon in today&amp;amp;rsquo;s cyber world, and is a severe threat to the mental and physical health of citizens. This opens the need to develop a robust system to prevent bullying content from online forums, blogs, and social media platforms to manage the impact in our society. Several machine learning (ML) algorithms have been proposed for this purpose. However, their performances are not consistent due to high class imbalance and generalisation issues. In recent years, large language models (LLMs) like BERT and RoBERTa have achieved state-of-the-art (SOTA) results in several natural language processing (NLP) tasks. Unfortunately, the LLMs have not been applied extensively for CB detection. In our paper, we explored the use of these models for cyberbullying (CB) detection. We have prepared a new dataset (D2) from existing studies (Formspring and Twitter). Our experimental results for dataset D1 and D2 showed that RoBERTa outperformed other models.</p>
	]]></content:encoded>

	<dc:title>The Use of a Large Language Model for Cyberbullying Detection</dc:title>
			<dc:creator>Bayode Ogunleye</dc:creator>
			<dc:creator>Babitha Dharmaraj</dc:creator>
		<dc:identifier>doi: 10.3390/analytics2030038</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2023-09-06</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2023-09-06</prism:publicationDate>
	<prism:volume>2</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>694</prism:startingPage>
		<prism:doi>10.3390/analytics2030038</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/2/3/38</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/2/3/37">

	<title>Analytics, Vol. 2, Pages 676-693: Heterogeneous Ensemble for Medical Data Classification</title>
	<link>https://www.mdpi.com/2813-2203/2/3/37</link>
	<description>For robust classification, selecting a proper classifier is of primary importance. However, selecting the best classifiers depends on the problem, as some classifiers work better at some tasks than on others. Despite the many results collected in the literature, the support vector machine (SVM) remains the leading adopted solution in many domains, thanks to its ease of use. In this paper, we propose a new method based on convolutional neural networks (CNNs) as an alternative to SVM. CNNs are specialized in processing data in a grid-like topology that usually represents images. To enable CNNs to work on different data types, we investigate reshaping one-dimensional vector representations into two-dimensional matrices and compared different approaches for feeding standard CNNs using two-dimensional feature vector representations. We evaluate the different techniques proposing a heterogeneous ensemble based on three classifiers: an SVM, a model based on random subspace of rotation boosting (RB), and a CNN. The robustness of our approach is tested across a set of benchmark datasets that represent a wide range of medical classification tasks. The proposed ensembles provide promising performance on all datasets.</description>
	<pubDate>2023-09-04</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 2, Pages 676-693: Heterogeneous Ensemble for Medical Data Classification</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/2/3/37">doi: 10.3390/analytics2030037</a></p>
	<p>Authors:
		Loris Nanni
		Sheryl Brahnam
		Andrea Loreggia
		Leonardo Barcellona
		</p>
	<p>For robust classification, selecting a proper classifier is of primary importance. However, selecting the best classifiers depends on the problem, as some classifiers work better at some tasks than on others. Despite the many results collected in the literature, the support vector machine (SVM) remains the leading adopted solution in many domains, thanks to its ease of use. In this paper, we propose a new method based on convolutional neural networks (CNNs) as an alternative to SVM. CNNs are specialized in processing data in a grid-like topology that usually represents images. To enable CNNs to work on different data types, we investigate reshaping one-dimensional vector representations into two-dimensional matrices and compared different approaches for feeding standard CNNs using two-dimensional feature vector representations. We evaluate the different techniques proposing a heterogeneous ensemble based on three classifiers: an SVM, a model based on random subspace of rotation boosting (RB), and a CNN. The robustness of our approach is tested across a set of benchmark datasets that represent a wide range of medical classification tasks. The proposed ensembles provide promising performance on all datasets.</p>
	]]></content:encoded>

	<dc:title>Heterogeneous Ensemble for Medical Data Classification</dc:title>
			<dc:creator>Loris Nanni</dc:creator>
			<dc:creator>Sheryl Brahnam</dc:creator>
			<dc:creator>Andrea Loreggia</dc:creator>
			<dc:creator>Leonardo Barcellona</dc:creator>
		<dc:identifier>doi: 10.3390/analytics2030037</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2023-09-04</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2023-09-04</prism:publicationDate>
	<prism:volume>2</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>676</prism:startingPage>
		<prism:doi>10.3390/analytics2030037</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/2/3/37</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/2/3/36">

	<title>Analytics, Vol. 2, Pages 656-675: Surgery Scheduling and Perioperative Care: Smoothing and Visualizing Elective Surgery and Recovery Patient Flow</title>
	<link>https://www.mdpi.com/2813-2203/2/3/36</link>
	<description>This paper addresses the practical problem of scheduling operating room (OR) elective surgeries to minimize the likelihood of surgical delays caused by the unavailability of capacity for patient recovery in a central post-anesthesia care unit (PACU). We segregate patients according to their patterns of flow through a multi-stage perioperative system and use characteristics of surgery type and surgeon booking times to predict time intervals for patient procedures and subsequent recoveries. Working with a hospital in which 50+ procedures are performed in 15+ ORs most weekdays, we develop a constraint programming (CP) model that takes the hospital&amp;amp;rsquo;s elective surgery pre-schedule as input and produces a recommended alternate schedule designed to minimize the expected peak number of patients in the PACU over the course of the day. Our model was developed from the hospital&amp;amp;rsquo;s data and evaluated through its application to daily schedules during a testing period. Schedules generated by our model indicated the potential to reduce the peak PACU load substantially, 20-30% during most days in our study period, or alternatively reduce average patient flow time by up to 15% given the same PACU peak load. We also developed tools for schedule visualization that can be used to aid management both before and after surgery day; plan PACU resources; propose critical schedule changes; identify the timing, location, and root causes of delay; and to discern the differences in surgical specialty case mixes and their potential impacts on the system. This work is especially timely given high surgical wait times in Ontario which even got worse due to the COVID-19 pandemic.</description>
	<pubDate>2023-08-21</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 2, Pages 656-675: Surgery Scheduling and Perioperative Care: Smoothing and Visualizing Elective Surgery and Recovery Patient Flow</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/2/3/36">doi: 10.3390/analytics2030036</a></p>
	<p>Authors:
		John S. F. Lyons
		Mehmet A. Begen
		Peter C. Bell
		</p>
	<p>This paper addresses the practical problem of scheduling operating room (OR) elective surgeries to minimize the likelihood of surgical delays caused by the unavailability of capacity for patient recovery in a central post-anesthesia care unit (PACU). We segregate patients according to their patterns of flow through a multi-stage perioperative system and use characteristics of surgery type and surgeon booking times to predict time intervals for patient procedures and subsequent recoveries. Working with a hospital in which 50+ procedures are performed in 15+ ORs most weekdays, we develop a constraint programming (CP) model that takes the hospital&amp;amp;rsquo;s elective surgery pre-schedule as input and produces a recommended alternate schedule designed to minimize the expected peak number of patients in the PACU over the course of the day. Our model was developed from the hospital&amp;amp;rsquo;s data and evaluated through its application to daily schedules during a testing period. Schedules generated by our model indicated the potential to reduce the peak PACU load substantially, 20-30% during most days in our study period, or alternatively reduce average patient flow time by up to 15% given the same PACU peak load. We also developed tools for schedule visualization that can be used to aid management both before and after surgery day; plan PACU resources; propose critical schedule changes; identify the timing, location, and root causes of delay; and to discern the differences in surgical specialty case mixes and their potential impacts on the system. This work is especially timely given high surgical wait times in Ontario which even got worse due to the COVID-19 pandemic.</p>
	]]></content:encoded>

	<dc:title>Surgery Scheduling and Perioperative Care: Smoothing and Visualizing Elective Surgery and Recovery Patient Flow</dc:title>
			<dc:creator>John S. F. Lyons</dc:creator>
			<dc:creator>Mehmet A. Begen</dc:creator>
			<dc:creator>Peter C. Bell</dc:creator>
		<dc:identifier>doi: 10.3390/analytics2030036</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2023-08-21</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2023-08-21</prism:publicationDate>
	<prism:volume>2</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>656</prism:startingPage>
		<prism:doi>10.3390/analytics2030036</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/2/3/36</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/2/3/35">

	<title>Analytics, Vol. 2, Pages 618-655: Cyberpsychology: A Longitudinal Analysis of Cyber Adversarial Tactics and Techniques</title>
	<link>https://www.mdpi.com/2813-2203/2/3/35</link>
	<description>The rapid proliferation of cyberthreats necessitates a robust understanding of their evolution and associated tactics, as found in this study. A longitudinal analysis of these threats was conducted, utilizing a six-year data set obtained from a deception network, which emphasized its significance in the study&amp;amp;rsquo;s primary aim: the exhaustive exploration of the tactics and strategies utilized by cybercriminals and how these tactics and techniques evolved in sophistication and target specificity over time. Different cyberattack instances were dissected and interpreted, with the patterns behind target selection shown. The focus was on unveiling patterns behind target selection and highlighting recurring techniques and emerging trends. The study&amp;amp;rsquo;s methodological design incorporated data preprocessing, exploratory data analysis, clustering and anomaly detection, temporal analysis, and cross-referencing. The validation process underscored the reliability and robustness of the findings, providing evidence of increasingly sophisticated, targeted cyberattacks. The work identified three distinct network traffic behavior clusters and temporal attack patterns. A validated scoring mechanism provided a benchmark for network anomalies, applicable for predictive analysis and facilitating comparative study of network behaviors. This benchmarking aids organizations in proactively identifying and responding to potential threats. The study significantly contributed to the cybersecurity discourse, offering insights that could guide the development of more effective defense strategies. The need for further investigation into the nature of detected anomalies was acknowledged, advocating for continuous research and proactive defense strategies in the face of the constantly evolving landscape of cyberthreats.</description>
	<pubDate>2023-08-11</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 2, Pages 618-655: Cyberpsychology: A Longitudinal Analysis of Cyber Adversarial Tactics and Techniques</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/2/3/35">doi: 10.3390/analytics2030035</a></p>
	<p>Authors:
		Marshall S. Rich
		</p>
	<p>The rapid proliferation of cyberthreats necessitates a robust understanding of their evolution and associated tactics, as found in this study. A longitudinal analysis of these threats was conducted, utilizing a six-year data set obtained from a deception network, which emphasized its significance in the study&amp;amp;rsquo;s primary aim: the exhaustive exploration of the tactics and strategies utilized by cybercriminals and how these tactics and techniques evolved in sophistication and target specificity over time. Different cyberattack instances were dissected and interpreted, with the patterns behind target selection shown. The focus was on unveiling patterns behind target selection and highlighting recurring techniques and emerging trends. The study&amp;amp;rsquo;s methodological design incorporated data preprocessing, exploratory data analysis, clustering and anomaly detection, temporal analysis, and cross-referencing. The validation process underscored the reliability and robustness of the findings, providing evidence of increasingly sophisticated, targeted cyberattacks. The work identified three distinct network traffic behavior clusters and temporal attack patterns. A validated scoring mechanism provided a benchmark for network anomalies, applicable for predictive analysis and facilitating comparative study of network behaviors. This benchmarking aids organizations in proactively identifying and responding to potential threats. The study significantly contributed to the cybersecurity discourse, offering insights that could guide the development of more effective defense strategies. The need for further investigation into the nature of detected anomalies was acknowledged, advocating for continuous research and proactive defense strategies in the face of the constantly evolving landscape of cyberthreats.</p>
	]]></content:encoded>

	<dc:title>Cyberpsychology: A Longitudinal Analysis of Cyber Adversarial Tactics and Techniques</dc:title>
			<dc:creator>Marshall S. Rich</dc:creator>
		<dc:identifier>doi: 10.3390/analytics2030035</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2023-08-11</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2023-08-11</prism:publicationDate>
	<prism:volume>2</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>618</prism:startingPage>
		<prism:doi>10.3390/analytics2030035</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/2/3/35</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/2/3/34">

	<title>Analytics, Vol. 2, Pages 604-617: Prediction of Stroke Disease with Demographic and Behavioural Data Using Random Forest Algorithm</title>
	<link>https://www.mdpi.com/2813-2203/2/3/34</link>
	<description>Stroke is a major cause of death worldwide, resulting from a blockage in the flow of blood to different parts of the brain. Many studies have proposed a stroke disease prediction model using medical features applied to deep learning (DL) algorithms to reduce its occurrence. However, these studies pay less attention to the predictors (both demographic and behavioural). Our study considers interpretability, robustness, and generalisation as key themes for deploying algorithms in the medical domain. Based on this background, we propose the use of random forest for stroke incidence prediction. Results from our experiment showed that random forest (RF) outperformed decision tree (DT) and logistic regression (LR) with a macro F1 score of 94%. Our findings indicated age and body mass index (BMI) as the most significant predictors of stroke disease incidence.</description>
	<pubDate>2023-08-02</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 2, Pages 604-617: Prediction of Stroke Disease with Demographic and Behavioural Data Using Random Forest Algorithm</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/2/3/34">doi: 10.3390/analytics2030034</a></p>
	<p>Authors:
		Olamilekan Shobayo
		Oluwafemi Zachariah
		Modupe Olufunke Odusami
		Bayode Ogunleye
		</p>
	<p>Stroke is a major cause of death worldwide, resulting from a blockage in the flow of blood to different parts of the brain. Many studies have proposed a stroke disease prediction model using medical features applied to deep learning (DL) algorithms to reduce its occurrence. However, these studies pay less attention to the predictors (both demographic and behavioural). Our study considers interpretability, robustness, and generalisation as key themes for deploying algorithms in the medical domain. Based on this background, we propose the use of random forest for stroke incidence prediction. Results from our experiment showed that random forest (RF) outperformed decision tree (DT) and logistic regression (LR) with a macro F1 score of 94%. Our findings indicated age and body mass index (BMI) as the most significant predictors of stroke disease incidence.</p>
	]]></content:encoded>

	<dc:title>Prediction of Stroke Disease with Demographic and Behavioural Data Using Random Forest Algorithm</dc:title>
			<dc:creator>Olamilekan Shobayo</dc:creator>
			<dc:creator>Oluwafemi Zachariah</dc:creator>
			<dc:creator>Modupe Olufunke Odusami</dc:creator>
			<dc:creator>Bayode Ogunleye</dc:creator>
		<dc:identifier>doi: 10.3390/analytics2030034</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2023-08-02</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2023-08-02</prism:publicationDate>
	<prism:volume>2</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>604</prism:startingPage>
		<prism:doi>10.3390/analytics2030034</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/2/3/34</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/2/3/33">

	<title>Analytics, Vol. 2, Pages 592-603: Identification of Patterns in the Stock Market through Unsupervised Algorithms</title>
	<link>https://www.mdpi.com/2813-2203/2/3/33</link>
	<description>Making predictions in the stock market is a challenging task. At the same time, several studies have focused on forecasting the future behavior of the market and classifying financial assets. A different approach is to classify correlated data to discover patterns and atypical behaviors in them. In this study, we propose applying unsupervised algorithms to process, model, and cluster related data from two different data sources, i.e., Google News and Yahoo Finance, to identify conditions in the stock market that might help to support the investment decision-making process. We applied principal component analysis (PCA) and a k-means clustering approach to group data according to their principal characteristics. We identified four conditions in the stock market, one comprising the least amount of data, characterized by high volatility. The main results show that, regularly, the stock market tends to have a steady performance. However, atypical conditions are conducive to higher volatility.</description>
	<pubDate>2023-07-27</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 2, Pages 592-603: Identification of Patterns in the Stock Market through Unsupervised Algorithms</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/2/3/33">doi: 10.3390/analytics2030033</a></p>
	<p>Authors:
		Adrian Barradas
		Rosa-Maria Canton-Croda
		Damian-Emilio Gibaja-Romero
		</p>
	<p>Making predictions in the stock market is a challenging task. At the same time, several studies have focused on forecasting the future behavior of the market and classifying financial assets. A different approach is to classify correlated data to discover patterns and atypical behaviors in them. In this study, we propose applying unsupervised algorithms to process, model, and cluster related data from two different data sources, i.e., Google News and Yahoo Finance, to identify conditions in the stock market that might help to support the investment decision-making process. We applied principal component analysis (PCA) and a k-means clustering approach to group data according to their principal characteristics. We identified four conditions in the stock market, one comprising the least amount of data, characterized by high volatility. The main results show that, regularly, the stock market tends to have a steady performance. However, atypical conditions are conducive to higher volatility.</p>
	]]></content:encoded>

	<dc:title>Identification of Patterns in the Stock Market through Unsupervised Algorithms</dc:title>
			<dc:creator>Adrian Barradas</dc:creator>
			<dc:creator>Rosa-Maria Canton-Croda</dc:creator>
			<dc:creator>Damian-Emilio Gibaja-Romero</dc:creator>
		<dc:identifier>doi: 10.3390/analytics2030033</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2023-07-27</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2023-07-27</prism:publicationDate>
	<prism:volume>2</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>592</prism:startingPage>
		<prism:doi>10.3390/analytics2030033</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/2/3/33</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/2/3/32">

	<title>Analytics, Vol. 2, Pages 577-591: Streamflow Estimation through Coupling of Hieararchical Clustering Analysis and Regression Analysis&amp;mdash;A Case Study in Euphrates-Tigris Basin</title>
	<link>https://www.mdpi.com/2813-2203/2/3/32</link>
	<description>In this study, the resilience of designed water systems in the face of limited streamflow gauging stations and escalating global warming impacts were investigated. By performing a regression analysis, simulated meteorological data with observed streamflow from 1971 to 2020 across 33 stream gauging stations in the Euphrates-Tigris Basin were correlated. Utilizing the Ordinary Least Squares regression method, streamflow for 2020&amp;amp;ndash;2100 using simulated meteorological data under RCP 4.5 and RCP 8.5 scenarios in CORDEX-EURO and CORDEX-MENA domains were also predicted. Streamflow variability was calculated based on meteorological variables and station morphological characteristics, particularly evapotranspiration. Hierarchical clustering analysis identified two clusters among the stream gauging stations, and for each cluster, two streamflow equations were derived. The regression analysis achieved robust streamflow predictions using six representative climate variables, with adj. R2 values of 0.7&amp;amp;ndash;0.85 across all models, primarily influenced by evapotranspiration. The use of a global model led to a 10% decrease in prediction capabilities for all CORDEX models based on R2 performance. This study emphasizes the importance of region homogeneity in estimating streamflow, encompassing both geographical and hydro-meteorological characteristics.</description>
	<pubDate>2023-07-13</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 2, Pages 577-591: Streamflow Estimation through Coupling of Hieararchical Clustering Analysis and Regression Analysis&amp;mdash;A Case Study in Euphrates-Tigris Basin</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/2/3/32">doi: 10.3390/analytics2030032</a></p>
	<p>Authors:
		Goksel Ezgi Guzey
		Bihrat Onoz
		</p>
	<p>In this study, the resilience of designed water systems in the face of limited streamflow gauging stations and escalating global warming impacts were investigated. By performing a regression analysis, simulated meteorological data with observed streamflow from 1971 to 2020 across 33 stream gauging stations in the Euphrates-Tigris Basin were correlated. Utilizing the Ordinary Least Squares regression method, streamflow for 2020&amp;amp;ndash;2100 using simulated meteorological data under RCP 4.5 and RCP 8.5 scenarios in CORDEX-EURO and CORDEX-MENA domains were also predicted. Streamflow variability was calculated based on meteorological variables and station morphological characteristics, particularly evapotranspiration. Hierarchical clustering analysis identified two clusters among the stream gauging stations, and for each cluster, two streamflow equations were derived. The regression analysis achieved robust streamflow predictions using six representative climate variables, with adj. R2 values of 0.7&amp;amp;ndash;0.85 across all models, primarily influenced by evapotranspiration. The use of a global model led to a 10% decrease in prediction capabilities for all CORDEX models based on R2 performance. This study emphasizes the importance of region homogeneity in estimating streamflow, encompassing both geographical and hydro-meteorological characteristics.</p>
	]]></content:encoded>

	<dc:title>Streamflow Estimation through Coupling of Hieararchical Clustering Analysis and Regression Analysis&amp;amp;mdash;A Case Study in Euphrates-Tigris Basin</dc:title>
			<dc:creator>Goksel Ezgi Guzey</dc:creator>
			<dc:creator>Bihrat Onoz</dc:creator>
		<dc:identifier>doi: 10.3390/analytics2030032</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2023-07-13</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2023-07-13</prism:publicationDate>
	<prism:volume>2</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>577</prism:startingPage>
		<prism:doi>10.3390/analytics2030032</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/2/3/32</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/2/3/31">

	<title>Analytics, Vol. 2, Pages 560-576: Hierarchical Model-Based Deep Reinforcement Learning for Single-Asset Trading</title>
	<link>https://www.mdpi.com/2813-2203/2/3/31</link>
	<description>We present a hierarchical reinforcement learning (RL) architecture that employs various low-level agents to act in the trading environment, i.e., the market. The highest-level agent selects from among a group of specialized agents, and then the selected agent decides when to sell or buy a single asset for a period of time. This period can be variable according to a termination function. We hypothesized that, due to different market regimes, more than one single agent is needed when trying to learn from such heterogeneous data, and instead, multiple agents will perform better, with each one specializing in a subset of the data. We use k-meansclustering to partition the data and train each agent with a different cluster. Partitioning the input data also helps model-based RL (MBRL), where models can be heterogeneous. We also add two simple decision-making models to the set of low-level agents, diversifying the pool of available agents, and thus increasing overall behavioral flexibility. We perform multiple experiments showing the strengths of a hierarchical approach and test various prediction models at both levels. We also use a risk-based reward at the high level, which transforms the overall problem into a risk-return optimization. This type of reward shows a significant reduction in risk while minimally reducing profits. Overall, the hierarchical approach shows significant promise, especially when the pool of low-level agents is highly diverse. The usefulness of such a system is clear, especially for human-devised strategies, which could be incorporated in a sound manner into larger, powerful automatic systems.</description>
	<pubDate>2023-07-11</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 2, Pages 560-576: Hierarchical Model-Based Deep Reinforcement Learning for Single-Asset Trading</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/2/3/31">doi: 10.3390/analytics2030031</a></p>
	<p>Authors:
		Adrian Millea
		</p>
	<p>We present a hierarchical reinforcement learning (RL) architecture that employs various low-level agents to act in the trading environment, i.e., the market. The highest-level agent selects from among a group of specialized agents, and then the selected agent decides when to sell or buy a single asset for a period of time. This period can be variable according to a termination function. We hypothesized that, due to different market regimes, more than one single agent is needed when trying to learn from such heterogeneous data, and instead, multiple agents will perform better, with each one specializing in a subset of the data. We use k-meansclustering to partition the data and train each agent with a different cluster. Partitioning the input data also helps model-based RL (MBRL), where models can be heterogeneous. We also add two simple decision-making models to the set of low-level agents, diversifying the pool of available agents, and thus increasing overall behavioral flexibility. We perform multiple experiments showing the strengths of a hierarchical approach and test various prediction models at both levels. We also use a risk-based reward at the high level, which transforms the overall problem into a risk-return optimization. This type of reward shows a significant reduction in risk while minimally reducing profits. Overall, the hierarchical approach shows significant promise, especially when the pool of low-level agents is highly diverse. The usefulness of such a system is clear, especially for human-devised strategies, which could be incorporated in a sound manner into larger, powerful automatic systems.</p>
	]]></content:encoded>

	<dc:title>Hierarchical Model-Based Deep Reinforcement Learning for Single-Asset Trading</dc:title>
			<dc:creator>Adrian Millea</dc:creator>
		<dc:identifier>doi: 10.3390/analytics2030031</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2023-07-11</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2023-07-11</prism:publicationDate>
	<prism:volume>2</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>560</prism:startingPage>
		<prism:doi>10.3390/analytics2030031</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/2/3/31</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/2/3/30">

	<title>Analytics, Vol. 2, Pages 546-559: occams: A Text Summarization Package</title>
	<link>https://www.mdpi.com/2813-2203/2/3/30</link>
	<description>Extractive text summarization selects asmall subset of sentences from a document, which gives good &amp;amp;ldquo;coverage&amp;amp;rdquo; of a document. When given a set of term weights indicating the importance of the terms, the concept of coverage may be formalized into a combinatorial optimization problem known as the budgeted maximum coverage problem. Extractive methods in this class are known to beamong the best of classic extractive summarization systems. This paper gives a synopsis of thesoftware package occams, which is a multilingual extractive single and multi-document summarization package based on an algorithm giving an optimal approximation to the budgeted maximum coverage problem. The occams package is written in Python and provides an easy-to-use modular interface, allowing it to work in conjunction with popular Python NLP packages, such as nltk, stanza or spacy.</description>
	<pubDate>2023-06-30</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 2, Pages 546-559: occams: A Text Summarization Package</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/2/3/30">doi: 10.3390/analytics2030030</a></p>
	<p>Authors:
		Clinton T. White
		Neil P. Molino
		Julia S. Yang
		John M. Conroy
		</p>
	<p>Extractive text summarization selects asmall subset of sentences from a document, which gives good &amp;amp;ldquo;coverage&amp;amp;rdquo; of a document. When given a set of term weights indicating the importance of the terms, the concept of coverage may be formalized into a combinatorial optimization problem known as the budgeted maximum coverage problem. Extractive methods in this class are known to beamong the best of classic extractive summarization systems. This paper gives a synopsis of thesoftware package occams, which is a multilingual extractive single and multi-document summarization package based on an algorithm giving an optimal approximation to the budgeted maximum coverage problem. The occams package is written in Python and provides an easy-to-use modular interface, allowing it to work in conjunction with popular Python NLP packages, such as nltk, stanza or spacy.</p>
	]]></content:encoded>

	<dc:title>occams: A Text Summarization Package</dc:title>
			<dc:creator>Clinton T. White</dc:creator>
			<dc:creator>Neil P. Molino</dc:creator>
			<dc:creator>Julia S. Yang</dc:creator>
			<dc:creator>John M. Conroy</dc:creator>
		<dc:identifier>doi: 10.3390/analytics2030030</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2023-06-30</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2023-06-30</prism:publicationDate>
	<prism:volume>2</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>546</prism:startingPage>
		<prism:doi>10.3390/analytics2030030</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/2/3/30</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/2/2/29">

	<title>Analytics, Vol. 2, Pages 530-545: Bayesian Mixture Copula Estimation and Selection with Applications</title>
	<link>https://www.mdpi.com/2813-2203/2/2/29</link>
	<description>Mixture copulas are popular and essential tools for studying complex dependencies among variables. However, selecting the correct mixture models often involves repeated testing and estimations using criteria such as AIC, which could require effort and time. In this paper, we propose a method that would enable us to select and estimate the correct mixture copulas simultaneously. This is accomplished by first overfitting the model and then conducting the Bayesian estimations. We verify the correctness of our approach by numerical simulations. Finally, the real data analysis is performed by studying the dependencies among three major financial markets.</description>
	<pubDate>2023-06-15</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 2, Pages 530-545: Bayesian Mixture Copula Estimation and Selection with Applications</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/2/2/29">doi: 10.3390/analytics2020029</a></p>
	<p>Authors:
		Yujian Liu
		Dejun Xie
		Siyi Yu
		</p>
	<p>Mixture copulas are popular and essential tools for studying complex dependencies among variables. However, selecting the correct mixture models often involves repeated testing and estimations using criteria such as AIC, which could require effort and time. In this paper, we propose a method that would enable us to select and estimate the correct mixture copulas simultaneously. This is accomplished by first overfitting the model and then conducting the Bayesian estimations. We verify the correctness of our approach by numerical simulations. Finally, the real data analysis is performed by studying the dependencies among three major financial markets.</p>
	]]></content:encoded>

	<dc:title>Bayesian Mixture Copula Estimation and Selection with Applications</dc:title>
			<dc:creator>Yujian Liu</dc:creator>
			<dc:creator>Dejun Xie</dc:creator>
			<dc:creator>Siyi Yu</dc:creator>
		<dc:identifier>doi: 10.3390/analytics2020029</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2023-06-15</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2023-06-15</prism:publicationDate>
	<prism:volume>2</prism:volume>
	<prism:number>2</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>530</prism:startingPage>
		<prism:doi>10.3390/analytics2020029</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/2/2/29</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/2/2/28">

	<title>Analytics, Vol. 2, Pages 509-529: Preliminary Perspectives on Information Passing in the Intelligence Community</title>
	<link>https://www.mdpi.com/2813-2203/2/2/28</link>
	<description>Analyst sensemaking research typically focuses on individual or small groups conducting intelligence tasks. This has helped understand information retrieval tasks and how people communicate information. As a part of the grand challenge of the Summer Conference on Applied Data Science (SCADS) to build a system that can generate tailored daily reports (TLDR) for intelligence analysts, we conducted a qualitative interview study with analysts to increase understanding of information passing in the intelligence community. While our results are preliminary, we expect that this work will contribute to a better understanding of the information ecosystem of the intelligence community, how institutional dynamics affect information passing, and what implications this has for a TLDR system. This work describes our involvement in and work completed during SCADS. Although preliminary, we identify that information passing is both a formal and informal process and often follows professional networks due especially to the small population and specialization of work. We call attention to the need for future analysis of information ecosystems to better support tailored information retrieval features.</description>
	<pubDate>2023-06-15</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 2, Pages 509-529: Preliminary Perspectives on Information Passing in the Intelligence Community</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/2/2/28">doi: 10.3390/analytics2020028</a></p>
	<p>Authors:
		Jeremy E. Block
		Ilana Bookner
		Sharon Lynn Chu
		R. Jordan Crouser
		Donald R. Honeycutt
		Rebecca M. Jonas
		Abhishek Kulkarni
		Yancy Vance Paredes
		Eric D. Ragan
		</p>
	<p>Analyst sensemaking research typically focuses on individual or small groups conducting intelligence tasks. This has helped understand information retrieval tasks and how people communicate information. As a part of the grand challenge of the Summer Conference on Applied Data Science (SCADS) to build a system that can generate tailored daily reports (TLDR) for intelligence analysts, we conducted a qualitative interview study with analysts to increase understanding of information passing in the intelligence community. While our results are preliminary, we expect that this work will contribute to a better understanding of the information ecosystem of the intelligence community, how institutional dynamics affect information passing, and what implications this has for a TLDR system. This work describes our involvement in and work completed during SCADS. Although preliminary, we identify that information passing is both a formal and informal process and often follows professional networks due especially to the small population and specialization of work. We call attention to the need for future analysis of information ecosystems to better support tailored information retrieval features.</p>
	]]></content:encoded>

	<dc:title>Preliminary Perspectives on Information Passing in the Intelligence Community</dc:title>
			<dc:creator>Jeremy E. Block</dc:creator>
			<dc:creator>Ilana Bookner</dc:creator>
			<dc:creator>Sharon Lynn Chu</dc:creator>
			<dc:creator>R. Jordan Crouser</dc:creator>
			<dc:creator>Donald R. Honeycutt</dc:creator>
			<dc:creator>Rebecca M. Jonas</dc:creator>
			<dc:creator>Abhishek Kulkarni</dc:creator>
			<dc:creator>Yancy Vance Paredes</dc:creator>
			<dc:creator>Eric D. Ragan</dc:creator>
		<dc:identifier>doi: 10.3390/analytics2020028</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2023-06-15</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2023-06-15</prism:publicationDate>
	<prism:volume>2</prism:volume>
	<prism:number>2</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>509</prism:startingPage>
		<prism:doi>10.3390/analytics2020028</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/2/2/28</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/2/2/27">

	<title>Analytics, Vol. 2, Pages 485-508: Spatiotemporal Data Mining Problems and Methods</title>
	<link>https://www.mdpi.com/2813-2203/2/2/27</link>
	<description>Many scientific fields show great interest in the extraction and processing of spatiotemporal data, such as medicine with an emphasis on epidemiology and neurology, geology, social sciences, meteorology, and a great interest is also observed in the study of transport. Spatiotemporal data differ significantly from spatial data, since spatiotemporal data refer to measurements, which take into account both the place and the time in which they are received, with their respective characteristics, while spatial data refer to and describe information related only to place. The innovation brought about by spatiotemporal data mining has caused a revolution in many scientific fields, and this is because through it we can now provide solutions and answers to complex problems, as well as provide useful and valuable predictions, through predictive learning. However, combining time and place in data mining presents significant challenges and difficulties that must be overcome. Spatiotemporal data mining and analysis is a relatively new approach to data mining which has been studied more systematically in the last decade. The purpose of this article is to provide a good introduction to spatiotemporal data, and through this detailed description, we attempt to introduce descriptive logic and gain a complete knowledge of these data. We aim to introduce a new way of describing them, aiming for future studies, by combining the expressions that arise by type of data, using descriptive logic, with new expressions, that can be derived, to describe future states of objects and environments with great precision, providing accurate predictions. In order to highlight the value of spatiotemporal data, we proceed to give a brief description of ST data in the introduction. We describe the relevant work carried out to date, the types of spatiotemporal (ST) data, their properties and the transformations that can be made between them, attempting, to a small extent, to introduce constraints and rules using descriptive logic, introducing descriptive logic into spatiotemporal data by type, when initially presenting the ST data. The data snapshots by species and similarities between the cases are then described. We describe methods, introducing clustering, dynamic ST clusters, predictive learning, pattern mining frequency, and pattern emergence, and problems such as anomaly detection, identifying time points of changes in the behavior of the observed object, and development of relationships between them. We describe the application of ST data in various fields today, as well as the future work. We finally conclude with our conclusions, with the representation and study of spatiotemporal data can, in combination with other properties which accompany all natural phenomena, through their appropriate processing, lead to safe conclusions regarding the study of problems, and also with great precision in the extraction of predictions by accurately determining future states of an environment or an object. Thus, the importance of ST data makes them particularly valuable today in various scientific fields, and their extraction is a particularly demanding challenge for the future.</description>
	<pubDate>2023-06-14</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 2, Pages 485-508: Spatiotemporal Data Mining Problems and Methods</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/2/2/27">doi: 10.3390/analytics2020027</a></p>
	<p>Authors:
		Eleftheria Koutsaki
		George Vardakis
		Nikolaos Papadakis
		</p>
	<p>Many scientific fields show great interest in the extraction and processing of spatiotemporal data, such as medicine with an emphasis on epidemiology and neurology, geology, social sciences, meteorology, and a great interest is also observed in the study of transport. Spatiotemporal data differ significantly from spatial data, since spatiotemporal data refer to measurements, which take into account both the place and the time in which they are received, with their respective characteristics, while spatial data refer to and describe information related only to place. The innovation brought about by spatiotemporal data mining has caused a revolution in many scientific fields, and this is because through it we can now provide solutions and answers to complex problems, as well as provide useful and valuable predictions, through predictive learning. However, combining time and place in data mining presents significant challenges and difficulties that must be overcome. Spatiotemporal data mining and analysis is a relatively new approach to data mining which has been studied more systematically in the last decade. The purpose of this article is to provide a good introduction to spatiotemporal data, and through this detailed description, we attempt to introduce descriptive logic and gain a complete knowledge of these data. We aim to introduce a new way of describing them, aiming for future studies, by combining the expressions that arise by type of data, using descriptive logic, with new expressions, that can be derived, to describe future states of objects and environments with great precision, providing accurate predictions. In order to highlight the value of spatiotemporal data, we proceed to give a brief description of ST data in the introduction. We describe the relevant work carried out to date, the types of spatiotemporal (ST) data, their properties and the transformations that can be made between them, attempting, to a small extent, to introduce constraints and rules using descriptive logic, introducing descriptive logic into spatiotemporal data by type, when initially presenting the ST data. The data snapshots by species and similarities between the cases are then described. We describe methods, introducing clustering, dynamic ST clusters, predictive learning, pattern mining frequency, and pattern emergence, and problems such as anomaly detection, identifying time points of changes in the behavior of the observed object, and development of relationships between them. We describe the application of ST data in various fields today, as well as the future work. We finally conclude with our conclusions, with the representation and study of spatiotemporal data can, in combination with other properties which accompany all natural phenomena, through their appropriate processing, lead to safe conclusions regarding the study of problems, and also with great precision in the extraction of predictions by accurately determining future states of an environment or an object. Thus, the importance of ST data makes them particularly valuable today in various scientific fields, and their extraction is a particularly demanding challenge for the future.</p>
	]]></content:encoded>

	<dc:title>Spatiotemporal Data Mining Problems and Methods</dc:title>
			<dc:creator>Eleftheria Koutsaki</dc:creator>
			<dc:creator>George Vardakis</dc:creator>
			<dc:creator>Nikolaos Papadakis</dc:creator>
		<dc:identifier>doi: 10.3390/analytics2020027</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2023-06-14</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2023-06-14</prism:publicationDate>
	<prism:volume>2</prism:volume>
	<prism:number>2</prism:number>
	<prism:section>Review</prism:section>
	<prism:startingPage>485</prism:startingPage>
		<prism:doi>10.3390/analytics2020027</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/2/2/27</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/2/2/26">

	<title>Analytics, Vol. 2, Pages 463-484: A Novel Zero-Truncated Katz Distribution by the Lagrange Expansion of the Second Kind with Associated Inferences</title>
	<link>https://www.mdpi.com/2813-2203/2/2/26</link>
	<description>In this article, the Lagrange expansion of the second kind is used to generate a novel zero-truncated Katz distribution; we refer to it as the Lagrangian zero-truncated Katz distribution (LZTKD). Notably, the zero-truncated Katz distribution is a special case of this distribution. Along with the closed form expression of all its statistical characteristics, the LZTKD is proven to provide an adequate model for both underdispersed and overdispersed zero-truncated count datasets. Specifically, we show that the associated hazard rate function has increasing, decreasing, bathtub, or upside-down bathtub shapes. Moreover, we demonstrate that the LZTKD belongs to the Lagrangian distribution of the first kind. Then, applications of the LZTKD in statistical scenarios are explored. The unknown parameters are estimated using the well-reputed method of the maximum likelihood. In addition, the generalized likelihood ratio test procedure is applied to test the significance of the additional parameter. In order to evaluate the performance of the maximum likelihood estimates, simulation studies are also conducted. The use of real-life datasets further highlights the relevance and applicability of the proposed model.</description>
	<pubDate>2023-06-01</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 2, Pages 463-484: A Novel Zero-Truncated Katz Distribution by the Lagrange Expansion of the Second Kind with Associated Inferences</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/2/2/26">doi: 10.3390/analytics2020026</a></p>
	<p>Authors:
		Damodaran Santhamani Shibu
		Christophe Chesneau
		Mohanan Monisha
		Radhakumari Maya
		Muhammed Rasheed Irshad
		</p>
	<p>In this article, the Lagrange expansion of the second kind is used to generate a novel zero-truncated Katz distribution; we refer to it as the Lagrangian zero-truncated Katz distribution (LZTKD). Notably, the zero-truncated Katz distribution is a special case of this distribution. Along with the closed form expression of all its statistical characteristics, the LZTKD is proven to provide an adequate model for both underdispersed and overdispersed zero-truncated count datasets. Specifically, we show that the associated hazard rate function has increasing, decreasing, bathtub, or upside-down bathtub shapes. Moreover, we demonstrate that the LZTKD belongs to the Lagrangian distribution of the first kind. Then, applications of the LZTKD in statistical scenarios are explored. The unknown parameters are estimated using the well-reputed method of the maximum likelihood. In addition, the generalized likelihood ratio test procedure is applied to test the significance of the additional parameter. In order to evaluate the performance of the maximum likelihood estimates, simulation studies are also conducted. The use of real-life datasets further highlights the relevance and applicability of the proposed model.</p>
	]]></content:encoded>

	<dc:title>A Novel Zero-Truncated Katz Distribution by the Lagrange Expansion of the Second Kind with Associated Inferences</dc:title>
			<dc:creator>Damodaran Santhamani Shibu</dc:creator>
			<dc:creator>Christophe Chesneau</dc:creator>
			<dc:creator>Mohanan Monisha</dc:creator>
			<dc:creator>Radhakumari Maya</dc:creator>
			<dc:creator>Muhammed Rasheed Irshad</dc:creator>
		<dc:identifier>doi: 10.3390/analytics2020026</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2023-06-01</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2023-06-01</prism:publicationDate>
	<prism:volume>2</prism:volume>
	<prism:number>2</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>463</prism:startingPage>
		<prism:doi>10.3390/analytics2020026</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/2/2/26</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2813-2203/2/2/25">

	<title>Analytics, Vol. 2, Pages 438-462: Generalized Unit Half-Logistic Geometric Distribution: Properties and Regression with Applications to Insurance</title>
	<link>https://www.mdpi.com/2813-2203/2/2/25</link>
	<description>The use of distributions to model and quantify risk is essential in risk assessment and management. In this study, the generalized unit half-logistic geometric (GUHLG) distribution is developed to model bounded insurance data on the unit interval. The corresponding probability density function plots indicate that the related distribution can handle data that exhibit left-skewed, right-skewed, symmetric, reversed-J, and bathtub shapes. The hazard rate function also suggests that the distribution can be applied to analyze data with bathtubs, N-shapes, and increasing failure rates. Subsequently, the inferential aspects of the proposed model are investigated. In particular, Monte Carlo simulation exercises are carried out to examine the performance of the estimation method by using an algorithm to generate random observations from the quantile function. The results of the simulation suggest that the considered estimation method is efficient. The univariate application of the distribution and the multivariate application of the associated regression using risk survey data reveal that the model provides a better fit than the other existing distributions and regression models. Under the multivariate application, we estimate the parameters of the regression model using both maximum likelihood and Bayesian estimations. The estimates of the parameters for the two methods are very close. Diagnostic plots of the Bayesian method using the trace, ergodic, and autocorrelation plots reveal that the chains converge to a stationary distribution.</description>
	<pubDate>2023-05-16</pubDate>

	<content:encoded><![CDATA[
	<p><b>Analytics, Vol. 2, Pages 438-462: Generalized Unit Half-Logistic Geometric Distribution: Properties and Regression with Applications to Insurance</b></p>
	<p>Analytics <a href="https://www.mdpi.com/2813-2203/2/2/25">doi: 10.3390/analytics2020025</a></p>
	<p>Authors:
		Suleman Nasiru
		Christophe Chesneau
		Abdul Ghaniyyu Abubakari
		Irene Dekomwine Angbing
		</p>
	<p>The use of distributions to model and quantify risk is essential in risk assessment and management. In this study, the generalized unit half-logistic geometric (GUHLG) distribution is developed to model bounded insurance data on the unit interval. The corresponding probability density function plots indicate that the related distribution can handle data that exhibit left-skewed, right-skewed, symmetric, reversed-J, and bathtub shapes. The hazard rate function also suggests that the distribution can be applied to analyze data with bathtubs, N-shapes, and increasing failure rates. Subsequently, the inferential aspects of the proposed model are investigated. In particular, Monte Carlo simulation exercises are carried out to examine the performance of the estimation method by using an algorithm to generate random observations from the quantile function. The results of the simulation suggest that the considered estimation method is efficient. The univariate application of the distribution and the multivariate application of the associated regression using risk survey data reveal that the model provides a better fit than the other existing distributions and regression models. Under the multivariate application, we estimate the parameters of the regression model using both maximum likelihood and Bayesian estimations. The estimates of the parameters for the two methods are very close. Diagnostic plots of the Bayesian method using the trace, ergodic, and autocorrelation plots reveal that the chains converge to a stationary distribution.</p>
	]]></content:encoded>

	<dc:title>Generalized Unit Half-Logistic Geometric Distribution: Properties and Regression with Applications to Insurance</dc:title>
			<dc:creator>Suleman Nasiru</dc:creator>
			<dc:creator>Christophe Chesneau</dc:creator>
			<dc:creator>Abdul Ghaniyyu Abubakari</dc:creator>
			<dc:creator>Irene Dekomwine Angbing</dc:creator>
		<dc:identifier>doi: 10.3390/analytics2020025</dc:identifier>
	<dc:source>Analytics</dc:source>
	<dc:date>2023-05-16</dc:date>

	<prism:publicationName>Analytics</prism:publicationName>
	<prism:publicationDate>2023-05-16</prism:publicationDate>
	<prism:volume>2</prism:volume>
	<prism:number>2</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>438</prism:startingPage>
		<prism:doi>10.3390/analytics2020025</prism:doi>
	<prism:url>https://www.mdpi.com/2813-2203/2/2/25</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
    
<cc:License rdf:about="https://creativecommons.org/licenses/by/4.0/">
	<cc:permits rdf:resource="https://creativecommons.org/ns#Reproduction" />
	<cc:permits rdf:resource="https://creativecommons.org/ns#Distribution" />
	<cc:permits rdf:resource="https://creativecommons.org/ns#DerivativeWorks" />
</cc:License>

</rdf:RDF>
