<?xml version="1.0" encoding="UTF-8"?>
<rdf:RDF xmlns="http://purl.org/rss/1.0/"
 xmlns:dc="http://purl.org/dc/elements/1.1/"
 xmlns:dcterms="http://purl.org/dc/terms/"
 xmlns:cc="http://web.resource.org/cc/"
 xmlns:prism="http://prismstandard.org/namespaces/basic/2.0/"
 xmlns:rdf="http://www.w3.org/1999/02/22-rdf-syntax-ns#"
 xmlns:admin="http://webns.net/mvcb/"
 xmlns:content="http://purl.org/rss/1.0/modules/content/">
    <channel rdf:about="https://www.mdpi.com/rss/journal/BDCC">
		<title>Big Data and Cognitive Computing</title>
		<description>Latest open access articles published in Big Data Cogn. Comput. at https://www.mdpi.com/journal/BDCC</description>
		<link>https://www.mdpi.com/journal/BDCC</link>
		<admin:generatorAgent rdf:resource="https://www.mdpi.com/journal/BDCC"/>
		<admin:errorReportsTo rdf:resource="mailto:support@mdpi.com"/>
		<dc:publisher>MDPI</dc:publisher>
		<dc:language>en</dc:language>
		<dc:rights>Creative Commons Attribution (CC-BY)</dc:rights>
						<prism:copyright>MDPI</prism:copyright>
		<prism:rightsAgent>support@mdpi.com</prism:rightsAgent>
		<image rdf:resource="https://pub.mdpi-res.com/img/design/mdpi-pub-logo.png?13cf3b5bd783e021?1778581344"/>
				<items>
			<rdf:Seq>
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/5/152" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/5/151" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/5/150" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/5/149" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/5/148" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/5/147" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/5/146" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/5/145" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/5/144" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/5/143" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/5/142" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/5/141" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/5/140" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/5/139" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/5/138" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/5/137" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/5/136" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/5/135" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/5/134" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/5/133" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/5/132" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/5/131" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/5/130" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/4/129" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/4/128" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/4/127" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/4/126" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/4/125" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/4/124" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/4/123" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/4/122" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/4/121" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/4/120" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/4/118" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/4/119" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/4/117" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/4/115" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/4/116" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/4/114" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/4/113" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/4/112" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/4/111" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/4/110" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/4/109" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/4/108" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/4/107" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/4/106" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/4/105" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/4/104" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/4/103" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/4/102" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/4/101" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/4/100" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/3/99" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/3/98" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/3/97" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/3/96" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/3/95" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/3/94" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/3/93" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/3/92" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/3/91" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/3/90" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/3/89" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/3/88" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/3/87" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/3/86" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/3/85" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/3/84" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/3/83" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/3/82" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/3/81" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/3/80" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/3/79" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/3/78" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/3/77" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/3/76" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/3/75" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/3/74" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/3/73" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/3/72" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/3/71" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/3/70" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/3/69" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/3/68" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/3/67" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/2/66" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/2/65" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/2/64" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/2/63" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/2/62" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/2/61" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/2/60" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/2/59" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/2/58" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/2/57" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/2/56" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/2/55" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/2/54" />
            				<rdf:li rdf:resource="https://www.mdpi.com/2504-2289/10/2/53" />
                    	</rdf:Seq>
		</items>
				<cc:license rdf:resource="https://creativecommons.org/licenses/by/4.0/" />
	</channel>

        <item rdf:about="https://www.mdpi.com/2504-2289/10/5/152">

	<title>BDCC, Vol. 10, Pages 152: Data-Driven Peak Demand Identification in Commercial Electricity Consumption for Load Curve Flattening</title>
	<link>https://www.mdpi.com/2504-2289/10/5/152</link>
	<description>Effective peak load management enables utilities to mitigate increased electricity demand and optimize the use of available resources during periods of maximum consumption. Accurate forecasting of the peak load is essential for ensuring the reliability, efficiency, and resilience of contemporary power systems. In this study, commercial customer-level data were employed to identify electricity peak demand within the Polish power system, drawing upon historical records of both energy consumption and meteorological variables. Departing from conventional time series forecasting approaches, the problem was intentionally reformulated as a pattern recognition task. Three classification techniques were systematically evaluated to identify individual customers&amp;amp;rsquo; peak load events, thereby offering a basis for demand-side management strategies and incentive mechanisms aimed at flattening load profiles and improving grid stability. The proposed approach demonstrates how data-driven analytics can support utilities in extracting actionable knowledge from large-scale energy datasets and enabling proactive demand response programs. Empirical results indicate that the proposed methods are capable of predicting up to 90% of electricity peak occurrences, with a forecasting horizon of 24 h leading to significant shifts in the load curve.</description>
	<pubDate>2026-05-12</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 152: Data-Driven Peak Demand Identification in Commercial Electricity Consumption for Load Curve Flattening</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/5/152">doi: 10.3390/bdcc10050152</a></p>
	<p>Authors:
		Michał Gostkowski
		Tomasz Ząbkowski
		Krzysztof Gajowniczek
		</p>
	<p>Effective peak load management enables utilities to mitigate increased electricity demand and optimize the use of available resources during periods of maximum consumption. Accurate forecasting of the peak load is essential for ensuring the reliability, efficiency, and resilience of contemporary power systems. In this study, commercial customer-level data were employed to identify electricity peak demand within the Polish power system, drawing upon historical records of both energy consumption and meteorological variables. Departing from conventional time series forecasting approaches, the problem was intentionally reformulated as a pattern recognition task. Three classification techniques were systematically evaluated to identify individual customers&amp;amp;rsquo; peak load events, thereby offering a basis for demand-side management strategies and incentive mechanisms aimed at flattening load profiles and improving grid stability. The proposed approach demonstrates how data-driven analytics can support utilities in extracting actionable knowledge from large-scale energy datasets and enabling proactive demand response programs. Empirical results indicate that the proposed methods are capable of predicting up to 90% of electricity peak occurrences, with a forecasting horizon of 24 h leading to significant shifts in the load curve.</p>
	]]></content:encoded>

	<dc:title>Data-Driven Peak Demand Identification in Commercial Electricity Consumption for Load Curve Flattening</dc:title>
			<dc:creator>Michał Gostkowski</dc:creator>
			<dc:creator>Tomasz Ząbkowski</dc:creator>
			<dc:creator>Krzysztof Gajowniczek</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10050152</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-05-12</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-05-12</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>5</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>152</prism:startingPage>
		<prism:doi>10.3390/bdcc10050152</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/5/152</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/5/151">

	<title>BDCC, Vol. 10, Pages 151: A Comparative Evaluation of AI Approaches to Large-Scale Scientific Subject Classification</title>
	<link>https://www.mdpi.com/2504-2289/10/5/151</link>
	<description>Background: The Hungarian Science Bibliography applies the OECD Frascati Fields of Science and Technology taxonomy for subject classification; however, approximately 80% of its records lack assigned categories. Automated large-scale classification could support retrospective completion and improve the quality of bibliographic data. Methods: We evaluated multiple artificial intelligence approaches to classifying publications into level 4 Frascati categories using only titles and keywords. Training datasets were compiled from bibliographic records and subjected to heuristic and large-language-model-based filtering to reduce noise and ambiguity. The approaches tested included statistical methods, classical machine learning classifiers, fine-tuned SciBERT models, zero-shot prompting with large language models, and a Mixture-of-Experts architecture. Results: Data quality had a stronger impact on performance than model complexity. Large-language-model-based filtering substantially improved classification results. The best-performing model, a Support Vector Classifier, achieved a weighted F1 score of 0.83, which is an outstanding result relative to state-of-the-art approaches from the literature. Conclusions: Our findings contribute new insights into classification research and may assist others in selecting appropriate solutions for real-world, large-scale bibliographic classification tasks.</description>
	<pubDate>2026-05-11</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 151: A Comparative Evaluation of AI Approaches to Large-Scale Scientific Subject Classification</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/5/151">doi: 10.3390/bdcc10050151</a></p>
	<p>Authors:
		Roland Tanácsi
		András Micsik
		</p>
	<p>Background: The Hungarian Science Bibliography applies the OECD Frascati Fields of Science and Technology taxonomy for subject classification; however, approximately 80% of its records lack assigned categories. Automated large-scale classification could support retrospective completion and improve the quality of bibliographic data. Methods: We evaluated multiple artificial intelligence approaches to classifying publications into level 4 Frascati categories using only titles and keywords. Training datasets were compiled from bibliographic records and subjected to heuristic and large-language-model-based filtering to reduce noise and ambiguity. The approaches tested included statistical methods, classical machine learning classifiers, fine-tuned SciBERT models, zero-shot prompting with large language models, and a Mixture-of-Experts architecture. Results: Data quality had a stronger impact on performance than model complexity. Large-language-model-based filtering substantially improved classification results. The best-performing model, a Support Vector Classifier, achieved a weighted F1 score of 0.83, which is an outstanding result relative to state-of-the-art approaches from the literature. Conclusions: Our findings contribute new insights into classification research and may assist others in selecting appropriate solutions for real-world, large-scale bibliographic classification tasks.</p>
	]]></content:encoded>

	<dc:title>A Comparative Evaluation of AI Approaches to Large-Scale Scientific Subject Classification</dc:title>
			<dc:creator>Roland Tanácsi</dc:creator>
			<dc:creator>András Micsik</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10050151</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-05-11</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-05-11</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>5</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>151</prism:startingPage>
		<prism:doi>10.3390/bdcc10050151</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/5/151</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/5/150">

	<title>BDCC, Vol. 10, Pages 150: BDERL: A Reinforcement Learning-Enhanced Differential Evolution for the Earliness&amp;ndash;Tardiness RCPSP</title>
	<link>https://www.mdpi.com/2504-2289/10/5/150</link>
	<description>This paper introduces the ETMS-RCPSP (Earliness&amp;amp;ndash;Tardiness Multi-Skill Resource-Constrained Scheduling Problem)&amp;amp;mdash;a novel problem derived from the MS-RCPSP by adding constraints on project completion time or actual production contracts. The goal of the new problem is to control the project completion time as closely as possible to reality&amp;amp;mdash;this differs from the original MS-RCPSP, which aimed to minimize project execution time. The objective of the problem is of greater practical significance in ensuring project completion on schedule while also addressing related issues, such as the ability to receive finished products on time as stipulated in the contract. The ETMS-RCPSP is an NP-hard problem whose result can be used for resource allocation in project execution or for resource arrangement in production lines to fulfill economic contracts. To address the ETMS-RCPSP, the paper proposes a new evolutionary algorithm, BDERL (Balanced Differential Evolution with Reinforcement Learning), that combines differential evolution with a problem-specific decoding mechanism and an adaptive parameter control strategy based on reinforcement learning (Q-learning). The proposed algorithm is evaluated on benchmark instances derived from the iMOPSE dataset and the TNG company dataset&amp;amp;mdash;a real-world dataset from manufacturing and contract-driven environments. Experimental results demonstrate that the approach consistently reduces total production costs compared to baseline heuristics while maintaining competitive computational efficiency. The findings underscore the efficacy of adaptive hybrid optimization techniques in solving intricate production scheduling problems characterized by limited resources and varied skill competencies.</description>
	<pubDate>2026-05-11</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 150: BDERL: A Reinforcement Learning-Enhanced Differential Evolution for the Earliness&amp;ndash;Tardiness RCPSP</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/5/150">doi: 10.3390/bdcc10050150</a></p>
	<p>Authors:
		Hao Nguyen Thi
		Loc Nguyen The
		Huu Dang Quoc
		</p>
	<p>This paper introduces the ETMS-RCPSP (Earliness&amp;amp;ndash;Tardiness Multi-Skill Resource-Constrained Scheduling Problem)&amp;amp;mdash;a novel problem derived from the MS-RCPSP by adding constraints on project completion time or actual production contracts. The goal of the new problem is to control the project completion time as closely as possible to reality&amp;amp;mdash;this differs from the original MS-RCPSP, which aimed to minimize project execution time. The objective of the problem is of greater practical significance in ensuring project completion on schedule while also addressing related issues, such as the ability to receive finished products on time as stipulated in the contract. The ETMS-RCPSP is an NP-hard problem whose result can be used for resource allocation in project execution or for resource arrangement in production lines to fulfill economic contracts. To address the ETMS-RCPSP, the paper proposes a new evolutionary algorithm, BDERL (Balanced Differential Evolution with Reinforcement Learning), that combines differential evolution with a problem-specific decoding mechanism and an adaptive parameter control strategy based on reinforcement learning (Q-learning). The proposed algorithm is evaluated on benchmark instances derived from the iMOPSE dataset and the TNG company dataset&amp;amp;mdash;a real-world dataset from manufacturing and contract-driven environments. Experimental results demonstrate that the approach consistently reduces total production costs compared to baseline heuristics while maintaining competitive computational efficiency. The findings underscore the efficacy of adaptive hybrid optimization techniques in solving intricate production scheduling problems characterized by limited resources and varied skill competencies.</p>
	]]></content:encoded>

	<dc:title>BDERL: A Reinforcement Learning-Enhanced Differential Evolution for the Earliness&amp;amp;ndash;Tardiness RCPSP</dc:title>
			<dc:creator>Hao Nguyen Thi</dc:creator>
			<dc:creator>Loc Nguyen The</dc:creator>
			<dc:creator>Huu Dang Quoc</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10050150</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-05-11</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-05-11</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>5</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>150</prism:startingPage>
		<prism:doi>10.3390/bdcc10050150</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/5/150</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/5/149">

	<title>BDCC, Vol. 10, Pages 149: MDA-Net: A Segmentation Network for Kidney Tumor Based on Enhanced Multi-Scale Feature Extraction and Attention Refinement</title>
	<link>https://www.mdpi.com/2504-2289/10/5/149</link>
	<description>Accurate kidney tumor segmentation from abdominal CT is essential for quantitative assessment and treatment planning. However, indistinct tumor boundaries and substantial inter-patient shape variability render traditional hand-crafted feature-based methods unreliable for precise delineation. Although deep learning has advanced this task, these methods still struggle with multi-scale tumor characteristics, complex morphological variations, and background noise in medical images. To address these challenges, we propose MDA-Net, an end-to-end segmentation method based on enhanced multi-scale feature extraction and attention refinement. Specifically, we introduce a Multi-Scale Feature Extraction (MSFE) module into encoder&amp;amp;ndash;decoder skip connections to aggregate dilated features across multiple receptive fields and learn branch-wise weights for adaptive refinement and fusion, thereby enhancing boundary details and semantic cues to reduce tumor-tissue ambiguity. At the bottleneck, a Deformable Pyramid Feature Refinement (DPFR) module combines deformable sampling with pyramid contextual modeling, thereby improving adaptability to variations in tumor shape and scale while preserving feature resolution. Moreover, a Channel and Spatial Attention (CASA) module is embedded in the decoder to suppress background interference and enhance boundary-sensitive structures during upsampling via coordinated channel and spatial reweighting, thereby improving the reconstruction of fine-grained tumor morphology and contours. Experiments on both KiTS19 and KiTS21 show that MDA-Net consistently improves tumor boundary delineation, lesion localization, and mask reconstruction, demonstrating stronger robustness and cross-dataset generalizability than representative baseline methods. Ablation studies further confirm the complementary effects of MSFE, DPFR, and CASA. In addition, Grad-CAM visualizations improve the clinical transparency and interpretability of the model. Overall, this method advances deep learning for medical image analysis and supports precise diagnosis and treatment of renal tumors.</description>
	<pubDate>2026-05-08</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 149: MDA-Net: A Segmentation Network for Kidney Tumor Based on Enhanced Multi-Scale Feature Extraction and Attention Refinement</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/5/149">doi: 10.3390/bdcc10050149</a></p>
	<p>Authors:
		Shaofu Lin
		Yumiao Chang
		Jianhui Chen
		Lianfang Ma
		</p>
	<p>Accurate kidney tumor segmentation from abdominal CT is essential for quantitative assessment and treatment planning. However, indistinct tumor boundaries and substantial inter-patient shape variability render traditional hand-crafted feature-based methods unreliable for precise delineation. Although deep learning has advanced this task, these methods still struggle with multi-scale tumor characteristics, complex morphological variations, and background noise in medical images. To address these challenges, we propose MDA-Net, an end-to-end segmentation method based on enhanced multi-scale feature extraction and attention refinement. Specifically, we introduce a Multi-Scale Feature Extraction (MSFE) module into encoder&amp;amp;ndash;decoder skip connections to aggregate dilated features across multiple receptive fields and learn branch-wise weights for adaptive refinement and fusion, thereby enhancing boundary details and semantic cues to reduce tumor-tissue ambiguity. At the bottleneck, a Deformable Pyramid Feature Refinement (DPFR) module combines deformable sampling with pyramid contextual modeling, thereby improving adaptability to variations in tumor shape and scale while preserving feature resolution. Moreover, a Channel and Spatial Attention (CASA) module is embedded in the decoder to suppress background interference and enhance boundary-sensitive structures during upsampling via coordinated channel and spatial reweighting, thereby improving the reconstruction of fine-grained tumor morphology and contours. Experiments on both KiTS19 and KiTS21 show that MDA-Net consistently improves tumor boundary delineation, lesion localization, and mask reconstruction, demonstrating stronger robustness and cross-dataset generalizability than representative baseline methods. Ablation studies further confirm the complementary effects of MSFE, DPFR, and CASA. In addition, Grad-CAM visualizations improve the clinical transparency and interpretability of the model. Overall, this method advances deep learning for medical image analysis and supports precise diagnosis and treatment of renal tumors.</p>
	]]></content:encoded>

	<dc:title>MDA-Net: A Segmentation Network for Kidney Tumor Based on Enhanced Multi-Scale Feature Extraction and Attention Refinement</dc:title>
			<dc:creator>Shaofu Lin</dc:creator>
			<dc:creator>Yumiao Chang</dc:creator>
			<dc:creator>Jianhui Chen</dc:creator>
			<dc:creator>Lianfang Ma</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10050149</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-05-08</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-05-08</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>5</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>149</prism:startingPage>
		<prism:doi>10.3390/bdcc10050149</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/5/149</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/5/148">

	<title>BDCC, Vol. 10, Pages 148: Adaptive Neural Network System for Preventing Violations of Personal Digital Rights as a National Security Factor</title>
	<link>https://www.mdpi.com/2504-2289/10/5/148</link>
	<description>The article develops a hybrid multimodal neural network for the automatic prevention of personal digital rights violations, focusing on improving security through anomaly detection and ensuring data confidentiality. The main aim is to integrate several innovative methods, such as federated learning, gating, latent competitive learning, and a variational autoencoder, to improve violation detection accuracy. The key contribution is the development of a training mixture that combines a probabilistic anomaly detector and an autoencoder reconstruction signal, which allows for effective detection of typical incidents and hidden anomalies. The experimental evaluation results showed high-performance indicators, with ROC-AUC at 0.96 and accuracy at 0.94, confirming the system&amp;amp;rsquo;s effectiveness on anonymized data. The results obtained have a significant practical contribution, as they can be integrated into national information security systems, including SOC and forensic reports, which will ensure a higher level of personal data protection and reduce privacy breach risks. The scope of the proposed system simultaneously covers cybersecurity, personal data protection, national security, SOC systems, and forensic analysis.</description>
	<pubDate>2026-05-08</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 148: Adaptive Neural Network System for Preventing Violations of Personal Digital Rights as a National Security Factor</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/5/148">doi: 10.3390/bdcc10050148</a></p>
	<p>Authors:
		Serhii Vladov
		Oksana Mulesa
		Maryana Marusinets
		Tiberiy Chegi
		Victoria Vysotska
		Anton Kazakov
		Iryna Kirieieva
		Maksym Korniienko
		Tetiana Morhunova
		</p>
	<p>The article develops a hybrid multimodal neural network for the automatic prevention of personal digital rights violations, focusing on improving security through anomaly detection and ensuring data confidentiality. The main aim is to integrate several innovative methods, such as federated learning, gating, latent competitive learning, and a variational autoencoder, to improve violation detection accuracy. The key contribution is the development of a training mixture that combines a probabilistic anomaly detector and an autoencoder reconstruction signal, which allows for effective detection of typical incidents and hidden anomalies. The experimental evaluation results showed high-performance indicators, with ROC-AUC at 0.96 and accuracy at 0.94, confirming the system&amp;amp;rsquo;s effectiveness on anonymized data. The results obtained have a significant practical contribution, as they can be integrated into national information security systems, including SOC and forensic reports, which will ensure a higher level of personal data protection and reduce privacy breach risks. The scope of the proposed system simultaneously covers cybersecurity, personal data protection, national security, SOC systems, and forensic analysis.</p>
	]]></content:encoded>

	<dc:title>Adaptive Neural Network System for Preventing Violations of Personal Digital Rights as a National Security Factor</dc:title>
			<dc:creator>Serhii Vladov</dc:creator>
			<dc:creator>Oksana Mulesa</dc:creator>
			<dc:creator>Maryana Marusinets</dc:creator>
			<dc:creator>Tiberiy Chegi</dc:creator>
			<dc:creator>Victoria Vysotska</dc:creator>
			<dc:creator>Anton Kazakov</dc:creator>
			<dc:creator>Iryna Kirieieva</dc:creator>
			<dc:creator>Maksym Korniienko</dc:creator>
			<dc:creator>Tetiana Morhunova</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10050148</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-05-08</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-05-08</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>5</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>148</prism:startingPage>
		<prism:doi>10.3390/bdcc10050148</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/5/148</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/5/147">

	<title>BDCC, Vol. 10, Pages 147: Federated Learning-Based Adaptive Multi-Head Attention Model for Wind Power Forecasting</title>
	<link>https://www.mdpi.com/2504-2289/10/5/147</link>
	<description>Enhancing the accuracy of short-term wind power forecasting helps mitigate the adverse impacts of prediction errors on grid dispatch. Wind power exhibits a significantly nonlinear dependence on multiple influencing factors. However, existing methods struggle to effectively resolve multi-dimensional feature redundancy and multi-scale non-stationary evolutionary characteristics inherent in far-offshore wind power forecasting tasks. This leads to bottlenecks such as insufficient feature discriminability and temporal dependency focus shift under complex marine environments, ultimately limiting further improvements in prediction accuracy. To address these challenges, this paper proposes a federated learning-based adaptive multi-head attention model for wind power forecasting (Fed-AMHA). The proposed framework operates as follows: First, each wind farm client utilizes a Bidirectional Long Short-Term Memory (BiLSTM) network to model input sequences bidirectionally, capturing long-term temporal dependencies. Subsequently, linear projection and parallel one-dimensional convolution operations are introduced to mine multi-scale local temporal features from each time step and its neighborhood. Building upon this, channel attention and multi-head temporal feature attention mechanisms are stacked. The model adaptively adjusts the weights of different time slices and feature channels by learning the importance of each channel to the forecasting task. The central server then aggregates the model parameters uploaded by the clients via averaging, enabling cross-site collaborative training without directly sharing raw data. Simulation results based on public datasets and actual wind farm data under various short-term forecasting scenarios demonstrate that the proposed model consistently achieves lower prediction errors and superior stability compared to existing forecasting models under identical settings.</description>
	<pubDate>2026-05-07</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 147: Federated Learning-Based Adaptive Multi-Head Attention Model for Wind Power Forecasting</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/5/147">doi: 10.3390/bdcc10050147</a></p>
	<p>Authors:
		Yihua Zhu
		Chao Luo
		Ke Wu
		Jiawei Yu
		Binjiang Hu
		Lei Huang
		Bitao Xiao
		</p>
	<p>Enhancing the accuracy of short-term wind power forecasting helps mitigate the adverse impacts of prediction errors on grid dispatch. Wind power exhibits a significantly nonlinear dependence on multiple influencing factors. However, existing methods struggle to effectively resolve multi-dimensional feature redundancy and multi-scale non-stationary evolutionary characteristics inherent in far-offshore wind power forecasting tasks. This leads to bottlenecks such as insufficient feature discriminability and temporal dependency focus shift under complex marine environments, ultimately limiting further improvements in prediction accuracy. To address these challenges, this paper proposes a federated learning-based adaptive multi-head attention model for wind power forecasting (Fed-AMHA). The proposed framework operates as follows: First, each wind farm client utilizes a Bidirectional Long Short-Term Memory (BiLSTM) network to model input sequences bidirectionally, capturing long-term temporal dependencies. Subsequently, linear projection and parallel one-dimensional convolution operations are introduced to mine multi-scale local temporal features from each time step and its neighborhood. Building upon this, channel attention and multi-head temporal feature attention mechanisms are stacked. The model adaptively adjusts the weights of different time slices and feature channels by learning the importance of each channel to the forecasting task. The central server then aggregates the model parameters uploaded by the clients via averaging, enabling cross-site collaborative training without directly sharing raw data. Simulation results based on public datasets and actual wind farm data under various short-term forecasting scenarios demonstrate that the proposed model consistently achieves lower prediction errors and superior stability compared to existing forecasting models under identical settings.</p>
	]]></content:encoded>

	<dc:title>Federated Learning-Based Adaptive Multi-Head Attention Model for Wind Power Forecasting</dc:title>
			<dc:creator>Yihua Zhu</dc:creator>
			<dc:creator>Chao Luo</dc:creator>
			<dc:creator>Ke Wu</dc:creator>
			<dc:creator>Jiawei Yu</dc:creator>
			<dc:creator>Binjiang Hu</dc:creator>
			<dc:creator>Lei Huang</dc:creator>
			<dc:creator>Bitao Xiao</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10050147</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-05-07</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-05-07</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>5</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>147</prism:startingPage>
		<prism:doi>10.3390/bdcc10050147</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/5/147</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/5/146">

	<title>BDCC, Vol. 10, Pages 146: Talent Identification and AI-Driven Decision Tools in Sport: A Policy-Oriented Perspective on Algorithmic Bias, Data Privacy, and Digital Determinism in Player Evaluation</title>
	<link>https://www.mdpi.com/2504-2289/10/5/146</link>
	<description>Big-data analytics are increasingly used in scouting and talent identification, with machine learning (ML) tools applied to evaluate and predict player performance based on match statistics, video tracking, physical and anthropometric tests, psychological assessments, social media data, and qualitative scouting reports. Advances in computer vision, together with the emergence of affordable automated broadcasting and data collection systems, have extended the deployment of ML-driven scouting from professional to youth sport. The use of algorithms in educational, employment, and healthcare settings has been shown to introduce biases and discrimination while wrongly assuming accuracy and objectivity because the decisions are made automatically and quantitatively. In this respect, we briefly describe the development of data-driven performance analysis and how ML-based technologies are currently applied for early screening and comparison of large player populations. Based on a narrative overview of the literature, we draw on evidence from education, employment, and healthcare to identify risks that may also emerge in ML-driven player evaluation, including algorithmic bias, non-representative training data, privacy concerns, and the persistence of model-based labels over time, especially in youth sport. Our main contribution is translating these threats into governance principles and operational safeguards for responsible use of AI in scouting and talent identification.</description>
	<pubDate>2026-05-07</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 146: Talent Identification and AI-Driven Decision Tools in Sport: A Policy-Oriented Perspective on Algorithmic Bias, Data Privacy, and Digital Determinism in Player Evaluation</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/5/146">doi: 10.3390/bdcc10050146</a></p>
	<p>Authors:
		Elia Morgulev
		Ofer H. Azar
		</p>
	<p>Big-data analytics are increasingly used in scouting and talent identification, with machine learning (ML) tools applied to evaluate and predict player performance based on match statistics, video tracking, physical and anthropometric tests, psychological assessments, social media data, and qualitative scouting reports. Advances in computer vision, together with the emergence of affordable automated broadcasting and data collection systems, have extended the deployment of ML-driven scouting from professional to youth sport. The use of algorithms in educational, employment, and healthcare settings has been shown to introduce biases and discrimination while wrongly assuming accuracy and objectivity because the decisions are made automatically and quantitatively. In this respect, we briefly describe the development of data-driven performance analysis and how ML-based technologies are currently applied for early screening and comparison of large player populations. Based on a narrative overview of the literature, we draw on evidence from education, employment, and healthcare to identify risks that may also emerge in ML-driven player evaluation, including algorithmic bias, non-representative training data, privacy concerns, and the persistence of model-based labels over time, especially in youth sport. Our main contribution is translating these threats into governance principles and operational safeguards for responsible use of AI in scouting and talent identification.</p>
	]]></content:encoded>

	<dc:title>Talent Identification and AI-Driven Decision Tools in Sport: A Policy-Oriented Perspective on Algorithmic Bias, Data Privacy, and Digital Determinism in Player Evaluation</dc:title>
			<dc:creator>Elia Morgulev</dc:creator>
			<dc:creator>Ofer H. Azar</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10050146</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-05-07</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-05-07</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>5</prism:number>
	<prism:section>Review</prism:section>
	<prism:startingPage>146</prism:startingPage>
		<prism:doi>10.3390/bdcc10050146</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/5/146</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/5/145">

	<title>BDCC, Vol. 10, Pages 145: AI-Driven Generation of Old English: A Framework for Low-Resource Languages</title>
	<link>https://www.mdpi.com/2504-2289/10/5/145</link>
	<description>Preserving ancient languages is essential for understanding the cultural and linguistic heritage of humanity. Old English, however, remains critically under-resourced, which limits its accessibility to modern natural language processing (NLP) techniques. We present a scalable framework that uses advanced large language models (LLMs) to generate high-quality Old English texts to address this gap. In this study, we specifically employ state-of-the-art models, including Llama-3.1-8B and Mistral-7B, as our foundation models, which are then adapted to the unique characteristics of Old English. Our approach combines parameter-efficient fine-tuning (Low-Rank Adaptation (LoRA)), data augmentation via back-translation, and a dual-agent pipeline that separates content generation (in English) and translation (into Old English). Evaluation with automated metrics (BLEU, METEOR, and CHRF) shows improvements over baseline models, with BLEU scores increasing from 26 to over 65 for English-to-Old English translation. Expert human assessment confirms high grammatical accuracy and stylistic fidelity in the generated texts, with average scores of 9.0/10 for inflection and word order, 9.1/10 for lexical authenticity, and 7.8 for semantic coherence. These results demonstrate that the framework can reliably expand limited historical corpora while maintaining linguistic integrity, with immediate practical applications in digital humanities research, computational philology, and the development of educational resources for Old English study. Beyond expanding the Old English corpus, our method offers a practical blueprint for revitalizing other endangered languages, thus linking AI innovation with the goals of cultural preservation.</description>
	<pubDate>2026-05-06</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 145: AI-Driven Generation of Old English: A Framework for Low-Resource Languages</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/5/145">doi: 10.3390/bdcc10050145</a></p>
	<p>Authors:
		Rodrigo Gabriel Salazar Alva
		Matías Núñez
		Cristian López Del Alamo
		Javier Martín Arista
		</p>
	<p>Preserving ancient languages is essential for understanding the cultural and linguistic heritage of humanity. Old English, however, remains critically under-resourced, which limits its accessibility to modern natural language processing (NLP) techniques. We present a scalable framework that uses advanced large language models (LLMs) to generate high-quality Old English texts to address this gap. In this study, we specifically employ state-of-the-art models, including Llama-3.1-8B and Mistral-7B, as our foundation models, which are then adapted to the unique characteristics of Old English. Our approach combines parameter-efficient fine-tuning (Low-Rank Adaptation (LoRA)), data augmentation via back-translation, and a dual-agent pipeline that separates content generation (in English) and translation (into Old English). Evaluation with automated metrics (BLEU, METEOR, and CHRF) shows improvements over baseline models, with BLEU scores increasing from 26 to over 65 for English-to-Old English translation. Expert human assessment confirms high grammatical accuracy and stylistic fidelity in the generated texts, with average scores of 9.0/10 for inflection and word order, 9.1/10 for lexical authenticity, and 7.8 for semantic coherence. These results demonstrate that the framework can reliably expand limited historical corpora while maintaining linguistic integrity, with immediate practical applications in digital humanities research, computational philology, and the development of educational resources for Old English study. Beyond expanding the Old English corpus, our method offers a practical blueprint for revitalizing other endangered languages, thus linking AI innovation with the goals of cultural preservation.</p>
	]]></content:encoded>

	<dc:title>AI-Driven Generation of Old English: A Framework for Low-Resource Languages</dc:title>
			<dc:creator>Rodrigo Gabriel Salazar Alva</dc:creator>
			<dc:creator>Matías Núñez</dc:creator>
			<dc:creator>Cristian López Del Alamo</dc:creator>
			<dc:creator>Javier Martín Arista</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10050145</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-05-06</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-05-06</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>5</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>145</prism:startingPage>
		<prism:doi>10.3390/bdcc10050145</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/5/145</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/5/144">

	<title>BDCC, Vol. 10, Pages 144: HYSARD: A Hybrid Feature-Fusion Model for Sarcasm Detection Using RoBERTa Embeddings and Linguistic Features</title>
	<link>https://www.mdpi.com/2504-2289/10/5/144</link>
	<description>Sarcasm detection remains a challenging task in natural language processing because sarcastic expressions often convey meanings that contradict their literal wording. Although transformer-based encoders such as RoBERTa capture contextual semantics effectively, sparse linguistic signals common in sarcastic user-generated text, such as exaggerated punctuation, elongated words, capitalization, and sentiment contrast, may not always remain explicitly accessible in the final sentence representation. To address this limitation, we propose HYSARD, a hybrid feature-fusion model that combines RoBERTa-based sentence embeddings with complementary linguistic features, including sentiment polarity, stylistic markers, syntactic patterns, and TF-IDF lexical cues. The resulting feature space is refined through Random Forest-based feature selection to reduce redundancy and improve robustness, while SMOTE mitigates class imbalance during training. We evaluate HYSARD on the SemEval-2022 iSarcasmEval dataset and the balanced Main and Political subsets of SARC 2.0. Results show strong and consistent performance across datasets, with an F1-score of 0.80 on iSarcasmEval, while held-out test-set error analysis further highlights strong class-wise discrimination. The ablation study further confirms that combining contextual embeddings with explicit linguistic cues improves sarcasm detection over reduced feature configurations. These findings show that hybrid feature fusion remains an effective and practical strategy for sarcasm detection in noisy social media text.</description>
	<pubDate>2026-05-06</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 144: HYSARD: A Hybrid Feature-Fusion Model for Sarcasm Detection Using RoBERTa Embeddings and Linguistic Features</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/5/144">doi: 10.3390/bdcc10050144</a></p>
	<p>Authors:
		Ismail Jabri
		Zine Eddine Louriga
		Aziza El Ouaazizi
		Abdelaziz Ahaitouf
		</p>
	<p>Sarcasm detection remains a challenging task in natural language processing because sarcastic expressions often convey meanings that contradict their literal wording. Although transformer-based encoders such as RoBERTa capture contextual semantics effectively, sparse linguistic signals common in sarcastic user-generated text, such as exaggerated punctuation, elongated words, capitalization, and sentiment contrast, may not always remain explicitly accessible in the final sentence representation. To address this limitation, we propose HYSARD, a hybrid feature-fusion model that combines RoBERTa-based sentence embeddings with complementary linguistic features, including sentiment polarity, stylistic markers, syntactic patterns, and TF-IDF lexical cues. The resulting feature space is refined through Random Forest-based feature selection to reduce redundancy and improve robustness, while SMOTE mitigates class imbalance during training. We evaluate HYSARD on the SemEval-2022 iSarcasmEval dataset and the balanced Main and Political subsets of SARC 2.0. Results show strong and consistent performance across datasets, with an F1-score of 0.80 on iSarcasmEval, while held-out test-set error analysis further highlights strong class-wise discrimination. The ablation study further confirms that combining contextual embeddings with explicit linguistic cues improves sarcasm detection over reduced feature configurations. These findings show that hybrid feature fusion remains an effective and practical strategy for sarcasm detection in noisy social media text.</p>
	]]></content:encoded>

	<dc:title>HYSARD: A Hybrid Feature-Fusion Model for Sarcasm Detection Using RoBERTa Embeddings and Linguistic Features</dc:title>
			<dc:creator>Ismail Jabri</dc:creator>
			<dc:creator>Zine Eddine Louriga</dc:creator>
			<dc:creator>Aziza El Ouaazizi</dc:creator>
			<dc:creator>Abdelaziz Ahaitouf</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10050144</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-05-06</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-05-06</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>5</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>144</prism:startingPage>
		<prism:doi>10.3390/bdcc10050144</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/5/144</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/5/143">

	<title>BDCC, Vol. 10, Pages 143: Evaluating Computational Approaches for Harmful Content Analysis: Promise, Pitfalls and Tools for Responsible Research</title>
	<link>https://www.mdpi.com/2504-2289/10/5/143</link>
	<description>This manuscript develops and demonstrates a practical framework for evaluating automated classifiers used in communication research, using harmful language detection as an illustrative case. We combine (a) a structured review of documentation practices for 27 publicly available classifiers and their associated annotation processes with (b) a cross-dataset evaluation that re-tests each model beyond its original training context. Across 27 datasets, we extract and compare reporting on construct definitions, annotator instructions, and inter-annotator agreement, and we quantify generalization by applying each model to multiple out-of-domain test sets. We also benchmark a contemporary large language model (GPT-5) under a consistent prompting protocol to illustrate how LLM-based classification compares to fine-tuned classifiers. Results show that documentation is uneven and often insufficient for theory-driven measurement, inter-annotator agreement varies widely across datasets, and cross-dataset performance frequently drops substantially relative to within-dataset evaluations. Building on these findings and existing validation guidance, we provide a reusable checklist and decision flow to help researchers select, justify, and report classifier-based measures in ways that support transparency and cumulative science. Recommendations for researchers, reviewers, and journal editors stress aligning model selection with standards of validity, reliability, and transparency.</description>
	<pubDate>2026-05-02</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 143: Evaluating Computational Approaches for Harmful Content Analysis: Promise, Pitfalls and Tools for Responsible Research</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/5/143">doi: 10.3390/bdcc10050143</a></p>
	<p>Authors:
		Itai Himelboim
		Mudit Baid
		</p>
	<p>This manuscript develops and demonstrates a practical framework for evaluating automated classifiers used in communication research, using harmful language detection as an illustrative case. We combine (a) a structured review of documentation practices for 27 publicly available classifiers and their associated annotation processes with (b) a cross-dataset evaluation that re-tests each model beyond its original training context. Across 27 datasets, we extract and compare reporting on construct definitions, annotator instructions, and inter-annotator agreement, and we quantify generalization by applying each model to multiple out-of-domain test sets. We also benchmark a contemporary large language model (GPT-5) under a consistent prompting protocol to illustrate how LLM-based classification compares to fine-tuned classifiers. Results show that documentation is uneven and often insufficient for theory-driven measurement, inter-annotator agreement varies widely across datasets, and cross-dataset performance frequently drops substantially relative to within-dataset evaluations. Building on these findings and existing validation guidance, we provide a reusable checklist and decision flow to help researchers select, justify, and report classifier-based measures in ways that support transparency and cumulative science. Recommendations for researchers, reviewers, and journal editors stress aligning model selection with standards of validity, reliability, and transparency.</p>
	]]></content:encoded>

	<dc:title>Evaluating Computational Approaches for Harmful Content Analysis: Promise, Pitfalls and Tools for Responsible Research</dc:title>
			<dc:creator>Itai Himelboim</dc:creator>
			<dc:creator>Mudit Baid</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10050143</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-05-02</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-05-02</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>5</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>143</prism:startingPage>
		<prism:doi>10.3390/bdcc10050143</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/5/143</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/5/142">

	<title>BDCC, Vol. 10, Pages 142: Towards Improved Clinical Adoption of AI Segmentation Models: Benchmarking High-Performance Models for Resource-Constrained Settings</title>
	<link>https://www.mdpi.com/2504-2289/10/5/142</link>
	<description>High-performance medical segmentation models are often benchmarked on high-end GPUs. Such benchmarks do not provide useful performance insights for point-of-care low-end devices. This work, firstly, posits that to achieve improved clinical adoption of AI-powered segmentation models, especially in reduced manpower settings like rural hospitals, we need benchmarks that provide actionable insights on the degree to which high-performance models address five deployment constraints viz: resource-effectiveness for low-end computing devices, clinically acceptable accuracy, clinically compatible execution times, localization of user data, and user-based finetuning. In this work, five state-of-the-art foundation segmentation models and one target-specific model were systematically evaluated on three multi-organ medical datasets. Furthermore, the best-ranking foundation model and target-specific model were benchmarked on three low-end devices. Our findings show that lightweight foundation models provided the best performance trade-off and are easily user-fine-tuned on custom datasets. Target-specific models provide high accuracy out-of-the-box, but may require significant optimisation to deliver comparably fast execution times and user-based finetuning on low-end devices. The methods and results from this research provide actionable insights on high-performance medical segmentation models for low-end computing devices, as a necessary step towards improved adoption in resource-limited clinical settings.</description>
	<pubDate>2026-05-02</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 142: Towards Improved Clinical Adoption of AI Segmentation Models: Benchmarking High-Performance Models for Resource-Constrained Settings</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/5/142">doi: 10.3390/bdcc10050142</a></p>
	<p>Authors:
		Emmanuel Chibuikem Nnadozie
		Susana Merino-Caviedes
		Daniel A. de Luis-Román
		Marcos Martín-Fernández
		Carlos Alberola-López
		</p>
	<p>High-performance medical segmentation models are often benchmarked on high-end GPUs. Such benchmarks do not provide useful performance insights for point-of-care low-end devices. This work, firstly, posits that to achieve improved clinical adoption of AI-powered segmentation models, especially in reduced manpower settings like rural hospitals, we need benchmarks that provide actionable insights on the degree to which high-performance models address five deployment constraints viz: resource-effectiveness for low-end computing devices, clinically acceptable accuracy, clinically compatible execution times, localization of user data, and user-based finetuning. In this work, five state-of-the-art foundation segmentation models and one target-specific model were systematically evaluated on three multi-organ medical datasets. Furthermore, the best-ranking foundation model and target-specific model were benchmarked on three low-end devices. Our findings show that lightweight foundation models provided the best performance trade-off and are easily user-fine-tuned on custom datasets. Target-specific models provide high accuracy out-of-the-box, but may require significant optimisation to deliver comparably fast execution times and user-based finetuning on low-end devices. The methods and results from this research provide actionable insights on high-performance medical segmentation models for low-end computing devices, as a necessary step towards improved adoption in resource-limited clinical settings.</p>
	]]></content:encoded>

	<dc:title>Towards Improved Clinical Adoption of AI Segmentation Models: Benchmarking High-Performance Models for Resource-Constrained Settings</dc:title>
			<dc:creator>Emmanuel Chibuikem Nnadozie</dc:creator>
			<dc:creator>Susana Merino-Caviedes</dc:creator>
			<dc:creator>Daniel A. de Luis-Román</dc:creator>
			<dc:creator>Marcos Martín-Fernández</dc:creator>
			<dc:creator>Carlos Alberola-López</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10050142</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-05-02</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-05-02</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>5</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>142</prism:startingPage>
		<prism:doi>10.3390/bdcc10050142</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/5/142</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/5/141">

	<title>BDCC, Vol. 10, Pages 141: BERT-Based Models for Normalization of Adverse Drug Event Expressions in Social Media to Standard Medical Terminology for Drug Safety Analysis</title>
	<link>https://www.mdpi.com/2504-2289/10/5/141</link>
	<description>Social media platforms host abundant and timely descriptions of medication experiences that can complement traditional pharmacovigilance systems. Yet the linguistic informality of these data presents a major challenge for mapping adverse drug event (ADE) expressions to standardized medical terminology. In this study, we developed BERT-based language models to classify ADE mentions from social media into MedDRA System Organ Classes (SOCs). Using the SMM4H and CADEC corpora, as well as their combination, we performed 20 iterations of 20% holdout validation for 3-, 6-, 22-, and 25-SOC classification tasks with a selected fixed training configuration (learning rate, batch size, and training epochs) based on training-loss convergence. The models achieved accuracies ranging from 75% to 94%, demonstrating strong performance for SOC-level classification of noisy and informal ADE expressions under the evaluated settings. These results are based on a controlled mention-level evaluation using deduplicated adverse drug event strings and do not establish document-level or real-world deployment generalization. This work provides a systematic evaluation of BERT-based models for SOC-level classification of ADEs and demonstrates consistent performance within the evaluated datasets and label granularities. While direct comparison with prior studies is limited by differences in datasets and evaluation protocols, the results demonstrate that transformer-based models can effectively classify ADEs into SOCs. These findings support the use of transformer-based normalization for SOC-level aggregation of user-reported adverse events and their integration into large-scale social media pharmacovigilance pipelines as a downstream component under controlled conditions.</description>
	<pubDate>2026-05-02</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 141: BERT-Based Models for Normalization of Adverse Drug Event Expressions in Social Media to Standard Medical Terminology for Drug Safety Analysis</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/5/141">doi: 10.3390/bdcc10050141</a></p>
	<p>Authors:
		Fan Dong
		Wenjing Guo
		Jie Liu
		Ann Varghese
		Weida Tong
		Tucker A. Patterson
		Huixiao Hong
		</p>
	<p>Social media platforms host abundant and timely descriptions of medication experiences that can complement traditional pharmacovigilance systems. Yet the linguistic informality of these data presents a major challenge for mapping adverse drug event (ADE) expressions to standardized medical terminology. In this study, we developed BERT-based language models to classify ADE mentions from social media into MedDRA System Organ Classes (SOCs). Using the SMM4H and CADEC corpora, as well as their combination, we performed 20 iterations of 20% holdout validation for 3-, 6-, 22-, and 25-SOC classification tasks with a selected fixed training configuration (learning rate, batch size, and training epochs) based on training-loss convergence. The models achieved accuracies ranging from 75% to 94%, demonstrating strong performance for SOC-level classification of noisy and informal ADE expressions under the evaluated settings. These results are based on a controlled mention-level evaluation using deduplicated adverse drug event strings and do not establish document-level or real-world deployment generalization. This work provides a systematic evaluation of BERT-based models for SOC-level classification of ADEs and demonstrates consistent performance within the evaluated datasets and label granularities. While direct comparison with prior studies is limited by differences in datasets and evaluation protocols, the results demonstrate that transformer-based models can effectively classify ADEs into SOCs. These findings support the use of transformer-based normalization for SOC-level aggregation of user-reported adverse events and their integration into large-scale social media pharmacovigilance pipelines as a downstream component under controlled conditions.</p>
	]]></content:encoded>

	<dc:title>BERT-Based Models for Normalization of Adverse Drug Event Expressions in Social Media to Standard Medical Terminology for Drug Safety Analysis</dc:title>
			<dc:creator>Fan Dong</dc:creator>
			<dc:creator>Wenjing Guo</dc:creator>
			<dc:creator>Jie Liu</dc:creator>
			<dc:creator>Ann Varghese</dc:creator>
			<dc:creator>Weida Tong</dc:creator>
			<dc:creator>Tucker A. Patterson</dc:creator>
			<dc:creator>Huixiao Hong</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10050141</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-05-02</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-05-02</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>5</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>141</prism:startingPage>
		<prism:doi>10.3390/bdcc10050141</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/5/141</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/5/140">

	<title>BDCC, Vol. 10, Pages 140: BWT-Enhanced Compression for GIS Raster Data: A Hybrid AV1-Inspired Approach with Burrows&amp;ndash;Wheeler Transform</title>
	<link>https://www.mdpi.com/2504-2289/10/5/140</link>
	<description>The AVIF (AV1 Image File Format) is a modern, royalty-free image format that leverages the AV1 video codec for superior compression efficiency, supporting both lossy and lossless modes. Its entropy encoding relies on a multi-symbol context-adaptive arithmetic coder (range coding with adaptive cumulative distribution functions (CDFs)), which is effective for general imagery but may not optimally exploit the repetitive structures common in Geographic Information System (GIS) maps/data. This paper proposes replacing AVIF&amp;amp;rsquo;s entropy encoder with the Burrows&amp;amp;ndash;Wheeler Transform (BWT), a reversible preprocessing algorithm that rearranges data to create runs of similar symbols, enhancing subsequent compression. We detail the technical steps for modification, drawing from AV1&amp;amp;rsquo;s open-source implementation, and explain why BWT is advantageous for GIS raster maps/data, which often feature large uniform areas, limited color palettes, and spatial redundancies. Empirical evidence from related studies on BWT-based image compression shows improvements in lossless scenarios, potentially considerably reducing file sizes over standard methods while preserving data integrity critical for geospatial analysis. This swap could improve storage, transmission, and processing efficiency in GIS applications, such as remote sensing and cartography. The discussion includes challenges like computational overhead and compatibility, with recommendations for implementations. The resulting BWT-AVIF hybrid produces a non-standard AV1 bit-stream that is not compliant with the AV1 or AVIF specifications and therefore requires custom decoders. It is presented here as a research prototype for GIS-specific compression rather than a compliant AVIF extension.</description>
	<pubDate>2026-05-01</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 140: BWT-Enhanced Compression for GIS Raster Data: A Hybrid AV1-Inspired Approach with Burrows&amp;ndash;Wheeler Transform</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/5/140">doi: 10.3390/bdcc10050140</a></p>
	<p>Authors:
		Yair Wiseman
		</p>
	<p>The AVIF (AV1 Image File Format) is a modern, royalty-free image format that leverages the AV1 video codec for superior compression efficiency, supporting both lossy and lossless modes. Its entropy encoding relies on a multi-symbol context-adaptive arithmetic coder (range coding with adaptive cumulative distribution functions (CDFs)), which is effective for general imagery but may not optimally exploit the repetitive structures common in Geographic Information System (GIS) maps/data. This paper proposes replacing AVIF&amp;amp;rsquo;s entropy encoder with the Burrows&amp;amp;ndash;Wheeler Transform (BWT), a reversible preprocessing algorithm that rearranges data to create runs of similar symbols, enhancing subsequent compression. We detail the technical steps for modification, drawing from AV1&amp;amp;rsquo;s open-source implementation, and explain why BWT is advantageous for GIS raster maps/data, which often feature large uniform areas, limited color palettes, and spatial redundancies. Empirical evidence from related studies on BWT-based image compression shows improvements in lossless scenarios, potentially considerably reducing file sizes over standard methods while preserving data integrity critical for geospatial analysis. This swap could improve storage, transmission, and processing efficiency in GIS applications, such as remote sensing and cartography. The discussion includes challenges like computational overhead and compatibility, with recommendations for implementations. The resulting BWT-AVIF hybrid produces a non-standard AV1 bit-stream that is not compliant with the AV1 or AVIF specifications and therefore requires custom decoders. It is presented here as a research prototype for GIS-specific compression rather than a compliant AVIF extension.</p>
	]]></content:encoded>

	<dc:title>BWT-Enhanced Compression for GIS Raster Data: A Hybrid AV1-Inspired Approach with Burrows&amp;amp;ndash;Wheeler Transform</dc:title>
			<dc:creator>Yair Wiseman</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10050140</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-05-01</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-05-01</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>5</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>140</prism:startingPage>
		<prism:doi>10.3390/bdcc10050140</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/5/140</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/5/139">

	<title>BDCC, Vol. 10, Pages 139: A Hybrid Artificial Intelligence Framework for Reliable and Seamless Vertical Handover in Next-Generation Heterogeneous Networks</title>
	<link>https://www.mdpi.com/2504-2289/10/5/139</link>
	<description>Next-generation heterogeneous wireless networks (HetNets) comprising LTE macro-cells, 5G New Radio (NR) small cells, and WiFi 6 access points aim to provide seamless connectivity under diverse mobility scenarios. However, vertical handover (VHO) remains a performance bottleneck because of the highly variable radio environments, dynamic user mobility, stringent quality of service (QoS) requirements, and the coexistence of multi-tier access technologies. Existing handover approaches based on deep learning and deep reinforcement learning (DRL) suffer from limitations: deep learning models lack decision-making capabilities, whereas DRL models, particularly deep Q-network (DQN)-based policies, face Q-value overestimation and unstable convergence. To overcome these limitations, this paper introduces a Hybrid deep double-Q networks (DDQN)&amp;amp;ndash;bidirectional long short-term memory (Bi-LSTM) Framework that integrates bi-directional mobility prediction and DRL-based adaptive decision-making. The Bi-LSTM module captures forward and backward temporal dependencies and predicts future Received Signal Strength (RSS) trajectories, mobility dynamics, and cell-edge transitions. The DDQN module stabilizes the action value estimation, mitigates overestimation bias, and enables context-aware handover decisions. A multi-tier simulation environment consisting of LTE, 5G NR, and WiFi 6 networks was developed using realistic path loss, shadowing, interference, and mobility models. Extensive evaluations demonstrated substantial improvements in mobility prediction accuracy, handover stability, radio link reliability, throughput efficiency, and latency reduction compared to conventional RSS-based and DQN-based schemes. The findings highlight the effectiveness of integrating predictive intelligence with reinforcement learning for reliable mobility management in 5G-Advanced and emerging 6G networks.</description>
	<pubDate>2026-04-29</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 139: A Hybrid Artificial Intelligence Framework for Reliable and Seamless Vertical Handover in Next-Generation Heterogeneous Networks</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/5/139">doi: 10.3390/bdcc10050139</a></p>
	<p>Authors:
		Sunisa Kunarak
		</p>
	<p>Next-generation heterogeneous wireless networks (HetNets) comprising LTE macro-cells, 5G New Radio (NR) small cells, and WiFi 6 access points aim to provide seamless connectivity under diverse mobility scenarios. However, vertical handover (VHO) remains a performance bottleneck because of the highly variable radio environments, dynamic user mobility, stringent quality of service (QoS) requirements, and the coexistence of multi-tier access technologies. Existing handover approaches based on deep learning and deep reinforcement learning (DRL) suffer from limitations: deep learning models lack decision-making capabilities, whereas DRL models, particularly deep Q-network (DQN)-based policies, face Q-value overestimation and unstable convergence. To overcome these limitations, this paper introduces a Hybrid deep double-Q networks (DDQN)&amp;amp;ndash;bidirectional long short-term memory (Bi-LSTM) Framework that integrates bi-directional mobility prediction and DRL-based adaptive decision-making. The Bi-LSTM module captures forward and backward temporal dependencies and predicts future Received Signal Strength (RSS) trajectories, mobility dynamics, and cell-edge transitions. The DDQN module stabilizes the action value estimation, mitigates overestimation bias, and enables context-aware handover decisions. A multi-tier simulation environment consisting of LTE, 5G NR, and WiFi 6 networks was developed using realistic path loss, shadowing, interference, and mobility models. Extensive evaluations demonstrated substantial improvements in mobility prediction accuracy, handover stability, radio link reliability, throughput efficiency, and latency reduction compared to conventional RSS-based and DQN-based schemes. The findings highlight the effectiveness of integrating predictive intelligence with reinforcement learning for reliable mobility management in 5G-Advanced and emerging 6G networks.</p>
	]]></content:encoded>

	<dc:title>A Hybrid Artificial Intelligence Framework for Reliable and Seamless Vertical Handover in Next-Generation Heterogeneous Networks</dc:title>
			<dc:creator>Sunisa Kunarak</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10050139</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-04-29</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-04-29</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>5</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>139</prism:startingPage>
		<prism:doi>10.3390/bdcc10050139</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/5/139</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/5/138">

	<title>BDCC, Vol. 10, Pages 138: GPU-TOPSIS: A Complete Vectorized and Parallel Reformulation of the TOPSIS Method for Large-Scale Multi-Criteria Decision Making</title>
	<link>https://www.mdpi.com/2504-2289/10/5/138</link>
	<description>The TOPSIS (Technique for Order Preference by Similarity to Ideal Solution) method is one of the most widely used multi-criteria decision-making (MCDM) approaches in industrial, financial, and scientific fields. However, its sequential computational cost of O(m &amp;amp;times; n), where m denotes the number of alternatives and n the number of criteria, becomes prohibitive when decision matrices have several million rows. Despite its geometric interpretability and simplicity, classical TOPSIS faces two key computational bottlenecks at scale: (i) Euclidean distance calculations O(m &amp;amp;times; n) dominating the total cost, and (ii) the O(m &amp;amp;times; log m) sorting step, both inherently sequential and memory-bound on CPUs. To overcome these limitations, we propose GPU-TOPSIS, a fully vectorized and parallel reformulation of TOPSIS based on tensor execution on graphics processing units (GPUs), whose main contributions are: (i) a formally correct reformulation of TOPSIS as a GPU tensor pipeline preserving mathematical fidelity to the original method; (ii) a two-pass fragment-processing algorithm guaranteeing exact mathematical equivalence with monolithic TOPSIS, while reducing the memory footprint from O(m &amp;amp;times; n) to O(mt &amp;amp;times; n), where mt &amp;amp;lt; m is the size of each independently processed fragment; (iii) three independent implementations on CuPy, PyTorch, and TensorFlow, ensuring the framework&amp;amp;rsquo;s portability and genericity. Experimental evaluations on real data from the Amazon Products 2023 dataset, using matrices of up to 200 million alternatives (via the 2-pass formulation), demonstrate speedups of up to 4.75&amp;amp;times; compared to the reference CPU implementation (NumPy), with inter-backend score differences below 5 &amp;amp;times; 10&amp;amp;minus;8 and 100% ranking overlap across all tested Top-K thresholds. A perturbation sensitivity analysis of the criteria weights and cross-backend consistency tests confirms that GPU acceleration fully preserves robustness and decision reliability, making GPU-TOPSIS a practical, open, and reproducible solution for large-scale multi-criteria decision making in Big Data environments.</description>
	<pubDate>2026-04-28</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 138: GPU-TOPSIS: A Complete Vectorized and Parallel Reformulation of the TOPSIS Method for Large-Scale Multi-Criteria Decision Making</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/5/138">doi: 10.3390/bdcc10050138</a></p>
	<p>Authors:
		Latifa Boubekri
		Hassnae Aberkane
		Mohammed Chaouki Abounaima
		Loubna Lamrini
		</p>
	<p>The TOPSIS (Technique for Order Preference by Similarity to Ideal Solution) method is one of the most widely used multi-criteria decision-making (MCDM) approaches in industrial, financial, and scientific fields. However, its sequential computational cost of O(m &amp;amp;times; n), where m denotes the number of alternatives and n the number of criteria, becomes prohibitive when decision matrices have several million rows. Despite its geometric interpretability and simplicity, classical TOPSIS faces two key computational bottlenecks at scale: (i) Euclidean distance calculations O(m &amp;amp;times; n) dominating the total cost, and (ii) the O(m &amp;amp;times; log m) sorting step, both inherently sequential and memory-bound on CPUs. To overcome these limitations, we propose GPU-TOPSIS, a fully vectorized and parallel reformulation of TOPSIS based on tensor execution on graphics processing units (GPUs), whose main contributions are: (i) a formally correct reformulation of TOPSIS as a GPU tensor pipeline preserving mathematical fidelity to the original method; (ii) a two-pass fragment-processing algorithm guaranteeing exact mathematical equivalence with monolithic TOPSIS, while reducing the memory footprint from O(m &amp;amp;times; n) to O(mt &amp;amp;times; n), where mt &amp;amp;lt; m is the size of each independently processed fragment; (iii) three independent implementations on CuPy, PyTorch, and TensorFlow, ensuring the framework&amp;amp;rsquo;s portability and genericity. Experimental evaluations on real data from the Amazon Products 2023 dataset, using matrices of up to 200 million alternatives (via the 2-pass formulation), demonstrate speedups of up to 4.75&amp;amp;times; compared to the reference CPU implementation (NumPy), with inter-backend score differences below 5 &amp;amp;times; 10&amp;amp;minus;8 and 100% ranking overlap across all tested Top-K thresholds. A perturbation sensitivity analysis of the criteria weights and cross-backend consistency tests confirms that GPU acceleration fully preserves robustness and decision reliability, making GPU-TOPSIS a practical, open, and reproducible solution for large-scale multi-criteria decision making in Big Data environments.</p>
	]]></content:encoded>

	<dc:title>GPU-TOPSIS: A Complete Vectorized and Parallel Reformulation of the TOPSIS Method for Large-Scale Multi-Criteria Decision Making</dc:title>
			<dc:creator>Latifa Boubekri</dc:creator>
			<dc:creator>Hassnae Aberkane</dc:creator>
			<dc:creator>Mohammed Chaouki Abounaima</dc:creator>
			<dc:creator>Loubna Lamrini</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10050138</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-04-28</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-04-28</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>5</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>138</prism:startingPage>
		<prism:doi>10.3390/bdcc10050138</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/5/138</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/5/137">

	<title>BDCC, Vol. 10, Pages 137: A Review of Key Technologies for Systems Based on Non-Volatile Memory</title>
	<link>https://www.mdpi.com/2504-2289/10/5/137</link>
	<description>With the continuous growth of data-intensive applications and artificial intelligence workloads, traditional dynamic random access memory (DRAM) is increasingly struggling to meet demands in terms of capacity scale, energy consumption constraints, and data retention after power failure. Consequently, non-volatile memory (NVM) has emerged as a crucial technology for bridging the gap between the memory and storage layers. However, due to inherent differences in write life, read&amp;amp;ndash;write performance variations, and consistency guarantee after failure, the systematic application of NVM still faces a series of challenges. Addressing these issues, this paper takes as its starting point the adaptation of medium characteristics and system design, and summarizes the research progress in aspects such as write optimization, consistency and security coordination mechanisms, data structure modification under hybrid memory architecture, and cross-layer resource collaboration. It also conducts an in-depth analysis of representative solutions and evaluation methods. The review results show that current research has shifted from improving a single performance bottleneck to multi-mechanism collaborative optimization. Various technical approaches have proven complementary in alleviating write amplification, enhancing persistence efficiency, and optimizing access patterns. This paper demonstrates that achieving stable and scalable application of NVM requires establishing a more systematic collaborative design concept between durability, security, and performance. As AI training workloads and big data analytics place increasing demands on memory bandwidth and persistence, the techniques surveyed here provide a foundational basis for next-generation memory-centric computing infrastructures.</description>
	<pubDate>2026-04-27</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 137: A Review of Key Technologies for Systems Based on Non-Volatile Memory</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/5/137">doi: 10.3390/bdcc10050137</a></p>
	<p>Authors:
		Yuhan Zhang
		Zehang Wang
		Yuanfang Chen
		Chunfeng Du
		Jing Chen
		</p>
	<p>With the continuous growth of data-intensive applications and artificial intelligence workloads, traditional dynamic random access memory (DRAM) is increasingly struggling to meet demands in terms of capacity scale, energy consumption constraints, and data retention after power failure. Consequently, non-volatile memory (NVM) has emerged as a crucial technology for bridging the gap between the memory and storage layers. However, due to inherent differences in write life, read&amp;amp;ndash;write performance variations, and consistency guarantee after failure, the systematic application of NVM still faces a series of challenges. Addressing these issues, this paper takes as its starting point the adaptation of medium characteristics and system design, and summarizes the research progress in aspects such as write optimization, consistency and security coordination mechanisms, data structure modification under hybrid memory architecture, and cross-layer resource collaboration. It also conducts an in-depth analysis of representative solutions and evaluation methods. The review results show that current research has shifted from improving a single performance bottleneck to multi-mechanism collaborative optimization. Various technical approaches have proven complementary in alleviating write amplification, enhancing persistence efficiency, and optimizing access patterns. This paper demonstrates that achieving stable and scalable application of NVM requires establishing a more systematic collaborative design concept between durability, security, and performance. As AI training workloads and big data analytics place increasing demands on memory bandwidth and persistence, the techniques surveyed here provide a foundational basis for next-generation memory-centric computing infrastructures.</p>
	]]></content:encoded>

	<dc:title>A Review of Key Technologies for Systems Based on Non-Volatile Memory</dc:title>
			<dc:creator>Yuhan Zhang</dc:creator>
			<dc:creator>Zehang Wang</dc:creator>
			<dc:creator>Yuanfang Chen</dc:creator>
			<dc:creator>Chunfeng Du</dc:creator>
			<dc:creator>Jing Chen</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10050137</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-04-27</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-04-27</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>5</prism:number>
	<prism:section>Review</prism:section>
	<prism:startingPage>137</prism:startingPage>
		<prism:doi>10.3390/bdcc10050137</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/5/137</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/5/136">

	<title>BDCC, Vol. 10, Pages 136: A Robust Ensemble Learning Approach to URL-Based Phishing Webpage Detection</title>
	<link>https://www.mdpi.com/2504-2289/10/5/136</link>
	<description>The proliferation of online fraud has resulted in substantial financial damage to individuals and organizations alike, with web phishing emerging as one of the most pervasive and harmful attack vectors. In response, this paper proposes the Stacking Ensemble Models Generator (SEMG), a URL-based phishing detection approach that leverages a multi-objective Genetic Algorithm to jointly optimize Precision and Recall in the selection and configuration of stacking ensemble models. An initial pool of base learners is trained on labeled datasets and subsequently evolved through genetic operators toward a globally optimal ensemble. Experimental evaluation across five datasets sourced from Mendeley and UCI repositories demonstrates that SEMG consistently surpasses individual base learners and compares favorably against existing methods, attaining 99.2% performance across all metrics on D2 while matching or exceeding state-of-the-art results on the remaining benchmarks. These outcomes underscore the framework&amp;amp;rsquo;s robustness and its potential for deployment in real-world phishing detection systems.</description>
	<pubDate>2026-04-27</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 136: A Robust Ensemble Learning Approach to URL-Based Phishing Webpage Detection</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/5/136">doi: 10.3390/bdcc10050136</a></p>
	<p>Authors:
		Abdellah Rezoug
		Mohamed Bader-el-den
		</p>
	<p>The proliferation of online fraud has resulted in substantial financial damage to individuals and organizations alike, with web phishing emerging as one of the most pervasive and harmful attack vectors. In response, this paper proposes the Stacking Ensemble Models Generator (SEMG), a URL-based phishing detection approach that leverages a multi-objective Genetic Algorithm to jointly optimize Precision and Recall in the selection and configuration of stacking ensemble models. An initial pool of base learners is trained on labeled datasets and subsequently evolved through genetic operators toward a globally optimal ensemble. Experimental evaluation across five datasets sourced from Mendeley and UCI repositories demonstrates that SEMG consistently surpasses individual base learners and compares favorably against existing methods, attaining 99.2% performance across all metrics on D2 while matching or exceeding state-of-the-art results on the remaining benchmarks. These outcomes underscore the framework&amp;amp;rsquo;s robustness and its potential for deployment in real-world phishing detection systems.</p>
	]]></content:encoded>

	<dc:title>A Robust Ensemble Learning Approach to URL-Based Phishing Webpage Detection</dc:title>
			<dc:creator>Abdellah Rezoug</dc:creator>
			<dc:creator>Mohamed Bader-el-den</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10050136</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-04-27</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-04-27</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>5</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>136</prism:startingPage>
		<prism:doi>10.3390/bdcc10050136</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/5/136</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/5/135">

	<title>BDCC, Vol. 10, Pages 135: Enhancing Adversarial Transferability via Fourier-Based Input Transformation</title>
	<link>https://www.mdpi.com/2504-2289/10/5/135</link>
	<description>Adversarial transferability makes black-box attacks practical and exposes weaknesses of deep neural networks for computer vision, image recognition, and visual understanding. Among various transferability-enhancing methods, input transformation is one of the most effective strategies. However, existing methods often ignore the decoupling of style and semantics in the input image, as well as the need for customized transformation strategies, resulting in limited performance gains or suboptimal outcomes. In this paper, we propose a novel Fourier-based perspective for input transformation generalization in the context of vision adversarial attacks. The main observations are that the Fourier amplitude captures stylistic information and the phase encompasses richer semantics which are crucial for visual understanding. Motivated by this, we develop a Fourier-based strategy, which performs a stylistic transform and semantic mixup on the input examples to improve transferability. To avoid inconsistent semantics of augmented images for the surrogate model, we mix the original images with the augmentations to maintain semantic consistency and mitigate imprecise gradients. Extensive experiments on ImageNet-compatible datasets demonstrate that our method consistently outperforms existing input transformation attacks.</description>
	<pubDate>2026-04-27</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 135: Enhancing Adversarial Transferability via Fourier-Based Input Transformation</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/5/135">doi: 10.3390/bdcc10050135</a></p>
	<p>Authors:
		Zilin Tian
		Xin Wang
		Yunfei Long
		Liguo Zhang
		</p>
	<p>Adversarial transferability makes black-box attacks practical and exposes weaknesses of deep neural networks for computer vision, image recognition, and visual understanding. Among various transferability-enhancing methods, input transformation is one of the most effective strategies. However, existing methods often ignore the decoupling of style and semantics in the input image, as well as the need for customized transformation strategies, resulting in limited performance gains or suboptimal outcomes. In this paper, we propose a novel Fourier-based perspective for input transformation generalization in the context of vision adversarial attacks. The main observations are that the Fourier amplitude captures stylistic information and the phase encompasses richer semantics which are crucial for visual understanding. Motivated by this, we develop a Fourier-based strategy, which performs a stylistic transform and semantic mixup on the input examples to improve transferability. To avoid inconsistent semantics of augmented images for the surrogate model, we mix the original images with the augmentations to maintain semantic consistency and mitigate imprecise gradients. Extensive experiments on ImageNet-compatible datasets demonstrate that our method consistently outperforms existing input transformation attacks.</p>
	]]></content:encoded>

	<dc:title>Enhancing Adversarial Transferability via Fourier-Based Input Transformation</dc:title>
			<dc:creator>Zilin Tian</dc:creator>
			<dc:creator>Xin Wang</dc:creator>
			<dc:creator>Yunfei Long</dc:creator>
			<dc:creator>Liguo Zhang</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10050135</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-04-27</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-04-27</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>5</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>135</prism:startingPage>
		<prism:doi>10.3390/bdcc10050135</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/5/135</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/5/134">

	<title>BDCC, Vol. 10, Pages 134: A Physically Regularized Control-Oriented State Model and Nonlinear Model Predictive Control Framework for an Ice Rink Refrigeration System</title>
	<link>https://www.mdpi.com/2504-2289/10/5/134</link>
	<description>Energy-intensive refrigeration systems require predictive models that remain informative under counterfactual control trajectories, not only on archived operation. This paper develops a control-oriented multi-step state model and a nonlinear model predictive control framework for an indoor ice-rink refrigeration system. Historical state, control, and exogenous variables are encoded jointly with an admissible future control trajectory, and a normalized thermal-balance residual is added to the training objective. A lightweight conditioned transformer predicts ice temperature, return-glycol temperature, supply-glycol temperature, and compressor power over a 30 min horizon. The selected weakly regularized model with regularization coefficient &amp;amp;lambda;phys = 0.001 decreases the normalized thermal-balance root-mean-square error on the horizon tail by 30.29% relative to the base model while increasing the average ice-temperature root-mean-square error by only 1.90%. In a surrogate-based counterfactual four-day evaluation, the resulting nonlinear model predictive controller reduces predicted daily energy by 4.84%, terminal violation share by 17.32%, mean absolute terminal ice-temperature deviation by 18.74%, and the mean objective value by 30.82% relative to historical admissible setpoint tracking. The mean full control cycle time is 0.0311 s, confirming real-time feasibility for a 5 min supervisory update interval. All controller results are surrogate-based rather than field-deployed and therefore represent receding-horizon benchmark results under learned-model evaluation, not realized field savings.</description>
	<pubDate>2026-04-26</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 134: A Physically Regularized Control-Oriented State Model and Nonlinear Model Predictive Control Framework for an Ice Rink Refrigeration System</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/5/134">doi: 10.3390/bdcc10050134</a></p>
	<p>Authors:
		Alexander A. Karmanov
		Petr V. Nikitin
		</p>
	<p>Energy-intensive refrigeration systems require predictive models that remain informative under counterfactual control trajectories, not only on archived operation. This paper develops a control-oriented multi-step state model and a nonlinear model predictive control framework for an indoor ice-rink refrigeration system. Historical state, control, and exogenous variables are encoded jointly with an admissible future control trajectory, and a normalized thermal-balance residual is added to the training objective. A lightweight conditioned transformer predicts ice temperature, return-glycol temperature, supply-glycol temperature, and compressor power over a 30 min horizon. The selected weakly regularized model with regularization coefficient &amp;amp;lambda;phys = 0.001 decreases the normalized thermal-balance root-mean-square error on the horizon tail by 30.29% relative to the base model while increasing the average ice-temperature root-mean-square error by only 1.90%. In a surrogate-based counterfactual four-day evaluation, the resulting nonlinear model predictive controller reduces predicted daily energy by 4.84%, terminal violation share by 17.32%, mean absolute terminal ice-temperature deviation by 18.74%, and the mean objective value by 30.82% relative to historical admissible setpoint tracking. The mean full control cycle time is 0.0311 s, confirming real-time feasibility for a 5 min supervisory update interval. All controller results are surrogate-based rather than field-deployed and therefore represent receding-horizon benchmark results under learned-model evaluation, not realized field savings.</p>
	]]></content:encoded>

	<dc:title>A Physically Regularized Control-Oriented State Model and Nonlinear Model Predictive Control Framework for an Ice Rink Refrigeration System</dc:title>
			<dc:creator>Alexander A. Karmanov</dc:creator>
			<dc:creator>Petr V. Nikitin</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10050134</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-04-26</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-04-26</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>5</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>134</prism:startingPage>
		<prism:doi>10.3390/bdcc10050134</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/5/134</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/5/133">

	<title>BDCC, Vol. 10, Pages 133: Mamba-Based Video Analysis for Blood Pressure Estimation</title>
	<link>https://www.mdpi.com/2504-2289/10/5/133</link>
	<description>Blood pressure monitoring is important for overall health assessment, yet traditional cuff-based methods are intrusive and unsuitable for continuous monitoring. This paper proposes a contactless approach for blood pressure estimation from facial videos using a bidirectional Mamba-based architecture with uncertainty quantification. Our method processes 64-frame video segments through a hierarchical 3D convolutional encoder to extract spatiotemporal features, then applies bidirectional state-space modeling to capture temporal dynamics efficiently. The model was evaluated on the Vitals for Vision (V4V) dataset, achieving mean absolute errors of 13.15 mmHg for systolic and 9.56 mmHg for diastolic blood pressure, outperforming prior methods while requiring significantly fewer computational resources than attention-based approaches. While these results do not meet clinical-grade diagnostic standards, they demonstrate the feasibility of contactless blood pressure estimation for non-clinical applications such as wellness monitoring, preliminary health screening, and continuous remote observation, where unobtrusive and computationally efficient monitoring is desirable.</description>
	<pubDate>2026-04-26</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 133: Mamba-Based Video Analysis for Blood Pressure Estimation</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/5/133">doi: 10.3390/bdcc10050133</a></p>
	<p>Authors:
		Walaa Othman
		Batol Hamoud
		Nikolay Shilov
		Alexey Kashevnik
		Alexander Mayatin
		</p>
	<p>Blood pressure monitoring is important for overall health assessment, yet traditional cuff-based methods are intrusive and unsuitable for continuous monitoring. This paper proposes a contactless approach for blood pressure estimation from facial videos using a bidirectional Mamba-based architecture with uncertainty quantification. Our method processes 64-frame video segments through a hierarchical 3D convolutional encoder to extract spatiotemporal features, then applies bidirectional state-space modeling to capture temporal dynamics efficiently. The model was evaluated on the Vitals for Vision (V4V) dataset, achieving mean absolute errors of 13.15 mmHg for systolic and 9.56 mmHg for diastolic blood pressure, outperforming prior methods while requiring significantly fewer computational resources than attention-based approaches. While these results do not meet clinical-grade diagnostic standards, they demonstrate the feasibility of contactless blood pressure estimation for non-clinical applications such as wellness monitoring, preliminary health screening, and continuous remote observation, where unobtrusive and computationally efficient monitoring is desirable.</p>
	]]></content:encoded>

	<dc:title>Mamba-Based Video Analysis for Blood Pressure Estimation</dc:title>
			<dc:creator>Walaa Othman</dc:creator>
			<dc:creator>Batol Hamoud</dc:creator>
			<dc:creator>Nikolay Shilov</dc:creator>
			<dc:creator>Alexey Kashevnik</dc:creator>
			<dc:creator>Alexander Mayatin</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10050133</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-04-26</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-04-26</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>5</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>133</prism:startingPage>
		<prism:doi>10.3390/bdcc10050133</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/5/133</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/5/132">

	<title>BDCC, Vol. 10, Pages 132: Adversarial Evaluation of Large Language Models for Building Robust Offensive Language Detection in Moroccan Arabic</title>
	<link>https://www.mdpi.com/2504-2289/10/5/132</link>
	<description>Offensive language detection is crucial for ensuring safe and inclusive digital environments. Identifying harmful content protects users and supports healthier online interactions. Despite advances in transformer-based models, particularly Large Language Models (LLMs), their application to this task remains underexplored for low-resource languages such as Moroccan Arabic, especially compared with high-resource languages. This study evaluates the performance of various open- and closed-source LLMs for offensive language detection in Moroccan Darija. The evaluated models include general-purpose LLMs such as LLaMA, Mistral, and Gemma, as well as Arabic-focused models such as ArabianGPT, Falcon Arabic, and Atlas-Chat. We also experiment with reasoning models such as DeepSeek and GPT-4. Beyond traditional evaluation metrics, we investigate the robustness of these LLMs and examine the impact of adversarial training on their performance. Moreover, we contribute to the field by creating a large, high-quality dataset. Our evaluation revealed that GPT-4o Mini achieved the best overall performance, reaching an F1-score of 88%. However, robustness testing under black-box and white-box adversarial attacks exposed notable vulnerabilities, with attack success rates reaching 30%, thereby highlighting the need for enhancement. Despite the complex morphology and linguistic variability of Moroccan Darija, adversarial training resulted in a notable improvement in both overall model performance and robustness against adversarial attacks, yielding an average increase of 20.89% in resistance to attacks. Furthermore, this approach enabled GPT-4o Mini to achieve an F1-score of 91%, surpassing the current state-of-the-art performance by 6%. These results highlight the importance of incorporating adversarial approaches in low-resource dialectal settings to effectively address linguistic variability and data scarcity.</description>
	<pubDate>2026-04-24</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 132: Adversarial Evaluation of Large Language Models for Building Robust Offensive Language Detection in Moroccan Arabic</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/5/132">doi: 10.3390/bdcc10050132</a></p>
	<p>Authors:
		Soufiyan Ouali
		Kanza Raisi
		Asmaa Mourhir
		El Habib Nfaoui
		Said El Garouani
		</p>
	<p>Offensive language detection is crucial for ensuring safe and inclusive digital environments. Identifying harmful content protects users and supports healthier online interactions. Despite advances in transformer-based models, particularly Large Language Models (LLMs), their application to this task remains underexplored for low-resource languages such as Moroccan Arabic, especially compared with high-resource languages. This study evaluates the performance of various open- and closed-source LLMs for offensive language detection in Moroccan Darija. The evaluated models include general-purpose LLMs such as LLaMA, Mistral, and Gemma, as well as Arabic-focused models such as ArabianGPT, Falcon Arabic, and Atlas-Chat. We also experiment with reasoning models such as DeepSeek and GPT-4. Beyond traditional evaluation metrics, we investigate the robustness of these LLMs and examine the impact of adversarial training on their performance. Moreover, we contribute to the field by creating a large, high-quality dataset. Our evaluation revealed that GPT-4o Mini achieved the best overall performance, reaching an F1-score of 88%. However, robustness testing under black-box and white-box adversarial attacks exposed notable vulnerabilities, with attack success rates reaching 30%, thereby highlighting the need for enhancement. Despite the complex morphology and linguistic variability of Moroccan Darija, adversarial training resulted in a notable improvement in both overall model performance and robustness against adversarial attacks, yielding an average increase of 20.89% in resistance to attacks. Furthermore, this approach enabled GPT-4o Mini to achieve an F1-score of 91%, surpassing the current state-of-the-art performance by 6%. These results highlight the importance of incorporating adversarial approaches in low-resource dialectal settings to effectively address linguistic variability and data scarcity.</p>
	]]></content:encoded>

	<dc:title>Adversarial Evaluation of Large Language Models for Building Robust Offensive Language Detection in Moroccan Arabic</dc:title>
			<dc:creator>Soufiyan Ouali</dc:creator>
			<dc:creator>Kanza Raisi</dc:creator>
			<dc:creator>Asmaa Mourhir</dc:creator>
			<dc:creator>El Habib Nfaoui</dc:creator>
			<dc:creator>Said El Garouani</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10050132</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-04-24</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-04-24</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>5</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>132</prism:startingPage>
		<prism:doi>10.3390/bdcc10050132</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/5/132</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/5/131">

	<title>BDCC, Vol. 10, Pages 131: FEM-Based Hybrid Compression Framework with Pipeline Implementation for Efficient Deep Neural Networks on Tiny ImageNet</title>
	<link>https://www.mdpi.com/2504-2289/10/5/131</link>
	<description>The high accuracy achieved by deep learning techniques has made them indispensable in computer vision applications. However, their substantial memory demands and high computational complexity limit their deployment in resource-constrained environments. To address this challenge, this study introduces a Feature Enhancement Module (FEM) as part of a unified hybrid compression framework that combines mixed-precision quantization and structured pruning to improve model efficiency. Experimental results on the Tiny ImageNet dataset using ResNet50 and MobileNetV3 architectures demonstrate the strong adaptability and scalability of the proposed approach. Compared with state-of-the-art compression methods, the proposed FEM-based framework achieves up to 6% improvement in Top-1 accuracy, while reducing memory usage by 32.26% and improving inference speed by 66%. Furthermore, the ablation study demonstrates that incorporating the FEM module leads to up to 24% improvement over the baseline model, highlighting its effectiveness. The results further show that FEM effectively preserves inter-channel feature representation stability even under aggressive compression, making it well suited for real-time processing and practical Artificial Intelligence (AI) applications. By maintaining semantic richness while significantly reducing computational cost, the proposed method bridges the gap between high-performance deep models and lightweight, deployable solutions. Overall, the FEM-based hybrid compression framework establishes a scalable and architecture-independent foundation for sustainable deep learning in resource-limited environments.</description>
	<pubDate>2026-04-22</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 131: FEM-Based Hybrid Compression Framework with Pipeline Implementation for Efficient Deep Neural Networks on Tiny ImageNet</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/5/131">doi: 10.3390/bdcc10050131</a></p>
	<p>Authors:
		Areej Hamza
		Amel Tuama
		Asraf Mohamed Moubark
		</p>
	<p>The high accuracy achieved by deep learning techniques has made them indispensable in computer vision applications. However, their substantial memory demands and high computational complexity limit their deployment in resource-constrained environments. To address this challenge, this study introduces a Feature Enhancement Module (FEM) as part of a unified hybrid compression framework that combines mixed-precision quantization and structured pruning to improve model efficiency. Experimental results on the Tiny ImageNet dataset using ResNet50 and MobileNetV3 architectures demonstrate the strong adaptability and scalability of the proposed approach. Compared with state-of-the-art compression methods, the proposed FEM-based framework achieves up to 6% improvement in Top-1 accuracy, while reducing memory usage by 32.26% and improving inference speed by 66%. Furthermore, the ablation study demonstrates that incorporating the FEM module leads to up to 24% improvement over the baseline model, highlighting its effectiveness. The results further show that FEM effectively preserves inter-channel feature representation stability even under aggressive compression, making it well suited for real-time processing and practical Artificial Intelligence (AI) applications. By maintaining semantic richness while significantly reducing computational cost, the proposed method bridges the gap between high-performance deep models and lightweight, deployable solutions. Overall, the FEM-based hybrid compression framework establishes a scalable and architecture-independent foundation for sustainable deep learning in resource-limited environments.</p>
	]]></content:encoded>

	<dc:title>FEM-Based Hybrid Compression Framework with Pipeline Implementation for Efficient Deep Neural Networks on Tiny ImageNet</dc:title>
			<dc:creator>Areej Hamza</dc:creator>
			<dc:creator>Amel Tuama</dc:creator>
			<dc:creator>Asraf Mohamed Moubark</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10050131</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-04-22</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-04-22</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>5</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>131</prism:startingPage>
		<prism:doi>10.3390/bdcc10050131</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/5/131</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/5/130">

	<title>BDCC, Vol. 10, Pages 130: A Data-Driven Machine Learning Framework for Multi-Criteria ESG Evaluation</title>
	<link>https://www.mdpi.com/2504-2289/10/5/130</link>
	<description>This study proposes a novel data-driven machine learning (ML) framework for multi-criteria environmental, social, and governance (ESG) evaluation. The framework aims to address the transparency, consistency, and subjectivity limitations of existing ESG evaluation systems by employing a fully data-driven, modular, and ML-supported architecture. It comprises three main modules: (i) ESG data preprocessing with missing-data imputation by the MissForest algorithm; (ii) a three-plane ESG feature selection workflow that integrates clustering, feature importance, and classification algorithms to identify representative ESG indicators; and (iii) a hybrid weighting and ranking procedure that combines unsupervised principal component analysis (PCA), criteria importance through inter-criteria correlation (CRITIC), and technique for order preference by similarity to ideal solution (TOPSIS) methods. A recent 2024 real-world application involving 57 listed Chinese pharmaceutical and biotechnology companies and 70 ESG indicators demonstrates the framework&amp;amp;rsquo;s practical utility in producing transparent and objective ESG rankings. The main contributions of this work are fourfold: (1) the development of an end-to-end, entirely data-driven ML framework for ESG evaluation; (2) the introduction of an innovative three-plane ESG feature selection workflow within the framework; (3) the first study for designing a hybrid PCA-CRITIC-TOPSIS approach in ESG weighting and ranking; (4) the validation of the framework through a real-world industry application using recent and authentic ESG data.</description>
	<pubDate>2026-04-22</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 130: A Data-Driven Machine Learning Framework for Multi-Criteria ESG Evaluation</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/5/130">doi: 10.3390/bdcc10050130</a></p>
	<p>Authors:
		Zhiyuan Wang
		Tristan Lim
		Yun Teng
		Chongwu Xia
		</p>
	<p>This study proposes a novel data-driven machine learning (ML) framework for multi-criteria environmental, social, and governance (ESG) evaluation. The framework aims to address the transparency, consistency, and subjectivity limitations of existing ESG evaluation systems by employing a fully data-driven, modular, and ML-supported architecture. It comprises three main modules: (i) ESG data preprocessing with missing-data imputation by the MissForest algorithm; (ii) a three-plane ESG feature selection workflow that integrates clustering, feature importance, and classification algorithms to identify representative ESG indicators; and (iii) a hybrid weighting and ranking procedure that combines unsupervised principal component analysis (PCA), criteria importance through inter-criteria correlation (CRITIC), and technique for order preference by similarity to ideal solution (TOPSIS) methods. A recent 2024 real-world application involving 57 listed Chinese pharmaceutical and biotechnology companies and 70 ESG indicators demonstrates the framework&amp;amp;rsquo;s practical utility in producing transparent and objective ESG rankings. The main contributions of this work are fourfold: (1) the development of an end-to-end, entirely data-driven ML framework for ESG evaluation; (2) the introduction of an innovative three-plane ESG feature selection workflow within the framework; (3) the first study for designing a hybrid PCA-CRITIC-TOPSIS approach in ESG weighting and ranking; (4) the validation of the framework through a real-world industry application using recent and authentic ESG data.</p>
	]]></content:encoded>

	<dc:title>A Data-Driven Machine Learning Framework for Multi-Criteria ESG Evaluation</dc:title>
			<dc:creator>Zhiyuan Wang</dc:creator>
			<dc:creator>Tristan Lim</dc:creator>
			<dc:creator>Yun Teng</dc:creator>
			<dc:creator>Chongwu Xia</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10050130</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-04-22</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-04-22</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>5</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>130</prism:startingPage>
		<prism:doi>10.3390/bdcc10050130</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/5/130</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/4/129">

	<title>BDCC, Vol. 10, Pages 129: Fuzz Driver Generation: A Survey and Outlook from the Perspective of Data Sources</title>
	<link>https://www.mdpi.com/2504-2289/10/4/129</link>
	<description>Fuzzing is an essential element of software supply chain security governance. Despite its importance, the widespread adoption of library fuzzing is limited by the significant costs associated with constructing fuzz drivers. Without a clear entry point, the reachable path space of the target library is determined by the interplay of API call sequences, parameter dependencies, and state constraints. As a result, fuzz drivers must achieve not only successful builds but also provide sufficient semantic context to enable exploration of deeper state machine interactions, thereby avoiding premature stagnation at superficial validation logic. To systematically assess advancements in automated fuzz driver generation, this paper develops a taxonomy organized around the primary data sources used to derive driver-generation constraints, categorizing existing approaches into four technological trajectories: Usage Artifact Mining, Source Code Constraint Inference, Binary Semantics Recovery, and Heterogeneous Data Fusion. Large language models are increasingly integrated into these workflows as generators and as components for constraint alignment and repair. To address inconsistencies in experimental methodologies, this paper introduces a bounded comparability-oriented evaluation perspective focused on three dimensions: validity, reachability-related evidence, and reproducibility and cost. Together with a disclosure and reporting protocol for metric comparability, this perspective clarifies the information needed for cross-study comparison and examines the unique features and inherent limitations of each technical trajectory. Based on these findings, three key directions for future research are identified: facilitating structural evolution in response to coverage plateaus to address deep logic unreachability; coordinating dynamic closed-loop orchestration that utilizes on-demand heterogeneous data retrieval to resolve context challenges; and developing language-agnostic driver representations with pluggable adaptation mechanisms to improve cross-ecosystem portability and scalability.</description>
	<pubDate>2026-04-21</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 129: Fuzz Driver Generation: A Survey and Outlook from the Perspective of Data Sources</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/4/129">doi: 10.3390/bdcc10040129</a></p>
	<p>Authors:
		Xiao Feng
		Shuaibing Lu
		Taotao Gu
		Yuanping Nie
		Qian Yan
		Mucheng Yang
		Jinyang Chen
		Xiaohui Kuang
		</p>
	<p>Fuzzing is an essential element of software supply chain security governance. Despite its importance, the widespread adoption of library fuzzing is limited by the significant costs associated with constructing fuzz drivers. Without a clear entry point, the reachable path space of the target library is determined by the interplay of API call sequences, parameter dependencies, and state constraints. As a result, fuzz drivers must achieve not only successful builds but also provide sufficient semantic context to enable exploration of deeper state machine interactions, thereby avoiding premature stagnation at superficial validation logic. To systematically assess advancements in automated fuzz driver generation, this paper develops a taxonomy organized around the primary data sources used to derive driver-generation constraints, categorizing existing approaches into four technological trajectories: Usage Artifact Mining, Source Code Constraint Inference, Binary Semantics Recovery, and Heterogeneous Data Fusion. Large language models are increasingly integrated into these workflows as generators and as components for constraint alignment and repair. To address inconsistencies in experimental methodologies, this paper introduces a bounded comparability-oriented evaluation perspective focused on three dimensions: validity, reachability-related evidence, and reproducibility and cost. Together with a disclosure and reporting protocol for metric comparability, this perspective clarifies the information needed for cross-study comparison and examines the unique features and inherent limitations of each technical trajectory. Based on these findings, three key directions for future research are identified: facilitating structural evolution in response to coverage plateaus to address deep logic unreachability; coordinating dynamic closed-loop orchestration that utilizes on-demand heterogeneous data retrieval to resolve context challenges; and developing language-agnostic driver representations with pluggable adaptation mechanisms to improve cross-ecosystem portability and scalability.</p>
	]]></content:encoded>

	<dc:title>Fuzz Driver Generation: A Survey and Outlook from the Perspective of Data Sources</dc:title>
			<dc:creator>Xiao Feng</dc:creator>
			<dc:creator>Shuaibing Lu</dc:creator>
			<dc:creator>Taotao Gu</dc:creator>
			<dc:creator>Yuanping Nie</dc:creator>
			<dc:creator>Qian Yan</dc:creator>
			<dc:creator>Mucheng Yang</dc:creator>
			<dc:creator>Jinyang Chen</dc:creator>
			<dc:creator>Xiaohui Kuang</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10040129</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-04-21</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-04-21</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>4</prism:number>
	<prism:section>Review</prism:section>
	<prism:startingPage>129</prism:startingPage>
		<prism:doi>10.3390/bdcc10040129</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/4/129</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/4/128">

	<title>BDCC, Vol. 10, Pages 128: A Reservoir Computing Approach for Synchronizing Discrete-Time 3D Chaotic Systems</title>
	<link>https://www.mdpi.com/2504-2289/10/4/128</link>
	<description>Reservoir computing (RC) is an efficient framework for processing time-series data. This work investigates the synchronization of two independently trained reservoir computers that, after training, operate without external input from the chaotic system and interact solely through symmetric linear coupling. This approach addresses a gap in existing reservoir computing-based synchronization studies, which predominantly rely on master&amp;amp;ndash;slave or system-driven configurations. In this work, we first build and train two reservoir computing models based on 3D nonlinear chaotic maps and hyperchaotic systems and then introduce a symmetric linear coupling mechanism between them. This study demonstrates that reservoir computing can accurately reproduce the short-term dynamics of chaotic systems and provides insight into the interactions between learned dynamical models, while also helping us understand how complex systems connect and operate collectively. We use this systematic approach to establish a framework for understanding how two trained reservoir computers interact under varying coupling strengths, enabling a detailed investigation of their synchronization behavior. To demonstrate the adaptability of the proposed framework to diverse dynamical behaviors, we systematically investigated three discrete chaotic and hyperchaotic systems: (1) discrete 3D sinusoidal map with discrete Lorenz attractor, (2) 3D sinusoidal map with conjoined Lorenz twin attractor, and (3) 3D quadratic hyperchaotic map. For performance evaluation, we trained coupled RCs and computed the synchronization error for different coupling strengths. We also present phase portraits and time-series plots of the attractors and RCs, along with the synchronization error as a function of the coupling strength, thereby demonstrating the possibility of synchronization of two linearly coupled RCs, which are independently trained on discrete, three-dimensional chaotic and hyperchaotic systems.</description>
	<pubDate>2026-04-21</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 128: A Reservoir Computing Approach for Synchronizing Discrete-Time 3D Chaotic Systems</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/4/128">doi: 10.3390/bdcc10040128</a></p>
	<p>Authors:
		Vismaya V. S
		Swetha P
		Jubin K. Babu
		Diya Gijo
		Varada M. T
		Adithya K. K
		Ekaterina Kopets
		Sishu Shankar Muni
		</p>
	<p>Reservoir computing (RC) is an efficient framework for processing time-series data. This work investigates the synchronization of two independently trained reservoir computers that, after training, operate without external input from the chaotic system and interact solely through symmetric linear coupling. This approach addresses a gap in existing reservoir computing-based synchronization studies, which predominantly rely on master&amp;amp;ndash;slave or system-driven configurations. In this work, we first build and train two reservoir computing models based on 3D nonlinear chaotic maps and hyperchaotic systems and then introduce a symmetric linear coupling mechanism between them. This study demonstrates that reservoir computing can accurately reproduce the short-term dynamics of chaotic systems and provides insight into the interactions between learned dynamical models, while also helping us understand how complex systems connect and operate collectively. We use this systematic approach to establish a framework for understanding how two trained reservoir computers interact under varying coupling strengths, enabling a detailed investigation of their synchronization behavior. To demonstrate the adaptability of the proposed framework to diverse dynamical behaviors, we systematically investigated three discrete chaotic and hyperchaotic systems: (1) discrete 3D sinusoidal map with discrete Lorenz attractor, (2) 3D sinusoidal map with conjoined Lorenz twin attractor, and (3) 3D quadratic hyperchaotic map. For performance evaluation, we trained coupled RCs and computed the synchronization error for different coupling strengths. We also present phase portraits and time-series plots of the attractors and RCs, along with the synchronization error as a function of the coupling strength, thereby demonstrating the possibility of synchronization of two linearly coupled RCs, which are independently trained on discrete, three-dimensional chaotic and hyperchaotic systems.</p>
	]]></content:encoded>

	<dc:title>A Reservoir Computing Approach for Synchronizing Discrete-Time 3D Chaotic Systems</dc:title>
			<dc:creator>Vismaya V. S</dc:creator>
			<dc:creator>Swetha P</dc:creator>
			<dc:creator>Jubin K. Babu</dc:creator>
			<dc:creator>Diya Gijo</dc:creator>
			<dc:creator>Varada M. T</dc:creator>
			<dc:creator>Adithya K. K</dc:creator>
			<dc:creator>Ekaterina Kopets</dc:creator>
			<dc:creator>Sishu Shankar Muni</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10040128</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-04-21</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-04-21</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>4</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>128</prism:startingPage>
		<prism:doi>10.3390/bdcc10040128</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/4/128</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/4/127">

	<title>BDCC, Vol. 10, Pages 127: Generative AI and Large Language Models</title>
	<link>https://www.mdpi.com/2504-2289/10/4/127</link>
	<description>In recent years, generative artificial intelligence and, in particular, large language models (LLMs) have rapidly transformed the landscape of data analysis, knowledge extraction, content generation, and intelligent decision support [...]</description>
	<pubDate>2026-04-21</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 127: Generative AI and Large Language Models</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/4/127">doi: 10.3390/bdcc10040127</a></p>
	<p>Authors:
		Fabrizio Marozzo
		Riccardo Cantini
		</p>
	<p>In recent years, generative artificial intelligence and, in particular, large language models (LLMs) have rapidly transformed the landscape of data analysis, knowledge extraction, content generation, and intelligent decision support [...]</p>
	]]></content:encoded>

	<dc:title>Generative AI and Large Language Models</dc:title>
			<dc:creator>Fabrizio Marozzo</dc:creator>
			<dc:creator>Riccardo Cantini</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10040127</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-04-21</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-04-21</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>4</prism:number>
	<prism:section>Editorial</prism:section>
	<prism:startingPage>127</prism:startingPage>
		<prism:doi>10.3390/bdcc10040127</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/4/127</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/4/126">

	<title>BDCC, Vol. 10, Pages 126: Edge Node Deployment for Turbidity Estimation in Farm Ponds</title>
	<link>https://www.mdpi.com/2504-2289/10/4/126</link>
	<description>Image-based AI offers a low-cost alternative to traditional turbidity sensors in farm ponds, yet the prevailing shift toward Vision Transformers (ViTs) critically overlooks two field realities: the chronic scarcity of annotated data (Small Data) and the strict computational limits of edge hardware. This study presents a frugal computer vision framework that challenges the need for complex architectures in environmental screening. By systematically benchmarking six deep learning models across a calibrated high-turbidity dataset (200&amp;amp;ndash;800 NTU, 700 images) under standardized capture conditions, we demonstrate that traditional Convolutional Neural Networks (CNNs) possess a crucial inductive bias for this task. Specifically, ResNet-50 significantly outperformed modern ViTs in both accuracy (96.3% vs. 80.0%) and data efficiency, effectively capturing spatial scattering patterns without the massive data requirements that hindered transformer convergence. Deployed on a resource-constrained Raspberry Pi 4, the CNN-based system achieved an inference latency of 46 ms, demonstrated in an initial hardware-in-the-loop field proof-of-concept (82.4% agreement under baseline, calm-weather conditions, n=17). This edge-native approach not only provides actionable spatial turbidity maps to guide on-farm filtration and livestock management decisions but also establishes a critical architectural baseline: under controlled capture protocols, mature CNNs consistently outperform ViTs, establishing them as the optimal architecture for frugal, small-data agricultural Internet of Things (IoT) deployments.</description>
	<pubDate>2026-04-18</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 126: Edge Node Deployment for Turbidity Estimation in Farm Ponds</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/4/126">doi: 10.3390/bdcc10040126</a></p>
	<p>Authors:
		Martin Moreno
		Iván Trejo-Zúñiga
		Víctor Alejandro González-Huitrón
		René Francisco Santana-Cruz
		Raúl García García
		Gabriela Pineda Chacón
		</p>
	<p>Image-based AI offers a low-cost alternative to traditional turbidity sensors in farm ponds, yet the prevailing shift toward Vision Transformers (ViTs) critically overlooks two field realities: the chronic scarcity of annotated data (Small Data) and the strict computational limits of edge hardware. This study presents a frugal computer vision framework that challenges the need for complex architectures in environmental screening. By systematically benchmarking six deep learning models across a calibrated high-turbidity dataset (200&amp;amp;ndash;800 NTU, 700 images) under standardized capture conditions, we demonstrate that traditional Convolutional Neural Networks (CNNs) possess a crucial inductive bias for this task. Specifically, ResNet-50 significantly outperformed modern ViTs in both accuracy (96.3% vs. 80.0%) and data efficiency, effectively capturing spatial scattering patterns without the massive data requirements that hindered transformer convergence. Deployed on a resource-constrained Raspberry Pi 4, the CNN-based system achieved an inference latency of 46 ms, demonstrated in an initial hardware-in-the-loop field proof-of-concept (82.4% agreement under baseline, calm-weather conditions, n=17). This edge-native approach not only provides actionable spatial turbidity maps to guide on-farm filtration and livestock management decisions but also establishes a critical architectural baseline: under controlled capture protocols, mature CNNs consistently outperform ViTs, establishing them as the optimal architecture for frugal, small-data agricultural Internet of Things (IoT) deployments.</p>
	]]></content:encoded>

	<dc:title>Edge Node Deployment for Turbidity Estimation in Farm Ponds</dc:title>
			<dc:creator>Martin Moreno</dc:creator>
			<dc:creator>Iván Trejo-Zúñiga</dc:creator>
			<dc:creator>Víctor Alejandro González-Huitrón</dc:creator>
			<dc:creator>René Francisco Santana-Cruz</dc:creator>
			<dc:creator>Raúl García García</dc:creator>
			<dc:creator>Gabriela Pineda Chacón</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10040126</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-04-18</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-04-18</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>4</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>126</prism:startingPage>
		<prism:doi>10.3390/bdcc10040126</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/4/126</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/4/125">

	<title>BDCC, Vol. 10, Pages 125: LST-AGCN: A Novel Unified Lightweight Attention Framework for Efficient Skeleton-Based Action Recognition</title>
	<link>https://www.mdpi.com/2504-2289/10/4/125</link>
	<description>While Graph Convolutional Networks (GCNs) have revolutionized skeleton-based action recognition, existing methods face a critical efficiency&amp;amp;ndash;accuracy dilemma: state-of-the-art approaches achieve high performance through computationally expensive multi-stream fusion (joint, bone, joint motion, and bone motion) and deep architectures, limiting real-world deployment on resource-constrained devices. We propose LST-AGCN (Lightweight Spatial&amp;amp;ndash;Temporal Attention Graph Convolutional Network), introducing three technical contributions that address this challenge: (1) Unified Attention Module (UAM)&amp;amp;mdash;a framework that integrates channel, spatial, and temporal attention through a single compact operation, significantly reducing attention parameters compared to separate attention mechanisms; (2) Depthwise Separable Attention Mechanism (DSAM)&amp;amp;mdash;a factorization using depthwise separable convolutions that achieves linear complexity reduction from O(C2) to O(C) in attention operations; and (3) Efficient Topology-Aware Fusion (ETAF)&amp;amp;mdash;an adaptive Joint-wise Attention strategy that captures fine-grained spatial relationships without quadratic complexity growth. Extensive experiments on NTU RGB+D 60 and NTU RGB+D 120 datasets demonstrate that LST-AGCN achieves strong performance using only joint modality (86.14%/94.0% and 79.5%/82.0% Top-1 accuracy with 99.0% Top-5 on cross-view) while requiring 14.11 M parameters and 19.02 GFLOPs, delivering efficient inference suitable for edge deployment.</description>
	<pubDate>2026-04-18</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 125: LST-AGCN: A Novel Unified Lightweight Attention Framework for Efficient Skeleton-Based Action Recognition</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/4/125">doi: 10.3390/bdcc10040125</a></p>
	<p>Authors:
		Khadija Lasri
		Khalid El Fazazy
		Adnane Mohamed Mahraz
		Hamid Tairi
		Jamal Riffi
		</p>
	<p>While Graph Convolutional Networks (GCNs) have revolutionized skeleton-based action recognition, existing methods face a critical efficiency&amp;amp;ndash;accuracy dilemma: state-of-the-art approaches achieve high performance through computationally expensive multi-stream fusion (joint, bone, joint motion, and bone motion) and deep architectures, limiting real-world deployment on resource-constrained devices. We propose LST-AGCN (Lightweight Spatial&amp;amp;ndash;Temporal Attention Graph Convolutional Network), introducing three technical contributions that address this challenge: (1) Unified Attention Module (UAM)&amp;amp;mdash;a framework that integrates channel, spatial, and temporal attention through a single compact operation, significantly reducing attention parameters compared to separate attention mechanisms; (2) Depthwise Separable Attention Mechanism (DSAM)&amp;amp;mdash;a factorization using depthwise separable convolutions that achieves linear complexity reduction from O(C2) to O(C) in attention operations; and (3) Efficient Topology-Aware Fusion (ETAF)&amp;amp;mdash;an adaptive Joint-wise Attention strategy that captures fine-grained spatial relationships without quadratic complexity growth. Extensive experiments on NTU RGB+D 60 and NTU RGB+D 120 datasets demonstrate that LST-AGCN achieves strong performance using only joint modality (86.14%/94.0% and 79.5%/82.0% Top-1 accuracy with 99.0% Top-5 on cross-view) while requiring 14.11 M parameters and 19.02 GFLOPs, delivering efficient inference suitable for edge deployment.</p>
	]]></content:encoded>

	<dc:title>LST-AGCN: A Novel Unified Lightweight Attention Framework for Efficient Skeleton-Based Action Recognition</dc:title>
			<dc:creator>Khadija Lasri</dc:creator>
			<dc:creator>Khalid El Fazazy</dc:creator>
			<dc:creator>Adnane Mohamed Mahraz</dc:creator>
			<dc:creator>Hamid Tairi</dc:creator>
			<dc:creator>Jamal Riffi</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10040125</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-04-18</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-04-18</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>4</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>125</prism:startingPage>
		<prism:doi>10.3390/bdcc10040125</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/4/125</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/4/124">

	<title>BDCC, Vol. 10, Pages 124: Understanding the Global Trends of 2025 Through the Defly Compass Methodology</title>
	<link>https://www.mdpi.com/2504-2289/10/4/124</link>
	<description>This study aims to identify and synthesize the major global trends that shaped 2025 by applying the DeflyCompass methodology to a curated corpus of strategic foresight reports. The study synthesizes insights from 23 strategic reports published by leading international organizations, including the World Economic Forum, Accenture, Euromonitor, and major technology firms. Methodologically, DeflyCompass operationalizes a structured hybrid human&amp;amp;ndash;AI pipeline comprising the deployment of multi-agent AI systems, automated knowledge graph construction, semantic clustering, and hybrid human&amp;amp;ndash;AI validation processes, reducing an initial set of 816 preliminary signals to a validated catalog of 50 high-priority trends across six PESTEL domains: Political, Economic, Social, Technological, Environmental, and Legal/Governance. Key findings indicate that artificial intelligence functions as a systemic enabling technology across all domains, climate and sustainability imperatives permeate multiple domains, geopolitical fragmentation introduces systemic tension, and trust deficits emerge as a critical vulnerability. The study contributes a replicable and scalable framework for global-level strategic foresight that operationalizes human&amp;amp;ndash;AI integration within a rigorous expert-driven validation process, complementing existing hybrid analytical approaches in the literature. Implications extend to decision-making in technology governance, sustainability strategy, social adaptation, and scenario planning, highlighting the necessity of integrating AI augmentation with human expertise for effective future-oriented planning.</description>
	<pubDate>2026-04-17</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 124: Understanding the Global Trends of 2025 Through the Defly Compass Methodology</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/4/124">doi: 10.3390/bdcc10040124</a></p>
	<p>Authors:
		Mabel López Bordao
		Antonia Ferrer Sapena
		Carlos A. Reyes Pérez
		Enrique A. Sánchez Pérez
		</p>
	<p>This study aims to identify and synthesize the major global trends that shaped 2025 by applying the DeflyCompass methodology to a curated corpus of strategic foresight reports. The study synthesizes insights from 23 strategic reports published by leading international organizations, including the World Economic Forum, Accenture, Euromonitor, and major technology firms. Methodologically, DeflyCompass operationalizes a structured hybrid human&amp;amp;ndash;AI pipeline comprising the deployment of multi-agent AI systems, automated knowledge graph construction, semantic clustering, and hybrid human&amp;amp;ndash;AI validation processes, reducing an initial set of 816 preliminary signals to a validated catalog of 50 high-priority trends across six PESTEL domains: Political, Economic, Social, Technological, Environmental, and Legal/Governance. Key findings indicate that artificial intelligence functions as a systemic enabling technology across all domains, climate and sustainability imperatives permeate multiple domains, geopolitical fragmentation introduces systemic tension, and trust deficits emerge as a critical vulnerability. The study contributes a replicable and scalable framework for global-level strategic foresight that operationalizes human&amp;amp;ndash;AI integration within a rigorous expert-driven validation process, complementing existing hybrid analytical approaches in the literature. Implications extend to decision-making in technology governance, sustainability strategy, social adaptation, and scenario planning, highlighting the necessity of integrating AI augmentation with human expertise for effective future-oriented planning.</p>
	]]></content:encoded>

	<dc:title>Understanding the Global Trends of 2025 Through the Defly Compass Methodology</dc:title>
			<dc:creator>Mabel López Bordao</dc:creator>
			<dc:creator>Antonia Ferrer Sapena</dc:creator>
			<dc:creator>Carlos A. Reyes Pérez</dc:creator>
			<dc:creator>Enrique A. Sánchez Pérez</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10040124</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-04-17</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-04-17</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>4</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>124</prism:startingPage>
		<prism:doi>10.3390/bdcc10040124</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/4/124</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/4/123">

	<title>BDCC, Vol. 10, Pages 123: Enhancing Collaborative AI Learning: A Blockchain-Secured, Edge-Enabled Platform for Multimodal Education in IIoT Environments</title>
	<link>https://www.mdpi.com/2504-2289/10/4/123</link>
	<description>As industries deploy more connected devices in factories, warehouses, and smart facilities, the need for artificial intelligence (AI) systems that can operate securely in distributed, data-intensive environments is growing. Traditional centralized learning and online education platforms struggle when students and systems have to process real-time streams (sensors, video, text) with strict latency and privacy requirements. To address this challenge, a blockchain-secured, edge-enabled multimodal federated learning framework tailored for Industrial IoT (IIoT) environments is proposed. The model integrates four key layers: (i) a blockchain layer that provides credentialing, transparency, and token-based incentives; (ii) a multimodal community layer that supports group formation, peer consensus, and cross-modal learning across text, images, audio, and sensor data; (iii) an edge computing layer that enables low-latency task offloading and secure training within Intel SGX enclaves; and (iv) a data layer that applies pre-processing, differential privacy, and synthetic augmentation to safeguard sensitive information. Experiments on industrial multimodal datasets demonstrate 42% faster model aggregation, 78.9% multimodal accuracy, and 1.9% accuracy loss under &amp;amp;epsilon; = 1.0 differential privacy. This shows a scalable and practical path for decentralized AI training in next-generation IIoT systems, confirming the possibility of technical support for educational processes. However, the conducted research requires a validation of pedagogical effectiveness.</description>
	<pubDate>2026-04-17</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 123: Enhancing Collaborative AI Learning: A Blockchain-Secured, Edge-Enabled Platform for Multimodal Education in IIoT Environments</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/4/123">doi: 10.3390/bdcc10040123</a></p>
	<p>Authors:
		Ahsan Rafiq
		Eduard Melnik
		Alexey Samoylov
		Alexander Kozlovskiy
		Irina Safronenkova
		</p>
	<p>As industries deploy more connected devices in factories, warehouses, and smart facilities, the need for artificial intelligence (AI) systems that can operate securely in distributed, data-intensive environments is growing. Traditional centralized learning and online education platforms struggle when students and systems have to process real-time streams (sensors, video, text) with strict latency and privacy requirements. To address this challenge, a blockchain-secured, edge-enabled multimodal federated learning framework tailored for Industrial IoT (IIoT) environments is proposed. The model integrates four key layers: (i) a blockchain layer that provides credentialing, transparency, and token-based incentives; (ii) a multimodal community layer that supports group formation, peer consensus, and cross-modal learning across text, images, audio, and sensor data; (iii) an edge computing layer that enables low-latency task offloading and secure training within Intel SGX enclaves; and (iv) a data layer that applies pre-processing, differential privacy, and synthetic augmentation to safeguard sensitive information. Experiments on industrial multimodal datasets demonstrate 42% faster model aggregation, 78.9% multimodal accuracy, and 1.9% accuracy loss under &amp;amp;epsilon; = 1.0 differential privacy. This shows a scalable and practical path for decentralized AI training in next-generation IIoT systems, confirming the possibility of technical support for educational processes. However, the conducted research requires a validation of pedagogical effectiveness.</p>
	]]></content:encoded>

	<dc:title>Enhancing Collaborative AI Learning: A Blockchain-Secured, Edge-Enabled Platform for Multimodal Education in IIoT Environments</dc:title>
			<dc:creator>Ahsan Rafiq</dc:creator>
			<dc:creator>Eduard Melnik</dc:creator>
			<dc:creator>Alexey Samoylov</dc:creator>
			<dc:creator>Alexander Kozlovskiy</dc:creator>
			<dc:creator>Irina Safronenkova</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10040123</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-04-17</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-04-17</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>4</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>123</prism:startingPage>
		<prism:doi>10.3390/bdcc10040123</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/4/123</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/4/122">

	<title>BDCC, Vol. 10, Pages 122: Ontology-Guided Multimodal Framework for Explainable Music Similarity and Recommendation</title>
	<link>https://www.mdpi.com/2504-2289/10/4/122</link>
	<description>Analyzing music similarity in large catalogs is challenging because people perceive music differently and important details are found in audio, text, and metadata. This article introduces a multimodal framework that uses an ontology to make music similarity and recommendation more explainable. The framework brings together learned features from audio, lyrics, and other text with structured metadata in a shared similarity space, and then improves ranking with a music ontology that captures relationships between songs, artists, genres, and moods. The design works with any encoder that creates fixed-size features. This study uses strong neural audio and text encoders, mainly based on transformers. This approach allows the system to handle different input types while staying reliable across datasets. This study tests the framework on several open music and audio datasets using content-based retrieval tasks and standard ranking measures. In addition to Configurations C1&amp;amp;ndash;C4, this study includes an external content-based reference baseline based on conventional MIR audio descriptors. This baseline represents a signal-level retrieval approach that models complementary aspects of the audio signal, such as timbre, harmony, and spectral characteristics, and is evaluated under the same retrieval protocol as the main framework. It is included to provide an external comparison point outside the proposed C1&amp;amp;ndash;C4 design. Compared to audio-only and non-ontological variants within the same framework, the proposed multimodal and ontology-guided configurations achieve better precision, recall, and mean average precision, and also cover more rare content. Visualizations and case studies show that combining different data types and using ontology-based reranking can improve performance and make results easier to interpret. This work lays the groundwork for explainable, cognitively informed music recommendation systems and points to future work in modeling user behavior over time and adapting to different cultures.</description>
	<pubDate>2026-04-15</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 122: Ontology-Guided Multimodal Framework for Explainable Music Similarity and Recommendation</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/4/122">doi: 10.3390/bdcc10040122</a></p>
	<p>Authors:
		Mikhail Rumiantcev
		</p>
	<p>Analyzing music similarity in large catalogs is challenging because people perceive music differently and important details are found in audio, text, and metadata. This article introduces a multimodal framework that uses an ontology to make music similarity and recommendation more explainable. The framework brings together learned features from audio, lyrics, and other text with structured metadata in a shared similarity space, and then improves ranking with a music ontology that captures relationships between songs, artists, genres, and moods. The design works with any encoder that creates fixed-size features. This study uses strong neural audio and text encoders, mainly based on transformers. This approach allows the system to handle different input types while staying reliable across datasets. This study tests the framework on several open music and audio datasets using content-based retrieval tasks and standard ranking measures. In addition to Configurations C1&amp;amp;ndash;C4, this study includes an external content-based reference baseline based on conventional MIR audio descriptors. This baseline represents a signal-level retrieval approach that models complementary aspects of the audio signal, such as timbre, harmony, and spectral characteristics, and is evaluated under the same retrieval protocol as the main framework. It is included to provide an external comparison point outside the proposed C1&amp;amp;ndash;C4 design. Compared to audio-only and non-ontological variants within the same framework, the proposed multimodal and ontology-guided configurations achieve better precision, recall, and mean average precision, and also cover more rare content. Visualizations and case studies show that combining different data types and using ontology-based reranking can improve performance and make results easier to interpret. This work lays the groundwork for explainable, cognitively informed music recommendation systems and points to future work in modeling user behavior over time and adapting to different cultures.</p>
	]]></content:encoded>

	<dc:title>Ontology-Guided Multimodal Framework for Explainable Music Similarity and Recommendation</dc:title>
			<dc:creator>Mikhail Rumiantcev</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10040122</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-04-15</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-04-15</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>4</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>122</prism:startingPage>
		<prism:doi>10.3390/bdcc10040122</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/4/122</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/4/121">

	<title>BDCC, Vol. 10, Pages 121: Distilling the Complexity of Agent-Based Simulations into Textual Explanations via Large Language Models</title>
	<link>https://www.mdpi.com/2504-2289/10/4/121</link>
	<description>Communicating the design and results of agent-based models (ABMs) to subject matter experts is challenging, which hinders participation and limits trust in simulation-based decision support. Large language models (LLMs) can communicate ABMs as textual summaries, thus complementing traditional disclosure through statistical and visualization techniques. While prior work translated the structure of conceptual models into narratives via LLMs, our extension covers the dynamics of simulation models via an automated simulation-to-text method that extracts contextual information from NetLogo ABMs, performs repeated simulations, and generates narrative descriptions (including the model&amp;amp;rsquo;s purpose, parameters, and simulation dynamics) using mutimodal LLMs. Furthermore, four summarization algorithms spanning abstractive and extractive methods provide shorter reports. Using Design-of-Experiments methods over three peer-reviewed ABMs, state-of-the-art multimodal LLMs from 2026 (Gemini 3.1 Pro, Qwen 3.5, Kimi K2.5, Claude Opus 4.6) and different prompt elements (e.g., roles, examples, generating insights, statistical analyses), we compare our results with several reference reports (e.g., from associate professors). We find that report quality is determined mainly (i.e., up to 34% of the variance) by the summarization algorithm and its interaction with the LLM, with abstractive summarizers (BART, T5) producing more coherent and readable reports, while Claude Opus 4.6 is the most robust LLM.</description>
	<pubDate>2026-04-15</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 121: Distilling the Complexity of Agent-Based Simulations into Textual Explanations via Large Language Models</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/4/121">doi: 10.3390/bdcc10040121</a></p>
	<p>Authors:
		Noé Y. Flandre
		Philippe J. Giabbanelli
		</p>
	<p>Communicating the design and results of agent-based models (ABMs) to subject matter experts is challenging, which hinders participation and limits trust in simulation-based decision support. Large language models (LLMs) can communicate ABMs as textual summaries, thus complementing traditional disclosure through statistical and visualization techniques. While prior work translated the structure of conceptual models into narratives via LLMs, our extension covers the dynamics of simulation models via an automated simulation-to-text method that extracts contextual information from NetLogo ABMs, performs repeated simulations, and generates narrative descriptions (including the model&amp;amp;rsquo;s purpose, parameters, and simulation dynamics) using mutimodal LLMs. Furthermore, four summarization algorithms spanning abstractive and extractive methods provide shorter reports. Using Design-of-Experiments methods over three peer-reviewed ABMs, state-of-the-art multimodal LLMs from 2026 (Gemini 3.1 Pro, Qwen 3.5, Kimi K2.5, Claude Opus 4.6) and different prompt elements (e.g., roles, examples, generating insights, statistical analyses), we compare our results with several reference reports (e.g., from associate professors). We find that report quality is determined mainly (i.e., up to 34% of the variance) by the summarization algorithm and its interaction with the LLM, with abstractive summarizers (BART, T5) producing more coherent and readable reports, while Claude Opus 4.6 is the most robust LLM.</p>
	]]></content:encoded>

	<dc:title>Distilling the Complexity of Agent-Based Simulations into Textual Explanations via Large Language Models</dc:title>
			<dc:creator>Noé Y. Flandre</dc:creator>
			<dc:creator>Philippe J. Giabbanelli</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10040121</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-04-15</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-04-15</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>4</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>121</prism:startingPage>
		<prism:doi>10.3390/bdcc10040121</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/4/121</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/4/120">

	<title>BDCC, Vol. 10, Pages 120: Enhancing Retrieval-Augmented Generation with Entity Linking for Educational Platforms</title>
	<link>https://www.mdpi.com/2504-2289/10/4/120</link>
	<description>In the era of Large Language Models (LLMs), Retrieval-Augmented Generation (RAG) architectures are gaining significant attention for their ability to ground language generation in reliable knowledge sources. Despite their effectiveness, RAG systems based solely on semantic similarity often fail to ensure factual accuracy in specialized domains, where terminological ambiguity can affect retrieval relevance. This study proposes Entity Linking Enhanced RAG (ELERAG), an enhanced RAG architecture that integrates a factual signal derived from Entity Linking to improve the accuracy of educational question-answering systems in Italian. The system includes a Wikidata-based Entity Linking module and implements a hybrid re-ranking strategy based on Reciprocal Rank Fusion (RRF). To validate our approach, we compared it against standard baselines and state-of-the-art methods, including a Weighted-Score Re-ranking, a standalone Cross-Encoder and a combined RRF + Cross-Encoder pipeline. Experiments were conducted on two benchmarks: a custom academic dataset and the standard SQuAD-it dataset. Results show that, in domain-specific contexts, ELERAG significantly outperforms both the baseline and the Cross-Encoder configurations. Conversely, the Cross-Encoder approaches achieve the best results on the general-domain dataset. These findings provide strong experimental evidence of the domain mismatch effect, highlighting the importance of domain-adapted hybrid strategies to enhance factual precision in educational RAG systems without relying on computationally expensive models trained on disparate data distributions. They also demonstrate the potential of entity-aware RAG systems in educational environments, fostering adaptive and reliable AI-based tutoring tools.</description>
	<pubDate>2026-04-13</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 120: Enhancing Retrieval-Augmented Generation with Entity Linking for Educational Platforms</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/4/120">doi: 10.3390/bdcc10040120</a></p>
	<p>Authors:
		Francesco Granata
		Francesco Poggi
		Misael Mongiovì
		</p>
	<p>In the era of Large Language Models (LLMs), Retrieval-Augmented Generation (RAG) architectures are gaining significant attention for their ability to ground language generation in reliable knowledge sources. Despite their effectiveness, RAG systems based solely on semantic similarity often fail to ensure factual accuracy in specialized domains, where terminological ambiguity can affect retrieval relevance. This study proposes Entity Linking Enhanced RAG (ELERAG), an enhanced RAG architecture that integrates a factual signal derived from Entity Linking to improve the accuracy of educational question-answering systems in Italian. The system includes a Wikidata-based Entity Linking module and implements a hybrid re-ranking strategy based on Reciprocal Rank Fusion (RRF). To validate our approach, we compared it against standard baselines and state-of-the-art methods, including a Weighted-Score Re-ranking, a standalone Cross-Encoder and a combined RRF + Cross-Encoder pipeline. Experiments were conducted on two benchmarks: a custom academic dataset and the standard SQuAD-it dataset. Results show that, in domain-specific contexts, ELERAG significantly outperforms both the baseline and the Cross-Encoder configurations. Conversely, the Cross-Encoder approaches achieve the best results on the general-domain dataset. These findings provide strong experimental evidence of the domain mismatch effect, highlighting the importance of domain-adapted hybrid strategies to enhance factual precision in educational RAG systems without relying on computationally expensive models trained on disparate data distributions. They also demonstrate the potential of entity-aware RAG systems in educational environments, fostering adaptive and reliable AI-based tutoring tools.</p>
	]]></content:encoded>

	<dc:title>Enhancing Retrieval-Augmented Generation with Entity Linking for Educational Platforms</dc:title>
			<dc:creator>Francesco Granata</dc:creator>
			<dc:creator>Francesco Poggi</dc:creator>
			<dc:creator>Misael Mongiovì</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10040120</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-04-13</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-04-13</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>4</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>120</prism:startingPage>
		<prism:doi>10.3390/bdcc10040120</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/4/120</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/4/118">

	<title>BDCC, Vol. 10, Pages 118: Spatio-Temporal Analysis of Handball Players&amp;rsquo; Actions from Broadcast Videos Using Deep Learning</title>
	<link>https://www.mdpi.com/2504-2289/10/4/118</link>
	<description>Handball performance analysis is still often conducted through the manual review of match videos, while automation on broadcast footage remains challenging due to camera motion, strong perspective effects, and frequent occlusions during dense interactions. This study presents a practical and reproducible monocular pipeline for extracting handball analytics from a single broadcast viewpoint. Players are detected per frame, tracked over time, and projected onto a standardized handball court via homography-based camera calibration. The resulting court-referenced trajectories in metric units enable motion indicators such as distance covered and speed, along with coaching-oriented visual summaries, including trajectory overlays and heatmaps. In addition, clip-level action recognition is performed using interpretable kinematic and scene-derived features and lightweight classifiers, with a comparative evaluation across multiple classical models. The modular design keeps the intermediate steps explicit, supports reproducibility, and facilitates interpretation of both intermediate outputs and final analytics. Experiments on the UNIRI handball dataset demonstrate that meaningful performance analytics and action understanding can be obtained from single-camera broadcast video using transparent intermediate representations. This work highlights the practical potential of interpretable trajectory-based modeling for under-instrumented sports and provides a reproducible baseline for future extensions incorporating richer contextual cues.</description>
	<pubDate>2026-04-12</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 118: Spatio-Temporal Analysis of Handball Players&amp;rsquo; Actions from Broadcast Videos Using Deep Learning</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/4/118">doi: 10.3390/bdcc10040118</a></p>
	<p>Authors:
		Kosmas Katsioulas
		Ilias Maglogiannis
		</p>
	<p>Handball performance analysis is still often conducted through the manual review of match videos, while automation on broadcast footage remains challenging due to camera motion, strong perspective effects, and frequent occlusions during dense interactions. This study presents a practical and reproducible monocular pipeline for extracting handball analytics from a single broadcast viewpoint. Players are detected per frame, tracked over time, and projected onto a standardized handball court via homography-based camera calibration. The resulting court-referenced trajectories in metric units enable motion indicators such as distance covered and speed, along with coaching-oriented visual summaries, including trajectory overlays and heatmaps. In addition, clip-level action recognition is performed using interpretable kinematic and scene-derived features and lightweight classifiers, with a comparative evaluation across multiple classical models. The modular design keeps the intermediate steps explicit, supports reproducibility, and facilitates interpretation of both intermediate outputs and final analytics. Experiments on the UNIRI handball dataset demonstrate that meaningful performance analytics and action understanding can be obtained from single-camera broadcast video using transparent intermediate representations. This work highlights the practical potential of interpretable trajectory-based modeling for under-instrumented sports and provides a reproducible baseline for future extensions incorporating richer contextual cues.</p>
	]]></content:encoded>

	<dc:title>Spatio-Temporal Analysis of Handball Players&amp;amp;rsquo; Actions from Broadcast Videos Using Deep Learning</dc:title>
			<dc:creator>Kosmas Katsioulas</dc:creator>
			<dc:creator>Ilias Maglogiannis</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10040118</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-04-12</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-04-12</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>4</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>118</prism:startingPage>
		<prism:doi>10.3390/bdcc10040118</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/4/118</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/4/119">

	<title>BDCC, Vol. 10, Pages 119: FMT-SVM: A Communication-Efficient Federated Multi-Task Support Vector Machine Framework for Healthcare</title>
	<link>https://www.mdpi.com/2504-2289/10/4/119</link>
	<description>Federated learning has become a promising paradigm in the training of decentralized machine learning models across clients without sharing raw data, thereby preserving privacy. Current federated support vector machine methods are mainly based on the learning of a single global model, which inadequately addresses the challenges presented by heterogeneous and non-IID client data distributions. To overcome these limitations, we propose FMT-SVM, a novel federated multi-task learning framework that jointly trains both binary and multi-class classification tasks within each client, where the client uses a unified convolutional neural network encoder to extract common features among tasks, which are passed to task-specific linear SVM heads dedicated to each classification task. By leveraging a primal optimization integrating task covariance and global consensus regularization, FMT-SVM explicitly models relationships between heterogeneous tasks and enforces alignment across clients, effectively handling the non-IID nature of data distributions. Unlike traditional FL methods that exchange entire model parameters or large support vector sets, our method communicates only the compact SVM heads during aggregation, greatly reducing communication overhead and enhancing scalability for clients with limited bandwidth. To further enhance privacy, Gaussian differential privacy mechanisms are applied to client updates, balancing privacy preservation with predictive performance. Experiments are performed on two medical image datasets: the Pediatric Pneumonia Dataset and the Breast Ultrasound dataset, demonstrating that the FMT-SVM framework achieves competitive accuracy on both binary and multi-class tasks while maintaining communication efficiency and privacy guarantees. These results highlight the capability of the proposed FMT-SVM framework as a practical, scalable, and privacy-aware solution for the federated true multi-task learning problem in sensitive healthcare applications.</description>
	<pubDate>2026-04-12</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 119: FMT-SVM: A Communication-Efficient Federated Multi-Task Support Vector Machine Framework for Healthcare</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/4/119">doi: 10.3390/bdcc10040119</a></p>
	<p>Authors:
		Naima Firdaus
		Sachin Balkrushna Jadhav
		Zahid Raza
		Maria Lapina
		Mikhail Babenko
		</p>
	<p>Federated learning has become a promising paradigm in the training of decentralized machine learning models across clients without sharing raw data, thereby preserving privacy. Current federated support vector machine methods are mainly based on the learning of a single global model, which inadequately addresses the challenges presented by heterogeneous and non-IID client data distributions. To overcome these limitations, we propose FMT-SVM, a novel federated multi-task learning framework that jointly trains both binary and multi-class classification tasks within each client, where the client uses a unified convolutional neural network encoder to extract common features among tasks, which are passed to task-specific linear SVM heads dedicated to each classification task. By leveraging a primal optimization integrating task covariance and global consensus regularization, FMT-SVM explicitly models relationships between heterogeneous tasks and enforces alignment across clients, effectively handling the non-IID nature of data distributions. Unlike traditional FL methods that exchange entire model parameters or large support vector sets, our method communicates only the compact SVM heads during aggregation, greatly reducing communication overhead and enhancing scalability for clients with limited bandwidth. To further enhance privacy, Gaussian differential privacy mechanisms are applied to client updates, balancing privacy preservation with predictive performance. Experiments are performed on two medical image datasets: the Pediatric Pneumonia Dataset and the Breast Ultrasound dataset, demonstrating that the FMT-SVM framework achieves competitive accuracy on both binary and multi-class tasks while maintaining communication efficiency and privacy guarantees. These results highlight the capability of the proposed FMT-SVM framework as a practical, scalable, and privacy-aware solution for the federated true multi-task learning problem in sensitive healthcare applications.</p>
	]]></content:encoded>

	<dc:title>FMT-SVM: A Communication-Efficient Federated Multi-Task Support Vector Machine Framework for Healthcare</dc:title>
			<dc:creator>Naima Firdaus</dc:creator>
			<dc:creator>Sachin Balkrushna Jadhav</dc:creator>
			<dc:creator>Zahid Raza</dc:creator>
			<dc:creator>Maria Lapina</dc:creator>
			<dc:creator>Mikhail Babenko</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10040119</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-04-12</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-04-12</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>4</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>119</prism:startingPage>
		<prism:doi>10.3390/bdcc10040119</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/4/119</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/4/117">

	<title>BDCC, Vol. 10, Pages 117: Understanding and Predicting Tourist Behavior Through Large Language Models</title>
	<link>https://www.mdpi.com/2504-2289/10/4/117</link>
	<description>Understanding and predicting how tourists move through a city is a challenging task, as it involves a complex interplay of spatial, temporal, and social factors. Traditional recommender systems often rely on structured data, trying to capture the nature of the problem. However, recent advances in Large Language Models (LLMs) open new possibilities for reasoning over richer, text-based representations of user context, even without a dedicated pre-training phase. In this study, we investigate the potential of LLMs to interpret and predict tourist movements in a real-world application scenario involving tourist visits to Verona, a municipality in Northern Italy, between 2014 and 2023. We propose an incremental prompt engineering approach that gradually enriches the model input, from spatial features alone to richer behavioral information, including visit histories, time information, and user cluster patterns. The approach is evaluated using six open-source models, enabling us to compare their accuracy and efficiency across various levels of contextual enrichment. The results provide a first insight about the abilities of LLMs to incorporate spatio-temporal contextual factors, thus improving predictions, while maintaining computational efficiency. The analysis of the model-generated explanations completes the picture by adding an interpretability dimension that most existing next-PoI prediction solutions lack. Overall, the study demonstrates the potential of LLMs to integrate multiple contextual dimensions in tourism mobility, highlighting the possibility of a more text-oriented, adaptive, and explainable T-RS.</description>
	<pubDate>2026-04-11</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 117: Understanding and Predicting Tourist Behavior Through Large Language Models</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/4/117">doi: 10.3390/bdcc10040117</a></p>
	<p>Authors:
		Anna Dalla Vecchia
		Simone Mattioli
		Sara Migliorini
		Elisa Quintarelli
		</p>
	<p>Understanding and predicting how tourists move through a city is a challenging task, as it involves a complex interplay of spatial, temporal, and social factors. Traditional recommender systems often rely on structured data, trying to capture the nature of the problem. However, recent advances in Large Language Models (LLMs) open new possibilities for reasoning over richer, text-based representations of user context, even without a dedicated pre-training phase. In this study, we investigate the potential of LLMs to interpret and predict tourist movements in a real-world application scenario involving tourist visits to Verona, a municipality in Northern Italy, between 2014 and 2023. We propose an incremental prompt engineering approach that gradually enriches the model input, from spatial features alone to richer behavioral information, including visit histories, time information, and user cluster patterns. The approach is evaluated using six open-source models, enabling us to compare their accuracy and efficiency across various levels of contextual enrichment. The results provide a first insight about the abilities of LLMs to incorporate spatio-temporal contextual factors, thus improving predictions, while maintaining computational efficiency. The analysis of the model-generated explanations completes the picture by adding an interpretability dimension that most existing next-PoI prediction solutions lack. Overall, the study demonstrates the potential of LLMs to integrate multiple contextual dimensions in tourism mobility, highlighting the possibility of a more text-oriented, adaptive, and explainable T-RS.</p>
	]]></content:encoded>

	<dc:title>Understanding and Predicting Tourist Behavior Through Large Language Models</dc:title>
			<dc:creator>Anna Dalla Vecchia</dc:creator>
			<dc:creator>Simone Mattioli</dc:creator>
			<dc:creator>Sara Migliorini</dc:creator>
			<dc:creator>Elisa Quintarelli</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10040117</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-04-11</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-04-11</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>4</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>117</prism:startingPage>
		<prism:doi>10.3390/bdcc10040117</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/4/117</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/4/115">

	<title>BDCC, Vol. 10, Pages 115: Experimental Validation and Reservoir Computing Capability of Spiking Neuron Based on Threshold Selector and Tunnel Diode</title>
	<link>https://www.mdpi.com/2504-2289/10/4/115</link>
	<description>Despite the success of artificial neural networks in solving numerous tasks, they face significant challenges, including difficulties in online adaptation and rapidly increasing energy consumption. As a biologically plausible alternative, spiking neural networks offer promising capabilities for efficient cognitive computing. Recently, a three-element spiking neuron model consisting of a threshold selector, a tunnel diode, and a capacitor was proposed. In this work, we experimentally validate this model using a threshold selector hardware emulator and demonstrate its dynamical equivalence to the biologically plausible Izhikevich neuron model. To evaluate the novel neuron&amp;amp;rsquo;s applicability for cognitive computing, we implement a liquid state machine (LSM) reservoir architecture with spatially dependent random topology for synaptic weight distribution. Our simulations on the MNIST and Fashion-MNIST benchmarks demonstrate competitive classification accuracy (97.9% and 89.5%, respectively) while offering estimated energy efficiency and processing speed enhancements compared to existing FPGA-based and memristor-based spiking reservoir implementations. The developed reservoir is feasible for processing neuromorphic sensors output, including visual perception tasks.</description>
	<pubDate>2026-04-10</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 115: Experimental Validation and Reservoir Computing Capability of Spiking Neuron Based on Threshold Selector and Tunnel Diode</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/4/115">doi: 10.3390/bdcc10040115</a></p>
	<p>Authors:
		Vasiliy Pchelko
		Vladislav Kholkin
		Vyacheslav Rybin
		Alexander Mikhailov
		Timur Karimov
		</p>
	<p>Despite the success of artificial neural networks in solving numerous tasks, they face significant challenges, including difficulties in online adaptation and rapidly increasing energy consumption. As a biologically plausible alternative, spiking neural networks offer promising capabilities for efficient cognitive computing. Recently, a three-element spiking neuron model consisting of a threshold selector, a tunnel diode, and a capacitor was proposed. In this work, we experimentally validate this model using a threshold selector hardware emulator and demonstrate its dynamical equivalence to the biologically plausible Izhikevich neuron model. To evaluate the novel neuron&amp;amp;rsquo;s applicability for cognitive computing, we implement a liquid state machine (LSM) reservoir architecture with spatially dependent random topology for synaptic weight distribution. Our simulations on the MNIST and Fashion-MNIST benchmarks demonstrate competitive classification accuracy (97.9% and 89.5%, respectively) while offering estimated energy efficiency and processing speed enhancements compared to existing FPGA-based and memristor-based spiking reservoir implementations. The developed reservoir is feasible for processing neuromorphic sensors output, including visual perception tasks.</p>
	]]></content:encoded>

	<dc:title>Experimental Validation and Reservoir Computing Capability of Spiking Neuron Based on Threshold Selector and Tunnel Diode</dc:title>
			<dc:creator>Vasiliy Pchelko</dc:creator>
			<dc:creator>Vladislav Kholkin</dc:creator>
			<dc:creator>Vyacheslav Rybin</dc:creator>
			<dc:creator>Alexander Mikhailov</dc:creator>
			<dc:creator>Timur Karimov</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10040115</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-04-10</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-04-10</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>4</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>115</prism:startingPage>
		<prism:doi>10.3390/bdcc10040115</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/4/115</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/4/116">

	<title>BDCC, Vol. 10, Pages 116: Non-Invasive Sleep Stage Classification with Imbalance-Aware Machine Learning for Healthcare Monitoring</title>
	<link>https://www.mdpi.com/2504-2289/10/4/116</link>
	<description>Sleep plays a fundamental role in human health and cognitive functioning, motivating the development of reliable and scalable methodologies for sleep stage classification (SSC). Recent advances in non-invasive and economically sustainable sensing technologies enable continuous sleep monitoring beyond laboratory settings. However, SSC remains a challenging data analytics task due to the intrinsic class imbalance among sleep stages. This study investigates the effectiveness of different imbalanced data management strategies within a machine learning framework for non-invasive SSC. The proposed approach relies exclusively on heart rate and motion signals, which can be acquired through wearable devices or contactless under-mattress sensors, making it suitable for longitudinal monitoring scenarios. Using the PhysioNet DREAMT dataset, 32 experimental scenarios are defined by combining data-level techniques (ADASYN oversampling with different balancing weights), algorithm-level strategies (cost-sensitive learning), and hybrid solutions. Four model families are evaluated&amp;amp;mdash;Decision Tree, k-Nearest Neighbors, Ensemble Classifiers, and Artificial Neural Networks&amp;amp;mdash;across classification tasks involving 2, 3, 4, and 5 sleep stages. The experimental results show that ensemble-based models provide robust and consistent performance under severe class imbalance, achieving macro accuracies of 82% for sleep&amp;amp;ndash;wake detection, 73% for 3-stage classification, 72% for 4-stage classification, and 64% for 5-stage classification. These findings confirm the relevance of imbalance-aware analytics and demonstrate the feasibility of accurate, minimally invasive SSC within big data and cognitive computing paradigms.</description>
	<pubDate>2026-04-10</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 116: Non-Invasive Sleep Stage Classification with Imbalance-Aware Machine Learning for Healthcare Monitoring</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/4/116">doi: 10.3390/bdcc10040116</a></p>
	<p>Authors:
		Luisiana Sabbatini
		Alberto Belli
		Sara Bruschi
		Marco Esposito
		Sara Raggiunto
		Paola Pierleoni
		</p>
	<p>Sleep plays a fundamental role in human health and cognitive functioning, motivating the development of reliable and scalable methodologies for sleep stage classification (SSC). Recent advances in non-invasive and economically sustainable sensing technologies enable continuous sleep monitoring beyond laboratory settings. However, SSC remains a challenging data analytics task due to the intrinsic class imbalance among sleep stages. This study investigates the effectiveness of different imbalanced data management strategies within a machine learning framework for non-invasive SSC. The proposed approach relies exclusively on heart rate and motion signals, which can be acquired through wearable devices or contactless under-mattress sensors, making it suitable for longitudinal monitoring scenarios. Using the PhysioNet DREAMT dataset, 32 experimental scenarios are defined by combining data-level techniques (ADASYN oversampling with different balancing weights), algorithm-level strategies (cost-sensitive learning), and hybrid solutions. Four model families are evaluated&amp;amp;mdash;Decision Tree, k-Nearest Neighbors, Ensemble Classifiers, and Artificial Neural Networks&amp;amp;mdash;across classification tasks involving 2, 3, 4, and 5 sleep stages. The experimental results show that ensemble-based models provide robust and consistent performance under severe class imbalance, achieving macro accuracies of 82% for sleep&amp;amp;ndash;wake detection, 73% for 3-stage classification, 72% for 4-stage classification, and 64% for 5-stage classification. These findings confirm the relevance of imbalance-aware analytics and demonstrate the feasibility of accurate, minimally invasive SSC within big data and cognitive computing paradigms.</p>
	]]></content:encoded>

	<dc:title>Non-Invasive Sleep Stage Classification with Imbalance-Aware Machine Learning for Healthcare Monitoring</dc:title>
			<dc:creator>Luisiana Sabbatini</dc:creator>
			<dc:creator>Alberto Belli</dc:creator>
			<dc:creator>Sara Bruschi</dc:creator>
			<dc:creator>Marco Esposito</dc:creator>
			<dc:creator>Sara Raggiunto</dc:creator>
			<dc:creator>Paola Pierleoni</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10040116</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-04-10</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-04-10</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>4</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>116</prism:startingPage>
		<prism:doi>10.3390/bdcc10040116</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/4/116</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/4/114">

	<title>BDCC, Vol. 10, Pages 114: Hybrid Approach to Patient Review Classification at Scale: From Expert Annotations to Production-Ready Machine Learning Models for Sustainable Healthcare</title>
	<link>https://www.mdpi.com/2504-2289/10/4/114</link>
	<description>Patients leave millions of medical reviews annually, providing critical data for quality management. However, manual processing is infeasible, and existing systems fail to distinguish medical from organizational problems&amp;amp;mdash;a distinction essential for complaint routing. The consequences of misrouting are significant: clinical issues may go unaddressed when medical complaints reach administrative staff, while systemic service problems remain unresolved when organizational complaints reach medical directors. We developed a hybrid approach combining expert annotation with Large Language Models (LLMs). Fifteen prompt iterations on 1500 reviews with expert validation (modified Cohen&amp;amp;rsquo;s kappa (&amp;amp;kappa;_mod), which weights errors hierarchically, reached 0.745) preceded the LLM annotation of 15,000 mixed-sentiment and positive reviews. These were combined with 7417 expert-annotated negative reviews to form a corpus of 22,417 reviews. Eight architectures, ranging from Logistic Regression to a BERT + TF-IDF + LightGBM ensemble, were compared using both standard metrics and domain-specific practical metrics tailored to complaint routing. The best model, scaled to 4.3 million Russian-language reviews from the Prodoctorov.ru platform, achieved 92.9% Practical Accuracy&amp;amp;mdash;the proportion of reviews classified without critical medical&amp;amp;ndash;organizational misclassification errors (M &amp;amp;harr; O)&amp;amp;mdash;compared to 68.0% standard accuracy, which treats all errors equally. Critical errors were reduced to 1.4%, yielding 144,000 more correctly processed complaints than traditional methods (TF-IDF + Logistic Regression). Analysis of the scaled data revealed the following: 46.1% M (medical), 21.0% O (organizational), and 32.9% C (combined) reviews; medical ratings were highest (4.75 vs. 4.59 for organizational, p &amp;amp;lt; 0.001); combined reviews were longest (802 characters); zero-star reviews comprised 3.8% of feedback, with organizational complaints dominating (38.2%) among extreme negatives; and average ratings rose by 1.24 points over 14 years. This hybrid approach yields expert-comparable corpora, automates 93% of feedback processing, ensures correct complaint routing, and contributes to healthcare sustainability by reducing administrative burden, accelerating resolution, and enabling data-driven quality management without proportional increases in human resources. All analyses were conducted on Russian-language patient reviews.</description>
	<pubDate>2026-04-09</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 114: Hybrid Approach to Patient Review Classification at Scale: From Expert Annotations to Production-Ready Machine Learning Models for Sustainable Healthcare</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/4/114">doi: 10.3390/bdcc10040114</a></p>
	<p>Authors:
		Irina Evgenievna Kalabikhina
		Anton Vasilyevich Kolotusha
		Vadim Sergeevich Moshkin
		</p>
	<p>Patients leave millions of medical reviews annually, providing critical data for quality management. However, manual processing is infeasible, and existing systems fail to distinguish medical from organizational problems&amp;amp;mdash;a distinction essential for complaint routing. The consequences of misrouting are significant: clinical issues may go unaddressed when medical complaints reach administrative staff, while systemic service problems remain unresolved when organizational complaints reach medical directors. We developed a hybrid approach combining expert annotation with Large Language Models (LLMs). Fifteen prompt iterations on 1500 reviews with expert validation (modified Cohen&amp;amp;rsquo;s kappa (&amp;amp;kappa;_mod), which weights errors hierarchically, reached 0.745) preceded the LLM annotation of 15,000 mixed-sentiment and positive reviews. These were combined with 7417 expert-annotated negative reviews to form a corpus of 22,417 reviews. Eight architectures, ranging from Logistic Regression to a BERT + TF-IDF + LightGBM ensemble, were compared using both standard metrics and domain-specific practical metrics tailored to complaint routing. The best model, scaled to 4.3 million Russian-language reviews from the Prodoctorov.ru platform, achieved 92.9% Practical Accuracy&amp;amp;mdash;the proportion of reviews classified without critical medical&amp;amp;ndash;organizational misclassification errors (M &amp;amp;harr; O)&amp;amp;mdash;compared to 68.0% standard accuracy, which treats all errors equally. Critical errors were reduced to 1.4%, yielding 144,000 more correctly processed complaints than traditional methods (TF-IDF + Logistic Regression). Analysis of the scaled data revealed the following: 46.1% M (medical), 21.0% O (organizational), and 32.9% C (combined) reviews; medical ratings were highest (4.75 vs. 4.59 for organizational, p &amp;amp;lt; 0.001); combined reviews were longest (802 characters); zero-star reviews comprised 3.8% of feedback, with organizational complaints dominating (38.2%) among extreme negatives; and average ratings rose by 1.24 points over 14 years. This hybrid approach yields expert-comparable corpora, automates 93% of feedback processing, ensures correct complaint routing, and contributes to healthcare sustainability by reducing administrative burden, accelerating resolution, and enabling data-driven quality management without proportional increases in human resources. All analyses were conducted on Russian-language patient reviews.</p>
	]]></content:encoded>

	<dc:title>Hybrid Approach to Patient Review Classification at Scale: From Expert Annotations to Production-Ready Machine Learning Models for Sustainable Healthcare</dc:title>
			<dc:creator>Irina Evgenievna Kalabikhina</dc:creator>
			<dc:creator>Anton Vasilyevich Kolotusha</dc:creator>
			<dc:creator>Vadim Sergeevich Moshkin</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10040114</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-04-09</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-04-09</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>4</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>114</prism:startingPage>
		<prism:doi>10.3390/bdcc10040114</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/4/114</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/4/113">

	<title>BDCC, Vol. 10, Pages 113: Adaptive Sensitivity-Aware Differential Privacy Accounting for Federated Smart-Meter Theft Detection</title>
	<link>https://www.mdpi.com/2504-2289/10/4/113</link>
	<description>Smart-meter theft detection requires learning from fine-grained electricity consumption data, whose centralized processing poses significant privacy risks. Federated learning (FL) mitigates these risks by decentralizing training, but providing rigorous user-level differential privacy (DP) under non-IID data and heterogeneous client behavior remains challenging. Existing DP-FL approaches rely on fixed global clipping bounds for client updates, which substantially overestimate sensitivity when privacy loss is composed using R&amp;amp;eacute;nyi Differential Privacy (RDP), zero-Concentrated DP (zCDP), or Moments Accountant (MA) frameworks, leading to excessive noise and degraded utility. This work proposes an adaptive clipping-based RDP accountant that incorporates empirical, round-wise update magnitudes into privacy accounting by rescaling each round&amp;amp;rsquo;s RDP contribution according to the observed clipping ratio. The method is optimizer-agnostic and is evaluated with FedAvg, FedProx, and SCAFFOLD on the SGCC smart-meter theft dataset under IID and Dirichlet non-IID partitions. Experimental results show consistently tighter privacy bounds and improved model utility compared to classical DP accountants, demonstrating the effectiveness of sensitivity-aware privacy accounting for practical differentially private FL.</description>
	<pubDate>2026-04-08</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 113: Adaptive Sensitivity-Aware Differential Privacy Accounting for Federated Smart-Meter Theft Detection</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/4/113">doi: 10.3390/bdcc10040113</a></p>
	<p>Authors:
		Diego Labate
		Dipanwita Thakur
		Giancarlo Fortino
		</p>
	<p>Smart-meter theft detection requires learning from fine-grained electricity consumption data, whose centralized processing poses significant privacy risks. Federated learning (FL) mitigates these risks by decentralizing training, but providing rigorous user-level differential privacy (DP) under non-IID data and heterogeneous client behavior remains challenging. Existing DP-FL approaches rely on fixed global clipping bounds for client updates, which substantially overestimate sensitivity when privacy loss is composed using R&amp;amp;eacute;nyi Differential Privacy (RDP), zero-Concentrated DP (zCDP), or Moments Accountant (MA) frameworks, leading to excessive noise and degraded utility. This work proposes an adaptive clipping-based RDP accountant that incorporates empirical, round-wise update magnitudes into privacy accounting by rescaling each round&amp;amp;rsquo;s RDP contribution according to the observed clipping ratio. The method is optimizer-agnostic and is evaluated with FedAvg, FedProx, and SCAFFOLD on the SGCC smart-meter theft dataset under IID and Dirichlet non-IID partitions. Experimental results show consistently tighter privacy bounds and improved model utility compared to classical DP accountants, demonstrating the effectiveness of sensitivity-aware privacy accounting for practical differentially private FL.</p>
	]]></content:encoded>

	<dc:title>Adaptive Sensitivity-Aware Differential Privacy Accounting for Federated Smart-Meter Theft Detection</dc:title>
			<dc:creator>Diego Labate</dc:creator>
			<dc:creator>Dipanwita Thakur</dc:creator>
			<dc:creator>Giancarlo Fortino</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10040113</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-04-08</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-04-08</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>4</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>113</prism:startingPage>
		<prism:doi>10.3390/bdcc10040113</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/4/113</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/4/112">

	<title>BDCC, Vol. 10, Pages 112: Interpretable Optimized Extreme Gradient Boosting for Prediction of Higher Heating Value from Elemental Composition of Coal Resource to Energy Conversion</title>
	<link>https://www.mdpi.com/2504-2289/10/4/112</link>
	<description>The higher heating value (HHV), sometimes referred to as the gross calorific value, is a crucial metric for determining a fuel&amp;amp;rsquo;s primary energy potential in energy production systems. By combining extreme gradient boosting (XGBoost) with the differential evolution (DE) optimizer, an innovative machine learning-based model was created in this study to forecast the HHV (dependent variable). As input variables, the model included the constituents of the coal&amp;amp;rsquo;s ultimate analysis: carbon (C), oxygen (O), hydrogen (H), nitrogen (N), and sulfur (S). For comparative purposes, random forest regression (RFR), M5 model tree, multivariate linear regression (MLR), and previously reported empirical correlations were also applied to the experimental dataset. The results showed that the XGBoost strategy produced the most accurate predictions. An initial XGBoost analysis was carried out to identify the relative contribution of the input variables to coal HHV prediction. In particular, for coal HHV estimates reliant on experimental samples, the XGBoost regression produced a correlation coefficient of 0.9858 and a coefficient of determination of 0.9691. The excellent agreement between observed and anticipated values shows that the DE/XGBoost-based approximation performed satisfactorily. Lastly, a synopsis of the investigation&amp;amp;rsquo;s key conclusions is provided.</description>
	<pubDate>2026-04-07</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 112: Interpretable Optimized Extreme Gradient Boosting for Prediction of Higher Heating Value from Elemental Composition of Coal Resource to Energy Conversion</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/4/112">doi: 10.3390/bdcc10040112</a></p>
	<p>Authors:
		Paulino José García-Nieto
		Esperanza García-Gonzalo
		José Pablo Paredes-Sánchez
		Luis Alfonso Menéndez-García
		</p>
	<p>The higher heating value (HHV), sometimes referred to as the gross calorific value, is a crucial metric for determining a fuel&amp;amp;rsquo;s primary energy potential in energy production systems. By combining extreme gradient boosting (XGBoost) with the differential evolution (DE) optimizer, an innovative machine learning-based model was created in this study to forecast the HHV (dependent variable). As input variables, the model included the constituents of the coal&amp;amp;rsquo;s ultimate analysis: carbon (C), oxygen (O), hydrogen (H), nitrogen (N), and sulfur (S). For comparative purposes, random forest regression (RFR), M5 model tree, multivariate linear regression (MLR), and previously reported empirical correlations were also applied to the experimental dataset. The results showed that the XGBoost strategy produced the most accurate predictions. An initial XGBoost analysis was carried out to identify the relative contribution of the input variables to coal HHV prediction. In particular, for coal HHV estimates reliant on experimental samples, the XGBoost regression produced a correlation coefficient of 0.9858 and a coefficient of determination of 0.9691. The excellent agreement between observed and anticipated values shows that the DE/XGBoost-based approximation performed satisfactorily. Lastly, a synopsis of the investigation&amp;amp;rsquo;s key conclusions is provided.</p>
	]]></content:encoded>

	<dc:title>Interpretable Optimized Extreme Gradient Boosting for Prediction of Higher Heating Value from Elemental Composition of Coal Resource to Energy Conversion</dc:title>
			<dc:creator>Paulino José García-Nieto</dc:creator>
			<dc:creator>Esperanza García-Gonzalo</dc:creator>
			<dc:creator>José Pablo Paredes-Sánchez</dc:creator>
			<dc:creator>Luis Alfonso Menéndez-García</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10040112</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-04-07</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-04-07</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>4</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>112</prism:startingPage>
		<prism:doi>10.3390/bdcc10040112</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/4/112</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/4/111">

	<title>BDCC, Vol. 10, Pages 111: A Comparative Study of Federated Learning and Amino Acid Encoding with IoT Malware Detection as a Case Study</title>
	<link>https://www.mdpi.com/2504-2289/10/4/111</link>
	<description>The increasing deployment of Internet of Things (IoT) devices introduces significant security challenges, while privacy concerns limit centralized data aggregation for intrusion detection. Federated learning (FL) offers a decentralized alternative, yet the interaction between feature representation, model architecture, and data heterogeneity remains insufficiently understood in IoT malware detection. This study provides a controlled comparative analysis of centralized and federated learning, optionally using amino acid encoding, under IID and Non-IID conditions using a 10,000-sample subset of the CTU&amp;amp;ndash;IoT&amp;amp;ndash;Malware&amp;amp;ndash;Capture dataset. First, we evaluate raw tabular features versus amino acid-based feature encoding, followed by a lightweight multi-layer perceptron (2882 parameters) versus a deeper residual network (70,532 parameters), across binary and multi-class classification tasks. In the binary setting, centralized training achieved up to 98.6% accuracy, while federated IID training reached 98.6%, with differences within statistical variance. Under Non-IID conditions, performance decreased modestly (0.1&amp;amp;ndash;0.5 percentage points), and accuracy was consistently lower when using encoded features compared with raw features. The degradation is smaller in deeper architectures and may offer improved stability under highly skewed federated conditions. In the four-class setting, the complex network achieved up to 97.8% accuracy with raw features, while amino acid encoding achieves up to 93.3%. The results show that federated learning can achieve performance comparable to centralized training under moderate heterogeneity, that lightweight architectures are sufficient for low-dimensional IoT traffic features, and that feature compression via amino acid encoding does not inherently mitigate Non-IID effects. These findings clarify the relative impact of representation, heterogeneity, and architectural capacity in practical FL-based IoT intrusion detection systems.</description>
	<pubDate>2026-04-06</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 111: A Comparative Study of Federated Learning and Amino Acid Encoding with IoT Malware Detection as a Case Study</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/4/111">doi: 10.3390/bdcc10040111</a></p>
	<p>Authors:
		Thaer AL Ibaisi
		Stefan Kuhn
		Muhammad Kazim
		Ismail Kara
		Turgay Altindag
		Mujeeb Ur Rehman
		</p>
	<p>The increasing deployment of Internet of Things (IoT) devices introduces significant security challenges, while privacy concerns limit centralized data aggregation for intrusion detection. Federated learning (FL) offers a decentralized alternative, yet the interaction between feature representation, model architecture, and data heterogeneity remains insufficiently understood in IoT malware detection. This study provides a controlled comparative analysis of centralized and federated learning, optionally using amino acid encoding, under IID and Non-IID conditions using a 10,000-sample subset of the CTU&amp;amp;ndash;IoT&amp;amp;ndash;Malware&amp;amp;ndash;Capture dataset. First, we evaluate raw tabular features versus amino acid-based feature encoding, followed by a lightweight multi-layer perceptron (2882 parameters) versus a deeper residual network (70,532 parameters), across binary and multi-class classification tasks. In the binary setting, centralized training achieved up to 98.6% accuracy, while federated IID training reached 98.6%, with differences within statistical variance. Under Non-IID conditions, performance decreased modestly (0.1&amp;amp;ndash;0.5 percentage points), and accuracy was consistently lower when using encoded features compared with raw features. The degradation is smaller in deeper architectures and may offer improved stability under highly skewed federated conditions. In the four-class setting, the complex network achieved up to 97.8% accuracy with raw features, while amino acid encoding achieves up to 93.3%. The results show that federated learning can achieve performance comparable to centralized training under moderate heterogeneity, that lightweight architectures are sufficient for low-dimensional IoT traffic features, and that feature compression via amino acid encoding does not inherently mitigate Non-IID effects. These findings clarify the relative impact of representation, heterogeneity, and architectural capacity in practical FL-based IoT intrusion detection systems.</p>
	]]></content:encoded>

	<dc:title>A Comparative Study of Federated Learning and Amino Acid Encoding with IoT Malware Detection as a Case Study</dc:title>
			<dc:creator>Thaer AL Ibaisi</dc:creator>
			<dc:creator>Stefan Kuhn</dc:creator>
			<dc:creator>Muhammad Kazim</dc:creator>
			<dc:creator>Ismail Kara</dc:creator>
			<dc:creator>Turgay Altindag</dc:creator>
			<dc:creator>Mujeeb Ur Rehman</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10040111</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-04-06</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-04-06</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>4</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>111</prism:startingPage>
		<prism:doi>10.3390/bdcc10040111</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/4/111</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/4/110">

	<title>BDCC, Vol. 10, Pages 110: LLMs for Integrated Business Intelligence: A Big Data-Driven Framework Integrating Marketing Optimization, Financial Performance, and Audit Quality</title>
	<link>https://www.mdpi.com/2504-2289/10/4/110</link>
	<description>Enterprise decision making in marketing, finance, and audit remains fragmented, leading to inefficient budget allocation and incomplete risk assessment. This study proposes an integrated, Big Data-driven decision-support framework that unifies Large Language Models (LLMs), attention-based marketing mix modeling, and multi-agent, game-theoretic optimization to coordinate cross-functional decisions. The architecture combines five modules: LLM-enhanced customer segmentation and customer lifetime value prediction, attention-weighted marketing mix modeling, multi-agent LLM systems for hierarchical budget optimization, attention-informed Markov multi-touch attribution, and LLM-augmented audit quality assessment. Empirical validation on a large-scale e-commerce dataset with 2.8 million customers and USD 156 million in marketing expenditure shows that marketing return on investment increases from 4.2 to 6.78 (61.4% relative improvement), financial forecasting error (MAPE) decreases from 12.8% to 4.7% (63.3% reduction), fraud detection accuracy improves by 29.8%, the Audit Quality Index reaches 0.951, and customer lifetime value prediction accuracy improves from 76.4% to 91.3%. By operationalizing the convergence of LLMs, attention mechanisms, and game-theoretic reasoning within a unified and empirically validated framework, the study delivers both theoretical advances and practically deployable tools for integrated business intelligence in digital economies.</description>
	<pubDate>2026-04-05</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 110: LLMs for Integrated Business Intelligence: A Big Data-Driven Framework Integrating Marketing Optimization, Financial Performance, and Audit Quality</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/4/110">doi: 10.3390/bdcc10040110</a></p>
	<p>Authors:
		Leonidas Theodorakopoulos
		Aristeidis Karras
		Alexandra Theodoropoulou
		Christos Klavdianos
		</p>
	<p>Enterprise decision making in marketing, finance, and audit remains fragmented, leading to inefficient budget allocation and incomplete risk assessment. This study proposes an integrated, Big Data-driven decision-support framework that unifies Large Language Models (LLMs), attention-based marketing mix modeling, and multi-agent, game-theoretic optimization to coordinate cross-functional decisions. The architecture combines five modules: LLM-enhanced customer segmentation and customer lifetime value prediction, attention-weighted marketing mix modeling, multi-agent LLM systems for hierarchical budget optimization, attention-informed Markov multi-touch attribution, and LLM-augmented audit quality assessment. Empirical validation on a large-scale e-commerce dataset with 2.8 million customers and USD 156 million in marketing expenditure shows that marketing return on investment increases from 4.2 to 6.78 (61.4% relative improvement), financial forecasting error (MAPE) decreases from 12.8% to 4.7% (63.3% reduction), fraud detection accuracy improves by 29.8%, the Audit Quality Index reaches 0.951, and customer lifetime value prediction accuracy improves from 76.4% to 91.3%. By operationalizing the convergence of LLMs, attention mechanisms, and game-theoretic reasoning within a unified and empirically validated framework, the study delivers both theoretical advances and practically deployable tools for integrated business intelligence in digital economies.</p>
	]]></content:encoded>

	<dc:title>LLMs for Integrated Business Intelligence: A Big Data-Driven Framework Integrating Marketing Optimization, Financial Performance, and Audit Quality</dc:title>
			<dc:creator>Leonidas Theodorakopoulos</dc:creator>
			<dc:creator>Aristeidis Karras</dc:creator>
			<dc:creator>Alexandra Theodoropoulou</dc:creator>
			<dc:creator>Christos Klavdianos</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10040110</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-04-05</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-04-05</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>4</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>110</prism:startingPage>
		<prism:doi>10.3390/bdcc10040110</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/4/110</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/4/109">

	<title>BDCC, Vol. 10, Pages 109: Stock Market Forecasting in Taiwan: A Radius Neighbors Regressor Approach</title>
	<link>https://www.mdpi.com/2504-2289/10/4/109</link>
	<description>This study proposes a machine learning framework tailored to the institutional characteristics of Taiwan&amp;amp;rsquo;s stock market, aiming to enhance forecasting accuracy for the Taiwan Stock Exchange Capitalization Weighted Stock Index (TAIEX). The model employs the Radius Neighbors Regressor with a dynamic radius-based similarity measure and integrates domain-specific features including technical indicators, volume&amp;amp;ndash;price relationships, and Qualified Foreign Institutional Investor (QFII) activity. A custom 60-day input window with a 20-day forecast horizon is applied to capture medium-term market dynamics. The framework was evaluated through extensive backtesting and real-world validation with the TAIEX Futures. The results demonstrate that the model achieves a peak directional accuracy of 85.1% under optimal parameter settings. Moreover, trading simulations confirm its practical viability, yielding a cumulative return on investment (ROI) of approximately 1600% during the short-term evaluation period (2023&amp;amp;ndash;2025) and nearly 2000% in the long-term evaluation (2019&amp;amp;ndash;2025), even after accounting for transaction costs and stop-loss mechanisms. These findings indicate that combining historical pattern similarity with institutional investor behavior substantially improves predictive power and profitability. Nevertheless, the framework remains constrained by its reliance on Taiwan-specific institutional features and historical trading data, limiting generalizability. Future research should extend applications to other markets while incorporating macroeconomic variables, corporate fundamentals, and news-driven signals to enhance adaptability.</description>
	<pubDate>2026-04-04</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 109: Stock Market Forecasting in Taiwan: A Radius Neighbors Regressor Approach</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/4/109">doi: 10.3390/bdcc10040109</a></p>
	<p>Authors:
		Yu-Kai Huang
		Chih-Hung Chen
		Yun-Cheng Tsai
		Shun-Shii Lin
		</p>
	<p>This study proposes a machine learning framework tailored to the institutional characteristics of Taiwan&amp;amp;rsquo;s stock market, aiming to enhance forecasting accuracy for the Taiwan Stock Exchange Capitalization Weighted Stock Index (TAIEX). The model employs the Radius Neighbors Regressor with a dynamic radius-based similarity measure and integrates domain-specific features including technical indicators, volume&amp;amp;ndash;price relationships, and Qualified Foreign Institutional Investor (QFII) activity. A custom 60-day input window with a 20-day forecast horizon is applied to capture medium-term market dynamics. The framework was evaluated through extensive backtesting and real-world validation with the TAIEX Futures. The results demonstrate that the model achieves a peak directional accuracy of 85.1% under optimal parameter settings. Moreover, trading simulations confirm its practical viability, yielding a cumulative return on investment (ROI) of approximately 1600% during the short-term evaluation period (2023&amp;amp;ndash;2025) and nearly 2000% in the long-term evaluation (2019&amp;amp;ndash;2025), even after accounting for transaction costs and stop-loss mechanisms. These findings indicate that combining historical pattern similarity with institutional investor behavior substantially improves predictive power and profitability. Nevertheless, the framework remains constrained by its reliance on Taiwan-specific institutional features and historical trading data, limiting generalizability. Future research should extend applications to other markets while incorporating macroeconomic variables, corporate fundamentals, and news-driven signals to enhance adaptability.</p>
	]]></content:encoded>

	<dc:title>Stock Market Forecasting in Taiwan: A Radius Neighbors Regressor Approach</dc:title>
			<dc:creator>Yu-Kai Huang</dc:creator>
			<dc:creator>Chih-Hung Chen</dc:creator>
			<dc:creator>Yun-Cheng Tsai</dc:creator>
			<dc:creator>Shun-Shii Lin</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10040109</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-04-04</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-04-04</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>4</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>109</prism:startingPage>
		<prism:doi>10.3390/bdcc10040109</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/4/109</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/4/108">

	<title>BDCC, Vol. 10, Pages 108: Exploring the Mechanisms Influencing Graduate Students&amp;rsquo; Adoption of Generative AI: Insights from the Technology Acceptance Model</title>
	<link>https://www.mdpi.com/2504-2289/10/4/108</link>
	<description>The rapid development of Generative Artificial Intelligence (GenAI) in graduate education has changed human&amp;amp;ndash;AI interaction within knowledge-intensive environments, leading to important questions about user-side cognitive adaptation in probabilistic AI systems. While many studies focus on ethical implications, limited attention has been paid to the cognitive mechanisms underlying graduate students&amp;amp;rsquo; adoption of GenAI. Drawing on the Technology Acceptance Model (TAM), this study explores the cognitive and interactional mechanisms shaping graduate students&amp;amp;rsquo; adoption and usage of GenAI. Using thematic analysis of in-depth interviews with 20 graduate students from diverse academic backgrounds, the study identifies seven interrelated constructs: perceived usefulness, perceived ease of use, external environment, risk perception, attitude, behavioral intention, and interaction subjectivity. This study demonstrates that the adoption of GenAI is not merely a result of perceived efficiency but is shaped by cognitive calibration between trust and risk evaluation. Moreover, interaction subjectivity emerges as a metacognitive factor that determines whether engagement results in human&amp;amp;ndash;AI collaboration or passive automation. By integrating external environment, risk perception, and interaction subjectivity, this study provides a cognitively grounded framework for understanding human&amp;amp;ndash;AI adoption and interaction dynamics. Practically, the findings provide design-relevant insights for developing GenAI systems that support calibrated trust, uncertainty awareness, and adaptive cognitive participation.</description>
	<pubDate>2026-04-03</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 108: Exploring the Mechanisms Influencing Graduate Students&amp;rsquo; Adoption of Generative AI: Insights from the Technology Acceptance Model</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/4/108">doi: 10.3390/bdcc10040108</a></p>
	<p>Authors:
		Qing Chen
		Yujie Xue
		Jie Lin
		Chang Zhu
		</p>
	<p>The rapid development of Generative Artificial Intelligence (GenAI) in graduate education has changed human&amp;amp;ndash;AI interaction within knowledge-intensive environments, leading to important questions about user-side cognitive adaptation in probabilistic AI systems. While many studies focus on ethical implications, limited attention has been paid to the cognitive mechanisms underlying graduate students&amp;amp;rsquo; adoption of GenAI. Drawing on the Technology Acceptance Model (TAM), this study explores the cognitive and interactional mechanisms shaping graduate students&amp;amp;rsquo; adoption and usage of GenAI. Using thematic analysis of in-depth interviews with 20 graduate students from diverse academic backgrounds, the study identifies seven interrelated constructs: perceived usefulness, perceived ease of use, external environment, risk perception, attitude, behavioral intention, and interaction subjectivity. This study demonstrates that the adoption of GenAI is not merely a result of perceived efficiency but is shaped by cognitive calibration between trust and risk evaluation. Moreover, interaction subjectivity emerges as a metacognitive factor that determines whether engagement results in human&amp;amp;ndash;AI collaboration or passive automation. By integrating external environment, risk perception, and interaction subjectivity, this study provides a cognitively grounded framework for understanding human&amp;amp;ndash;AI adoption and interaction dynamics. Practically, the findings provide design-relevant insights for developing GenAI systems that support calibrated trust, uncertainty awareness, and adaptive cognitive participation.</p>
	]]></content:encoded>

	<dc:title>Exploring the Mechanisms Influencing Graduate Students&amp;amp;rsquo; Adoption of Generative AI: Insights from the Technology Acceptance Model</dc:title>
			<dc:creator>Qing Chen</dc:creator>
			<dc:creator>Yujie Xue</dc:creator>
			<dc:creator>Jie Lin</dc:creator>
			<dc:creator>Chang Zhu</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10040108</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-04-03</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-04-03</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>4</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>108</prism:startingPage>
		<prism:doi>10.3390/bdcc10040108</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/4/108</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/4/107">

	<title>BDCC, Vol. 10, Pages 107: Multi-Scale Optimal Transport Transformer for Efficient Exemplar-Based Image Translation</title>
	<link>https://www.mdpi.com/2504-2289/10/4/107</link>
	<description>Exemplar-based image translation generates an output image by transferring appearance from a reference exemplar to a content image. Existing works only consider the local correspondences between two modalities, and ignore the global distributions in each modality, struggling to obtain fine-grained details with efficient computation. In this paper, we propose OTFormer, a multi-scale Optimal Transport transformer for exemplarbased image translation. We formulate cross-modal alignment as a multi-scale optimal transport problem, which progressively provides a globally coherent matching. In addition, we design a lightweight multi-scale fusion block to extract and fuse features efficiently. Experiments on CelebA-HQ and DeepFashion demonstrate that OTFormer improves both image fidelity and style adherence, while reducing model parameters by 62% and achieving faster inference compared with strong baselines. These results highlight OTguided global alignment as an effective and deployable solution for high-fidelity exemplarbased image translation.</description>
	<pubDate>2026-04-01</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 107: Multi-Scale Optimal Transport Transformer for Efficient Exemplar-Based Image Translation</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/4/107">doi: 10.3390/bdcc10040107</a></p>
	<p>Authors:
		Jinsong Zhang
		Xiongzheng Li
		Yuqin Lin
		</p>
	<p>Exemplar-based image translation generates an output image by transferring appearance from a reference exemplar to a content image. Existing works only consider the local correspondences between two modalities, and ignore the global distributions in each modality, struggling to obtain fine-grained details with efficient computation. In this paper, we propose OTFormer, a multi-scale Optimal Transport transformer for exemplarbased image translation. We formulate cross-modal alignment as a multi-scale optimal transport problem, which progressively provides a globally coherent matching. In addition, we design a lightweight multi-scale fusion block to extract and fuse features efficiently. Experiments on CelebA-HQ and DeepFashion demonstrate that OTFormer improves both image fidelity and style adherence, while reducing model parameters by 62% and achieving faster inference compared with strong baselines. These results highlight OTguided global alignment as an effective and deployable solution for high-fidelity exemplarbased image translation.</p>
	]]></content:encoded>

	<dc:title>Multi-Scale Optimal Transport Transformer for Efficient Exemplar-Based Image Translation</dc:title>
			<dc:creator>Jinsong Zhang</dc:creator>
			<dc:creator>Xiongzheng Li</dc:creator>
			<dc:creator>Yuqin Lin</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10040107</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-04-01</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-04-01</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>4</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>107</prism:startingPage>
		<prism:doi>10.3390/bdcc10040107</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/4/107</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/4/106">

	<title>BDCC, Vol. 10, Pages 106: Multi-Modal Method for Candidate Interview Assessment Based on Computer Vision and Large Language Models</title>
	<link>https://www.mdpi.com/2504-2289/10/4/106</link>
	<description>Candidate interview assessment is primarily reliant on subjective human judgment, while existing AI-based methods rely on end-to-end predictions with no psychometric basis. In this paper, we propose an interpretable multi-modal framework that combines nonverbal behavior, LLM-based verbal analysis, and Big Five personality traits into three theory-based constructs: professional-cognitive competence, observed leadership behavior, and leadership disposition. The proposed method utilizes computer vision and larger language models to extract features from video interviews. Rather than targeting predictive accuracy, the proposed method prioritizes construct validity and transparent aggregation under severe label scarcity. The proposed method aggregates the constructs into a Top Potential Score that reflects the executive abilities of the candidate. Experiments on the method show its ability to significantly differentiate top candidates from others (Cliff&amp;amp;rsquo;s delta = 0.91 for the composite Top Potential Score, permutation p = 0.0002). Leave-one-out analysis verifies robustness, while rank-based evaluation yields 100% recall of executive candidates in the top 20% of rated applications. The findings justify the use of the proposed multi-modal method as an interpretable decision-support tool for candidate interview assessment.</description>
	<pubDate>2026-04-01</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 106: Multi-Modal Method for Candidate Interview Assessment Based on Computer Vision and Large Language Models</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/4/106">doi: 10.3390/bdcc10040106</a></p>
	<p>Authors:
		Kenan Kassab
		Alexey Kashevnik
		Irina Shoshina
		</p>
	<p>Candidate interview assessment is primarily reliant on subjective human judgment, while existing AI-based methods rely on end-to-end predictions with no psychometric basis. In this paper, we propose an interpretable multi-modal framework that combines nonverbal behavior, LLM-based verbal analysis, and Big Five personality traits into three theory-based constructs: professional-cognitive competence, observed leadership behavior, and leadership disposition. The proposed method utilizes computer vision and larger language models to extract features from video interviews. Rather than targeting predictive accuracy, the proposed method prioritizes construct validity and transparent aggregation under severe label scarcity. The proposed method aggregates the constructs into a Top Potential Score that reflects the executive abilities of the candidate. Experiments on the method show its ability to significantly differentiate top candidates from others (Cliff&amp;amp;rsquo;s delta = 0.91 for the composite Top Potential Score, permutation p = 0.0002). Leave-one-out analysis verifies robustness, while rank-based evaluation yields 100% recall of executive candidates in the top 20% of rated applications. The findings justify the use of the proposed multi-modal method as an interpretable decision-support tool for candidate interview assessment.</p>
	]]></content:encoded>

	<dc:title>Multi-Modal Method for Candidate Interview Assessment Based on Computer Vision and Large Language Models</dc:title>
			<dc:creator>Kenan Kassab</dc:creator>
			<dc:creator>Alexey Kashevnik</dc:creator>
			<dc:creator>Irina Shoshina</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10040106</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-04-01</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-04-01</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>4</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>106</prism:startingPage>
		<prism:doi>10.3390/bdcc10040106</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/4/106</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/4/105">

	<title>BDCC, Vol. 10, Pages 105: H2Avatar: Expressive Whole-Body Avatars from Monocular Video via Hierarchical Geometry and Hybrid Rendering</title>
	<link>https://www.mdpi.com/2504-2289/10/4/105</link>
	<description>Reconstructing photorealistic and animatable whole-body avatars from monocular videos is a hot topic in computer vision and computer graphics. However, existing methods still face challenges due to the limited frequency response of single-scale geometry encodings and the instability of appearance modeling without an explicit surface anchor. In this paper, we present H2Avatar, a real-time framework that builds on a mesh-embedded 3D Gaussian representation guided by SMPL-X and disentangles geometry and appearance into hierarchical and hybrid components. For geometry, we propose a semantic-aware hierarchical encoding based on a multi-scale tri-plane pyramid, where features at different resolutions capture both global structure and high-frequency surface details such as clothing wrinkles. For appearance, we introduce a hybrid rendering strategy that anchors canonical colors using a learnable UV texture map, and complements it with a neural residual color branch conditioned on tri-plane features, pose embedding, and surface normals to model pose- and view-dependent shading variations. This design improves temporal stability and preserves identity details while enhancing photorealism under complex motions. Experiments on the NeuMan dataset demonstrate that H2Avatar consistently outperforms representative baselines across multiple sequences, outperforming ExAvatar by up to 0.66 dB in PSNR and reducing LPIPS by up to 16.3%. These results validate the effectiveness of hierarchical geometry encoding and texture-anchored hybrid appearance modeling.</description>
	<pubDate>2026-04-01</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 105: H2Avatar: Expressive Whole-Body Avatars from Monocular Video via Hierarchical Geometry and Hybrid Rendering</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/4/105">doi: 10.3390/bdcc10040105</a></p>
	<p>Authors:
		Jinsong Zhang
		Cheng Guan
		Zhihua Lin
		Yuqin Lin
		</p>
	<p>Reconstructing photorealistic and animatable whole-body avatars from monocular videos is a hot topic in computer vision and computer graphics. However, existing methods still face challenges due to the limited frequency response of single-scale geometry encodings and the instability of appearance modeling without an explicit surface anchor. In this paper, we present H2Avatar, a real-time framework that builds on a mesh-embedded 3D Gaussian representation guided by SMPL-X and disentangles geometry and appearance into hierarchical and hybrid components. For geometry, we propose a semantic-aware hierarchical encoding based on a multi-scale tri-plane pyramid, where features at different resolutions capture both global structure and high-frequency surface details such as clothing wrinkles. For appearance, we introduce a hybrid rendering strategy that anchors canonical colors using a learnable UV texture map, and complements it with a neural residual color branch conditioned on tri-plane features, pose embedding, and surface normals to model pose- and view-dependent shading variations. This design improves temporal stability and preserves identity details while enhancing photorealism under complex motions. Experiments on the NeuMan dataset demonstrate that H2Avatar consistently outperforms representative baselines across multiple sequences, outperforming ExAvatar by up to 0.66 dB in PSNR and reducing LPIPS by up to 16.3%. These results validate the effectiveness of hierarchical geometry encoding and texture-anchored hybrid appearance modeling.</p>
	]]></content:encoded>

	<dc:title>H2Avatar: Expressive Whole-Body Avatars from Monocular Video via Hierarchical Geometry and Hybrid Rendering</dc:title>
			<dc:creator>Jinsong Zhang</dc:creator>
			<dc:creator>Cheng Guan</dc:creator>
			<dc:creator>Zhihua Lin</dc:creator>
			<dc:creator>Yuqin Lin</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10040105</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-04-01</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-04-01</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>4</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>105</prism:startingPage>
		<prism:doi>10.3390/bdcc10040105</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/4/105</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/4/104">

	<title>BDCC, Vol. 10, Pages 104: Enhanced Schema Linking with Large Language Models via Self-Verification and Value Hints</title>
	<link>https://www.mdpi.com/2504-2289/10/4/104</link>
	<description>Schema linking, the task of identifying relevant database schema elements (tables and columns) for natural language queries, is a critical component in database-driven natural language interfaces. While existing approaches rely on question decomposition to handle complex queries, they often suffer from error propagation and low precision. In this paper, we propose a novel schema linking framework enhanced by self-verification (SV) and value hints (VHs) that significantly improves both precision and recall. Our approach introduces two key components: (1) self-verification (SV), an iterative refinement mechanism that validates and corrects initial predictions through explicit verification prompts, and (2) value hints (VHs), which explicitly guide the model to recognize database values mentioned in queries. We conduct comprehensive experiments on two benchmark datasets, Spider and BIRD, using two language models of 4B and 80B parameters. Our results demonstrate that SV + VH consistently improves performance across datasets, models, and method configurations, outperforming both decomposition-based approaches and compute-matched alternatives such as self-consistency under equivalent inference budgets.</description>
	<pubDate>2026-03-31</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 104: Enhanced Schema Linking with Large Language Models via Self-Verification and Value Hints</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/4/104">doi: 10.3390/bdcc10040104</a></p>
	<p>Authors:
		Linfei Ma
		Dexing Wei
		Xiangpeng Li
		Feng Wen
		Haisu Zhang
		</p>
	<p>Schema linking, the task of identifying relevant database schema elements (tables and columns) for natural language queries, is a critical component in database-driven natural language interfaces. While existing approaches rely on question decomposition to handle complex queries, they often suffer from error propagation and low precision. In this paper, we propose a novel schema linking framework enhanced by self-verification (SV) and value hints (VHs) that significantly improves both precision and recall. Our approach introduces two key components: (1) self-verification (SV), an iterative refinement mechanism that validates and corrects initial predictions through explicit verification prompts, and (2) value hints (VHs), which explicitly guide the model to recognize database values mentioned in queries. We conduct comprehensive experiments on two benchmark datasets, Spider and BIRD, using two language models of 4B and 80B parameters. Our results demonstrate that SV + VH consistently improves performance across datasets, models, and method configurations, outperforming both decomposition-based approaches and compute-matched alternatives such as self-consistency under equivalent inference budgets.</p>
	]]></content:encoded>

	<dc:title>Enhanced Schema Linking with Large Language Models via Self-Verification and Value Hints</dc:title>
			<dc:creator>Linfei Ma</dc:creator>
			<dc:creator>Dexing Wei</dc:creator>
			<dc:creator>Xiangpeng Li</dc:creator>
			<dc:creator>Feng Wen</dc:creator>
			<dc:creator>Haisu Zhang</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10040104</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-03-31</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-03-31</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>4</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>104</prism:startingPage>
		<prism:doi>10.3390/bdcc10040104</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/4/104</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/4/103">

	<title>BDCC, Vol. 10, Pages 103: Semantic Agent-Based Intelligent Digital Twins Integrating Demand, Production and Product Through Asset Administration Shells</title>
	<link>https://www.mdpi.com/2504-2289/10/4/103</link>
	<description>Complex products and production processes are intertwined and demand expressive, lifecycle-wide digital representations. The Asset Administration Shell emerged as a standard for Digital Twins (DTs), structuring heterogeneous data across cloud-based Industrial Internet of Things (IIoT) infrastructures. However, today&amp;amp;rsquo;s deployments predominantly realize passive or reactive DTs, while intelligent behavior remains underexploited. This paper addresses this gap, proposing an end-to-end architecture operationalizing the DT Reference Model through the integration of machine-interpretable granulated industrial skills, which are semantically accumulated into a knowledge graph enabling discovery and reasoning, while a multi-agent system provides autonomous, utility-based negotiation via machine-to-machine interactions within a federated marketplace. The approach is applied in a real smart manufacturing demonstrator, combining order processes, production orchestration, and lifecycle documentation into a unified execution pipeline spanning IIoT-connected shopfloor assets and cloud-based services. Quantitative experiments evaluating negotiation latency, renegotiation robustness, and utility variation demonstrate stable, predictable behavior even under concurrent demand and failure scenarios. The architecture lays a foundation for interoperable, sovereign collaboration across value chains to realize shared production. The results underline the effectiveness of the tightly coupled enabler technologies realizing proactive, reconfigurable, and semantically enriched intelligent DTs.</description>
	<pubDate>2026-03-26</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 103: Semantic Agent-Based Intelligent Digital Twins Integrating Demand, Production and Product Through Asset Administration Shells</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/4/103">doi: 10.3390/bdcc10040103</a></p>
	<p>Authors:
		Joel Lehmann
		Tim Markus Häußermann
		Julian Reichwald
		</p>
	<p>Complex products and production processes are intertwined and demand expressive, lifecycle-wide digital representations. The Asset Administration Shell emerged as a standard for Digital Twins (DTs), structuring heterogeneous data across cloud-based Industrial Internet of Things (IIoT) infrastructures. However, today&amp;amp;rsquo;s deployments predominantly realize passive or reactive DTs, while intelligent behavior remains underexploited. This paper addresses this gap, proposing an end-to-end architecture operationalizing the DT Reference Model through the integration of machine-interpretable granulated industrial skills, which are semantically accumulated into a knowledge graph enabling discovery and reasoning, while a multi-agent system provides autonomous, utility-based negotiation via machine-to-machine interactions within a federated marketplace. The approach is applied in a real smart manufacturing demonstrator, combining order processes, production orchestration, and lifecycle documentation into a unified execution pipeline spanning IIoT-connected shopfloor assets and cloud-based services. Quantitative experiments evaluating negotiation latency, renegotiation robustness, and utility variation demonstrate stable, predictable behavior even under concurrent demand and failure scenarios. The architecture lays a foundation for interoperable, sovereign collaboration across value chains to realize shared production. The results underline the effectiveness of the tightly coupled enabler technologies realizing proactive, reconfigurable, and semantically enriched intelligent DTs.</p>
	]]></content:encoded>

	<dc:title>Semantic Agent-Based Intelligent Digital Twins Integrating Demand, Production and Product Through Asset Administration Shells</dc:title>
			<dc:creator>Joel Lehmann</dc:creator>
			<dc:creator>Tim Markus Häußermann</dc:creator>
			<dc:creator>Julian Reichwald</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10040103</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-03-26</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-03-26</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>4</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>103</prism:startingPage>
		<prism:doi>10.3390/bdcc10040103</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/4/103</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/4/102">

	<title>BDCC, Vol. 10, Pages 102: Emotional Framing in Prompts Modulates Large Language Model Performance</title>
	<link>https://www.mdpi.com/2504-2289/10/4/102</link>
	<description>Large Language Models (LLMs) demonstrate remarkable performance across a variety of natural language understanding tasks, yet their sensitivity to emotional framing in user prompts remains underexplored. This paper presents an empirical study investigating how four emotional tones&amp;amp;mdash;joy, apathy, anger, and fear&amp;amp;mdash;affect LLM performance on the SuperGLUE benchmark. We evaluate five instruction-tuned, open-weight models across eight diverse tasks, systematically modulating input prompts with affective cues while keeping semantic content constant. Results reveal that prompts framed with joy and apathy lead to consistently higher accuracy, with gains of up to 4.5 percentage points compared to fear-framed inputs, which yield the lowest performance. These findings demonstrate that affective modulation in user prompts measurably impacts LLM reasoning and task outcomes, suggesting that emotional framing is not merely stylistic but functionally relevant to model behavior. Our study provides a reproducible experimental framework and an open-source prompt set, offering a foundation for future research on affect-aware prompting strategies and their implications in human&amp;amp;ndash;AI interaction.</description>
	<pubDate>2026-03-24</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 102: Emotional Framing in Prompts Modulates Large Language Model Performance</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/4/102">doi: 10.3390/bdcc10040102</a></p>
	<p>Authors:
		Manuel Gozzi
		Francesca Fallucchi
		</p>
	<p>Large Language Models (LLMs) demonstrate remarkable performance across a variety of natural language understanding tasks, yet their sensitivity to emotional framing in user prompts remains underexplored. This paper presents an empirical study investigating how four emotional tones&amp;amp;mdash;joy, apathy, anger, and fear&amp;amp;mdash;affect LLM performance on the SuperGLUE benchmark. We evaluate five instruction-tuned, open-weight models across eight diverse tasks, systematically modulating input prompts with affective cues while keeping semantic content constant. Results reveal that prompts framed with joy and apathy lead to consistently higher accuracy, with gains of up to 4.5 percentage points compared to fear-framed inputs, which yield the lowest performance. These findings demonstrate that affective modulation in user prompts measurably impacts LLM reasoning and task outcomes, suggesting that emotional framing is not merely stylistic but functionally relevant to model behavior. Our study provides a reproducible experimental framework and an open-source prompt set, offering a foundation for future research on affect-aware prompting strategies and their implications in human&amp;amp;ndash;AI interaction.</p>
	]]></content:encoded>

	<dc:title>Emotional Framing in Prompts Modulates Large Language Model Performance</dc:title>
			<dc:creator>Manuel Gozzi</dc:creator>
			<dc:creator>Francesca Fallucchi</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10040102</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-03-24</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-03-24</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>4</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>102</prism:startingPage>
		<prism:doi>10.3390/bdcc10040102</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/4/102</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/4/101">

	<title>BDCC, Vol. 10, Pages 101: Integrative Machine Learning Framework for Epigenetic Biomarker Discovery and Disease Severity Prediction in Childhood Atopic Dermatitis</title>
	<link>https://www.mdpi.com/2504-2289/10/4/101</link>
	<description>Atopic dermatitis (AD) is a chronic inflammatory skin disorder that is significantly contributed to by epigenetics. We developed a machine learning-based framework to identify DNA methylation biomarkers associated with AD classification and severity. Genome-wide methylation data from peripheral blood were processed using four feature selection algorithms: coarse approximation linear function (CALF), elastic net (EN), minimum redundancy maximum relevance (mRMR), and recursive feature elimination with cross-validation (RFECV). The integrative framework identified a central panel of 8 CpG sites that achieved an area under the curve (AUC) of 1.00 in the test set. This panel demonstrated high disease specificity, showing poor classification performance for systemic lupus erythematosus (AUC = 0.46), Crohn&amp;amp;rsquo;s disease (AUC = 0.50), and oral squamous cell carcinoma (AUC = 0.58). Severity prediction using RFECV-selected 63 CpG sites (RFE63) achieved high accuracy across classifiers, with Random Forest (accuracy = 0.94) outperforming the others. The functional enrichment of CpG-associated genes highlighted key immune-related transcriptional regulators, including STAT5A, RUNX1, MEIS1, and PAX4. These genes are linked to chromatin remodeling, T helper cell differentiation, and interleukin-2 regulation, which are critical in AD pathogenesis and severity. Our findings demonstrate the utility of machine learning-integrated epigenomics in identifying robust, disease-specific biomarkers for AD diagnosis and monitoring, offering new insights into the molecular mechanisms underlying childhood AD. However, further validation in large-scale independent cohorts is required to confirm their clinical robustness and generalizability.</description>
	<pubDate>2026-03-24</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 101: Integrative Machine Learning Framework for Epigenetic Biomarker Discovery and Disease Severity Prediction in Childhood Atopic Dermatitis</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/4/101">doi: 10.3390/bdcc10040101</a></p>
	<p>Authors:
		Ding-Wei Chen
		Yun-Nan Chang
		</p>
	<p>Atopic dermatitis (AD) is a chronic inflammatory skin disorder that is significantly contributed to by epigenetics. We developed a machine learning-based framework to identify DNA methylation biomarkers associated with AD classification and severity. Genome-wide methylation data from peripheral blood were processed using four feature selection algorithms: coarse approximation linear function (CALF), elastic net (EN), minimum redundancy maximum relevance (mRMR), and recursive feature elimination with cross-validation (RFECV). The integrative framework identified a central panel of 8 CpG sites that achieved an area under the curve (AUC) of 1.00 in the test set. This panel demonstrated high disease specificity, showing poor classification performance for systemic lupus erythematosus (AUC = 0.46), Crohn&amp;amp;rsquo;s disease (AUC = 0.50), and oral squamous cell carcinoma (AUC = 0.58). Severity prediction using RFECV-selected 63 CpG sites (RFE63) achieved high accuracy across classifiers, with Random Forest (accuracy = 0.94) outperforming the others. The functional enrichment of CpG-associated genes highlighted key immune-related transcriptional regulators, including STAT5A, RUNX1, MEIS1, and PAX4. These genes are linked to chromatin remodeling, T helper cell differentiation, and interleukin-2 regulation, which are critical in AD pathogenesis and severity. Our findings demonstrate the utility of machine learning-integrated epigenomics in identifying robust, disease-specific biomarkers for AD diagnosis and monitoring, offering new insights into the molecular mechanisms underlying childhood AD. However, further validation in large-scale independent cohorts is required to confirm their clinical robustness and generalizability.</p>
	]]></content:encoded>

	<dc:title>Integrative Machine Learning Framework for Epigenetic Biomarker Discovery and Disease Severity Prediction in Childhood Atopic Dermatitis</dc:title>
			<dc:creator>Ding-Wei Chen</dc:creator>
			<dc:creator>Yun-Nan Chang</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10040101</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-03-24</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-03-24</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>4</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>101</prism:startingPage>
		<prism:doi>10.3390/bdcc10040101</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/4/101</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/4/100">

	<title>BDCC, Vol. 10, Pages 100: An Experimental Study on Harassment Moderation in Llama and Alpaca</title>
	<link>https://www.mdpi.com/2504-2289/10/4/100</link>
	<description>The growing integration of chatbots and large language models (LLMs) into society raises important concerns about their potential to reproduce toxic human behaviors. As a result, it is essential to investigate these models to mitigate or eliminate such risks. This paper presents an experimental study evaluating the responses of the Llama and Alpaca models to scenarios involving verbal harassment. The methodology involved using harassment dialogues generated by an LLM as prompts to elicit responses from both models. The responses were then analyzed for levels of toxicity, sexually explicit content, and flirtatiousness. The results indicate that although both models reduce explicit offensive terms, they exhibit limitations in identifying and intercepting abusive behavior from users. Statistical analysis reveals that general-purpose instruction tuning in Alpaca does not provide a robust safety barrier compared to the Llama base model for most variables investigated in the experiment. However, a significant difference was observed concerning flirting, where Llama proved more prone to validation and encouragement than Alpaca. Furthermore, the study identifies critical vulnerabilities, such as a &amp;amp;ldquo;self-deprecation&amp;amp;rdquo; bias in Llama and &amp;amp;ldquo;mirroring&amp;amp;rdquo; behavior in Alpaca. We also report a complementary triangulation with GPT-family models as a secondary point of reference. This paper discusses and contains content that can be offensive or upsetting.</description>
	<pubDate>2026-03-24</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 100: An Experimental Study on Harassment Moderation in Llama and Alpaca</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/4/100">doi: 10.3390/bdcc10040100</a></p>
	<p>Authors:
		Henrique Tostes de Sousa
		Leo Natan Paschoal
		</p>
	<p>The growing integration of chatbots and large language models (LLMs) into society raises important concerns about their potential to reproduce toxic human behaviors. As a result, it is essential to investigate these models to mitigate or eliminate such risks. This paper presents an experimental study evaluating the responses of the Llama and Alpaca models to scenarios involving verbal harassment. The methodology involved using harassment dialogues generated by an LLM as prompts to elicit responses from both models. The responses were then analyzed for levels of toxicity, sexually explicit content, and flirtatiousness. The results indicate that although both models reduce explicit offensive terms, they exhibit limitations in identifying and intercepting abusive behavior from users. Statistical analysis reveals that general-purpose instruction tuning in Alpaca does not provide a robust safety barrier compared to the Llama base model for most variables investigated in the experiment. However, a significant difference was observed concerning flirting, where Llama proved more prone to validation and encouragement than Alpaca. Furthermore, the study identifies critical vulnerabilities, such as a &amp;amp;ldquo;self-deprecation&amp;amp;rdquo; bias in Llama and &amp;amp;ldquo;mirroring&amp;amp;rdquo; behavior in Alpaca. We also report a complementary triangulation with GPT-family models as a secondary point of reference. This paper discusses and contains content that can be offensive or upsetting.</p>
	]]></content:encoded>

	<dc:title>An Experimental Study on Harassment Moderation in Llama and Alpaca</dc:title>
			<dc:creator>Henrique Tostes de Sousa</dc:creator>
			<dc:creator>Leo Natan Paschoal</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10040100</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-03-24</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-03-24</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>4</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>100</prism:startingPage>
		<prism:doi>10.3390/bdcc10040100</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/4/100</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/3/99">

	<title>BDCC, Vol. 10, Pages 99: A Multi-Feature Transition-Aware Framework for Next POI Recommendation</title>
	<link>https://www.mdpi.com/2504-2289/10/3/99</link>
	<description>Next Point-of-Interest (POI) recommendation focuses on predicting a user&amp;amp;rsquo;s subsequent location based on historical check-in data. In practice, however, check-in logs frequently contain uncertain records in which ambiguous spatial, temporal, or behavioral information obscures the underlying mobility regularities, thereby degrading prediction performance. To address this challenge, this study first infers user preferences from historical trajectories and reweights transition importance based on temporal and spatial proximity. It then models transition relationships using three complementary feature dimensions: POI category, spatial area, and routine versus non-routine behavioral patterns. Using transition probability analysis, feature-level dependencies in user mobility are systematically investigated. The findings demonstrate that these transition features contribute unevenly to predictive performance, with area-based transitions yielding the strongest results when used in isolation. Nonetheless, their joint integration consistently achieves the highest accuracy, underscoring the critical role of transition-aware modeling. Across two real-world datasets, the proposed framework consistently achieves state-of-the-art performance in top-ranked accuracy (Recall@1) and ranking quality (NDCG@1), while delivering competitive effectiveness at higher cutoff values (k=3 and k=5). Notably, on the NYC dataset, MTF-POI achieves the highest Recall@1 (+19.01% over the strongest baseline) with a marginal trade-off at Recall@3, reflecting the framework&amp;amp;rsquo;s design emphasis on precise next-step prediction.</description>
	<pubDate>2026-03-23</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 99: A Multi-Feature Transition-Aware Framework for Next POI Recommendation</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/3/99">doi: 10.3390/bdcc10030099</a></p>
	<p>Authors:
		Oraya Sooknit
		Jakkarin Suksawatchon
		Ureerat Suksawatchon
		</p>
	<p>Next Point-of-Interest (POI) recommendation focuses on predicting a user&amp;amp;rsquo;s subsequent location based on historical check-in data. In practice, however, check-in logs frequently contain uncertain records in which ambiguous spatial, temporal, or behavioral information obscures the underlying mobility regularities, thereby degrading prediction performance. To address this challenge, this study first infers user preferences from historical trajectories and reweights transition importance based on temporal and spatial proximity. It then models transition relationships using three complementary feature dimensions: POI category, spatial area, and routine versus non-routine behavioral patterns. Using transition probability analysis, feature-level dependencies in user mobility are systematically investigated. The findings demonstrate that these transition features contribute unevenly to predictive performance, with area-based transitions yielding the strongest results when used in isolation. Nonetheless, their joint integration consistently achieves the highest accuracy, underscoring the critical role of transition-aware modeling. Across two real-world datasets, the proposed framework consistently achieves state-of-the-art performance in top-ranked accuracy (Recall@1) and ranking quality (NDCG@1), while delivering competitive effectiveness at higher cutoff values (k=3 and k=5). Notably, on the NYC dataset, MTF-POI achieves the highest Recall@1 (+19.01% over the strongest baseline) with a marginal trade-off at Recall@3, reflecting the framework&amp;amp;rsquo;s design emphasis on precise next-step prediction.</p>
	]]></content:encoded>

	<dc:title>A Multi-Feature Transition-Aware Framework for Next POI Recommendation</dc:title>
			<dc:creator>Oraya Sooknit</dc:creator>
			<dc:creator>Jakkarin Suksawatchon</dc:creator>
			<dc:creator>Ureerat Suksawatchon</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10030099</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-03-23</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-03-23</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>99</prism:startingPage>
		<prism:doi>10.3390/bdcc10030099</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/3/99</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/3/98">

	<title>BDCC, Vol. 10, Pages 98: A Comparative Analysis of Deep-Learning-Based Speech Enhancement Models: Assessing Biometric Speaker Verification in Real-World Noisy Environments</title>
	<link>https://www.mdpi.com/2504-2289/10/3/98</link>
	<description>Speech enhancement through denoising is essential for maintaining signal intelligibility and quality in biometric speaker verification pipelines that operate in acoustically adverse conditions. Despite the proliferation of deep learning (DL) architectures for speech denoising, simultaneously optimizing noise attenuation, perceptual fidelity, and speaker-identity preservation remains an open problem. We address this gap by benchmarking three architecturally distinct DL-based enhancement models&amp;amp;mdash;Wave-U-Net, CMGAN, and U-Net&amp;amp;mdash;on three independent, domain-diverse corpora (SpEAR, VPQAD, and Clarkson) that the models never encountered during training and by introducing commercial-grade VeriSpeak speaker-verification scores as a biometric evaluation dimension absent from prior comparative studies. Our experiments reveal a clear three-way trade-off: U-Net achieves the highest signal-to-noise ratio (SNR) gains (+61.44% on SpEAR, +67.05% on VPQAD, +235.3% on Clarkson) but sacrifices naturalness; CMGAN yields the best perceptual evaluation of speech quality (PESQ) values (3.33, 1.35, and 2.50, respectively), favoring listening-comfort applications; and Wave-U-Net delivers the strongest biometric fidelity (VeriSpeak improvements of +11.63%, +30.22%, and +29.24%) while offering competitive perceptual quality. These results highlight that model selection must be driven by the target deployment scenario and provide actionable guidance for improving biometric verification robustness under real-world noise.</description>
	<pubDate>2026-03-23</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 98: A Comparative Analysis of Deep-Learning-Based Speech Enhancement Models: Assessing Biometric Speaker Verification in Real-World Noisy Environments</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/3/98">doi: 10.3390/bdcc10030098</a></p>
	<p>Authors:
		Md Jahangir Alam Khondkar
		Ajan Ahmed
		Stephanie Schuckers
		Masudul H. Imtiaz
		</p>
	<p>Speech enhancement through denoising is essential for maintaining signal intelligibility and quality in biometric speaker verification pipelines that operate in acoustically adverse conditions. Despite the proliferation of deep learning (DL) architectures for speech denoising, simultaneously optimizing noise attenuation, perceptual fidelity, and speaker-identity preservation remains an open problem. We address this gap by benchmarking three architecturally distinct DL-based enhancement models&amp;amp;mdash;Wave-U-Net, CMGAN, and U-Net&amp;amp;mdash;on three independent, domain-diverse corpora (SpEAR, VPQAD, and Clarkson) that the models never encountered during training and by introducing commercial-grade VeriSpeak speaker-verification scores as a biometric evaluation dimension absent from prior comparative studies. Our experiments reveal a clear three-way trade-off: U-Net achieves the highest signal-to-noise ratio (SNR) gains (+61.44% on SpEAR, +67.05% on VPQAD, +235.3% on Clarkson) but sacrifices naturalness; CMGAN yields the best perceptual evaluation of speech quality (PESQ) values (3.33, 1.35, and 2.50, respectively), favoring listening-comfort applications; and Wave-U-Net delivers the strongest biometric fidelity (VeriSpeak improvements of +11.63%, +30.22%, and +29.24%) while offering competitive perceptual quality. These results highlight that model selection must be driven by the target deployment scenario and provide actionable guidance for improving biometric verification robustness under real-world noise.</p>
	]]></content:encoded>

	<dc:title>A Comparative Analysis of Deep-Learning-Based Speech Enhancement Models: Assessing Biometric Speaker Verification in Real-World Noisy Environments</dc:title>
			<dc:creator>Md Jahangir Alam Khondkar</dc:creator>
			<dc:creator>Ajan Ahmed</dc:creator>
			<dc:creator>Stephanie Schuckers</dc:creator>
			<dc:creator>Masudul H. Imtiaz</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10030098</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-03-23</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-03-23</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>98</prism:startingPage>
		<prism:doi>10.3390/bdcc10030098</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/3/98</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/3/97">

	<title>BDCC, Vol. 10, Pages 97: Predicting Mortality and Readmission in Obstructive Sleep Apnea via LLM-Expanded Clinical Concepts</title>
	<link>https://www.mdpi.com/2504-2289/10/3/97</link>
	<description>Obstructive Sleep Apnea (OSA) is a common sleep disorder associated with serious health risks. This study leverages large language models (LLMs) to process and interpret clinical narratives in electronic health records. It develops clinically meaningful lexicons for predicting mortality and readmission risk, as well as for multiclass diagnostic classification in OSA patients. Using LLM-expanded lexicons, logistic regression models achieved ROC&amp;amp;ndash;AUC scores of 0.844 for 6-month all-cause post-discharge mortality, 0.817 for 1-year all-cause post-discharge mortality, and 0.729 for all-cause hospital readmissions following the first discharge. Diagnostic performance was highest with smaller n-gram representations, indicating that additional contextual length did not improve performance. Compared with frequency-based n-gram models, LLM-expanded lexicons yielded sparser feature sets with lower computational cost and comparable performance. Our findings highlight the potential of LLM-expanded lexicons to enhance OSA diagnosis and clinical risk stratification.</description>
	<pubDate>2026-03-21</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 97: Predicting Mortality and Readmission in Obstructive Sleep Apnea via LLM-Expanded Clinical Concepts</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/3/97">doi: 10.3390/bdcc10030097</a></p>
	<p>Authors:
		Awwal Ahmed
		Anthony Rispoli
		Carrie Wasieloski
		Ifrah Khurram
		Rafael Zamora-Resendiz
		Destinee Morrow
		Aijuan Dong
		Silvia Crivelli
		</p>
	<p>Obstructive Sleep Apnea (OSA) is a common sleep disorder associated with serious health risks. This study leverages large language models (LLMs) to process and interpret clinical narratives in electronic health records. It develops clinically meaningful lexicons for predicting mortality and readmission risk, as well as for multiclass diagnostic classification in OSA patients. Using LLM-expanded lexicons, logistic regression models achieved ROC&amp;amp;ndash;AUC scores of 0.844 for 6-month all-cause post-discharge mortality, 0.817 for 1-year all-cause post-discharge mortality, and 0.729 for all-cause hospital readmissions following the first discharge. Diagnostic performance was highest with smaller n-gram representations, indicating that additional contextual length did not improve performance. Compared with frequency-based n-gram models, LLM-expanded lexicons yielded sparser feature sets with lower computational cost and comparable performance. Our findings highlight the potential of LLM-expanded lexicons to enhance OSA diagnosis and clinical risk stratification.</p>
	]]></content:encoded>

	<dc:title>Predicting Mortality and Readmission in Obstructive Sleep Apnea via LLM-Expanded Clinical Concepts</dc:title>
			<dc:creator>Awwal Ahmed</dc:creator>
			<dc:creator>Anthony Rispoli</dc:creator>
			<dc:creator>Carrie Wasieloski</dc:creator>
			<dc:creator>Ifrah Khurram</dc:creator>
			<dc:creator>Rafael Zamora-Resendiz</dc:creator>
			<dc:creator>Destinee Morrow</dc:creator>
			<dc:creator>Aijuan Dong</dc:creator>
			<dc:creator>Silvia Crivelli</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10030097</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-03-21</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-03-21</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>97</prism:startingPage>
		<prism:doi>10.3390/bdcc10030097</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/3/97</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/3/96">

	<title>BDCC, Vol. 10, Pages 96: Hybrid Music Similarity with Hypergraph and Siamese Network</title>
	<link>https://www.mdpi.com/2504-2289/10/3/96</link>
	<description>This paper proposes a novel method for measuring music similarity. Existing music similarity measurements have often been used for music appreciation, but this paper proposes a method for measuring the similarity between music samples which are used for music production. Conventional music recommendation approaches often rely on either metadata-based similarity or audio-based feature similarity in isolation, which limits their effectiveness in sample-based recommendation scenarios where both compositional context and acoustic characteristics are important. To address this limitation, the proposed framework combines a hypergraph-based information similarity module with a feature-based similarity module learned using Siamese networks and triplet loss. In the information-based module, metadata attributes such as beats per minute (BPM), genre, chord, key, and instrument are modeled as vertices in a hypergraph, and Random Walk&amp;amp;ndash;Word2Vec embeddings are learned to capture structural relationships between music samples and their attributes. In parallel, the feature-based module employs vertex-specific Siamese networks trained on instrument and key classification tasks to learn perceptual similarity directly from audio signals. The two modules are trained independently and jointly utilized at the recommendation stage to provide attribute-specific similarity results for a given query sample. Results show that the proposed system achieves high Precision@k across multiple attributes and forms stable similarity structures in the embedding space, even without relying on user interaction data. These results reflect embedding consistency evaluated over the entire dataset where training and retrieval are performed on the same sample pool, rather than generalization to unseen samples. These results demonstrate that the proposed hybrid framework effectively captures both structural and perceptual similarity among music samples and is well suited for sample-based music recommendation in music production environments.</description>
	<pubDate>2026-03-21</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 96: Hybrid Music Similarity with Hypergraph and Siamese Network</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/3/96">doi: 10.3390/bdcc10030096</a></p>
	<p>Authors:
		Sera Kim
		Youngjun Kim
		Jaewon Lee
		Dalwon Jang
		</p>
	<p>This paper proposes a novel method for measuring music similarity. Existing music similarity measurements have often been used for music appreciation, but this paper proposes a method for measuring the similarity between music samples which are used for music production. Conventional music recommendation approaches often rely on either metadata-based similarity or audio-based feature similarity in isolation, which limits their effectiveness in sample-based recommendation scenarios where both compositional context and acoustic characteristics are important. To address this limitation, the proposed framework combines a hypergraph-based information similarity module with a feature-based similarity module learned using Siamese networks and triplet loss. In the information-based module, metadata attributes such as beats per minute (BPM), genre, chord, key, and instrument are modeled as vertices in a hypergraph, and Random Walk&amp;amp;ndash;Word2Vec embeddings are learned to capture structural relationships between music samples and their attributes. In parallel, the feature-based module employs vertex-specific Siamese networks trained on instrument and key classification tasks to learn perceptual similarity directly from audio signals. The two modules are trained independently and jointly utilized at the recommendation stage to provide attribute-specific similarity results for a given query sample. Results show that the proposed system achieves high Precision@k across multiple attributes and forms stable similarity structures in the embedding space, even without relying on user interaction data. These results reflect embedding consistency evaluated over the entire dataset where training and retrieval are performed on the same sample pool, rather than generalization to unseen samples. These results demonstrate that the proposed hybrid framework effectively captures both structural and perceptual similarity among music samples and is well suited for sample-based music recommendation in music production environments.</p>
	]]></content:encoded>

	<dc:title>Hybrid Music Similarity with Hypergraph and Siamese Network</dc:title>
			<dc:creator>Sera Kim</dc:creator>
			<dc:creator>Youngjun Kim</dc:creator>
			<dc:creator>Jaewon Lee</dc:creator>
			<dc:creator>Dalwon Jang</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10030096</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-03-21</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-03-21</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>96</prism:startingPage>
		<prism:doi>10.3390/bdcc10030096</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/3/96</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/3/95">

	<title>BDCC, Vol. 10, Pages 95: A Dynamic Prompt-Based Logic-Aided Compliance Checker</title>
	<link>https://www.mdpi.com/2504-2289/10/3/95</link>
	<description>Text-based automatic compliance checking (ACC) employs natural language processing technologies to scrutinize a corporation&amp;amp;rsquo;s business documents, ensuring adherence to related normative texts. The current methods fall into two primary categories: symbol-based and embedding-based approaches. Symbol-based methods, noted for their accuracy and transparent processing, suffer from limited versatility. Conversely, embedding-based methods operate independently of expert knowledge yet often yield challenging-to-interpret results and require substantial volumes of annotated data. While both types of methods exhibit advantages in different aspects, the current research fails to combine these advantages effectively. Therefore, the existing methods fail to balance interpretability, generalization ability, and accuracy, which are key requirements for practical compliance systems. To address this problem, we introduce a novel approach termed the Dynamic Prompt-based Logic-Aided Compliance Checker (DPLACC), which is grounded in the prompt learning framework. This method initially parses target texts, transforming the results into first-order logical expressions. It subsequently retrieves pertinent knowledge from a knowledge graph, converting the knowledge into analogous first-order logical expressions. These expressions are then encoded into a global semantic vector via a pre-trained first-order logistic encoder. Ultimately, the semantics of expressions and initial texts are amalgamated within the prompt template, facilitating the logical knowledge enhancement of model reasoning. Experiments on Chinese and English datasets demonstrate that DPLACC comprehensively outperforms existing methods based solely on symbols or embeddings in terms of accuracy, precision, recall, and F1 score and significantly surpasses current mainstream large language models. Furthermore, DPLACC exhibits enhanced interpretability and reduced data dependence, maintaining 70% checking accuracy with as few as ten training samples. This capability allows DPLACC to be rapidly deployed in data-scarce real-world scenarios with minimal annotation overhead, thus offering a practical pathway toward the scalable implementation of compliance inspection systems.</description>
	<pubDate>2026-03-21</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 95: A Dynamic Prompt-Based Logic-Aided Compliance Checker</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/3/95">doi: 10.3390/bdcc10030095</a></p>
	<p>Authors:
		Wenxi Sheng
		Chi Wei
		Yinuo Zhang
		Bowen Zhang
		Jingyun Sun
		</p>
	<p>Text-based automatic compliance checking (ACC) employs natural language processing technologies to scrutinize a corporation&amp;amp;rsquo;s business documents, ensuring adherence to related normative texts. The current methods fall into two primary categories: symbol-based and embedding-based approaches. Symbol-based methods, noted for their accuracy and transparent processing, suffer from limited versatility. Conversely, embedding-based methods operate independently of expert knowledge yet often yield challenging-to-interpret results and require substantial volumes of annotated data. While both types of methods exhibit advantages in different aspects, the current research fails to combine these advantages effectively. Therefore, the existing methods fail to balance interpretability, generalization ability, and accuracy, which are key requirements for practical compliance systems. To address this problem, we introduce a novel approach termed the Dynamic Prompt-based Logic-Aided Compliance Checker (DPLACC), which is grounded in the prompt learning framework. This method initially parses target texts, transforming the results into first-order logical expressions. It subsequently retrieves pertinent knowledge from a knowledge graph, converting the knowledge into analogous first-order logical expressions. These expressions are then encoded into a global semantic vector via a pre-trained first-order logistic encoder. Ultimately, the semantics of expressions and initial texts are amalgamated within the prompt template, facilitating the logical knowledge enhancement of model reasoning. Experiments on Chinese and English datasets demonstrate that DPLACC comprehensively outperforms existing methods based solely on symbols or embeddings in terms of accuracy, precision, recall, and F1 score and significantly surpasses current mainstream large language models. Furthermore, DPLACC exhibits enhanced interpretability and reduced data dependence, maintaining 70% checking accuracy with as few as ten training samples. This capability allows DPLACC to be rapidly deployed in data-scarce real-world scenarios with minimal annotation overhead, thus offering a practical pathway toward the scalable implementation of compliance inspection systems.</p>
	]]></content:encoded>

	<dc:title>A Dynamic Prompt-Based Logic-Aided Compliance Checker</dc:title>
			<dc:creator>Wenxi Sheng</dc:creator>
			<dc:creator>Chi Wei</dc:creator>
			<dc:creator>Yinuo Zhang</dc:creator>
			<dc:creator>Bowen Zhang</dc:creator>
			<dc:creator>Jingyun Sun</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10030095</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-03-21</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-03-21</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>95</prism:startingPage>
		<prism:doi>10.3390/bdcc10030095</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/3/95</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/3/94">

	<title>BDCC, Vol. 10, Pages 94: Generative AI and the Foundation Model Era: A Comprehensive Review</title>
	<link>https://www.mdpi.com/2504-2289/10/3/94</link>
	<description>Generative artificial intelligence and foundation models have changed machine learning by allowing systems to produce readable text, realistic images, and other multimodal content with little direct input from a user. Foundation models are large neural networks trained on very large and varied datasets, and they form the core of many current generative AI (GenAI) systems. Their rapid development has led to major advances in areas like natural language processing, computer vision, multimodal learning, and robotics. Examples include GPT, LLaMA, and diffusion-based architectures, such as models often used for image generation. Systems such as Stable Diffusion show this shift by illustrating how AI can interpret information, draw basic inferences, and produce new outputs using more than one type of data. This review surveys common foundation model architectures and examines what they can do in generative tasks. It reviews Transformer, diffusion, and multimodal architectures, focusing on methods that support scaling and transfer across domains. The paper also reviews key approaches to pretraining and fine-tuning, including self-supervised learning, instruction tuning, and parameter-efficient adaptation, which support these systems&amp;amp;rsquo; ability to generalize across tasks. In addition to the technical details, this review discusses how GenAI is being used for text generation, image synthesis, robotics, and biomedical research. The study also notes continuing challenges, such as the high computing and energy demands of large models, ethical concerns about data bias and misinformation, and worries about privacy, reliability, and responsible use of AI in real settings. This review brings together ideas about model design, training methods, and social implications to point future research toward GenAI systems that are efficient, easy to interpret, and reliable, while supporting scientific progress and ethical responsibility.</description>
	<pubDate>2026-03-20</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 94: Generative AI and the Foundation Model Era: A Comprehensive Review</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/3/94">doi: 10.3390/bdcc10030094</a></p>
	<p>Authors:
		Abdussalam Elhanashi
		Siham Essahraui
		Pierpaolo Dini
		Davide Paolini
		Qinghe Zheng
		Sergio Saponara
		</p>
	<p>Generative artificial intelligence and foundation models have changed machine learning by allowing systems to produce readable text, realistic images, and other multimodal content with little direct input from a user. Foundation models are large neural networks trained on very large and varied datasets, and they form the core of many current generative AI (GenAI) systems. Their rapid development has led to major advances in areas like natural language processing, computer vision, multimodal learning, and robotics. Examples include GPT, LLaMA, and diffusion-based architectures, such as models often used for image generation. Systems such as Stable Diffusion show this shift by illustrating how AI can interpret information, draw basic inferences, and produce new outputs using more than one type of data. This review surveys common foundation model architectures and examines what they can do in generative tasks. It reviews Transformer, diffusion, and multimodal architectures, focusing on methods that support scaling and transfer across domains. The paper also reviews key approaches to pretraining and fine-tuning, including self-supervised learning, instruction tuning, and parameter-efficient adaptation, which support these systems&amp;amp;rsquo; ability to generalize across tasks. In addition to the technical details, this review discusses how GenAI is being used for text generation, image synthesis, robotics, and biomedical research. The study also notes continuing challenges, such as the high computing and energy demands of large models, ethical concerns about data bias and misinformation, and worries about privacy, reliability, and responsible use of AI in real settings. This review brings together ideas about model design, training methods, and social implications to point future research toward GenAI systems that are efficient, easy to interpret, and reliable, while supporting scientific progress and ethical responsibility.</p>
	]]></content:encoded>

	<dc:title>Generative AI and the Foundation Model Era: A Comprehensive Review</dc:title>
			<dc:creator>Abdussalam Elhanashi</dc:creator>
			<dc:creator>Siham Essahraui</dc:creator>
			<dc:creator>Pierpaolo Dini</dc:creator>
			<dc:creator>Davide Paolini</dc:creator>
			<dc:creator>Qinghe Zheng</dc:creator>
			<dc:creator>Sergio Saponara</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10030094</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-03-20</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-03-20</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Review</prism:section>
	<prism:startingPage>94</prism:startingPage>
		<prism:doi>10.3390/bdcc10030094</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/3/94</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/3/93">

	<title>BDCC, Vol. 10, Pages 93: Uncertainty-First Forecasting of the South African Equity Market Using Deep Learning and Temporal Conformal Prediction</title>
	<link>https://www.mdpi.com/2504-2289/10/3/93</link>
	<description>Accurate forecasting of equity returns remains fundamentally constrained by weak short-horizon predictability, pronounced noise, and structural non-stationarity. While deep learning models have been widely applied to financial time series, most studies prioritize point prediction and provide limited guidance on reliable uncertainty quantification, particularly in emerging markets. This study developed an uncertainty-aware forecasting framework for the South African equity market by integrating variational mode decomposition (VMD), gated recurrent units (GRUs), and temporal conformal prediction (TCP) to construct distribution-free prediction intervals with finite-sample coverage guarantees. Using daily returns from the FTSE/JSE All Share Index, we first confirmed that baseline recurrent models applied directly to raw returns exhibited negligible out-of-sample explanatory power, consistent with weak-form market efficiency. Incorporating VMD enhanced representation learning and improved point forecast accuracy by isolating latent frequency components. However, model-based predictive variance alone proved insufficient for reliable calibration. Embedding the models within a rolling conformal prediction framework restored near-nominal coverage across multiple confidence levels while allowing interval widths to adapt dynamically to changing volatility regimes. Robustness analyses, including walk-forward validation, stress-regime evaluation, and block permutation negative control experiments, indicated that the observed performance was not driven by temporal leakage or alignment artifacts. The results further highlight a trade-off between interval sharpness and tail-risk protection, particularly during extreme market events. Overall, the findings support a shift from return-level prediction toward calibrated uncertainty estimation as a more stable and economically meaningful objective in non-stationary financial environments.</description>
	<pubDate>2026-03-20</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 93: Uncertainty-First Forecasting of the South African Equity Market Using Deep Learning and Temporal Conformal Prediction</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/3/93">doi: 10.3390/bdcc10030093</a></p>
	<p>Authors:
		Phumudzo Lloyd Seabe
		Claude Rodrigue Bambe Moutsinga
		Maggie Aphane
		</p>
	<p>Accurate forecasting of equity returns remains fundamentally constrained by weak short-horizon predictability, pronounced noise, and structural non-stationarity. While deep learning models have been widely applied to financial time series, most studies prioritize point prediction and provide limited guidance on reliable uncertainty quantification, particularly in emerging markets. This study developed an uncertainty-aware forecasting framework for the South African equity market by integrating variational mode decomposition (VMD), gated recurrent units (GRUs), and temporal conformal prediction (TCP) to construct distribution-free prediction intervals with finite-sample coverage guarantees. Using daily returns from the FTSE/JSE All Share Index, we first confirmed that baseline recurrent models applied directly to raw returns exhibited negligible out-of-sample explanatory power, consistent with weak-form market efficiency. Incorporating VMD enhanced representation learning and improved point forecast accuracy by isolating latent frequency components. However, model-based predictive variance alone proved insufficient for reliable calibration. Embedding the models within a rolling conformal prediction framework restored near-nominal coverage across multiple confidence levels while allowing interval widths to adapt dynamically to changing volatility regimes. Robustness analyses, including walk-forward validation, stress-regime evaluation, and block permutation negative control experiments, indicated that the observed performance was not driven by temporal leakage or alignment artifacts. The results further highlight a trade-off between interval sharpness and tail-risk protection, particularly during extreme market events. Overall, the findings support a shift from return-level prediction toward calibrated uncertainty estimation as a more stable and economically meaningful objective in non-stationary financial environments.</p>
	]]></content:encoded>

	<dc:title>Uncertainty-First Forecasting of the South African Equity Market Using Deep Learning and Temporal Conformal Prediction</dc:title>
			<dc:creator>Phumudzo Lloyd Seabe</dc:creator>
			<dc:creator>Claude Rodrigue Bambe Moutsinga</dc:creator>
			<dc:creator>Maggie Aphane</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10030093</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-03-20</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-03-20</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>93</prism:startingPage>
		<prism:doi>10.3390/bdcc10030093</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/3/93</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/3/92">

	<title>BDCC, Vol. 10, Pages 92: A Hybrid NER&amp;ndash;Sentiment Model for Uzbek Texts: Integrating Lexical, Deep Learning, and Entity-Based Approaches</title>
	<link>https://www.mdpi.com/2504-2289/10/3/92</link>
	<description>This work proposes a hybrid Uzbek sentiment analysis model (sometimes referred to as tonality analysis in the local literature) that integrates contextual text representations with named-entity information from an NER module and emoji-based emotional cues that are common in short online messages. To provide a comprehensive baseline comparison, we evaluate seven approaches&amp;amp;mdash;SVM, LSTM, mBERT, XLM-RoBERTa-base, mDeBERTa-v3, LaBSE, and the proposed hybrid model&amp;amp;mdash;covering both classical machine learning and modern multilingual transformer architectures for low-resource sentiment tasks. The overall pipeline begins with Uzbek-specific text normalization to reduce noise from informal spellings, transliteration variants, and inconsistent apostrophe usage. In parallel, the system performs explicit emoji extraction to capture affective signals that are often expressed non-verbally in social media texts. Next, we construct three complementary feature streams: a context encoder for sentence-level semantics, NER-driven entity features that encode entity mentions and types, and an emotion module that models emoji priors and their interaction with contextual meaning. These streams are fused into a unified representation and fed to a final classifier to predict sentiment polarity. Experiments on an Uzbek test set demonstrate that the hybrid model reaches an F1-score of 0.92, consistently outperforming text-only baselines. The results indicate that entity-aware and emoji-informed features improve robustness under sarcasm/irony, mixed sentiment with multiple targets, and orthographic noise, making the approach suitable for social media analytics, public opinion monitoring, customer feedback triage, and recommendation-oriented text mining.</description>
	<pubDate>2026-03-19</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 92: A Hybrid NER&amp;ndash;Sentiment Model for Uzbek Texts: Integrating Lexical, Deep Learning, and Entity-Based Approaches</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/3/92">doi: 10.3390/bdcc10030092</a></p>
	<p>Authors:
		Bobur Saidov
		Vladimir Barakhnin
		Rakhmon Saparbaev
		Zayniddin Narmuratov
		Rustamova Manzura
		Ruzmetova Zilolakhon
		Anorgul Atajanova
		</p>
	<p>This work proposes a hybrid Uzbek sentiment analysis model (sometimes referred to as tonality analysis in the local literature) that integrates contextual text representations with named-entity information from an NER module and emoji-based emotional cues that are common in short online messages. To provide a comprehensive baseline comparison, we evaluate seven approaches&amp;amp;mdash;SVM, LSTM, mBERT, XLM-RoBERTa-base, mDeBERTa-v3, LaBSE, and the proposed hybrid model&amp;amp;mdash;covering both classical machine learning and modern multilingual transformer architectures for low-resource sentiment tasks. The overall pipeline begins with Uzbek-specific text normalization to reduce noise from informal spellings, transliteration variants, and inconsistent apostrophe usage. In parallel, the system performs explicit emoji extraction to capture affective signals that are often expressed non-verbally in social media texts. Next, we construct three complementary feature streams: a context encoder for sentence-level semantics, NER-driven entity features that encode entity mentions and types, and an emotion module that models emoji priors and their interaction with contextual meaning. These streams are fused into a unified representation and fed to a final classifier to predict sentiment polarity. Experiments on an Uzbek test set demonstrate that the hybrid model reaches an F1-score of 0.92, consistently outperforming text-only baselines. The results indicate that entity-aware and emoji-informed features improve robustness under sarcasm/irony, mixed sentiment with multiple targets, and orthographic noise, making the approach suitable for social media analytics, public opinion monitoring, customer feedback triage, and recommendation-oriented text mining.</p>
	]]></content:encoded>

	<dc:title>A Hybrid NER&amp;amp;ndash;Sentiment Model for Uzbek Texts: Integrating Lexical, Deep Learning, and Entity-Based Approaches</dc:title>
			<dc:creator>Bobur Saidov</dc:creator>
			<dc:creator>Vladimir Barakhnin</dc:creator>
			<dc:creator>Rakhmon Saparbaev</dc:creator>
			<dc:creator>Zayniddin Narmuratov</dc:creator>
			<dc:creator>Rustamova Manzura</dc:creator>
			<dc:creator>Ruzmetova Zilolakhon</dc:creator>
			<dc:creator>Anorgul Atajanova</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10030092</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-03-19</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-03-19</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>92</prism:startingPage>
		<prism:doi>10.3390/bdcc10030092</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/3/92</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/3/91">

	<title>BDCC, Vol. 10, Pages 91: Data-Driven Cognitive Early Warning for Goaf Spontaneous Combustion: An Edge-Deployed RBF Network with Real-Time Multisensor Analytics</title>
	<link>https://www.mdpi.com/2504-2289/10/3/91</link>
	<description>Spontaneous combustion in goaf areas poses a significant threat to coal mine safety. Traditional safety management systems, reliant on passive response and single-indicator thresholds, often suffer from delayed warnings and lack cognitive decision support. To address this challenge, this study proposes a big-data-driven cognitive computing framework for dynamic risk prediction of goaf spontaneous combustion, based on a &amp;amp;ldquo;Cloud-Edge-End&amp;amp;rdquo; collaborative architecture. The method leverages multi-sensor big data streams (CO, C2H4, O2, etc.) and deploys a lightweight Radial Basis Function (RBF) neural network on underground edge computing nodes (STM32) for real-time analytics. The model demonstrates excellent predictive performance on imbalanced datasets, with a PR-AUC of 0.910 and a recall of 99.7%. The edge-deployed RBF model achieves a single-pass inference time of only 0.62 ms, enabling real-time cognitive risk mapping. Field application at Z Coal Mine validated the system&amp;amp;rsquo;s effectiveness, providing an average pre-warning time of 48.5 h, achieving zero spontaneous combustion accidents, and reducing the Total Recordable Injury Rate (TRIR) by 15.2%. This work illustrates how edge-based cognitive computing can transform safety management from passive response to proactive prevention, offering a scalable and interpretable framework for intelligent mine safety.</description>
	<pubDate>2026-03-19</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 91: Data-Driven Cognitive Early Warning for Goaf Spontaneous Combustion: An Edge-Deployed RBF Network with Real-Time Multisensor Analytics</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/3/91">doi: 10.3390/bdcc10030091</a></p>
	<p>Authors:
		Gang Cheng
		Hailin Pei
		Xiaokang Chen
		Xiaorong Pang
		Renzheng Sun
		</p>
	<p>Spontaneous combustion in goaf areas poses a significant threat to coal mine safety. Traditional safety management systems, reliant on passive response and single-indicator thresholds, often suffer from delayed warnings and lack cognitive decision support. To address this challenge, this study proposes a big-data-driven cognitive computing framework for dynamic risk prediction of goaf spontaneous combustion, based on a &amp;amp;ldquo;Cloud-Edge-End&amp;amp;rdquo; collaborative architecture. The method leverages multi-sensor big data streams (CO, C2H4, O2, etc.) and deploys a lightweight Radial Basis Function (RBF) neural network on underground edge computing nodes (STM32) for real-time analytics. The model demonstrates excellent predictive performance on imbalanced datasets, with a PR-AUC of 0.910 and a recall of 99.7%. The edge-deployed RBF model achieves a single-pass inference time of only 0.62 ms, enabling real-time cognitive risk mapping. Field application at Z Coal Mine validated the system&amp;amp;rsquo;s effectiveness, providing an average pre-warning time of 48.5 h, achieving zero spontaneous combustion accidents, and reducing the Total Recordable Injury Rate (TRIR) by 15.2%. This work illustrates how edge-based cognitive computing can transform safety management from passive response to proactive prevention, offering a scalable and interpretable framework for intelligent mine safety.</p>
	]]></content:encoded>

	<dc:title>Data-Driven Cognitive Early Warning for Goaf Spontaneous Combustion: An Edge-Deployed RBF Network with Real-Time Multisensor Analytics</dc:title>
			<dc:creator>Gang Cheng</dc:creator>
			<dc:creator>Hailin Pei</dc:creator>
			<dc:creator>Xiaokang Chen</dc:creator>
			<dc:creator>Xiaorong Pang</dc:creator>
			<dc:creator>Renzheng Sun</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10030091</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-03-19</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-03-19</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>91</prism:startingPage>
		<prism:doi>10.3390/bdcc10030091</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/3/91</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/3/90">

	<title>BDCC, Vol. 10, Pages 90: Dual-Stream Transformer with Kalman-Based Sensor Fusion for Wearable Fall Detection</title>
	<link>https://www.mdpi.com/2504-2289/10/3/90</link>
	<description>Wearable fall detection systems face a fundamental challenge: while gyroscope data provide valuable orientation cues, naively combining raw gyroscope and accelerometer signals can degrade performance due to noise contamination. To overcome this challenge, we present a dual-stream transformer architecture that incorporates (i) Kalman-based sensor fusion to convert noisy gyroscope angular velocities into stable orientation estimates (roll, pitch, yaw), maintaining an internal state of body pose, and (ii) processing accelerometer and orientation streams in separate encoder pathways before fusion to prevent cross-modal interference. Our architecture further integrates Squeeze-and-Excitation channel attention and Temporal Attention Pooling to focus on fall-critical temporal patterns. Evaluated on the SmartFallMM dataset using 21-fold leave-one-subject-out cross-validation, the dual-stream Kalman transformer achieves 91.10% F1, outperforming single-stream Kalman transformers (89.80% F1) by 1.30% and single-stream baseline transformers (88.96% F1) by 2.14%. We further evaluate the model in real time using a watch-based SmartFall App on five participants, maintaining an average F1 score of 83% and an accuracy of 90%. These results indicate robust performance in both offline and real-world deployment settings, establishing a new state-of-the-art for inertial-measurement-unit-based fall detection on commodity smartwatch devices.</description>
	<pubDate>2026-03-17</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 90: Dual-Stream Transformer with Kalman-Based Sensor Fusion for Wearable Fall Detection</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/3/90">doi: 10.3390/bdcc10030090</a></p>
	<p>Authors:
		Abheek Pradhan
		Sana Alamgeer
		Rakesh Suvvari
		Syed Tousiful Haque
		Anne H. H. Ngu
		</p>
	<p>Wearable fall detection systems face a fundamental challenge: while gyroscope data provide valuable orientation cues, naively combining raw gyroscope and accelerometer signals can degrade performance due to noise contamination. To overcome this challenge, we present a dual-stream transformer architecture that incorporates (i) Kalman-based sensor fusion to convert noisy gyroscope angular velocities into stable orientation estimates (roll, pitch, yaw), maintaining an internal state of body pose, and (ii) processing accelerometer and orientation streams in separate encoder pathways before fusion to prevent cross-modal interference. Our architecture further integrates Squeeze-and-Excitation channel attention and Temporal Attention Pooling to focus on fall-critical temporal patterns. Evaluated on the SmartFallMM dataset using 21-fold leave-one-subject-out cross-validation, the dual-stream Kalman transformer achieves 91.10% F1, outperforming single-stream Kalman transformers (89.80% F1) by 1.30% and single-stream baseline transformers (88.96% F1) by 2.14%. We further evaluate the model in real time using a watch-based SmartFall App on five participants, maintaining an average F1 score of 83% and an accuracy of 90%. These results indicate robust performance in both offline and real-world deployment settings, establishing a new state-of-the-art for inertial-measurement-unit-based fall detection on commodity smartwatch devices.</p>
	]]></content:encoded>

	<dc:title>Dual-Stream Transformer with Kalman-Based Sensor Fusion for Wearable Fall Detection</dc:title>
			<dc:creator>Abheek Pradhan</dc:creator>
			<dc:creator>Sana Alamgeer</dc:creator>
			<dc:creator>Rakesh Suvvari</dc:creator>
			<dc:creator>Syed Tousiful Haque</dc:creator>
			<dc:creator>Anne H. H. Ngu</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10030090</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-03-17</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-03-17</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>90</prism:startingPage>
		<prism:doi>10.3390/bdcc10030090</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/3/90</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/3/89">

	<title>BDCC, Vol. 10, Pages 89: DEPART: Multi-Task Interpretable Depression and Parkinson&amp;rsquo;s Disease Detection from In-the-Wild Video Data</title>
	<link>https://www.mdpi.com/2504-2289/10/3/89</link>
	<description>Automated video-based detection of cognitive disorders can enable a scalable non-invasive health monitoring. However, existing methods focus on a single disease and provide limited interpretability, whereas real-world videos often contain co-occurring conditions. We propose a novel unified multi-task method to detect depression and Parkinson&amp;amp;rsquo;s disease (PD) from in-the-wild video data called DEPART (DEpression and PArkinson&amp;amp;rsquo;s Recognition Technique). It performs body region extraction, Contrastive Language-Image Pre-training (CLIP)-based visual encoding, Transformer-based temporal modeling, and prototype-aware classification with a gated fusion technique. Gradient-based attention maps are used to visualize task-specific regions that drive predictions. Experiments on the In-the-Wild Speech Medical (WSM) corpus demonstrate competitive performance: the multi-task model achieves Recall of 82.39% for depression and 78.20% for PD, compared with 87.76% and 78.20%, for the best single-task models. The multi-task learning initially increases false positives for healthy persons in the PD subset, mainly due to annotation&amp;amp;ndash;modality mismatches, static visual content misinterpreted as motor impairments, and occasional body detection failures. After cleaning the test data, Recall for healthy individuals becomes comparable across models; the multi-task model improves Recall for both depression (from 82.39% to 87.50%) and PD (from 78.20% to 86.14%), suggesting better robustness for real-life clinical applications.</description>
	<pubDate>2026-03-16</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 89: DEPART: Multi-Task Interpretable Depression and Parkinson&amp;rsquo;s Disease Detection from In-the-Wild Video Data</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/3/89">doi: 10.3390/bdcc10030089</a></p>
	<p>Authors:
		Elena Ryumina
		Alexandr Axyonov
		Mikhail Dolgushin
		Dmitry Ryumin
		Alexey Karpov
		</p>
	<p>Automated video-based detection of cognitive disorders can enable a scalable non-invasive health monitoring. However, existing methods focus on a single disease and provide limited interpretability, whereas real-world videos often contain co-occurring conditions. We propose a novel unified multi-task method to detect depression and Parkinson&amp;amp;rsquo;s disease (PD) from in-the-wild video data called DEPART (DEpression and PArkinson&amp;amp;rsquo;s Recognition Technique). It performs body region extraction, Contrastive Language-Image Pre-training (CLIP)-based visual encoding, Transformer-based temporal modeling, and prototype-aware classification with a gated fusion technique. Gradient-based attention maps are used to visualize task-specific regions that drive predictions. Experiments on the In-the-Wild Speech Medical (WSM) corpus demonstrate competitive performance: the multi-task model achieves Recall of 82.39% for depression and 78.20% for PD, compared with 87.76% and 78.20%, for the best single-task models. The multi-task learning initially increases false positives for healthy persons in the PD subset, mainly due to annotation&amp;amp;ndash;modality mismatches, static visual content misinterpreted as motor impairments, and occasional body detection failures. After cleaning the test data, Recall for healthy individuals becomes comparable across models; the multi-task model improves Recall for both depression (from 82.39% to 87.50%) and PD (from 78.20% to 86.14%), suggesting better robustness for real-life clinical applications.</p>
	]]></content:encoded>

	<dc:title>DEPART: Multi-Task Interpretable Depression and Parkinson&amp;amp;rsquo;s Disease Detection from In-the-Wild Video Data</dc:title>
			<dc:creator>Elena Ryumina</dc:creator>
			<dc:creator>Alexandr Axyonov</dc:creator>
			<dc:creator>Mikhail Dolgushin</dc:creator>
			<dc:creator>Dmitry Ryumin</dc:creator>
			<dc:creator>Alexey Karpov</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10030089</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-03-16</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-03-16</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>89</prism:startingPage>
		<prism:doi>10.3390/bdcc10030089</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/3/89</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/3/88">

	<title>BDCC, Vol. 10, Pages 88: Unified Visual Synchrony: A Framework for Face&amp;ndash;Gesture Coherence in Multimodal Human&amp;ndash;AI Interaction</title>
	<link>https://www.mdpi.com/2504-2289/10/3/88</link>
	<description>Multimodal human&amp;amp;ndash;AI systems generally consider facial expressions and body motions as separate input streams, leading to disjointed interpretations and diminished emotional coherence. To overcome this issue, we offer the Engagement-Safe Expressive Alignment (ESEA) paradigm and the Unified Visual Synchrony (UVS) framework as its computational implementation. UVS models the coherence between facial expressions and gestures, offering an interpretable visual synchrony signal that can function as adaptive feedback in human&amp;amp;ndash;AI interactions. The framework&amp;amp;rsquo;s key component is the Consistency Index for Affective Synchrony (CIAS), which correlates brief visual segments with scalar synchrony scores through a common latent representation. Facial and gestural signals are processed by modality-specific projection networks into a unified latent space, and CIAS is derived from the similarity and short-term temporal consistency of these latent trajectories. The synchrony index is regarded as an estimation of affective visual coherence within the ESEA paradigm. We formalize the UVS/CIAS framework and conduct a comparative experimental evaluation utilizing matched and mismatched face&amp;amp;ndash;gesture segments derived from rendered dialog footage. Utilizing ROC analysis, score distribution comparisons, temporal visualizations, and negative control tests, we illustrate that CIAS effectively captures structured face&amp;amp;ndash;gesture alignment that surpasses similarity-based baselines, while also delivering a persistent, time-resolved synchronization signal. These findings establish CIAS as a principled and interpretable feedback signal for future affect-aware, engagement-focused multimodal agents.</description>
	<pubDate>2026-03-12</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 88: Unified Visual Synchrony: A Framework for Face&amp;ndash;Gesture Coherence in Multimodal Human&amp;ndash;AI Interaction</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/3/88">doi: 10.3390/bdcc10030088</a></p>
	<p>Authors:
		Saule Kudubayeva
		Yernar Seksenbayev
		Aigerim Yerimbetova
		Elmira Daiyrbayeva
		Bakzhan Sakenov
		Duman Telman
		Mussa Turdalyuly
		</p>
	<p>Multimodal human&amp;amp;ndash;AI systems generally consider facial expressions and body motions as separate input streams, leading to disjointed interpretations and diminished emotional coherence. To overcome this issue, we offer the Engagement-Safe Expressive Alignment (ESEA) paradigm and the Unified Visual Synchrony (UVS) framework as its computational implementation. UVS models the coherence between facial expressions and gestures, offering an interpretable visual synchrony signal that can function as adaptive feedback in human&amp;amp;ndash;AI interactions. The framework&amp;amp;rsquo;s key component is the Consistency Index for Affective Synchrony (CIAS), which correlates brief visual segments with scalar synchrony scores through a common latent representation. Facial and gestural signals are processed by modality-specific projection networks into a unified latent space, and CIAS is derived from the similarity and short-term temporal consistency of these latent trajectories. The synchrony index is regarded as an estimation of affective visual coherence within the ESEA paradigm. We formalize the UVS/CIAS framework and conduct a comparative experimental evaluation utilizing matched and mismatched face&amp;amp;ndash;gesture segments derived from rendered dialog footage. Utilizing ROC analysis, score distribution comparisons, temporal visualizations, and negative control tests, we illustrate that CIAS effectively captures structured face&amp;amp;ndash;gesture alignment that surpasses similarity-based baselines, while also delivering a persistent, time-resolved synchronization signal. These findings establish CIAS as a principled and interpretable feedback signal for future affect-aware, engagement-focused multimodal agents.</p>
	]]></content:encoded>

	<dc:title>Unified Visual Synchrony: A Framework for Face&amp;amp;ndash;Gesture Coherence in Multimodal Human&amp;amp;ndash;AI Interaction</dc:title>
			<dc:creator>Saule Kudubayeva</dc:creator>
			<dc:creator>Yernar Seksenbayev</dc:creator>
			<dc:creator>Aigerim Yerimbetova</dc:creator>
			<dc:creator>Elmira Daiyrbayeva</dc:creator>
			<dc:creator>Bakzhan Sakenov</dc:creator>
			<dc:creator>Duman Telman</dc:creator>
			<dc:creator>Mussa Turdalyuly</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10030088</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-03-12</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-03-12</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>88</prism:startingPage>
		<prism:doi>10.3390/bdcc10030088</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/3/88</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/3/87">

	<title>BDCC, Vol. 10, Pages 87: An Intelligent Evaluation Method for Slope Stability Based on a Database Integrating Real Cases and Numerical Simulations</title>
	<link>https://www.mdpi.com/2504-2289/10/3/87</link>
	<description>Slope instability can cause severe disasters, making stability prediction essential. Machine learning has become a key tool for this purpose, as it avoids complex mechanical calculations and efficiently handles high-dimensional data. Currently, the data used in machine learning primarily originate from real-world cases. However, such cases are inherently limited in quantity and often fail to comprehensively represent all potential slope conditions. To address these limitations, this study proposes a method for constructing numerical simulation databases. Based on this, we develop a model establishment method for rapid evaluation of slope stability integrating numerical simulation with engineering cases. This study uses six characteristic parameters to assess slope stability, including unit weight &amp;amp;gamma;, cohesion c, internal friction angle &amp;amp;phi;, slope angle &amp;amp;alpha;, slope height H, and pore pressure ratio ru. Through extensive literature mining, we established a database of 684 engineering cases. Based on statistical analysis of input parameters, a numerical simulation scheme was designed. Batch calculations were performed using MATLAB to determine simulation results. The engineering case database was then partitioned into training and testing sets for model development and validation. Subsequently, the numerical simulation database was incorporated into the training set for retesting. Results demonstrate that when considering all predictive indicators, the prediction accuracy of the GRNN-based model improved from 85% to 88.3%, while the PNN-based model showed an increase from 69% to 88.3%. This study offers new insights for optimizing numerical simulation design and enhancing machine learning performance in slope stability prediction.</description>
	<pubDate>2026-03-12</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 87: An Intelligent Evaluation Method for Slope Stability Based on a Database Integrating Real Cases and Numerical Simulations</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/3/87">doi: 10.3390/bdcc10030087</a></p>
	<p>Authors:
		Junyi Jiang
		Dong Li
		Qingyi Yang
		Zhenhua Zhang
		Lei Wang
		Wenru Zhao
		Mingliang Chen
		</p>
	<p>Slope instability can cause severe disasters, making stability prediction essential. Machine learning has become a key tool for this purpose, as it avoids complex mechanical calculations and efficiently handles high-dimensional data. Currently, the data used in machine learning primarily originate from real-world cases. However, such cases are inherently limited in quantity and often fail to comprehensively represent all potential slope conditions. To address these limitations, this study proposes a method for constructing numerical simulation databases. Based on this, we develop a model establishment method for rapid evaluation of slope stability integrating numerical simulation with engineering cases. This study uses six characteristic parameters to assess slope stability, including unit weight &amp;amp;gamma;, cohesion c, internal friction angle &amp;amp;phi;, slope angle &amp;amp;alpha;, slope height H, and pore pressure ratio ru. Through extensive literature mining, we established a database of 684 engineering cases. Based on statistical analysis of input parameters, a numerical simulation scheme was designed. Batch calculations were performed using MATLAB to determine simulation results. The engineering case database was then partitioned into training and testing sets for model development and validation. Subsequently, the numerical simulation database was incorporated into the training set for retesting. Results demonstrate that when considering all predictive indicators, the prediction accuracy of the GRNN-based model improved from 85% to 88.3%, while the PNN-based model showed an increase from 69% to 88.3%. This study offers new insights for optimizing numerical simulation design and enhancing machine learning performance in slope stability prediction.</p>
	]]></content:encoded>

	<dc:title>An Intelligent Evaluation Method for Slope Stability Based on a Database Integrating Real Cases and Numerical Simulations</dc:title>
			<dc:creator>Junyi Jiang</dc:creator>
			<dc:creator>Dong Li</dc:creator>
			<dc:creator>Qingyi Yang</dc:creator>
			<dc:creator>Zhenhua Zhang</dc:creator>
			<dc:creator>Lei Wang</dc:creator>
			<dc:creator>Wenru Zhao</dc:creator>
			<dc:creator>Mingliang Chen</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10030087</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-03-12</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-03-12</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>87</prism:startingPage>
		<prism:doi>10.3390/bdcc10030087</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/3/87</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/3/86">

	<title>BDCC, Vol. 10, Pages 86: SBT-Rec: A Structured Behavioral Tokenization Framework for LLM-Based Sequential Recommendation</title>
	<link>https://www.mdpi.com/2504-2289/10/3/86</link>
	<description>Generative recommendation systems based on Large Language Models leverage their reasoning capabilities to capture users&amp;amp;rsquo; latent interests. However, aligning continuous user behavioral embeddings with the discrete semantic space of LLMs remains a challenge. Direct alignment often leads to semantic mismatch and hallucination issues. Furthermore, existing methods typically rely on multi-stage training strategies to adapt to variations in feature distributions, thereby limiting training efficiency. To address the aforementioned issues, we propose SBT-Rec, a structured behavioral tokenization framework. Specifically, we first design a hierarchical discrete structure discovery module, utilizing a recursive residual quantization mechanism to decompose continuous behavioral vectors into discrete behavioral atoms to resolve modality discrepancies. Second, the multi-scale behavioral semantic reconstruction module reconstructs behavioral representations via residual superposition, thereby reducing data noise. Third, a residual-aware modality distribution aligner is introduced to transform behavioral features into input tokens compatible with the LLM via non-linear mapping. Finally, based on structured discrete representations, we propose a single-stage behavioral-semantic adaptive optimization strategy, achieving end-to-end parameter-efficient fine-tuning. Experiments on the MovieLens, LastFM, and Steam datasets demonstrate that SBT-Rec outperforms existing baseline models in terms of recommendation accuracy, training efficiency, and noise robustness.</description>
	<pubDate>2026-03-10</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 86: SBT-Rec: A Structured Behavioral Tokenization Framework for LLM-Based Sequential Recommendation</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/3/86">doi: 10.3390/bdcc10030086</a></p>
	<p>Authors:
		Langgao Cheng
		Yanying Mao
		Guowang Li
		Honghui Chen
		</p>
	<p>Generative recommendation systems based on Large Language Models leverage their reasoning capabilities to capture users&amp;amp;rsquo; latent interests. However, aligning continuous user behavioral embeddings with the discrete semantic space of LLMs remains a challenge. Direct alignment often leads to semantic mismatch and hallucination issues. Furthermore, existing methods typically rely on multi-stage training strategies to adapt to variations in feature distributions, thereby limiting training efficiency. To address the aforementioned issues, we propose SBT-Rec, a structured behavioral tokenization framework. Specifically, we first design a hierarchical discrete structure discovery module, utilizing a recursive residual quantization mechanism to decompose continuous behavioral vectors into discrete behavioral atoms to resolve modality discrepancies. Second, the multi-scale behavioral semantic reconstruction module reconstructs behavioral representations via residual superposition, thereby reducing data noise. Third, a residual-aware modality distribution aligner is introduced to transform behavioral features into input tokens compatible with the LLM via non-linear mapping. Finally, based on structured discrete representations, we propose a single-stage behavioral-semantic adaptive optimization strategy, achieving end-to-end parameter-efficient fine-tuning. Experiments on the MovieLens, LastFM, and Steam datasets demonstrate that SBT-Rec outperforms existing baseline models in terms of recommendation accuracy, training efficiency, and noise robustness.</p>
	]]></content:encoded>

	<dc:title>SBT-Rec: A Structured Behavioral Tokenization Framework for LLM-Based Sequential Recommendation</dc:title>
			<dc:creator>Langgao Cheng</dc:creator>
			<dc:creator>Yanying Mao</dc:creator>
			<dc:creator>Guowang Li</dc:creator>
			<dc:creator>Honghui Chen</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10030086</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-03-10</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-03-10</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>86</prism:startingPage>
		<prism:doi>10.3390/bdcc10030086</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/3/86</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/3/85">

	<title>BDCC, Vol. 10, Pages 85: A Reference Model for the Analysis and Indexing of Metaverse Recordings for Information Retrieval</title>
	<link>https://www.mdpi.com/2504-2289/10/3/85</link>
	<description>After the peak of the recent hype wave of interest surrounding the metaverse, virtual world applications remained in areas such as gaming, VR training, simulations, and collaboration. In this context, recordings are created which subsequently evolve into extensive collections that users may wish to access, search through, and retrieve items from. In order to facilitate searchability of metaverse recordings, it is necessary to adapt content analysis and indexing techniques to the specific characteristics of these recordings. This paper presents a reference model, the Processing Framework for Metaverse Recordings (PFMR), which details the phases of structural analysis, feature extraction, data mining, and feature fusion. The objective is to facilitate efficient retrieval of metaverse content. Our evaluation, based on a prototypical implementation, demonstrates the applicability and effectiveness of PFMR. This lays the groundwork for further integration of metaverse-specific content into Multimedia Information Retrieval systems. The evaluation of the 256 Metaverse Recording dataset shows that PFMRs&amp;amp;rsquo; domain-specific adaptability and integratability allows effective metaverse recording information retrieval for metaverse-specific features such as avatar detection, dialog mining, and toxicity classification.</description>
	<pubDate>2026-03-09</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 85: A Reference Model for the Analysis and Indexing of Metaverse Recordings for Information Retrieval</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/3/85">doi: 10.3390/bdcc10030085</a></p>
	<p>Authors:
		Patrick Steinert
		Stefan Wagenpfeil
		Ingo Frommholz
		Matthias L. Hemmje
		</p>
	<p>After the peak of the recent hype wave of interest surrounding the metaverse, virtual world applications remained in areas such as gaming, VR training, simulations, and collaboration. In this context, recordings are created which subsequently evolve into extensive collections that users may wish to access, search through, and retrieve items from. In order to facilitate searchability of metaverse recordings, it is necessary to adapt content analysis and indexing techniques to the specific characteristics of these recordings. This paper presents a reference model, the Processing Framework for Metaverse Recordings (PFMR), which details the phases of structural analysis, feature extraction, data mining, and feature fusion. The objective is to facilitate efficient retrieval of metaverse content. Our evaluation, based on a prototypical implementation, demonstrates the applicability and effectiveness of PFMR. This lays the groundwork for further integration of metaverse-specific content into Multimedia Information Retrieval systems. The evaluation of the 256 Metaverse Recording dataset shows that PFMRs&amp;amp;rsquo; domain-specific adaptability and integratability allows effective metaverse recording information retrieval for metaverse-specific features such as avatar detection, dialog mining, and toxicity classification.</p>
	]]></content:encoded>

	<dc:title>A Reference Model for the Analysis and Indexing of Metaverse Recordings for Information Retrieval</dc:title>
			<dc:creator>Patrick Steinert</dc:creator>
			<dc:creator>Stefan Wagenpfeil</dc:creator>
			<dc:creator>Ingo Frommholz</dc:creator>
			<dc:creator>Matthias L. Hemmje</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10030085</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-03-09</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-03-09</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>85</prism:startingPage>
		<prism:doi>10.3390/bdcc10030085</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/3/85</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/3/84">

	<title>BDCC, Vol. 10, Pages 84: Home-Based Telerehabilitation Through a Modular, Sensor-Integrated Virtual Monitoring System</title>
	<link>https://www.mdpi.com/2504-2289/10/3/84</link>
	<description>Home based telerehabilitation has expanded after COVID-19, but delivering timely guidance and monitoring exercise performance outside the clinic remains difficult. Traditional physiotherapy often relies on repeated execution of simple routines, yet clinicians have limited visibility into adherence and movement quality during unsupervised sessions. From a systems perspective, many telerehabilitation approaches also face constraints in accessibility, bandwidth, and computational cost that can limit practical deployment. This paper presents a modular telerehabilitation framework and prototype that captures and records rehabilitation exercise sessions for asynchronous clinician review in a 3D visualisation environment. The system integrates skeletal motion capture with plantar pressure sensing, and stores sessions as portable artefacts to support replay, inspection, and downstream analysis. A connector-based architecture enables extension to additional sensors without redesigning the core application, and the design aims to support deployment under constrained home computing and networking conditions. The manuscript contributes an implementation blueprint and reference architecture for multimodal capture and replay. Clinical effectiveness, usability outcomes, and quantitative sensor accuracy benchmarking are outside the scope of this work and are identified as necessary future evaluation.</description>
	<pubDate>2026-03-08</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 84: Home-Based Telerehabilitation Through a Modular, Sensor-Integrated Virtual Monitoring System</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/3/84">doi: 10.3390/bdcc10030084</a></p>
	<p>Authors:
		Zoltán Mészáros
		M. A. Hannan Bin Azhar
		Tasmina Islam
		Soumya Kanti Manna
		</p>
	<p>Home based telerehabilitation has expanded after COVID-19, but delivering timely guidance and monitoring exercise performance outside the clinic remains difficult. Traditional physiotherapy often relies on repeated execution of simple routines, yet clinicians have limited visibility into adherence and movement quality during unsupervised sessions. From a systems perspective, many telerehabilitation approaches also face constraints in accessibility, bandwidth, and computational cost that can limit practical deployment. This paper presents a modular telerehabilitation framework and prototype that captures and records rehabilitation exercise sessions for asynchronous clinician review in a 3D visualisation environment. The system integrates skeletal motion capture with plantar pressure sensing, and stores sessions as portable artefacts to support replay, inspection, and downstream analysis. A connector-based architecture enables extension to additional sensors without redesigning the core application, and the design aims to support deployment under constrained home computing and networking conditions. The manuscript contributes an implementation blueprint and reference architecture for multimodal capture and replay. Clinical effectiveness, usability outcomes, and quantitative sensor accuracy benchmarking are outside the scope of this work and are identified as necessary future evaluation.</p>
	]]></content:encoded>

	<dc:title>Home-Based Telerehabilitation Through a Modular, Sensor-Integrated Virtual Monitoring System</dc:title>
			<dc:creator>Zoltán Mészáros</dc:creator>
			<dc:creator>M. A. Hannan Bin Azhar</dc:creator>
			<dc:creator>Tasmina Islam</dc:creator>
			<dc:creator>Soumya Kanti Manna</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10030084</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-03-08</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-03-08</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>84</prism:startingPage>
		<prism:doi>10.3390/bdcc10030084</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/3/84</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/3/83">

	<title>BDCC, Vol. 10, Pages 83: Sound Event Detection in Smart Cities: A Systematic Review of Methods, Datasets, and Applications</title>
	<link>https://www.mdpi.com/2504-2289/10/3/83</link>
	<description>Sound Event Detection (SED) is a growing area with vast prospects for understanding and designing the sonic fabric of smart cities. In this paper, the latest advances in SED are summarized, focusing on models, datasets, and applications from scientific papers listed on Scopus and Web of Science. The paper provides a clear view of how SED is being used in smart cities, public safety, environment monitoring, and home security. The paper also addresses the challenges of SED, including dataset representativeness, model robustness under noisy or complex acoustic scenes, event rarity detection, as well as the ethics of using automatic listening. The paper also provides a view of future work to be undertaken in SED. The focus of the paper is on self-supervised learning, multi-modal fusion, neuro-inspired approaches, as well as privacy-preserving analytics. The paper provides a view of SED as a key technology to make smart cities safe, secure, and sustainable. SED has vast prospects as a key technology to enable artificial perception of smart cities.</description>
	<pubDate>2026-03-08</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 83: Sound Event Detection in Smart Cities: A Systematic Review of Methods, Datasets, and Applications</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/3/83">doi: 10.3390/bdcc10030083</a></p>
	<p>Authors:
		Giuseppe Ciaburro
		Virginia Puyana-Romero
		</p>
	<p>Sound Event Detection (SED) is a growing area with vast prospects for understanding and designing the sonic fabric of smart cities. In this paper, the latest advances in SED are summarized, focusing on models, datasets, and applications from scientific papers listed on Scopus and Web of Science. The paper provides a clear view of how SED is being used in smart cities, public safety, environment monitoring, and home security. The paper also addresses the challenges of SED, including dataset representativeness, model robustness under noisy or complex acoustic scenes, event rarity detection, as well as the ethics of using automatic listening. The paper also provides a view of future work to be undertaken in SED. The focus of the paper is on self-supervised learning, multi-modal fusion, neuro-inspired approaches, as well as privacy-preserving analytics. The paper provides a view of SED as a key technology to make smart cities safe, secure, and sustainable. SED has vast prospects as a key technology to enable artificial perception of smart cities.</p>
	]]></content:encoded>

	<dc:title>Sound Event Detection in Smart Cities: A Systematic Review of Methods, Datasets, and Applications</dc:title>
			<dc:creator>Giuseppe Ciaburro</dc:creator>
			<dc:creator>Virginia Puyana-Romero</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10030083</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-03-08</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-03-08</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Systematic Review</prism:section>
	<prism:startingPage>83</prism:startingPage>
		<prism:doi>10.3390/bdcc10030083</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/3/83</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/3/82">

	<title>BDCC, Vol. 10, Pages 82: Carbon Price Forecasting via a CNN-BiLSTM Model Integrating VMD and Classified News Sentiment</title>
	<link>https://www.mdpi.com/2504-2289/10/3/82</link>
	<description>Accurate carbon price forecasting is vital for risk management but is hindered by high volatility and sensitivity to external shocks. Existing multivariate models typically overlook unstructured news sentiment, failing to capture irrational fluctuations driven by market public opinion. To address this, this paper proposes VBN-Net, a hybrid model integrating carbon-specific news sentiment with Variational Mode Decomposition (VMD). Two core innovations are presented: First, a multi-modal input mechanism combines structured financial data with unstructured carbon news sentiment to effectively capture policy-driven shocks. Second, a Sequential Beluga Whale Optimization strategy is designed to adaptively optimize feature engineering in steps. Unlike conventional approaches, the VBN-Net first employs VMD for denoising and frequency decomposition, and then optimizes the fusion weights of news sentiment across different frequency components derived from multi-source news. This strategy effectively overcomes the subjectivity of manual parameter selection, providing high-quality features for a fixed CNN-BiLSTM backbone. By integrating VMD-based denoising with optimized multi-source news fusion, the model achieves consistent performance improvements across multiple evaluation metrics. The empirical findings validate the effectiveness of the proposed model in enhancing forecasting performance, thereby providing a reliable analytical tool for participants in the carbon market.</description>
	<pubDate>2026-03-06</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 82: Carbon Price Forecasting via a CNN-BiLSTM Model Integrating VMD and Classified News Sentiment</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/3/82">doi: 10.3390/bdcc10030082</a></p>
	<p>Authors:
		Xiyun Yang
		Han Chen
		Xiangjun Li
		Xiaoyu Liu
		</p>
	<p>Accurate carbon price forecasting is vital for risk management but is hindered by high volatility and sensitivity to external shocks. Existing multivariate models typically overlook unstructured news sentiment, failing to capture irrational fluctuations driven by market public opinion. To address this, this paper proposes VBN-Net, a hybrid model integrating carbon-specific news sentiment with Variational Mode Decomposition (VMD). Two core innovations are presented: First, a multi-modal input mechanism combines structured financial data with unstructured carbon news sentiment to effectively capture policy-driven shocks. Second, a Sequential Beluga Whale Optimization strategy is designed to adaptively optimize feature engineering in steps. Unlike conventional approaches, the VBN-Net first employs VMD for denoising and frequency decomposition, and then optimizes the fusion weights of news sentiment across different frequency components derived from multi-source news. This strategy effectively overcomes the subjectivity of manual parameter selection, providing high-quality features for a fixed CNN-BiLSTM backbone. By integrating VMD-based denoising with optimized multi-source news fusion, the model achieves consistent performance improvements across multiple evaluation metrics. The empirical findings validate the effectiveness of the proposed model in enhancing forecasting performance, thereby providing a reliable analytical tool for participants in the carbon market.</p>
	]]></content:encoded>

	<dc:title>Carbon Price Forecasting via a CNN-BiLSTM Model Integrating VMD and Classified News Sentiment</dc:title>
			<dc:creator>Xiyun Yang</dc:creator>
			<dc:creator>Han Chen</dc:creator>
			<dc:creator>Xiangjun Li</dc:creator>
			<dc:creator>Xiaoyu Liu</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10030082</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-03-06</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-03-06</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>82</prism:startingPage>
		<prism:doi>10.3390/bdcc10030082</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/3/82</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/3/81">

	<title>BDCC, Vol. 10, Pages 81: Predicting Bond Defaults in China: A Double-Ensemble Model Leveraging SMOTE for Class Imbalance</title>
	<link>https://www.mdpi.com/2504-2289/10/3/81</link>
	<description>This study proposes the Double-Ensemble Learning Classification with SMOTE (DELC-SMOTE), a novel hierarchical framework designed to address the critical challenge of severe class imbalance in financial bond default prediction. The model integrates the Synthetic Minority Over-sampling Technique (SMOTE) into a two-phase ensemble architecture. The first phase employs introspective stacking, where six heterogeneous base learners are individually enhanced through algorithm-specific balancing and meta-learning. The second phase fuses these optimized experts via performance-weighted voting. Empirical analysis utilizes a comprehensive dataset of 10,440 Chinese corporate bonds (522 defaults, ~5% default rate) sourced from Wind and CSMAR databases. Given the high cost of both false negatives and false positives in risk assessment, the Geometric Mean (G-mean) and Specificity are employed as primary evaluation metrics. Results demonstrate that the proposed DELC-SMOTE model significantly outperforms individual base classifiers and benchmark ensemble variants, achieving a G-mean of 0.9152 and a Specificity of 0.8715 under the primary experimental setting. The model exhibits robust performance across varying imbalance ratios (2%, 10%, 20%) and strong resilience against data noise, perturbations, and outliers. These findings indicate that the synergistic integration of data-level resampling within a diversified, two-tiered ensemble structure effectively mitigates class imbalance bias and enhances predictive reliability. The framework offers a robust and generalizable tool for actionable default risk assessment in imbalanced financial datasets.</description>
	<pubDate>2026-03-06</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 81: Predicting Bond Defaults in China: A Double-Ensemble Model Leveraging SMOTE for Class Imbalance</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/3/81">doi: 10.3390/bdcc10030081</a></p>
	<p>Authors:
		Chongwen Tian
		Rong Li
		</p>
	<p>This study proposes the Double-Ensemble Learning Classification with SMOTE (DELC-SMOTE), a novel hierarchical framework designed to address the critical challenge of severe class imbalance in financial bond default prediction. The model integrates the Synthetic Minority Over-sampling Technique (SMOTE) into a two-phase ensemble architecture. The first phase employs introspective stacking, where six heterogeneous base learners are individually enhanced through algorithm-specific balancing and meta-learning. The second phase fuses these optimized experts via performance-weighted voting. Empirical analysis utilizes a comprehensive dataset of 10,440 Chinese corporate bonds (522 defaults, ~5% default rate) sourced from Wind and CSMAR databases. Given the high cost of both false negatives and false positives in risk assessment, the Geometric Mean (G-mean) and Specificity are employed as primary evaluation metrics. Results demonstrate that the proposed DELC-SMOTE model significantly outperforms individual base classifiers and benchmark ensemble variants, achieving a G-mean of 0.9152 and a Specificity of 0.8715 under the primary experimental setting. The model exhibits robust performance across varying imbalance ratios (2%, 10%, 20%) and strong resilience against data noise, perturbations, and outliers. These findings indicate that the synergistic integration of data-level resampling within a diversified, two-tiered ensemble structure effectively mitigates class imbalance bias and enhances predictive reliability. The framework offers a robust and generalizable tool for actionable default risk assessment in imbalanced financial datasets.</p>
	]]></content:encoded>

	<dc:title>Predicting Bond Defaults in China: A Double-Ensemble Model Leveraging SMOTE for Class Imbalance</dc:title>
			<dc:creator>Chongwen Tian</dc:creator>
			<dc:creator>Rong Li</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10030081</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-03-06</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-03-06</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>81</prism:startingPage>
		<prism:doi>10.3390/bdcc10030081</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/3/81</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/3/80">

	<title>BDCC, Vol. 10, Pages 80: Effective Flow Ratio: A Novel Efficiency Metric for Heterogeneous Traffic in a Signalized Urban Intersection with Aerial Computer Vision</title>
	<link>https://www.mdpi.com/2504-2289/10/3/80</link>
	<description>Intelligent Transportation Systems (ITS) primarily rely on flow rate and occupancy to estimate traffic states. However, in heterogeneous traffic conditions characterized by weak lane discipline and diverse vehicle classes, these conventional metrics fail to capture the true operational efficiency of signalized intersections. High flow rates can mask underlying inefficiencies, while low flow rates do not necessarily indicate free-flow conditions. This paper introduces a novel computer vision-based metric, the Effective Flow Ratio (EFR), designed to quantify the actual discharge efficiency of mixed traffic. By leveraging Bird&amp;amp;rsquo;s-Eye View (BEV) vehicle tracking using You Only Look Once version 11 (YOLOv11) and ByteTrack, EFR distinguishes between kinematic movement and effective discharge, resolving the ambiguity of &amp;amp;ldquo;moving but not clearing&amp;amp;rdquo; states. We analyze 21 days of continuous footage from a rooftop-mounted camera overlooking a congested intersection in Dhaka, Bangladesh, exhibiting distinct non-linear behaviors compared to raw flow counts. Our results demonstrate that: (i) Flow rate and discharge efficiency are dynamically decoupled, evidenced by significant variance in EFR within identical flow bins; (ii) Temporal rolling correlations reveal transient regimes where traditional signal control logic would misinterpret congestion severity; and (iii) EFR provides a more robust proxy for intersection performance than occupancy or volume alone. The proposed metric offers a granular, physics-informed input for next-generation adaptive traffic signal control in developing urban environments.</description>
	<pubDate>2026-03-06</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 80: Effective Flow Ratio: A Novel Efficiency Metric for Heterogeneous Traffic in a Signalized Urban Intersection with Aerial Computer Vision</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/3/80">doi: 10.3390/bdcc10030080</a></p>
	<p>Authors:
		Abu Anas Ibn Samad
		Tanvir Ahmed
		Md Nazmul Huda
		</p>
	<p>Intelligent Transportation Systems (ITS) primarily rely on flow rate and occupancy to estimate traffic states. However, in heterogeneous traffic conditions characterized by weak lane discipline and diverse vehicle classes, these conventional metrics fail to capture the true operational efficiency of signalized intersections. High flow rates can mask underlying inefficiencies, while low flow rates do not necessarily indicate free-flow conditions. This paper introduces a novel computer vision-based metric, the Effective Flow Ratio (EFR), designed to quantify the actual discharge efficiency of mixed traffic. By leveraging Bird&amp;amp;rsquo;s-Eye View (BEV) vehicle tracking using You Only Look Once version 11 (YOLOv11) and ByteTrack, EFR distinguishes between kinematic movement and effective discharge, resolving the ambiguity of &amp;amp;ldquo;moving but not clearing&amp;amp;rdquo; states. We analyze 21 days of continuous footage from a rooftop-mounted camera overlooking a congested intersection in Dhaka, Bangladesh, exhibiting distinct non-linear behaviors compared to raw flow counts. Our results demonstrate that: (i) Flow rate and discharge efficiency are dynamically decoupled, evidenced by significant variance in EFR within identical flow bins; (ii) Temporal rolling correlations reveal transient regimes where traditional signal control logic would misinterpret congestion severity; and (iii) EFR provides a more robust proxy for intersection performance than occupancy or volume alone. The proposed metric offers a granular, physics-informed input for next-generation adaptive traffic signal control in developing urban environments.</p>
	]]></content:encoded>

	<dc:title>Effective Flow Ratio: A Novel Efficiency Metric for Heterogeneous Traffic in a Signalized Urban Intersection with Aerial Computer Vision</dc:title>
			<dc:creator>Abu Anas Ibn Samad</dc:creator>
			<dc:creator>Tanvir Ahmed</dc:creator>
			<dc:creator>Md Nazmul Huda</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10030080</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-03-06</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-03-06</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>80</prism:startingPage>
		<prism:doi>10.3390/bdcc10030080</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/3/80</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/3/79">

	<title>BDCC, Vol. 10, Pages 79: Feasibility Study of CUDA-Accelerated Homomorphic Encryption and Benchmarking on Consumer-Grade and Embedded GPUs</title>
	<link>https://www.mdpi.com/2504-2289/10/3/79</link>
	<description>Fully Homomorphic Encryption (FHE) provides strong data confidentiality during computation but often suffers from high latency on Central Processing Units (CPUs). This study evaluates Graphics Processing Unit (GPU) acceleration for modern FHE libraries across a laptop (NVIDIA GTX 1650 Ti), a server (NVIDIA RTX 4060), and a Jetson Nano 2 GB embedded GPU. We benchmark key generation, arithmetic operations, Boolean-gate evaluation and scheme-specific tasks such as relinearization and key switching, using library-provided benchmarks with an explicit baseline (operation scope, timing boundaries, and parameter tuples). Moreover, we compare GPU-native libraries (NuFHE, Phantom-FHE, and Troy-Nova) with CPU-oriented ones (Microsoft SEAL, HElib, OpenFHE, Cupcake, and TFHE-rs). Results show GPUs deliver significant speedups for targeted operations. For example, NuFHE&amp;amp;rsquo;s NVIDIA CUDA (Compute Unified Device Architecture) backend achieves about 1.4&amp;amp;times; faster Boolean-gate evaluation on the laptop and 3.4&amp;amp;times; faster on the server compared to its OpenCL backend. Likewise, RLWE (Ring Learning With Errors)-based schemes (BFV, CKKS, and BGV) see marked gains for polynomial arithmetic such as Number Theoretic Transform (NTT) when executed via Phantom-FHE. However, attempts to add CUDA support to Microsoft SEAL reveal four main challenges: high-precision modular arithmetic on GPUs, sequential dependencies in SEAL&amp;amp;rsquo;s design, limited GPU memory and complex build-system changes. In light of these findings, we propose revised guidelines for GPU-first FHE libraries and practical recommendations for deploying high-throughput, privacy-preserving solutions on modern GPUs.</description>
	<pubDate>2026-03-06</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 79: Feasibility Study of CUDA-Accelerated Homomorphic Encryption and Benchmarking on Consumer-Grade and Embedded GPUs</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/3/79">doi: 10.3390/bdcc10030079</a></p>
	<p>Authors:
		Volodymyr Dubetskyy
		Maria-Dolores Cano
		</p>
	<p>Fully Homomorphic Encryption (FHE) provides strong data confidentiality during computation but often suffers from high latency on Central Processing Units (CPUs). This study evaluates Graphics Processing Unit (GPU) acceleration for modern FHE libraries across a laptop (NVIDIA GTX 1650 Ti), a server (NVIDIA RTX 4060), and a Jetson Nano 2 GB embedded GPU. We benchmark key generation, arithmetic operations, Boolean-gate evaluation and scheme-specific tasks such as relinearization and key switching, using library-provided benchmarks with an explicit baseline (operation scope, timing boundaries, and parameter tuples). Moreover, we compare GPU-native libraries (NuFHE, Phantom-FHE, and Troy-Nova) with CPU-oriented ones (Microsoft SEAL, HElib, OpenFHE, Cupcake, and TFHE-rs). Results show GPUs deliver significant speedups for targeted operations. For example, NuFHE&amp;amp;rsquo;s NVIDIA CUDA (Compute Unified Device Architecture) backend achieves about 1.4&amp;amp;times; faster Boolean-gate evaluation on the laptop and 3.4&amp;amp;times; faster on the server compared to its OpenCL backend. Likewise, RLWE (Ring Learning With Errors)-based schemes (BFV, CKKS, and BGV) see marked gains for polynomial arithmetic such as Number Theoretic Transform (NTT) when executed via Phantom-FHE. However, attempts to add CUDA support to Microsoft SEAL reveal four main challenges: high-precision modular arithmetic on GPUs, sequential dependencies in SEAL&amp;amp;rsquo;s design, limited GPU memory and complex build-system changes. In light of these findings, we propose revised guidelines for GPU-first FHE libraries and practical recommendations for deploying high-throughput, privacy-preserving solutions on modern GPUs.</p>
	]]></content:encoded>

	<dc:title>Feasibility Study of CUDA-Accelerated Homomorphic Encryption and Benchmarking on Consumer-Grade and Embedded GPUs</dc:title>
			<dc:creator>Volodymyr Dubetskyy</dc:creator>
			<dc:creator>Maria-Dolores Cano</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10030079</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-03-06</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-03-06</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>79</prism:startingPage>
		<prism:doi>10.3390/bdcc10030079</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/3/79</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/3/78">

	<title>BDCC, Vol. 10, Pages 78: Kernel VICReg for Self-Supervised Learning in Reproducing Kernel Hilbert Space</title>
	<link>https://www.mdpi.com/2504-2289/10/3/78</link>
	<description>Self-supervised learning (SSL) has emerged as a powerful paradigm for representation learning by optimizing geometric objectives, such as invariance to augmentations, variance preservation, and feature decorrelation, without requiring labels. However, most existing methods operate in Euclidean space, limiting their ability to capture nonlinear dependencies and geometric structures. In this work, we propose Kernel VICReg, a novel self-supervised learning framework that pulls the VICReg objective into a Reproducing Kernel Hilbert Space (RKHS). By kernelizing each term of the loss, variance, invariance, and covariance, we obtain a general formulation that operates on double-centered kernel matrices and Hilbert&amp;amp;ndash;Schmidt norms, enabling nonlinear feature learning without explicit mappings. We demonstrate that Kernel VICReg mitigates the risk of representational collapse under challenging conditions and improves performance on datasets exhibiting nonlinear structure or limited sample regimes. Empirical evaluations across MNIST, CIFAR-10, STL-10, TinyImageNet, and ImageNet100 show consistent gains over Euclidean VICReg, with particularly strong improvements on datasets where nonlinear structures are prominent. UMAP visualizations are provided only as a qualitative illustration of embedding geometry and are not used as a calibration or statistical validation. Our results suggest that kernelizing SSL objectives is a promising direction for bridging classical kernel methods with modern representation learning.</description>
	<pubDate>2026-03-05</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 78: Kernel VICReg for Self-Supervised Learning in Reproducing Kernel Hilbert Space</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/3/78">doi: 10.3390/bdcc10030078</a></p>
	<p>Authors:
		M. Hadi Sepanj
		Benyamin Ghojogh
		Saed Moradi
		Paul Fieguth
		</p>
	<p>Self-supervised learning (SSL) has emerged as a powerful paradigm for representation learning by optimizing geometric objectives, such as invariance to augmentations, variance preservation, and feature decorrelation, without requiring labels. However, most existing methods operate in Euclidean space, limiting their ability to capture nonlinear dependencies and geometric structures. In this work, we propose Kernel VICReg, a novel self-supervised learning framework that pulls the VICReg objective into a Reproducing Kernel Hilbert Space (RKHS). By kernelizing each term of the loss, variance, invariance, and covariance, we obtain a general formulation that operates on double-centered kernel matrices and Hilbert&amp;amp;ndash;Schmidt norms, enabling nonlinear feature learning without explicit mappings. We demonstrate that Kernel VICReg mitigates the risk of representational collapse under challenging conditions and improves performance on datasets exhibiting nonlinear structure or limited sample regimes. Empirical evaluations across MNIST, CIFAR-10, STL-10, TinyImageNet, and ImageNet100 show consistent gains over Euclidean VICReg, with particularly strong improvements on datasets where nonlinear structures are prominent. UMAP visualizations are provided only as a qualitative illustration of embedding geometry and are not used as a calibration or statistical validation. Our results suggest that kernelizing SSL objectives is a promising direction for bridging classical kernel methods with modern representation learning.</p>
	]]></content:encoded>

	<dc:title>Kernel VICReg for Self-Supervised Learning in Reproducing Kernel Hilbert Space</dc:title>
			<dc:creator>M. Hadi Sepanj</dc:creator>
			<dc:creator>Benyamin Ghojogh</dc:creator>
			<dc:creator>Saed Moradi</dc:creator>
			<dc:creator>Paul Fieguth</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10030078</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-03-05</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-03-05</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>78</prism:startingPage>
		<prism:doi>10.3390/bdcc10030078</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/3/78</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/3/77">

	<title>BDCC, Vol. 10, Pages 77: Exploring Public Health Perspectives on Travel Behavior Using a Machine Learning Approach: Thailand Case Study</title>
	<link>https://www.mdpi.com/2504-2289/10/3/77</link>
	<description>Hospital transport services represent a vital alternative for addressing inequities in access to medical care, particularly in countries where public transportation systems are inadequate, such as Thailand. This approach enables equitable and widespread access to healthcare services for residents in underserved areas. The objective of this study is to analyze the factors influencing the choice of hospital transport travel mode by comparing various machine learning algorithms. The findings reveal that the categorical boosting model outperformed the other models across all performance metrics. The model results indicate that waiting time, travel time, travel cost, and comfortability significantly influence the decision to use hospital transport services. Furthermore, demographic data analysis highlights critical factors such as age, gender, income, travel frequency, occupation, and time of travel, all of which significantly affect the choice of hospital transport service. To maximize the practical implications of this study, policy recommendations and implementation strategies are proposed to support decision-makers in promoting equitable travel options and eliminating barriers to fair access to healthcare services.</description>
	<pubDate>2026-03-05</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 77: Exploring Public Health Perspectives on Travel Behavior Using a Machine Learning Approach: Thailand Case Study</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/3/77">doi: 10.3390/bdcc10030077</a></p>
	<p>Authors:
		Manlika Seefong
		Panuwat Wisutwattanasak
		Kestsirin Theerathitichaipa
		Pattarawadee Prasomsab
		Nisa Dackuntod
		Thanapong Champahom
		Rattanaporn Kasemsri
		</p>
	<p>Hospital transport services represent a vital alternative for addressing inequities in access to medical care, particularly in countries where public transportation systems are inadequate, such as Thailand. This approach enables equitable and widespread access to healthcare services for residents in underserved areas. The objective of this study is to analyze the factors influencing the choice of hospital transport travel mode by comparing various machine learning algorithms. The findings reveal that the categorical boosting model outperformed the other models across all performance metrics. The model results indicate that waiting time, travel time, travel cost, and comfortability significantly influence the decision to use hospital transport services. Furthermore, demographic data analysis highlights critical factors such as age, gender, income, travel frequency, occupation, and time of travel, all of which significantly affect the choice of hospital transport service. To maximize the practical implications of this study, policy recommendations and implementation strategies are proposed to support decision-makers in promoting equitable travel options and eliminating barriers to fair access to healthcare services.</p>
	]]></content:encoded>

	<dc:title>Exploring Public Health Perspectives on Travel Behavior Using a Machine Learning Approach: Thailand Case Study</dc:title>
			<dc:creator>Manlika Seefong</dc:creator>
			<dc:creator>Panuwat Wisutwattanasak</dc:creator>
			<dc:creator>Kestsirin Theerathitichaipa</dc:creator>
			<dc:creator>Pattarawadee Prasomsab</dc:creator>
			<dc:creator>Nisa Dackuntod</dc:creator>
			<dc:creator>Thanapong Champahom</dc:creator>
			<dc:creator>Rattanaporn Kasemsri</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10030077</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-03-05</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-03-05</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>77</prism:startingPage>
		<prism:doi>10.3390/bdcc10030077</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/3/77</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/3/76">

	<title>BDCC, Vol. 10, Pages 76: A Systematic Review of Cross-Population Shifts in Medical Imaging Analysis with Deep Learning</title>
	<link>https://www.mdpi.com/2504-2289/10/3/76</link>
	<description>Deep learning has achieved expert-level performance in medical imaging analysis. However, models often fail to generalize across patient populations due to cross-population domain shifts, distributional differences arising from demographic variability, variations in imaging protocols, scanner hardware, and differences in disease prevalence. This challenge limits the real-world deployment and can increase health inequities. This review systematically examines the nature, causes, and impact of cross-population domain shift in deep learning-based medical imaging analysis. We analyzed 50 peer-reviewed studies from 2020 to 2025, evaluating the proposed methodologies for handling population shifts, the datasets employed, and the metrics used to assess performance. Our findings demonstrate that performance degradation ranged from 10&amp;amp;ndash;25% when models were tested on unseen populations, emphasizing the substantial impact of domain shifts on model generalizability. The literature reveals that mitigation strategies broadly fall into two categories: data-centric approaches, such as augmentation and harmonization, and model-centric approaches, including domain adaptation, transfer learning, adversarial learning, multi-task learning, and continual learning. While domain adaptation and transfer learning are the most widely used, their performance gains across populations remain modest, ranging from 5&amp;amp;ndash;15%, and are not supported by external validation. Our synthesis reveals a significant reliance on large, publicly available datasets from limited regions, with an underrepresentation of data from low- and middle-income countries. Evaluation practices are inconsistent, with few studies employing standardized external test sets. This review provides a structured taxonomy of mitigation techniques, a refined analysis of domain shift characteristics, and an in-depth critique of methodological challenges. We highlight the urgent need for more geographically and demographically inclusive datasets, adaptable modeling techniques, and standardized evaluation protocols to enable accurate and equitable AI-driven diagnostics across diverse populations. Finally, we outline future research directions to guide the development of robust, generalizable, and fair models for medical imaging analysis.</description>
	<pubDate>2026-03-04</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 76: A Systematic Review of Cross-Population Shifts in Medical Imaging Analysis with Deep Learning</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/3/76">doi: 10.3390/bdcc10030076</a></p>
	<p>Authors:
		Aminu Musa
		Rajesh Prasad
		Peter Onwualu
		Monica Hernandez
		</p>
	<p>Deep learning has achieved expert-level performance in medical imaging analysis. However, models often fail to generalize across patient populations due to cross-population domain shifts, distributional differences arising from demographic variability, variations in imaging protocols, scanner hardware, and differences in disease prevalence. This challenge limits the real-world deployment and can increase health inequities. This review systematically examines the nature, causes, and impact of cross-population domain shift in deep learning-based medical imaging analysis. We analyzed 50 peer-reviewed studies from 2020 to 2025, evaluating the proposed methodologies for handling population shifts, the datasets employed, and the metrics used to assess performance. Our findings demonstrate that performance degradation ranged from 10&amp;amp;ndash;25% when models were tested on unseen populations, emphasizing the substantial impact of domain shifts on model generalizability. The literature reveals that mitigation strategies broadly fall into two categories: data-centric approaches, such as augmentation and harmonization, and model-centric approaches, including domain adaptation, transfer learning, adversarial learning, multi-task learning, and continual learning. While domain adaptation and transfer learning are the most widely used, their performance gains across populations remain modest, ranging from 5&amp;amp;ndash;15%, and are not supported by external validation. Our synthesis reveals a significant reliance on large, publicly available datasets from limited regions, with an underrepresentation of data from low- and middle-income countries. Evaluation practices are inconsistent, with few studies employing standardized external test sets. This review provides a structured taxonomy of mitigation techniques, a refined analysis of domain shift characteristics, and an in-depth critique of methodological challenges. We highlight the urgent need for more geographically and demographically inclusive datasets, adaptable modeling techniques, and standardized evaluation protocols to enable accurate and equitable AI-driven diagnostics across diverse populations. Finally, we outline future research directions to guide the development of robust, generalizable, and fair models for medical imaging analysis.</p>
	]]></content:encoded>

	<dc:title>A Systematic Review of Cross-Population Shifts in Medical Imaging Analysis with Deep Learning</dc:title>
			<dc:creator>Aminu Musa</dc:creator>
			<dc:creator>Rajesh Prasad</dc:creator>
			<dc:creator>Peter Onwualu</dc:creator>
			<dc:creator>Monica Hernandez</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10030076</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-03-04</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-03-04</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Systematic Review</prism:section>
	<prism:startingPage>76</prism:startingPage>
		<prism:doi>10.3390/bdcc10030076</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/3/76</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/3/75">

	<title>BDCC, Vol. 10, Pages 75: Evaluating Architecture Scalability and Transfer Learning in Urban Scene Segmentation Using Explainable AI</title>
	<link>https://www.mdpi.com/2504-2289/10/3/75</link>
	<description>Semantic segmentation plays a pivotal role in autonomous driving, enabling pixel-level understanding of road scenes. Although transformer-based models such as SegFormer have shown exceptional performance on large datasets, their generalization to smaller and geographically diverse datasets remains underexplored. In this work, we analyze the scalability and transferability of SegFormer variants (B3, B4, B5) using CamVid as the base dataset. We perform cross-dataset transfer learning to KITTI and IDD, evaluate class-level performance, and explore explainable AI via confidence heatmaps. Our findings show that SegFormer-B5 achieves the highest accuracy (82.4% mIoU) on CamVid, while transfer learning from CamVid improves mIoU on KITTI by 2.57% and enhances class-specific predictions in IDD by over 70%. These results highlight the practical potential of SegFormer in real-world segmentation systems and the interpretability benefits of confidence-based visual analysis.</description>
	<pubDate>2026-03-01</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 75: Evaluating Architecture Scalability and Transfer Learning in Urban Scene Segmentation Using Explainable AI</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/3/75">doi: 10.3390/bdcc10030075</a></p>
	<p>Authors:
		Tanmay Sunil Hatkar
		Abhinav Pandey
		Saad B. Ahmed
		</p>
	<p>Semantic segmentation plays a pivotal role in autonomous driving, enabling pixel-level understanding of road scenes. Although transformer-based models such as SegFormer have shown exceptional performance on large datasets, their generalization to smaller and geographically diverse datasets remains underexplored. In this work, we analyze the scalability and transferability of SegFormer variants (B3, B4, B5) using CamVid as the base dataset. We perform cross-dataset transfer learning to KITTI and IDD, evaluate class-level performance, and explore explainable AI via confidence heatmaps. Our findings show that SegFormer-B5 achieves the highest accuracy (82.4% mIoU) on CamVid, while transfer learning from CamVid improves mIoU on KITTI by 2.57% and enhances class-specific predictions in IDD by over 70%. These results highlight the practical potential of SegFormer in real-world segmentation systems and the interpretability benefits of confidence-based visual analysis.</p>
	]]></content:encoded>

	<dc:title>Evaluating Architecture Scalability and Transfer Learning in Urban Scene Segmentation Using Explainable AI</dc:title>
			<dc:creator>Tanmay Sunil Hatkar</dc:creator>
			<dc:creator>Abhinav Pandey</dc:creator>
			<dc:creator>Saad B. Ahmed</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10030075</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-03-01</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-03-01</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>75</prism:startingPage>
		<prism:doi>10.3390/bdcc10030075</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/3/75</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/3/74">

	<title>BDCC, Vol. 10, Pages 74: Data-Driven Ergonomic Load Dynamics for Human&amp;ndash;Autonomy Teams</title>
	<link>https://www.mdpi.com/2504-2289/10/3/74</link>
	<description>Ergonomic load in human&amp;amp;ndash;autonomy teams is commonly treated as a static score or a post-hoc audit, even though modern sensing and communication enable real-time regulation of operator effort. We model ergonomic load as a dissipative dynamical state inferred online from multimodal effort proxies and task context, and couple it to autonomy through load-dependent gain moderation and compliance shaping. The method is evaluated on public human&amp;amp;ndash;swarm and human&amp;amp;ndash;robot interaction traces together with effort-proximal wearable and myographic datasets using a unified, windowed pipeline and controlled stress tests that emulate latency, downsampling, packet loss, and channel dropouts. On a large human&amp;amp;ndash;swarm benchmark, the estimator achieves strong discrimination and calibration for rare high-load events (up to AUROC 0.87, AUPRC 0.41, ECE 0.031 at q=0.90) and degrades predictably under delay, with a knee around 300&amp;amp;ndash;400ms (AUROC 0.87&amp;amp;rarr;0.80, ECE 0.031&amp;amp;rarr;0.061 at 500ms). Embedding the estimate in the adaptation schedule reduces overload incidence and oscillatory redistribution while preserving coordination proxies in surrogate closed-loop simulation: overload time drops from 7.8% to 4.1% (relative reduction &amp;amp;asymp;&amp;amp;nbsp;47%) with throughput maintained near baseline (1.00&amp;amp;rarr;0.97) and oscillation power reduced (0.26&amp;amp;rarr;0.14) under nominal timing. These results provide a reproducible pathway for making ergonomics a control-relevant feedback signal, together with explicit operational constraints on estimator calibration (target ECE &amp;amp;le;0.05) and end-to-end latency (effective &amp;amp;tau;&amp;amp;le;300ms) required to avoid regime switching and maintain stable, interpretable adaptation.</description>
	<pubDate>2026-02-28</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 74: Data-Driven Ergonomic Load Dynamics for Human&amp;ndash;Autonomy Teams</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/3/74">doi: 10.3390/bdcc10030074</a></p>
	<p>Authors:
		Nikitas Gerolimos
		Vasileios Alevizos
		Georgios Priniotakis
		</p>
	<p>Ergonomic load in human&amp;amp;ndash;autonomy teams is commonly treated as a static score or a post-hoc audit, even though modern sensing and communication enable real-time regulation of operator effort. We model ergonomic load as a dissipative dynamical state inferred online from multimodal effort proxies and task context, and couple it to autonomy through load-dependent gain moderation and compliance shaping. The method is evaluated on public human&amp;amp;ndash;swarm and human&amp;amp;ndash;robot interaction traces together with effort-proximal wearable and myographic datasets using a unified, windowed pipeline and controlled stress tests that emulate latency, downsampling, packet loss, and channel dropouts. On a large human&amp;amp;ndash;swarm benchmark, the estimator achieves strong discrimination and calibration for rare high-load events (up to AUROC 0.87, AUPRC 0.41, ECE 0.031 at q=0.90) and degrades predictably under delay, with a knee around 300&amp;amp;ndash;400ms (AUROC 0.87&amp;amp;rarr;0.80, ECE 0.031&amp;amp;rarr;0.061 at 500ms). Embedding the estimate in the adaptation schedule reduces overload incidence and oscillatory redistribution while preserving coordination proxies in surrogate closed-loop simulation: overload time drops from 7.8% to 4.1% (relative reduction &amp;amp;asymp;&amp;amp;nbsp;47%) with throughput maintained near baseline (1.00&amp;amp;rarr;0.97) and oscillation power reduced (0.26&amp;amp;rarr;0.14) under nominal timing. These results provide a reproducible pathway for making ergonomics a control-relevant feedback signal, together with explicit operational constraints on estimator calibration (target ECE &amp;amp;le;0.05) and end-to-end latency (effective &amp;amp;tau;&amp;amp;le;300ms) required to avoid regime switching and maintain stable, interpretable adaptation.</p>
	]]></content:encoded>

	<dc:title>Data-Driven Ergonomic Load Dynamics for Human&amp;amp;ndash;Autonomy Teams</dc:title>
			<dc:creator>Nikitas Gerolimos</dc:creator>
			<dc:creator>Vasileios Alevizos</dc:creator>
			<dc:creator>Georgios Priniotakis</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10030074</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-02-28</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-02-28</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>74</prism:startingPage>
		<prism:doi>10.3390/bdcc10030074</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/3/74</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/3/73">

	<title>BDCC, Vol. 10, Pages 73: Building Prototype Evolution Pathway for Emotion Recognition in User-Generated Videos</title>
	<link>https://www.mdpi.com/2504-2289/10/3/73</link>
	<description>Large-scale pretrained foundation models are increasingly essential for affective analysis in user-generated videos. However, current approaches typically reuse generic multi-modal representations directly with task-specific adapters learned from scratch, and their performance is limited by the large affective domain gap and scarce emotion annotations. To address these issues, we introduce a novel paradigm that leverages auxiliary cross-modal priors to enhance unimodal emotion modeling, effectively exploiting modality-shared semantics and modality-specific inductive biases. Specifically, we propose a progressive prototype evolution framework that gradually transforms a neutral prototype into discriminative emotional representations through fine-grained cross-modal interactions with visual cues. The auxiliary prior serves as a structural constraint, reframing the adaptation challenge from a difficult domain shift problem into a more tractable prototype shift within the affective space. To ensure robust prototype construction and guided evolution, we further design category-aggregated prompting and bidirectional supervision mechanisms. Extensive experiments on VideoEmotion-8, Ekman-6, and MusicVideo-6 validate the superiority of our approach, achieving state-of-the-art results and demonstrating the effectiveness of leveraging auxiliary modality priors for foundation-model-based emotion recognition.</description>
	<pubDate>2026-02-28</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 73: Building Prototype Evolution Pathway for Emotion Recognition in User-Generated Videos</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/3/73">doi: 10.3390/bdcc10030073</a></p>
	<p>Authors:
		Yujie Liu
		Zhenyang Dong
		Yante Li
		Guoying Zhao
		</p>
	<p>Large-scale pretrained foundation models are increasingly essential for affective analysis in user-generated videos. However, current approaches typically reuse generic multi-modal representations directly with task-specific adapters learned from scratch, and their performance is limited by the large affective domain gap and scarce emotion annotations. To address these issues, we introduce a novel paradigm that leverages auxiliary cross-modal priors to enhance unimodal emotion modeling, effectively exploiting modality-shared semantics and modality-specific inductive biases. Specifically, we propose a progressive prototype evolution framework that gradually transforms a neutral prototype into discriminative emotional representations through fine-grained cross-modal interactions with visual cues. The auxiliary prior serves as a structural constraint, reframing the adaptation challenge from a difficult domain shift problem into a more tractable prototype shift within the affective space. To ensure robust prototype construction and guided evolution, we further design category-aggregated prompting and bidirectional supervision mechanisms. Extensive experiments on VideoEmotion-8, Ekman-6, and MusicVideo-6 validate the superiority of our approach, achieving state-of-the-art results and demonstrating the effectiveness of leveraging auxiliary modality priors for foundation-model-based emotion recognition.</p>
	]]></content:encoded>

	<dc:title>Building Prototype Evolution Pathway for Emotion Recognition in User-Generated Videos</dc:title>
			<dc:creator>Yujie Liu</dc:creator>
			<dc:creator>Zhenyang Dong</dc:creator>
			<dc:creator>Yante Li</dc:creator>
			<dc:creator>Guoying Zhao</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10030073</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-02-28</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-02-28</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>73</prism:startingPage>
		<prism:doi>10.3390/bdcc10030073</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/3/73</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/3/72">

	<title>BDCC, Vol. 10, Pages 72: Automating Data Product Discovery with Large Language Models and Metadata Reasoning</title>
	<link>https://www.mdpi.com/2504-2289/10/3/72</link>
	<description>The exponential growth of data over the past decade has created new challenges in transforming raw information into actionable knowledge, particularly through the development of data products. The latter is essentially the result of querying and retrieving specific portions of data from a data storage architecture at various levels of granularity. Traditionally, this transformation depends on domain experts manually analyzing datasets and providing feedback to effectively describe or annotate data that facilitates data retrieval. Nevertheless, this is a very time-consuming process that highlights the need for its potential automation. To address this challenge, the present paper proposes a framework which utilizes Large Language Models to support data product discovery through semantic metadata reasoning and executable query prototyping. The framework is evaluated across two domains and three levels of concept complexity to assess the LLM&amp;amp;rsquo;s ability to identify relevant datasets and generate executable data product queries under varying analytical demands. The findings indicate that LLMs perform effectively in simpler scenarios, but their performance declines as conceptual complexity and dataset volume increase.</description>
	<pubDate>2026-02-28</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 72: Automating Data Product Discovery with Large Language Models and Metadata Reasoning</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/3/72">doi: 10.3390/bdcc10030072</a></p>
	<p>Authors:
		Michalis Pingos
		Artemis Photiou
		Andreas S. Andreou
		</p>
	<p>The exponential growth of data over the past decade has created new challenges in transforming raw information into actionable knowledge, particularly through the development of data products. The latter is essentially the result of querying and retrieving specific portions of data from a data storage architecture at various levels of granularity. Traditionally, this transformation depends on domain experts manually analyzing datasets and providing feedback to effectively describe or annotate data that facilitates data retrieval. Nevertheless, this is a very time-consuming process that highlights the need for its potential automation. To address this challenge, the present paper proposes a framework which utilizes Large Language Models to support data product discovery through semantic metadata reasoning and executable query prototyping. The framework is evaluated across two domains and three levels of concept complexity to assess the LLM&amp;amp;rsquo;s ability to identify relevant datasets and generate executable data product queries under varying analytical demands. The findings indicate that LLMs perform effectively in simpler scenarios, but their performance declines as conceptual complexity and dataset volume increase.</p>
	]]></content:encoded>

	<dc:title>Automating Data Product Discovery with Large Language Models and Metadata Reasoning</dc:title>
			<dc:creator>Michalis Pingos</dc:creator>
			<dc:creator>Artemis Photiou</dc:creator>
			<dc:creator>Andreas S. Andreou</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10030072</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-02-28</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-02-28</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>72</prism:startingPage>
		<prism:doi>10.3390/bdcc10030072</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/3/72</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/3/71">

	<title>BDCC, Vol. 10, Pages 71: A Convergent Method for Energy Optimization in Modern Hopfield Networks</title>
	<link>https://www.mdpi.com/2504-2289/10/3/71</link>
	<description>Modern Hopfield networks are energy-based associative memory models whose performance critically depends on the structure and optimization of their energy functions. While recent formulations substantially improve storage capacity, the resulting non-convex energy landscapes are often optimized using heuristic update rules that can be sensitive to initialization and may not provide monotonic energy descent or rigorous convergence guarantees. In this work, we propose a new energy formulation for modern Hopfield networks together with a principled iterative optimization scheme. The proposed energy admits a natural decomposition that allows optimization via the concave&amp;amp;ndash;convex procedure (CCCP), yielding well-defined network dynamics with guaranteed energy descent beyond classical Hopfield updates. We establish fundamental theoretical properties of the proposed framework, including non-negativity, boundedness, and monotonic decrease in the energy along iterations. In particular, we prove that the induced dynamics converge to a stationary point of the energy function, providing explicit convergence guarantees for the resulting Hopfield-type model. We further evaluate the proposed approach on synthetic classification tasks and compare its optimization behavior with that of the original Hopfield network and several standard machine learning baselines. Experimental results demonstrate improved stability, convergence behavior, and competitive classification performance. We also validate the approach on real-world benchmark datasets to demonstrate utility beyond controlled experiments. Overall, this work provides a theoretically grounded energy-based optimization framework for modern Hopfield networks, clarifying the role of principled optimization in achieving stable and convergent associative memory dynamics.</description>
	<pubDate>2026-02-28</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 71: A Convergent Method for Energy Optimization in Modern Hopfield Networks</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/3/71">doi: 10.3390/bdcc10030071</a></p>
	<p>Authors:
		Yida Bao
		Mohammad Arifuzzaman
		Tran Duc Le
		Tao Jiang
		Jing Hou
		Yuan Xing
		Dongfang Hou
		</p>
	<p>Modern Hopfield networks are energy-based associative memory models whose performance critically depends on the structure and optimization of their energy functions. While recent formulations substantially improve storage capacity, the resulting non-convex energy landscapes are often optimized using heuristic update rules that can be sensitive to initialization and may not provide monotonic energy descent or rigorous convergence guarantees. In this work, we propose a new energy formulation for modern Hopfield networks together with a principled iterative optimization scheme. The proposed energy admits a natural decomposition that allows optimization via the concave&amp;amp;ndash;convex procedure (CCCP), yielding well-defined network dynamics with guaranteed energy descent beyond classical Hopfield updates. We establish fundamental theoretical properties of the proposed framework, including non-negativity, boundedness, and monotonic decrease in the energy along iterations. In particular, we prove that the induced dynamics converge to a stationary point of the energy function, providing explicit convergence guarantees for the resulting Hopfield-type model. We further evaluate the proposed approach on synthetic classification tasks and compare its optimization behavior with that of the original Hopfield network and several standard machine learning baselines. Experimental results demonstrate improved stability, convergence behavior, and competitive classification performance. We also validate the approach on real-world benchmark datasets to demonstrate utility beyond controlled experiments. Overall, this work provides a theoretically grounded energy-based optimization framework for modern Hopfield networks, clarifying the role of principled optimization in achieving stable and convergent associative memory dynamics.</p>
	]]></content:encoded>

	<dc:title>A Convergent Method for Energy Optimization in Modern Hopfield Networks</dc:title>
			<dc:creator>Yida Bao</dc:creator>
			<dc:creator>Mohammad Arifuzzaman</dc:creator>
			<dc:creator>Tran Duc Le</dc:creator>
			<dc:creator>Tao Jiang</dc:creator>
			<dc:creator>Jing Hou</dc:creator>
			<dc:creator>Yuan Xing</dc:creator>
			<dc:creator>Dongfang Hou</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10030071</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-02-28</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-02-28</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>71</prism:startingPage>
		<prism:doi>10.3390/bdcc10030071</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/3/71</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/3/70">

	<title>BDCC, Vol. 10, Pages 70: TransGoT: Structured Graph-of-Thoughts Reasoning for Machine Translation with Large Language Models</title>
	<link>https://www.mdpi.com/2504-2289/10/3/70</link>
	<description>Machine translation with large language models has recently attracted growing attention due to its flexibility and strong zero-shot and few-shot capabilities. However, most prompt-based LLM translation methods rely on linear generation or shallow self-refinement, implicitly committing to a single reasoning path. Such designs are brittle when translating long and syntactically complex sources, where reliable translation often requires structured planning and hypothesis exploration. In this paper, we propose TransGoT, a novel machine translation framework inspired by the graph-of-thoughts paradigm, which formulates translation as a structured, multi-stage reasoning process over a graph of intermediate thoughts. TransGoT explicitly decomposes translation into constraint identification, draft generation, and culture- and style-aware refinement, enabling systematic exploration and aggregation of alternative translation hypotheses. To better adapt graph-based reasoning to translation, we design two key mechanisms: (1) Uncertainty-driven thought transformation. Unlike general reasoning tasks, translation uncertainty is often localized and unevenly distributed across tokens, making holistic regeneration inefficient. We therefore design uncertainty-driven thought transformation, which leverages model-internal confidence signals to guide targeted token-level revision; (2) Dispersion-adaptive thought scoring. It emphasizes evaluation criteria with stronger inter-candidate variance to enable robust multi-criteria thought selection. We evaluate TransGoT on the WMT22 benchmarks and experimental results demonstrate that TransGoT consistently outperforms strong LLM-based translation baselines, validating the effectiveness of structured graph-based reasoning for machine translation.</description>
	<pubDate>2026-02-27</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 70: TransGoT: Structured Graph-of-Thoughts Reasoning for Machine Translation with Large Language Models</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/3/70">doi: 10.3390/bdcc10030070</a></p>
	<p>Authors:
		Danying Zhang
		Yixin Liu
		Jie Zhao
		Cai Xu
		</p>
	<p>Machine translation with large language models has recently attracted growing attention due to its flexibility and strong zero-shot and few-shot capabilities. However, most prompt-based LLM translation methods rely on linear generation or shallow self-refinement, implicitly committing to a single reasoning path. Such designs are brittle when translating long and syntactically complex sources, where reliable translation often requires structured planning and hypothesis exploration. In this paper, we propose TransGoT, a novel machine translation framework inspired by the graph-of-thoughts paradigm, which formulates translation as a structured, multi-stage reasoning process over a graph of intermediate thoughts. TransGoT explicitly decomposes translation into constraint identification, draft generation, and culture- and style-aware refinement, enabling systematic exploration and aggregation of alternative translation hypotheses. To better adapt graph-based reasoning to translation, we design two key mechanisms: (1) Uncertainty-driven thought transformation. Unlike general reasoning tasks, translation uncertainty is often localized and unevenly distributed across tokens, making holistic regeneration inefficient. We therefore design uncertainty-driven thought transformation, which leverages model-internal confidence signals to guide targeted token-level revision; (2) Dispersion-adaptive thought scoring. It emphasizes evaluation criteria with stronger inter-candidate variance to enable robust multi-criteria thought selection. We evaluate TransGoT on the WMT22 benchmarks and experimental results demonstrate that TransGoT consistently outperforms strong LLM-based translation baselines, validating the effectiveness of structured graph-based reasoning for machine translation.</p>
	]]></content:encoded>

	<dc:title>TransGoT: Structured Graph-of-Thoughts Reasoning for Machine Translation with Large Language Models</dc:title>
			<dc:creator>Danying Zhang</dc:creator>
			<dc:creator>Yixin Liu</dc:creator>
			<dc:creator>Jie Zhao</dc:creator>
			<dc:creator>Cai Xu</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10030070</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-02-27</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-02-27</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>70</prism:startingPage>
		<prism:doi>10.3390/bdcc10030070</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/3/70</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/3/69">

	<title>BDCC, Vol. 10, Pages 69: Generative AI as a General-Purpose Technology: Foundations, Applications, and Labor Market Implications Through 2030</title>
	<link>https://www.mdpi.com/2504-2289/10/3/69</link>
	<description>Generative Artificial Intelligence (AI) has transitioned from a research milestone to a general-purpose technology with wide-ranging implications for organizations, labor markets, and information systems. Thanks to improvements in deep learning, generative adversarial networks (GANs), variational autoencoders (VAEs), diffusion models, transformer-based language models, and reinforcement learning from human feedback (RLHF), generative AI can now create high-quality text, images, audio, code, and other types of content. This review synthesizes the core technical foundations and best practices for training, evaluation, and governance, with an emphasis on scalability and human oversight. The paper examines applications across customer service, marketing, software development, healthcare, finance, law, logistics, and the creative industries, and assesses the labor implications of generative AI using a sociotechnical lens. This study also develops a disruption index that integrates task exposure, adoption rates, time savings, and skill complementarity. The paper concludes with actionable recommendations for policymakers, organizations, and workers, emphasizing the importance of reskilling, algorithmic transparency, and inclusive innovation. Taken together, these contributions situate generative AI within broader debates about automation, augmentation, and the future of work.</description>
	<pubDate>2026-02-27</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 69: Generative AI as a General-Purpose Technology: Foundations, Applications, and Labor Market Implications Through 2030</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/3/69">doi: 10.3390/bdcc10030069</a></p>
	<p>Authors:
		Maikel Leon
		</p>
	<p>Generative Artificial Intelligence (AI) has transitioned from a research milestone to a general-purpose technology with wide-ranging implications for organizations, labor markets, and information systems. Thanks to improvements in deep learning, generative adversarial networks (GANs), variational autoencoders (VAEs), diffusion models, transformer-based language models, and reinforcement learning from human feedback (RLHF), generative AI can now create high-quality text, images, audio, code, and other types of content. This review synthesizes the core technical foundations and best practices for training, evaluation, and governance, with an emphasis on scalability and human oversight. The paper examines applications across customer service, marketing, software development, healthcare, finance, law, logistics, and the creative industries, and assesses the labor implications of generative AI using a sociotechnical lens. This study also develops a disruption index that integrates task exposure, adoption rates, time savings, and skill complementarity. The paper concludes with actionable recommendations for policymakers, organizations, and workers, emphasizing the importance of reskilling, algorithmic transparency, and inclusive innovation. Taken together, these contributions situate generative AI within broader debates about automation, augmentation, and the future of work.</p>
	]]></content:encoded>

	<dc:title>Generative AI as a General-Purpose Technology: Foundations, Applications, and Labor Market Implications Through 2030</dc:title>
			<dc:creator>Maikel Leon</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10030069</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-02-27</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-02-27</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Review</prism:section>
	<prism:startingPage>69</prism:startingPage>
		<prism:doi>10.3390/bdcc10030069</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/3/69</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/3/68">

	<title>BDCC, Vol. 10, Pages 68: An Intelligent Simulation Training System for Power Grid Control and Operations</title>
	<link>https://www.mdpi.com/2504-2289/10/3/68</link>
	<description>With the increasing complexity of power grid operations, operator training requires timely feedback and objective assessment. Traditional approaches based on lectures and scripted simulations provide limited personalization and weak explainability. This paper presents AI Instructors, an intelligent simulation training system for power-grid control and dispatching. The system is organized into learning, training, assessment, and analysis modules, and is built around two core technical components: (i) parameterized item generation from rule/knowledge bases using a phrase-enhanced transformer (PET), and (ii) solver-grounded, topology-aware grading with hierarchical feedback for both numeric and free-text responses. A voice interaction module is integrated to simulate telephone-based dispatch orders. We validate the system through a pilot deployment with licensed dispatch operators and scenario experiments on benchmark cases. Compared with a conventional scripted DTS workflow, AI Instructors achieves higher stepwise procedure accuracy (68%&amp;amp;rarr;90%), a lower topology-violation rate (32%&amp;amp;rarr;11%), and shorter response time (120 s&amp;amp;rarr;72 s), while increasing the proportion of parameterized questions and accelerating skill acquisition. These results suggest that combining adaptive sequencing with topology-safe, explainable evaluation can improve training effectiveness and operational safety.</description>
	<pubDate>2026-02-27</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 68: An Intelligent Simulation Training System for Power Grid Control and Operations</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/3/68">doi: 10.3390/bdcc10030068</a></p>
	<p>Authors:
		Sheng Yang
		Shengyuan Li
		Yuan Fu
		Wei Jiang
		Wenlong You
		Min Chen
		</p>
	<p>With the increasing complexity of power grid operations, operator training requires timely feedback and objective assessment. Traditional approaches based on lectures and scripted simulations provide limited personalization and weak explainability. This paper presents AI Instructors, an intelligent simulation training system for power-grid control and dispatching. The system is organized into learning, training, assessment, and analysis modules, and is built around two core technical components: (i) parameterized item generation from rule/knowledge bases using a phrase-enhanced transformer (PET), and (ii) solver-grounded, topology-aware grading with hierarchical feedback for both numeric and free-text responses. A voice interaction module is integrated to simulate telephone-based dispatch orders. We validate the system through a pilot deployment with licensed dispatch operators and scenario experiments on benchmark cases. Compared with a conventional scripted DTS workflow, AI Instructors achieves higher stepwise procedure accuracy (68%&amp;amp;rarr;90%), a lower topology-violation rate (32%&amp;amp;rarr;11%), and shorter response time (120 s&amp;amp;rarr;72 s), while increasing the proportion of parameterized questions and accelerating skill acquisition. These results suggest that combining adaptive sequencing with topology-safe, explainable evaluation can improve training effectiveness and operational safety.</p>
	]]></content:encoded>

	<dc:title>An Intelligent Simulation Training System for Power Grid Control and Operations</dc:title>
			<dc:creator>Sheng Yang</dc:creator>
			<dc:creator>Shengyuan Li</dc:creator>
			<dc:creator>Yuan Fu</dc:creator>
			<dc:creator>Wei Jiang</dc:creator>
			<dc:creator>Wenlong You</dc:creator>
			<dc:creator>Min Chen</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10030068</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-02-27</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-02-27</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>68</prism:startingPage>
		<prism:doi>10.3390/bdcc10030068</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/3/68</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/3/67">

	<title>BDCC, Vol. 10, Pages 67: Validating the Effectiveness of Fine-Tuning for Semantic Classification of Japanese Katakana Words: An Analysis of Frequency and Polysemy Effects on Accuracy</title>
	<link>https://www.mdpi.com/2504-2289/10/3/67</link>
	<description>In semantic classification of katakana words using large language models and pre-trained language models, semantic divergences from original English meanings, such as those found in Wasei-Eigo which is Japanese-made English, and the inherent sense ambiguity in katakana words may affect model accuracy. To analyze the impact of these loanword semantic characteristics on classification accuracy, we created a large-scale dataset from the Balanced Corpus of Contemporary Written Japanese. We extracted 403,819 sentences covering 230 katakana words defined in dictionaries and suitable for word sense disambiguation tasks, and used the gpt-4.1-mini model to predict the meaning of the target words based on their context, to create annotation data. We then fine-tuned the pre-trained language model DeBERTa V3 with this data. We compared baseline and fine-tuned model accuracy, dividing data into four quadrants based on frequency and polysemy to conduct statistical analysis and explore strategies for improving accuracy. We also tested the hypothesis that high-frequency, low-polysemy words would achieve the highest accuracy, while low-frequency, high-polysemy words would achieve the lowest. As a result, the fine-tuned model showed an average accuracy improvement of approximately 53% compared to the baseline model. As hypothesized, high-frequency, low-polysemy words achieved the highest accuracy (93.93%), while low-frequency, high-polysemy words achieved the lowest (81.14%). Our analysis quantitatively revealed that both frequency and polysemy contributed to accuracy improvement, but polysemy had a greater impact on accuracy than frequency.</description>
	<pubDate>2026-02-26</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 67: Validating the Effectiveness of Fine-Tuning for Semantic Classification of Japanese Katakana Words: An Analysis of Frequency and Polysemy Effects on Accuracy</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/3/67">doi: 10.3390/bdcc10030067</a></p>
	<p>Authors:
		Kazuki Kodaki
		Minoru Sasaki
		</p>
	<p>In semantic classification of katakana words using large language models and pre-trained language models, semantic divergences from original English meanings, such as those found in Wasei-Eigo which is Japanese-made English, and the inherent sense ambiguity in katakana words may affect model accuracy. To analyze the impact of these loanword semantic characteristics on classification accuracy, we created a large-scale dataset from the Balanced Corpus of Contemporary Written Japanese. We extracted 403,819 sentences covering 230 katakana words defined in dictionaries and suitable for word sense disambiguation tasks, and used the gpt-4.1-mini model to predict the meaning of the target words based on their context, to create annotation data. We then fine-tuned the pre-trained language model DeBERTa V3 with this data. We compared baseline and fine-tuned model accuracy, dividing data into four quadrants based on frequency and polysemy to conduct statistical analysis and explore strategies for improving accuracy. We also tested the hypothesis that high-frequency, low-polysemy words would achieve the highest accuracy, while low-frequency, high-polysemy words would achieve the lowest. As a result, the fine-tuned model showed an average accuracy improvement of approximately 53% compared to the baseline model. As hypothesized, high-frequency, low-polysemy words achieved the highest accuracy (93.93%), while low-frequency, high-polysemy words achieved the lowest (81.14%). Our analysis quantitatively revealed that both frequency and polysemy contributed to accuracy improvement, but polysemy had a greater impact on accuracy than frequency.</p>
	]]></content:encoded>

	<dc:title>Validating the Effectiveness of Fine-Tuning for Semantic Classification of Japanese Katakana Words: An Analysis of Frequency and Polysemy Effects on Accuracy</dc:title>
			<dc:creator>Kazuki Kodaki</dc:creator>
			<dc:creator>Minoru Sasaki</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10030067</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-02-26</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-02-26</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>3</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>67</prism:startingPage>
		<prism:doi>10.3390/bdcc10030067</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/3/67</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/2/66">

	<title>BDCC, Vol. 10, Pages 66: Comparative Read Performance Analysis of PostgreSQL and MongoDB in E-Commerce: An Empirical Study of Filtering and Analytical Queries</title>
	<link>https://www.mdpi.com/2504-2289/10/2/66</link>
	<description>This paper presents a comparative analysis of read performance for PostgreSQL and MongoDB in e-commerce scenarios, using identical datasets in a resource-constrained single-host environment. The results demonstrate that PostgreSQL executes complex analytical queries 1.6&amp;amp;ndash;15.1 times faster, depending on the query type and data volume. The study employed synthetic data generation with the Faker library across three stages, processing up to 300,000 products and executing each of 6 query types 15 times. Both filtering and analytical queries were tested on non-indexed data in a controlled localhost environment with PostgreSQL 17.5 and MongoDB 7.0.14, using default configurations. PostgreSQL showed 65&amp;amp;ndash;80% shorter execution times for multi-criteria queries, while MongoDB required approximately 33% less disk space. These findings suggest that normalized relational schemas are advantageous for transactional e-commerce systems where analytical queries dominate the workload. The results are directly applicable to small and medium e-commerce developers operating in budget-constrained, single-host deployment environments when choosing between relational and document-oriented databases for structured transactional data with read-heavy analytical workloads. A minimal indexed validation confirms that the baseline trends remain consistent under a simple indexing configuration. Future work will examine broader indexing strategies, write-intensive workloads, and distributed deployment scenarios.</description>
	<pubDate>2026-02-19</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 66: Comparative Read Performance Analysis of PostgreSQL and MongoDB in E-Commerce: An Empirical Study of Filtering and Analytical Queries</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/2/66">doi: 10.3390/bdcc10020066</a></p>
	<p>Authors:
		Jovita Urnikienė
		Vaida Steponavičienė
		Svetoslav Atanasov
		</p>
	<p>This paper presents a comparative analysis of read performance for PostgreSQL and MongoDB in e-commerce scenarios, using identical datasets in a resource-constrained single-host environment. The results demonstrate that PostgreSQL executes complex analytical queries 1.6&amp;amp;ndash;15.1 times faster, depending on the query type and data volume. The study employed synthetic data generation with the Faker library across three stages, processing up to 300,000 products and executing each of 6 query types 15 times. Both filtering and analytical queries were tested on non-indexed data in a controlled localhost environment with PostgreSQL 17.5 and MongoDB 7.0.14, using default configurations. PostgreSQL showed 65&amp;amp;ndash;80% shorter execution times for multi-criteria queries, while MongoDB required approximately 33% less disk space. These findings suggest that normalized relational schemas are advantageous for transactional e-commerce systems where analytical queries dominate the workload. The results are directly applicable to small and medium e-commerce developers operating in budget-constrained, single-host deployment environments when choosing between relational and document-oriented databases for structured transactional data with read-heavy analytical workloads. A minimal indexed validation confirms that the baseline trends remain consistent under a simple indexing configuration. Future work will examine broader indexing strategies, write-intensive workloads, and distributed deployment scenarios.</p>
	]]></content:encoded>

	<dc:title>Comparative Read Performance Analysis of PostgreSQL and MongoDB in E-Commerce: An Empirical Study of Filtering and Analytical Queries</dc:title>
			<dc:creator>Jovita Urnikienė</dc:creator>
			<dc:creator>Vaida Steponavičienė</dc:creator>
			<dc:creator>Svetoslav Atanasov</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10020066</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-02-19</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-02-19</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>2</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>66</prism:startingPage>
		<prism:doi>10.3390/bdcc10020066</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/2/66</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/2/65">

	<title>BDCC, Vol. 10, Pages 65: Integration of Lean Analytics and Industry 6.0: A Novel Meta-Theoretical Framework for Antifragile, Generative AI-Orchestrated, Circular&amp;ndash;Regenerative, and Hyper-Connected Manufacturing Ecosystems</title>
	<link>https://www.mdpi.com/2504-2289/10/2/65</link>
	<description>The convergence of Lean manufacturing principles with Industry 4.0 has yielded significant operational improvements, yet the emerging paradigm of Industry 6.0&amp;amp;mdash;characterized by antifragile, autonomous, and sustainable systems&amp;amp;mdash;demands a fundamental rethinking of existing analytical frameworks. This paper introduces the Industry 6.0 Lean Analytics (I6LA) Framework, a novel meta-theoretical approach that integrates Lean principles with the core concepts of Industry 6.0. By systematically analyzing the limitations of current Lean analytics in the context of Industry 6.0 requirements, we identify critical gaps in areas such as system resilience, AI-driven autonomy, and circular economy integration. The I6LA Framework addresses these gaps through four new theoretical pillars: Antifragile Lean Systems Theory, generative AI-Orchestrated Value Streams, Circular&amp;amp;ndash;Regenerative Analytics, and Hyper-Connected Ecosystem Integration. This research provides a new set of mathematical models for measuring antifragility, generative orchestration efficiency, and circularity, offering a comprehensive analytical toolkit for the next generation of manufacturing. The framework&amp;amp;rsquo;s primary contribution is a paradigm shift from optimizing stable, human-in-the-loop systems to managing dynamic, autonomous ecosystems that thrive on volatility and are regenerative by design. This paper provides both a robust theoretical foundation and practical implementation guidance for organizations navigating the transition to Industry 6.0.</description>
	<pubDate>2026-02-17</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 65: Integration of Lean Analytics and Industry 6.0: A Novel Meta-Theoretical Framework for Antifragile, Generative AI-Orchestrated, Circular&amp;ndash;Regenerative, and Hyper-Connected Manufacturing Ecosystems</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/2/65">doi: 10.3390/bdcc10020065</a></p>
	<p>Authors:
		Mohammad Shahin
		Mazdak Maghanaki
		F. Frank Chen
		</p>
	<p>The convergence of Lean manufacturing principles with Industry 4.0 has yielded significant operational improvements, yet the emerging paradigm of Industry 6.0&amp;amp;mdash;characterized by antifragile, autonomous, and sustainable systems&amp;amp;mdash;demands a fundamental rethinking of existing analytical frameworks. This paper introduces the Industry 6.0 Lean Analytics (I6LA) Framework, a novel meta-theoretical approach that integrates Lean principles with the core concepts of Industry 6.0. By systematically analyzing the limitations of current Lean analytics in the context of Industry 6.0 requirements, we identify critical gaps in areas such as system resilience, AI-driven autonomy, and circular economy integration. The I6LA Framework addresses these gaps through four new theoretical pillars: Antifragile Lean Systems Theory, generative AI-Orchestrated Value Streams, Circular&amp;amp;ndash;Regenerative Analytics, and Hyper-Connected Ecosystem Integration. This research provides a new set of mathematical models for measuring antifragility, generative orchestration efficiency, and circularity, offering a comprehensive analytical toolkit for the next generation of manufacturing. The framework&amp;amp;rsquo;s primary contribution is a paradigm shift from optimizing stable, human-in-the-loop systems to managing dynamic, autonomous ecosystems that thrive on volatility and are regenerative by design. This paper provides both a robust theoretical foundation and practical implementation guidance for organizations navigating the transition to Industry 6.0.</p>
	]]></content:encoded>

	<dc:title>Integration of Lean Analytics and Industry 6.0: A Novel Meta-Theoretical Framework for Antifragile, Generative AI-Orchestrated, Circular&amp;amp;ndash;Regenerative, and Hyper-Connected Manufacturing Ecosystems</dc:title>
			<dc:creator>Mohammad Shahin</dc:creator>
			<dc:creator>Mazdak Maghanaki</dc:creator>
			<dc:creator>F. Frank Chen</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10020065</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-02-17</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-02-17</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>2</prism:number>
	<prism:section>Perspective</prism:section>
	<prism:startingPage>65</prism:startingPage>
		<prism:doi>10.3390/bdcc10020065</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/2/65</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/2/64">

	<title>BDCC, Vol. 10, Pages 64: Efficient Time Series Visual Exploration for Insight Discovery</title>
	<link>https://www.mdpi.com/2504-2289/10/2/64</link>
	<description>Visual exploration of time series data is essential for uncovering meaningful insights in domains such as healthcare monitoring and financial analysis, yet it remains computationally challenging due to the combinatorial explosion of potential subsequence comparisons. For long time series, an exhaustive comparison of all possible subsequence pairs becomes prohibitively expensive, limiting interactive exploration. This paper presents the TiVEx (Time Series Visual Exploration) family of algorithms for efficiently discovering the top-k most dissimilar subsequence pairs in comparative time series analysis. TiVEx achieves scalability through three complementary strategies: TiVEx-sharing exploits computational reuse across overlapping subsequence windows, eliminating redundant distance calculations; TiVEx-pruning employs distance-based upper bounds to eliminate unpromising candidates without exhaustive evaluation; and TiVEx-hybrid integrates both mechanisms to maximize efficiency gains. The key observation is that overlapping subsequences share a substantial computational structure, which can be systematically exploited while maintaining result optimality through provably correct pruning bounds. Extensive experiments on six diverse datasets demonstrate that TiVEx-hybrid achieves up to 84% reduction in distance calculations compared to exhaustive search while producing identical top-k results. Compared to state-of-the-art subsequence comparison methods, TiVEx-hybrid achieves 2.3&amp;amp;times; improvement in computational efficiency. Our effectiveness analysis confirms that TiVEx achieves result quality within 5% of exhaustive search even when exploring only a subset of candidate positions, enabling scalable visual exploration without compromising insight quality.</description>
	<pubDate>2026-02-16</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 64: Efficient Time Series Visual Exploration for Insight Discovery</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/2/64">doi: 10.3390/bdcc10020064</a></p>
	<p>Authors:
		Heba Helal
		Mohamed A. Sharaf
		</p>
	<p>Visual exploration of time series data is essential for uncovering meaningful insights in domains such as healthcare monitoring and financial analysis, yet it remains computationally challenging due to the combinatorial explosion of potential subsequence comparisons. For long time series, an exhaustive comparison of all possible subsequence pairs becomes prohibitively expensive, limiting interactive exploration. This paper presents the TiVEx (Time Series Visual Exploration) family of algorithms for efficiently discovering the top-k most dissimilar subsequence pairs in comparative time series analysis. TiVEx achieves scalability through three complementary strategies: TiVEx-sharing exploits computational reuse across overlapping subsequence windows, eliminating redundant distance calculations; TiVEx-pruning employs distance-based upper bounds to eliminate unpromising candidates without exhaustive evaluation; and TiVEx-hybrid integrates both mechanisms to maximize efficiency gains. The key observation is that overlapping subsequences share a substantial computational structure, which can be systematically exploited while maintaining result optimality through provably correct pruning bounds. Extensive experiments on six diverse datasets demonstrate that TiVEx-hybrid achieves up to 84% reduction in distance calculations compared to exhaustive search while producing identical top-k results. Compared to state-of-the-art subsequence comparison methods, TiVEx-hybrid achieves 2.3&amp;amp;times; improvement in computational efficiency. Our effectiveness analysis confirms that TiVEx achieves result quality within 5% of exhaustive search even when exploring only a subset of candidate positions, enabling scalable visual exploration without compromising insight quality.</p>
	]]></content:encoded>

	<dc:title>Efficient Time Series Visual Exploration for Insight Discovery</dc:title>
			<dc:creator>Heba Helal</dc:creator>
			<dc:creator>Mohamed A. Sharaf</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10020064</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-02-16</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-02-16</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>2</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>64</prism:startingPage>
		<prism:doi>10.3390/bdcc10020064</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/2/64</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/2/63">

	<title>BDCC, Vol. 10, Pages 63: Cognitive Assemblages: Living with Algorithms</title>
	<link>https://www.mdpi.com/2504-2289/10/2/63</link>
	<description>The rapid expansion of algorithmic systems has transformed cognition into an increasingly distributed and collective enterprise, giving rise to what can be described as cognitive assemblages, dynamic constellations of humans, institutions, data infrastructures, and artificial agents. This paper traces the historical and conceptual evolution that has led to this shift. First, we show how cognition, once conceived as the property of autonomous individuals, has progressively become embedded in socio-technical networks in which algorithmic processes participate as co-agents. Second, we revisit the progressive awareness of human cognitive limits, from bounded rationality to contemporary theories of extended mind. These frameworks anticipate and help explain today&amp;amp;rsquo;s hybrid cognitive ecologies. Third, we assess the philosophical implications for Enlightenment ideals of autonomy, rationality, and self-governance, showing how these concepts must be reinterpreted in light of pervasive algorithmic intermediation. Finally, we examine global initiatives that seek to integrate augmented cognitive capacities into large-scale cybernetic forms of societal coordination, ranging from digital platforms and data spaces to AI-driven governance systems. These developments offer new opportunities for steering complex societies under conditions of globalization, environmental disruption, and the rise of autonomous intelligent systems, yet they also raise profound questions regarding control, accountability, and democratic legitimacy. We argue that understanding cognitive assemblages is essential to designing socio-technical systems capable of supporting collective intelligence while preserving human values in an era of accelerating complexity.</description>
	<pubDate>2026-02-16</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 63: Cognitive Assemblages: Living with Algorithms</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/2/63">doi: 10.3390/bdcc10020063</a></p>
	<p>Authors:
		Stéphane Grumbach
		</p>
	<p>The rapid expansion of algorithmic systems has transformed cognition into an increasingly distributed and collective enterprise, giving rise to what can be described as cognitive assemblages, dynamic constellations of humans, institutions, data infrastructures, and artificial agents. This paper traces the historical and conceptual evolution that has led to this shift. First, we show how cognition, once conceived as the property of autonomous individuals, has progressively become embedded in socio-technical networks in which algorithmic processes participate as co-agents. Second, we revisit the progressive awareness of human cognitive limits, from bounded rationality to contemporary theories of extended mind. These frameworks anticipate and help explain today&amp;amp;rsquo;s hybrid cognitive ecologies. Third, we assess the philosophical implications for Enlightenment ideals of autonomy, rationality, and self-governance, showing how these concepts must be reinterpreted in light of pervasive algorithmic intermediation. Finally, we examine global initiatives that seek to integrate augmented cognitive capacities into large-scale cybernetic forms of societal coordination, ranging from digital platforms and data spaces to AI-driven governance systems. These developments offer new opportunities for steering complex societies under conditions of globalization, environmental disruption, and the rise of autonomous intelligent systems, yet they also raise profound questions regarding control, accountability, and democratic legitimacy. We argue that understanding cognitive assemblages is essential to designing socio-technical systems capable of supporting collective intelligence while preserving human values in an era of accelerating complexity.</p>
	]]></content:encoded>

	<dc:title>Cognitive Assemblages: Living with Algorithms</dc:title>
			<dc:creator>Stéphane Grumbach</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10020063</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-02-16</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-02-16</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>2</prism:number>
	<prism:section>Review</prism:section>
	<prism:startingPage>63</prism:startingPage>
		<prism:doi>10.3390/bdcc10020063</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/2/63</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/2/62">

	<title>BDCC, Vol. 10, Pages 62: Skill Classification of Youth Table Tennis Players Using Sensor Fusion and the Random Forest Algorithm</title>
	<link>https://www.mdpi.com/2504-2289/10/2/62</link>
	<description>This study addresses the issue of inaccurate results in traditional table tennis player classification, which is often influenced by subjective judgment and environmental factors, by proposing a youth table tennis player classification system based on sensor fusion and the random forest algorithm. The system utilizes an embedded intelligent table tennis racket equipped with an ICM20948 nine-axis sensor and a wireless transmission module to capture real-time acceleration and angular velocity data during players&amp;amp;rsquo; strokes while synchronously employing a camera with OpenPose to extract joint angle variations. A total of 40 players&amp;amp;rsquo; stroke data were collected. Due to the limited sample size of top-tier players, the Synthetic Minority Over-sampling Technique (SMOTE) was applied, resulting in a final dataset of 360 records. Multiple key motion indicators were then computed and stored in a dedicated database. Experimental results showed that the proposed system, powered by the random forest algorithm, achieved a classification accuracy of 91.3% under conventional cross-validation, while subject-independent LOSO validation yielded a more conservative accuracy of 70.89%, making it a valuable reference for coaches and referees in conducting objective player classification. Future work will focus on expanding the dataset of domestic high-performance athletes and integrating precise sports science resources to further enhance the system&amp;amp;rsquo;s performance and algorithmic models, thereby promoting the scientific selection of national team players and advancing the intelligent development of table tennis.</description>
	<pubDate>2026-02-15</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 62: Skill Classification of Youth Table Tennis Players Using Sensor Fusion and the Random Forest Algorithm</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/2/62">doi: 10.3390/bdcc10020062</a></p>
	<p>Authors:
		Yung-Hoh Sheu
		Cheng-Yu Huang
		Li-Wei Tai
		Tzu-Hsuan Tai
		Sheng K. Wu
		</p>
	<p>This study addresses the issue of inaccurate results in traditional table tennis player classification, which is often influenced by subjective judgment and environmental factors, by proposing a youth table tennis player classification system based on sensor fusion and the random forest algorithm. The system utilizes an embedded intelligent table tennis racket equipped with an ICM20948 nine-axis sensor and a wireless transmission module to capture real-time acceleration and angular velocity data during players&amp;amp;rsquo; strokes while synchronously employing a camera with OpenPose to extract joint angle variations. A total of 40 players&amp;amp;rsquo; stroke data were collected. Due to the limited sample size of top-tier players, the Synthetic Minority Over-sampling Technique (SMOTE) was applied, resulting in a final dataset of 360 records. Multiple key motion indicators were then computed and stored in a dedicated database. Experimental results showed that the proposed system, powered by the random forest algorithm, achieved a classification accuracy of 91.3% under conventional cross-validation, while subject-independent LOSO validation yielded a more conservative accuracy of 70.89%, making it a valuable reference for coaches and referees in conducting objective player classification. Future work will focus on expanding the dataset of domestic high-performance athletes and integrating precise sports science resources to further enhance the system&amp;amp;rsquo;s performance and algorithmic models, thereby promoting the scientific selection of national team players and advancing the intelligent development of table tennis.</p>
	]]></content:encoded>

	<dc:title>Skill Classification of Youth Table Tennis Players Using Sensor Fusion and the Random Forest Algorithm</dc:title>
			<dc:creator>Yung-Hoh Sheu</dc:creator>
			<dc:creator>Cheng-Yu Huang</dc:creator>
			<dc:creator>Li-Wei Tai</dc:creator>
			<dc:creator>Tzu-Hsuan Tai</dc:creator>
			<dc:creator>Sheng K. Wu</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10020062</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-02-15</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-02-15</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>2</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>62</prism:startingPage>
		<prism:doi>10.3390/bdcc10020062</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/2/62</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/2/61">

	<title>BDCC, Vol. 10, Pages 61: Underwater Visual-Servo Alignment Control Integrating Geometric Cognition Compensation and Confidence Assessment</title>
	<link>https://www.mdpi.com/2504-2289/10/2/61</link>
	<description>To meet the requirements for the automatic alignment, insertion, and inspection of guide-tube opening pins on the upper core plate in a component pool during refueling outages of nuclear power units, this paper proposes a cognition-enhanced visual-servoing framework that integrates geometric cognition-based compensation, observation-confidence modeling, and constraint-aware optimal control. The framework addresses the key challenge posed by the coexistence of long-term geometric drift and underwater observation uncertainty. Specifically, historical closed-loop data are leveraged to learn and compensate for systematic geometric errors online, substantially improving coarse-positioning accuracy. In addition, an explicit confidence model is introduced to quantitatively assess the reliability of visual measurements. Building on these components, a confidence-driven, finite-horizon, constrained model predictive control strategy is designed to achieve safe and efficient finite-step convergence while strictly respecting actuator physical constraints. Ground experiments and deep-water component-pool validations demonstrate that the proposed method reduces coarse-positioning error by approximately 75%, achieves stable sub-millimeter alignment with an ample engineering safety margin, and effectively decreases erroneous insertions and the need for manual intervention. These results confirm the engineering applicability and safety advantages of the proposed cognition-enhanced visual-servoing framework for underwater alignment tasks in nuclear component pools.</description>
	<pubDate>2026-02-14</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 61: Underwater Visual-Servo Alignment Control Integrating Geometric Cognition Compensation and Confidence Assessment</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/2/61">doi: 10.3390/bdcc10020061</a></p>
	<p>Authors:
		Jinkun Li
		Lingyu Sun
		Minglu Zhang
		Xinbao Li
		</p>
	<p>To meet the requirements for the automatic alignment, insertion, and inspection of guide-tube opening pins on the upper core plate in a component pool during refueling outages of nuclear power units, this paper proposes a cognition-enhanced visual-servoing framework that integrates geometric cognition-based compensation, observation-confidence modeling, and constraint-aware optimal control. The framework addresses the key challenge posed by the coexistence of long-term geometric drift and underwater observation uncertainty. Specifically, historical closed-loop data are leveraged to learn and compensate for systematic geometric errors online, substantially improving coarse-positioning accuracy. In addition, an explicit confidence model is introduced to quantitatively assess the reliability of visual measurements. Building on these components, a confidence-driven, finite-horizon, constrained model predictive control strategy is designed to achieve safe and efficient finite-step convergence while strictly respecting actuator physical constraints. Ground experiments and deep-water component-pool validations demonstrate that the proposed method reduces coarse-positioning error by approximately 75%, achieves stable sub-millimeter alignment with an ample engineering safety margin, and effectively decreases erroneous insertions and the need for manual intervention. These results confirm the engineering applicability and safety advantages of the proposed cognition-enhanced visual-servoing framework for underwater alignment tasks in nuclear component pools.</p>
	]]></content:encoded>

	<dc:title>Underwater Visual-Servo Alignment Control Integrating Geometric Cognition Compensation and Confidence Assessment</dc:title>
			<dc:creator>Jinkun Li</dc:creator>
			<dc:creator>Lingyu Sun</dc:creator>
			<dc:creator>Minglu Zhang</dc:creator>
			<dc:creator>Xinbao Li</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10020061</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-02-14</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-02-14</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>2</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>61</prism:startingPage>
		<prism:doi>10.3390/bdcc10020061</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/2/61</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/2/60">

	<title>BDCC, Vol. 10, Pages 60: Reliability of LLM Inference Engines from a Static Perspective: Root Cause Analysis and Repair Suggestion via Natural Language Reports</title>
	<link>https://www.mdpi.com/2504-2289/10/2/60</link>
	<description>Large Language Model (LLM) inference engines are becoming critical system infrastructure, yet their increasing architectural complexity makes defects difficult to be diagnosed and repaired. Existing reliability studies predominantly focus on model behavior or training frameworks, leaving inference engine bugs underexplored, especially in settings where execution-based debugging is impractical. We present a static, issue-centric approach for automated root cause analysis and repair suggestion generation for LLM inference engines. Based solely on issue reports and developer discussions, we construct a real-world defect dataset and annotate each issue with a semantic root cause category and affected system module. Leveraging text-based representations, our framework performs root cause classification and coarse-grained module localization without requiring code execution or specialized runtime environments. We further integrate structured repair patterns with a large language model to generate interpretable and actionable repair suggestions. Experiments on real-world issues concerning vLLMs demonstrate that our approach achieves effective root cause identification and module localization under limited and imbalanced data. A cross-engine evaluation further shows promising generalization to TensorRT-LLM. Human evaluation confirms that the generated repair suggestions are correct, useful, and clearly expressed. Our results indicate that static, issue-level analysis is a viable foundation for scalable debugging assistance in LLM inference engines. This work highlights the feasibility of static, issue-level defect analysis for complex LLM inference engines and explores automated debugging assistance techniques. The dataset and implementation will be publicly released to facilitate future research.</description>
	<pubDate>2026-02-13</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 60: Reliability of LLM Inference Engines from a Static Perspective: Root Cause Analysis and Repair Suggestion via Natural Language Reports</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/2/60">doi: 10.3390/bdcc10020060</a></p>
	<p>Authors:
		Hongwei Li
		Yongjun Wang
		</p>
	<p>Large Language Model (LLM) inference engines are becoming critical system infrastructure, yet their increasing architectural complexity makes defects difficult to be diagnosed and repaired. Existing reliability studies predominantly focus on model behavior or training frameworks, leaving inference engine bugs underexplored, especially in settings where execution-based debugging is impractical. We present a static, issue-centric approach for automated root cause analysis and repair suggestion generation for LLM inference engines. Based solely on issue reports and developer discussions, we construct a real-world defect dataset and annotate each issue with a semantic root cause category and affected system module. Leveraging text-based representations, our framework performs root cause classification and coarse-grained module localization without requiring code execution or specialized runtime environments. We further integrate structured repair patterns with a large language model to generate interpretable and actionable repair suggestions. Experiments on real-world issues concerning vLLMs demonstrate that our approach achieves effective root cause identification and module localization under limited and imbalanced data. A cross-engine evaluation further shows promising generalization to TensorRT-LLM. Human evaluation confirms that the generated repair suggestions are correct, useful, and clearly expressed. Our results indicate that static, issue-level analysis is a viable foundation for scalable debugging assistance in LLM inference engines. This work highlights the feasibility of static, issue-level defect analysis for complex LLM inference engines and explores automated debugging assistance techniques. The dataset and implementation will be publicly released to facilitate future research.</p>
	]]></content:encoded>

	<dc:title>Reliability of LLM Inference Engines from a Static Perspective: Root Cause Analysis and Repair Suggestion via Natural Language Reports</dc:title>
			<dc:creator>Hongwei Li</dc:creator>
			<dc:creator>Yongjun Wang</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10020060</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-02-13</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-02-13</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>2</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>60</prism:startingPage>
		<prism:doi>10.3390/bdcc10020060</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/2/60</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/2/59">

	<title>BDCC, Vol. 10, Pages 59: PLTA-FinBERT: Pseudo-Label Generation-Based Test-Time Adaptation for Financial Sentiment Analysis</title>
	<link>https://www.mdpi.com/2504-2289/10/2/59</link>
	<description>Financial sentiment analysis leverages natural language processing techniques to quantitatively assess sentiment polarity and emotional tendencies in financial texts. Its practical application in investment decision-making and risk management faces two major challenges: the scarcity of high-quality labeled data due to expert annotation costs, and semantic drift caused by the continuous evolution of market language. To address these issues, this study proposes PLTA-FinBERT, a pseudo-label generation-based test-time adaptation framework that enables dynamic self-learning without requiring additional labeled data. The framework consists of two modules: a multi-perturbation pseudo-label generation mechanism that enhances label reliability through consistency voting and confidence-based filtering, and a test-time dynamic adaptation strategy that iteratively updates model parameters based on high-confidence pseudo-labels, allowing the model to continuously adapt to new linguistic patterns. PLTA-FinBERT achieves 0.8288 accuracy on the sentiment classification dataset of financial sentiment analysis, representing an absolute improvement of 2.37 percentage points over the benchmark. On the FiQA sentiment intensity prediction task, it obtains an R2 of 0.58, surpassing the previous state-of-the-art by 3 percentage points.</description>
	<pubDate>2026-02-11</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 59: PLTA-FinBERT: Pseudo-Label Generation-Based Test-Time Adaptation for Financial Sentiment Analysis</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/2/59">doi: 10.3390/bdcc10020059</a></p>
	<p>Authors:
		Hai Yang
		Hainan Chen
		Chang Jiang
		Juntao He
		Pengyang Li
		</p>
	<p>Financial sentiment analysis leverages natural language processing techniques to quantitatively assess sentiment polarity and emotional tendencies in financial texts. Its practical application in investment decision-making and risk management faces two major challenges: the scarcity of high-quality labeled data due to expert annotation costs, and semantic drift caused by the continuous evolution of market language. To address these issues, this study proposes PLTA-FinBERT, a pseudo-label generation-based test-time adaptation framework that enables dynamic self-learning without requiring additional labeled data. The framework consists of two modules: a multi-perturbation pseudo-label generation mechanism that enhances label reliability through consistency voting and confidence-based filtering, and a test-time dynamic adaptation strategy that iteratively updates model parameters based on high-confidence pseudo-labels, allowing the model to continuously adapt to new linguistic patterns. PLTA-FinBERT achieves 0.8288 accuracy on the sentiment classification dataset of financial sentiment analysis, representing an absolute improvement of 2.37 percentage points over the benchmark. On the FiQA sentiment intensity prediction task, it obtains an R2 of 0.58, surpassing the previous state-of-the-art by 3 percentage points.</p>
	]]></content:encoded>

	<dc:title>PLTA-FinBERT: Pseudo-Label Generation-Based Test-Time Adaptation for Financial Sentiment Analysis</dc:title>
			<dc:creator>Hai Yang</dc:creator>
			<dc:creator>Hainan Chen</dc:creator>
			<dc:creator>Chang Jiang</dc:creator>
			<dc:creator>Juntao He</dc:creator>
			<dc:creator>Pengyang Li</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10020059</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-02-11</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-02-11</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>2</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>59</prism:startingPage>
		<prism:doi>10.3390/bdcc10020059</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/2/59</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/2/58">

	<title>BDCC, Vol. 10, Pages 58: Bias Correction and Explainability Framework for Large Language Models: A Knowledge-Driven Approach</title>
	<link>https://www.mdpi.com/2504-2289/10/2/58</link>
	<description>Large Language Models (LLMs) have demonstrated extraordinary capabilities in natural language generation; however, their real-world deployment is frequently hindered by the generation of factually incorrect or biased content, along with an inherent deficiency in transparency. To address these critical limitations and thereby enhance the reliability and explainability of LLM outputs, this study proposes a novel integrated framework, namely the Adaptive Knowledge-Driven Correction Network (AKDC-Net), which incorporates three core algorithmic innovations. Firstly, the Hierarchical Uncertainty-Aware Bias Detector (HUABD) performs multi-level linguistic analysis (lexical, syntactic, semantic, and pragmatic) and, for the first time, decomposes predictive uncertainty into epistemic and aleatoric components. This decomposition enables principled, interpretable bias detection with clear theoretical underpinnings. Secondly, the Neural-Symbolic Knowledge Graph Enhanced Corrector (NSKGEC) integrates a temporal graph neural network with a differentiable symbolic reasoning module, facilitating logically consistent and factually grounded corrections based on dynamically updated knowledge sources. Thirdly, the Contrastive Learning-driven Multimodal Explanation Generator (CLMEG) leverages a cross-modal attention mechanism within a contrastive learning paradigm to generate coherent, high-quality textual and visual explanations that enhance the interpretability of LLM outputs. Extensive evaluations were conducted on a challenging medical domain dataset to validate the effectiveness of the proposed AKDC-Net framework. Experimental results demonstrate significant improvements over state-of-the-art baselines: specifically, a 14.1% increase in the F1-score for bias detection, a 19.4% enhancement in correction quality, and a 31.4% rise in user trust scores. These findings establish a new benchmark for the development of more trustworthy and transparent artificial intelligence (AI) systems, laying a solid foundation for the broader and more reliable application of LLMs in high-stakes domains.</description>
	<pubDate>2026-02-10</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 58: Bias Correction and Explainability Framework for Large Language Models: A Knowledge-Driven Approach</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/2/58">doi: 10.3390/bdcc10020058</a></p>
	<p>Authors:
		Xianming Yang
		Qi Li
		Chengdong Qian
		Haitao Wang
		Yonghui Wu
		Wei Wang
		</p>
	<p>Large Language Models (LLMs) have demonstrated extraordinary capabilities in natural language generation; however, their real-world deployment is frequently hindered by the generation of factually incorrect or biased content, along with an inherent deficiency in transparency. To address these critical limitations and thereby enhance the reliability and explainability of LLM outputs, this study proposes a novel integrated framework, namely the Adaptive Knowledge-Driven Correction Network (AKDC-Net), which incorporates three core algorithmic innovations. Firstly, the Hierarchical Uncertainty-Aware Bias Detector (HUABD) performs multi-level linguistic analysis (lexical, syntactic, semantic, and pragmatic) and, for the first time, decomposes predictive uncertainty into epistemic and aleatoric components. This decomposition enables principled, interpretable bias detection with clear theoretical underpinnings. Secondly, the Neural-Symbolic Knowledge Graph Enhanced Corrector (NSKGEC) integrates a temporal graph neural network with a differentiable symbolic reasoning module, facilitating logically consistent and factually grounded corrections based on dynamically updated knowledge sources. Thirdly, the Contrastive Learning-driven Multimodal Explanation Generator (CLMEG) leverages a cross-modal attention mechanism within a contrastive learning paradigm to generate coherent, high-quality textual and visual explanations that enhance the interpretability of LLM outputs. Extensive evaluations were conducted on a challenging medical domain dataset to validate the effectiveness of the proposed AKDC-Net framework. Experimental results demonstrate significant improvements over state-of-the-art baselines: specifically, a 14.1% increase in the F1-score for bias detection, a 19.4% enhancement in correction quality, and a 31.4% rise in user trust scores. These findings establish a new benchmark for the development of more trustworthy and transparent artificial intelligence (AI) systems, laying a solid foundation for the broader and more reliable application of LLMs in high-stakes domains.</p>
	]]></content:encoded>

	<dc:title>Bias Correction and Explainability Framework for Large Language Models: A Knowledge-Driven Approach</dc:title>
			<dc:creator>Xianming Yang</dc:creator>
			<dc:creator>Qi Li</dc:creator>
			<dc:creator>Chengdong Qian</dc:creator>
			<dc:creator>Haitao Wang</dc:creator>
			<dc:creator>Yonghui Wu</dc:creator>
			<dc:creator>Wei Wang</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10020058</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-02-10</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-02-10</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>2</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>58</prism:startingPage>
		<prism:doi>10.3390/bdcc10020058</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/2/58</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/2/57">

	<title>BDCC, Vol. 10, Pages 57: Enhancing the Artificial Rabbit Optimizer Using Fuzzy Rule Interpolation</title>
	<link>https://www.mdpi.com/2504-2289/10/2/57</link>
	<description>Metaheuristic optimization algorithms have demonstrated their effectiveness in solving complex optimization tasks, such as those related to Intrusion Detection Systems (IDSs). It was widely used to enhance the detection rate of various types of cyber attacks by reducing the feature space or tuning the model&amp;amp;rsquo;s hyperparameters. The Artificial Rabbit Optimizer (ARO) mimics rabbits&amp;amp;rsquo; intelligent foraging and hiding behavior. The ARO algorithm has seen widespread adoption in the optimization field. The widespread use of the ARO algorithm occurs due to its simple design and ease of implementation. However, ARO can get trapped in local optima due to its limited diversity in population dynamics. Although the transition between phases is managed via an energy shrink factor, fine-tuning this balance remains challenging and unexplored. These limitations could limit the ARO algorithm&amp;amp;rsquo;s effectiveness in high-dimensional space, as with IDS systems. This paper introduces a novel enhancement of the original ARO by integrating Fuzzy Rule Interpolation (FRI) to compute the energy factor during the optimization process dynamically. In this work, we integrate the FRI along with the ARO algorithm to improve solution accuracy, maintain population diversity, and accelerate convergence, particularly in high-dimensional and complex problems such as IDS. The integration of the FRI and ARO aimed to control the exploration-exploitation balance in the IDS application area. To validate our proposed hybrid approach, we tested it on a diverse set of intrusion datasets, covering eight different benchmark intrusion detection datasets. The suggested hybrid approach has been demonstrated to be effective in handling various intrusion classification tasks. For binary intrusion classification tasks, it achieved accuracy rates ranging from 96% to 99.9%. In the case of multiclass intrusion classification tasks, the accuracy was slightly more consistent, falling between 91.6% and 98.9%. The suggested approach effectively reduced the number of feature spaces, achieving reduction rates from 56% up to 96%. Furthermore, the proposed approach outperformed other state-of-the-art methods in terms of detection rate.</description>
	<pubDate>2026-02-10</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 57: Enhancing the Artificial Rabbit Optimizer Using Fuzzy Rule Interpolation</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/2/57">doi: 10.3390/bdcc10020057</a></p>
	<p>Authors:
		Mohammad Almseidin
		</p>
	<p>Metaheuristic optimization algorithms have demonstrated their effectiveness in solving complex optimization tasks, such as those related to Intrusion Detection Systems (IDSs). It was widely used to enhance the detection rate of various types of cyber attacks by reducing the feature space or tuning the model&amp;amp;rsquo;s hyperparameters. The Artificial Rabbit Optimizer (ARO) mimics rabbits&amp;amp;rsquo; intelligent foraging and hiding behavior. The ARO algorithm has seen widespread adoption in the optimization field. The widespread use of the ARO algorithm occurs due to its simple design and ease of implementation. However, ARO can get trapped in local optima due to its limited diversity in population dynamics. Although the transition between phases is managed via an energy shrink factor, fine-tuning this balance remains challenging and unexplored. These limitations could limit the ARO algorithm&amp;amp;rsquo;s effectiveness in high-dimensional space, as with IDS systems. This paper introduces a novel enhancement of the original ARO by integrating Fuzzy Rule Interpolation (FRI) to compute the energy factor during the optimization process dynamically. In this work, we integrate the FRI along with the ARO algorithm to improve solution accuracy, maintain population diversity, and accelerate convergence, particularly in high-dimensional and complex problems such as IDS. The integration of the FRI and ARO aimed to control the exploration-exploitation balance in the IDS application area. To validate our proposed hybrid approach, we tested it on a diverse set of intrusion datasets, covering eight different benchmark intrusion detection datasets. The suggested hybrid approach has been demonstrated to be effective in handling various intrusion classification tasks. For binary intrusion classification tasks, it achieved accuracy rates ranging from 96% to 99.9%. In the case of multiclass intrusion classification tasks, the accuracy was slightly more consistent, falling between 91.6% and 98.9%. The suggested approach effectively reduced the number of feature spaces, achieving reduction rates from 56% up to 96%. Furthermore, the proposed approach outperformed other state-of-the-art methods in terms of detection rate.</p>
	]]></content:encoded>

	<dc:title>Enhancing the Artificial Rabbit Optimizer Using Fuzzy Rule Interpolation</dc:title>
			<dc:creator>Mohammad Almseidin</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10020057</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-02-10</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-02-10</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>2</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>57</prism:startingPage>
		<prism:doi>10.3390/bdcc10020057</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/2/57</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/2/56">

	<title>BDCC, Vol. 10, Pages 56: ISFJ-RAG: Interventional Suppression of Hallucinations via Counter-Factual Joint Decoding Retrieval-Augment Generation</title>
	<link>https://www.mdpi.com/2504-2289/10/2/56</link>
	<description>Although retrieval-augmented generation (RAG) technology mitigates the hallucination issue in large language models (LLMs) by incorporating external knowledge, and combining reasoning models can further enhance RAG system performance, retrieval noise and attention bias still lead to the diffusion of factual errors in problems such as factual queries, multi-hop questions, and unanswerable questions. Existing methods struggle to effectively suppress &amp;amp;ldquo;high-confidence hallucinations&amp;amp;rdquo; in long-chain reasoning due to their failure to decouple knowledge bias effects from causal reasoning paths. To address this, this paper proposes the ISFJ-RAG framework, which dynamically intervenes in hallucinations through counterfactual joint decoding. First, a structural causal model (SCM) reveals three root causes of hallucinations in RAG systems: irrelevant knowledge interference, reasoning path bias, and spurious correlations in self-attention mechanisms. A dual-decoder architecture is further designed: the total causal effect decoder models the global relationship between user queries and knowledge, while the knowledge bias effect decoder captures potential biases induced by external knowledge. A dynamic modulation module converts the latter&amp;amp;rsquo;s output into a proxy measure of hallucination bias. By computing individual treatment effects (ITEs), the bias component is removed from the full generation distribution, achieving simultaneous suppression of knowledge-irrelevant and reasoning-irrelevant hallucinations. Ablation experiments validate the robustness of average token log-probability as a confidence metric. Experiments demonstrate that on the RAGEval benchmark, ISFJ-RAG improves generation completeness to 86.89% (+5.49%) while reducing hallucination rates to 10.39% (&amp;amp;minus;2.5%) and irrelevance rates to 4.44% (&amp;amp;minus;2.99%).</description>
	<pubDate>2026-02-09</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 56: ISFJ-RAG: Interventional Suppression of Hallucinations via Counter-Factual Joint Decoding Retrieval-Augment Generation</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/2/56">doi: 10.3390/bdcc10020056</a></p>
	<p>Authors:
		Yuezhao Liu
		Wei Li
		Yijie Wang
		Ningtong Chen
		Min Chen
		</p>
	<p>Although retrieval-augmented generation (RAG) technology mitigates the hallucination issue in large language models (LLMs) by incorporating external knowledge, and combining reasoning models can further enhance RAG system performance, retrieval noise and attention bias still lead to the diffusion of factual errors in problems such as factual queries, multi-hop questions, and unanswerable questions. Existing methods struggle to effectively suppress &amp;amp;ldquo;high-confidence hallucinations&amp;amp;rdquo; in long-chain reasoning due to their failure to decouple knowledge bias effects from causal reasoning paths. To address this, this paper proposes the ISFJ-RAG framework, which dynamically intervenes in hallucinations through counterfactual joint decoding. First, a structural causal model (SCM) reveals three root causes of hallucinations in RAG systems: irrelevant knowledge interference, reasoning path bias, and spurious correlations in self-attention mechanisms. A dual-decoder architecture is further designed: the total causal effect decoder models the global relationship between user queries and knowledge, while the knowledge bias effect decoder captures potential biases induced by external knowledge. A dynamic modulation module converts the latter&amp;amp;rsquo;s output into a proxy measure of hallucination bias. By computing individual treatment effects (ITEs), the bias component is removed from the full generation distribution, achieving simultaneous suppression of knowledge-irrelevant and reasoning-irrelevant hallucinations. Ablation experiments validate the robustness of average token log-probability as a confidence metric. Experiments demonstrate that on the RAGEval benchmark, ISFJ-RAG improves generation completeness to 86.89% (+5.49%) while reducing hallucination rates to 10.39% (&amp;amp;minus;2.5%) and irrelevance rates to 4.44% (&amp;amp;minus;2.99%).</p>
	]]></content:encoded>

	<dc:title>ISFJ-RAG: Interventional Suppression of Hallucinations via Counter-Factual Joint Decoding Retrieval-Augment Generation</dc:title>
			<dc:creator>Yuezhao Liu</dc:creator>
			<dc:creator>Wei Li</dc:creator>
			<dc:creator>Yijie Wang</dc:creator>
			<dc:creator>Ningtong Chen</dc:creator>
			<dc:creator>Min Chen</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10020056</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-02-09</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-02-09</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>2</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>56</prism:startingPage>
		<prism:doi>10.3390/bdcc10020056</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/2/56</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/2/55">

	<title>BDCC, Vol. 10, Pages 55: What Distinguishes AI-Generated from Human Writing? A Rapid Review of the Literature</title>
	<link>https://www.mdpi.com/2504-2289/10/2/55</link>
	<description>Large language models (LLMs) are now routine writing tools across various domains, intensifying questions about when text should be treated as human-authored, artificial intelligence (AI)-generated, or collaboratively produced. This rapid review aims to identify cue families reported in empirical studies as distinguishing AI from human-authored text and to assess how stable these cues are across genres/tasks, text lengths, and revision conditions. Following the Preferred Reporting Items for Systematic Reviews and Meta-Analysis (PRISMA) guidelines, we searched four online databases for peer-reviewed empirical articles (1 January 2022&amp;amp;ndash;1 January 2026). After deduplication and screening, 40 studies were included. Evidence converged on five cue families: surface, discourse/pragmatic, epistemic/content, predictability/probabilistic, and provenance. Surface cues dominated the literature and were the most consistently operationalized. Discourse/pragmatic cues followed, particularly in discipline-bound academic genres where stance and metadiscourse differentiated AI from human writing. Predictability/probabilistic cues were central in detector-focused studies, while epistemic/content cues emerged primarily in tasks where grounding and authenticity were salient. Provenance cues were concentrated in watermarking research. Across studies, cue stability was consistently conditional rather than universal. Specifically, surface and discourse cues often remained discriminative within constrained genres, but shifted with register and discipline; probabilistic cues were powerful yet fragile under paraphrasing, post-editing, and evasion; and provenance signals required robustness to editing, mixing, and span localization. Overall, the literature indicates that AI&amp;amp;ndash;human distinction emerges from layered and context-dependent cue profiles rather than from any single reliable marker. High-stakes decisions, therefore, require condition-aware interpretation, triangulation across multiple cue families, and human oversight rather than automated classification in isolation.</description>
	<pubDate>2026-02-08</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 55: What Distinguishes AI-Generated from Human Writing? A Rapid Review of the Literature</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/2/55">doi: 10.3390/bdcc10020055</a></p>
	<p>Authors:
		Georgios P. Georgiou
		</p>
	<p>Large language models (LLMs) are now routine writing tools across various domains, intensifying questions about when text should be treated as human-authored, artificial intelligence (AI)-generated, or collaboratively produced. This rapid review aims to identify cue families reported in empirical studies as distinguishing AI from human-authored text and to assess how stable these cues are across genres/tasks, text lengths, and revision conditions. Following the Preferred Reporting Items for Systematic Reviews and Meta-Analysis (PRISMA) guidelines, we searched four online databases for peer-reviewed empirical articles (1 January 2022&amp;amp;ndash;1 January 2026). After deduplication and screening, 40 studies were included. Evidence converged on five cue families: surface, discourse/pragmatic, epistemic/content, predictability/probabilistic, and provenance. Surface cues dominated the literature and were the most consistently operationalized. Discourse/pragmatic cues followed, particularly in discipline-bound academic genres where stance and metadiscourse differentiated AI from human writing. Predictability/probabilistic cues were central in detector-focused studies, while epistemic/content cues emerged primarily in tasks where grounding and authenticity were salient. Provenance cues were concentrated in watermarking research. Across studies, cue stability was consistently conditional rather than universal. Specifically, surface and discourse cues often remained discriminative within constrained genres, but shifted with register and discipline; probabilistic cues were powerful yet fragile under paraphrasing, post-editing, and evasion; and provenance signals required robustness to editing, mixing, and span localization. Overall, the literature indicates that AI&amp;amp;ndash;human distinction emerges from layered and context-dependent cue profiles rather than from any single reliable marker. High-stakes decisions, therefore, require condition-aware interpretation, triangulation across multiple cue families, and human oversight rather than automated classification in isolation.</p>
	]]></content:encoded>

	<dc:title>What Distinguishes AI-Generated from Human Writing? A Rapid Review of the Literature</dc:title>
			<dc:creator>Georgios P. Georgiou</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10020055</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-02-08</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-02-08</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>2</prism:number>
	<prism:section>Review</prism:section>
	<prism:startingPage>55</prism:startingPage>
		<prism:doi>10.3390/bdcc10020055</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/2/55</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/2/54">

	<title>BDCC, Vol. 10, Pages 54: Research on Modeling Method of eLoran Signal Propagation Delay Prediction Model: Integrating Path-Weighted Meteorological Data and Propagation Delay Data in Long-Distance Scenarios</title>
	<link>https://www.mdpi.com/2504-2289/10/2/54</link>
	<description>The enhanced long-range navigation (eLoran) system serves as an important backup method for the global navigation satellite system (GNSS) system. In long-distance transmission scenarios, the signal propagation delay of the eLoran system is affected by fluctuations in meteorological factors along the path. Regarding these issues, such as the potential timing system errors caused by meteorological factors and the limitation on the accuracy of the timing system, in this paper, an innovative prediction model is proposed to predict the propagation delay data by fusing the propagation delay data of multiple differential reference stations on the path and the path-weighted meteorological data. By collecting and processing actual data, four types of prediction tasks were designed. Comparative analyses of the prediction performance of eight common models were conducted on a unified dataset. The results show that the Pucheng&amp;amp;ndash;Zhengzhou path-weighted ten-factor back-propagation neural network (PZWT-BPNN) model performs the best, achieving a balance between prediction accuracy and training efficiency. This model effectively suppresses the timing errors caused by meteorological fluctuations and improves the prediction accuracy of the propagation delay of the system, providing corresponding technical support for key fields such as low-altitude economy and transportation.</description>
	<pubDate>2026-02-07</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 54: Research on Modeling Method of eLoran Signal Propagation Delay Prediction Model: Integrating Path-Weighted Meteorological Data and Propagation Delay Data in Long-Distance Scenarios</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/2/54">doi: 10.3390/bdcc10020054</a></p>
	<p>Authors:
		Tao Jin
		Shiyao Liu
		Baorong Yan
		Xiang Jiang
		Wei Guo
		Yu Hua
		Shougang Zhang
		Lu Xu
		</p>
	<p>The enhanced long-range navigation (eLoran) system serves as an important backup method for the global navigation satellite system (GNSS) system. In long-distance transmission scenarios, the signal propagation delay of the eLoran system is affected by fluctuations in meteorological factors along the path. Regarding these issues, such as the potential timing system errors caused by meteorological factors and the limitation on the accuracy of the timing system, in this paper, an innovative prediction model is proposed to predict the propagation delay data by fusing the propagation delay data of multiple differential reference stations on the path and the path-weighted meteorological data. By collecting and processing actual data, four types of prediction tasks were designed. Comparative analyses of the prediction performance of eight common models were conducted on a unified dataset. The results show that the Pucheng&amp;amp;ndash;Zhengzhou path-weighted ten-factor back-propagation neural network (PZWT-BPNN) model performs the best, achieving a balance between prediction accuracy and training efficiency. This model effectively suppresses the timing errors caused by meteorological fluctuations and improves the prediction accuracy of the propagation delay of the system, providing corresponding technical support for key fields such as low-altitude economy and transportation.</p>
	]]></content:encoded>

	<dc:title>Research on Modeling Method of eLoran Signal Propagation Delay Prediction Model: Integrating Path-Weighted Meteorological Data and Propagation Delay Data in Long-Distance Scenarios</dc:title>
			<dc:creator>Tao Jin</dc:creator>
			<dc:creator>Shiyao Liu</dc:creator>
			<dc:creator>Baorong Yan</dc:creator>
			<dc:creator>Xiang Jiang</dc:creator>
			<dc:creator>Wei Guo</dc:creator>
			<dc:creator>Yu Hua</dc:creator>
			<dc:creator>Shougang Zhang</dc:creator>
			<dc:creator>Lu Xu</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10020054</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-02-07</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-02-07</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>2</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>54</prism:startingPage>
		<prism:doi>10.3390/bdcc10020054</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/2/54</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
        <item rdf:about="https://www.mdpi.com/2504-2289/10/2/53">

	<title>BDCC, Vol. 10, Pages 53: Modern ICT Tools and Video Content in Athletes&amp;rsquo; Education&amp;mdash;Inspiration from Corporate Learning and Development</title>
	<link>https://www.mdpi.com/2504-2289/10/2/53</link>
	<description>Active athletes represent a specific target for learning and development. Their schedules, including training sessions and competitions, leave little time for education. However, athletes still need skills beyond sports to ensure they are prepared for future employment. Our study approaches this issue by identifying appropriate settings for athletes&amp;amp;rsquo; learning and development. (1) Based on the background of current athletes&amp;amp;rsquo; education, it addresses the gap of not enough attention being paid to transferable practices from corporate attitudes to learning and development. (2) The study&amp;amp;rsquo;s methodology primarily uses the case study concept because this conveys the video content we created for the athletes&amp;amp;rsquo; learning and development. This is combined with the method of content analysis of selected examples from corporate learning and development and the design thinking workshop, with the engagement of important stakeholder groups: athletes (2 participants), lecturers (2 participants), and representatives of sports organizations (1 participant). The other 9 workshop participants were master&amp;amp;rsquo;s students in a managerial study programme because of their age similarities with the current athletes and the applicability of the courses they were studying to athletes&amp;amp;rsquo; education. (3) The designed process was created as a digital twin using haptic artefacts and the S2M technology (version 1.0) within the OMiLAB platform (version 1.6). Our results show that video content tailored to the athletes&amp;amp;rsquo; constraints is a viable solution that improves their career prospects. (4) The study&amp;amp;rsquo;s practical implications are supported by the expert validation of the model provided by the inside of the large sports organizations&amp;amp;rsquo; management.</description>
	<pubDate>2026-02-06</pubDate>

	<content:encoded><![CDATA[
	<p><b>BDCC, Vol. 10, Pages 53: Modern ICT Tools and Video Content in Athletes&amp;rsquo; Education&amp;mdash;Inspiration from Corporate Learning and Development</b></p>
	<p>Big Data and Cognitive Computing <a href="https://www.mdpi.com/2504-2289/10/2/53">doi: 10.3390/bdcc10020053</a></p>
	<p>Authors:
		Martin Mičiak
		Dominika Toman
		Milan Kubina
		Tatiana Poljaková
		Klaudia Ivanovič
		Kvetoslava Šimová
		Anna Majchráková
		Ivana Bystrická
		Linda Kováčik
		Tibor Furmánek
		</p>
	<p>Active athletes represent a specific target for learning and development. Their schedules, including training sessions and competitions, leave little time for education. However, athletes still need skills beyond sports to ensure they are prepared for future employment. Our study approaches this issue by identifying appropriate settings for athletes&amp;amp;rsquo; learning and development. (1) Based on the background of current athletes&amp;amp;rsquo; education, it addresses the gap of not enough attention being paid to transferable practices from corporate attitudes to learning and development. (2) The study&amp;amp;rsquo;s methodology primarily uses the case study concept because this conveys the video content we created for the athletes&amp;amp;rsquo; learning and development. This is combined with the method of content analysis of selected examples from corporate learning and development and the design thinking workshop, with the engagement of important stakeholder groups: athletes (2 participants), lecturers (2 participants), and representatives of sports organizations (1 participant). The other 9 workshop participants were master&amp;amp;rsquo;s students in a managerial study programme because of their age similarities with the current athletes and the applicability of the courses they were studying to athletes&amp;amp;rsquo; education. (3) The designed process was created as a digital twin using haptic artefacts and the S2M technology (version 1.0) within the OMiLAB platform (version 1.6). Our results show that video content tailored to the athletes&amp;amp;rsquo; constraints is a viable solution that improves their career prospects. (4) The study&amp;amp;rsquo;s practical implications are supported by the expert validation of the model provided by the inside of the large sports organizations&amp;amp;rsquo; management.</p>
	]]></content:encoded>

	<dc:title>Modern ICT Tools and Video Content in Athletes&amp;amp;rsquo; Education&amp;amp;mdash;Inspiration from Corporate Learning and Development</dc:title>
			<dc:creator>Martin Mičiak</dc:creator>
			<dc:creator>Dominika Toman</dc:creator>
			<dc:creator>Milan Kubina</dc:creator>
			<dc:creator>Tatiana Poljaková</dc:creator>
			<dc:creator>Klaudia Ivanovič</dc:creator>
			<dc:creator>Kvetoslava Šimová</dc:creator>
			<dc:creator>Anna Majchráková</dc:creator>
			<dc:creator>Ivana Bystrická</dc:creator>
			<dc:creator>Linda Kováčik</dc:creator>
			<dc:creator>Tibor Furmánek</dc:creator>
		<dc:identifier>doi: 10.3390/bdcc10020053</dc:identifier>
	<dc:source>Big Data and Cognitive Computing</dc:source>
	<dc:date>2026-02-06</dc:date>

	<prism:publicationName>Big Data and Cognitive Computing</prism:publicationName>
	<prism:publicationDate>2026-02-06</prism:publicationDate>
	<prism:volume>10</prism:volume>
	<prism:number>2</prism:number>
	<prism:section>Article</prism:section>
	<prism:startingPage>53</prism:startingPage>
		<prism:doi>10.3390/bdcc10020053</prism:doi>
	<prism:url>https://www.mdpi.com/2504-2289/10/2/53</prism:url>
	
	<cc:license rdf:resource="CC BY 4.0"/>
</item>
    
<cc:License rdf:about="https://creativecommons.org/licenses/by/4.0/">
	<cc:permits rdf:resource="https://creativecommons.org/ns#Reproduction" />
	<cc:permits rdf:resource="https://creativecommons.org/ns#Distribution" />
	<cc:permits rdf:resource="https://creativecommons.org/ns#DerivativeWorks" />
</cc:License>

</rdf:RDF>
