Next Issue
Volume 15, August
Previous Issue
Volume 15, June
 
 

Algorithms, Volume 15, Issue 7 (July 2022) – 35 articles

Cover Story (view full-size image): Conflict detection ensures the correctness of packet classification and has received considerable attention in recent years. However, most conflict detection algorithms are implemented on a CPU. Compared with a CPU, a GPU exhibits higher computing power with parallel computing. In this study, we employed a GPU to develop two efficient algorithms for parallel conflict detection. In the first algorithm, we demonstrate how to perform conflict detection through parallel execution on GPU cores. In the second algorithm, we analyze the critical procedure in conflict detection so as to reduce the number of matches required for each filter. In addition, the second algorithm adopts a workload balance method to enable the load balancing of GPU execution threads, thereby significantly improving performance. View this paper
  • Issues are regarded as officially published after their release is announced to the table of contents alert mailing list.
  • You may sign up for e-mail alerts to receive table of contents of newly released issues.
  • PDF is the official format for papers published in both, html and pdf forms. To view the papers in pdf format, click on the "PDF Full-text" link, and use the free Adobe Reader to open them.
Order results
Result details
Section
Select all
Export citation of selected articles as:
15 pages, 607 KiB  
Article
LTU Attacker for Membership Inference
by Joseph Pedersen, Rafael Muñoz-Gómez, Jiangnan Huang, Haozhe Sun, Wei-Wei Tu and Isabelle Guyon
Algorithms 2022, 15(7), 254; https://doi.org/10.3390/a15070254 - 20 Jul 2022
Cited by 1 | Viewed by 1556
Abstract
We address the problem of defending predictive models, such as machine learning classifiers (Defender models), against membership inference attacks, in both the black-box and white-box setting, when the trainer and the trained model are publicly released. The Defender aims at optimizing a dual [...] Read more.
We address the problem of defending predictive models, such as machine learning classifiers (Defender models), against membership inference attacks, in both the black-box and white-box setting, when the trainer and the trained model are publicly released. The Defender aims at optimizing a dual objective: utility and privacy. Privacy is evaluated with the membership prediction error of a so-called “Leave-Two-Unlabeled” LTU Attacker, having access to all of the Defender and Reserved data, except for the membership label of one sample from each, giving the strongest possible attack scenario. We prove that, under certain conditions, even a “naïve” LTU Attacker can achieve lower bounds on privacy loss with simple attack strategies, leading to concrete necessary conditions to protect privacy, including: preventing over-fitting and adding some amount of randomness. This attack is straightforward to implement against any model trainer, and we demonstrate its performance against MemGaurd. However, we also show that such a naïve LTU Attacker can fail to attack the privacy of models known to be vulnerable in the literature, demonstrating that knowledge must be complemented with strong attack strategies to turn the LTU Attacker into a powerful means of evaluating privacy. The LTU Attacker can incorporate any existing attack strategy to compute individual privacy scores for each training sample. Our experiments on the QMNIST, CIFAR-10, and Location-30 datasets validate our theoretical results and confirm the roles of over-fitting prevention and randomness in the algorithms to protect against privacy attacks. Full article
(This article belongs to the Special Issue Black-Box Algorithms and Their Applications)
Show Figures

Figure 1

27 pages, 582 KiB  
Article
Linear Computation Coding: A Framework for Joint Quantization and Computing
by Ralf Reiner Müller, Bernhard Martin Wilhelm Gäde and Ali Bereyhi
Algorithms 2022, 15(7), 253; https://doi.org/10.3390/a15070253 - 20 Jul 2022
Cited by 5 | Viewed by 1969
Abstract
Here we introduce the new concept of computation coding. Similar to how rate-distortion theory is concerned with the lossy compression of data, computation coding deals with the lossy computation of functions. Particularizing to linear functions, we present an algorithmic approach to reduce the [...] Read more.
Here we introduce the new concept of computation coding. Similar to how rate-distortion theory is concerned with the lossy compression of data, computation coding deals with the lossy computation of functions. Particularizing to linear functions, we present an algorithmic approach to reduce the computational cost of multiplying a constant matrix with a variable vector, which requires neither a matrix nor vector having any particular structure or statistical properties. The algorithm decomposes the constant matrix into the product of codebook and wiring matrices whose entries are either zero or signed integer powers of two. For a typical application like the implementation of a deep neural network, the proposed algorithm reduces the number of required addition units several times. To achieve the accuracy of 16-bit signed integer arithmetic for 4k-vectors, no multipliers and only 1.5 adders per matrix entry are needed. Full article
(This article belongs to the Special Issue Algorithms in Reconfigurable Computing)
Show Figures

Figure 1

14 pages, 555 KiB  
Article
Informer-WGAN: High Missing Rate Time Series Imputation Based on Adversarial Training and a Self-Attention Mechanism
by Yufan Qian, Limei Tian, Baichen Zhai, Shufan Zhang and Rui Wu
Algorithms 2022, 15(7), 252; https://doi.org/10.3390/a15070252 - 20 Jul 2022
Cited by 4 | Viewed by 1869
Abstract
Missing observations in time series will distort the data characteristics, change the dataset expectations, high-order distances, and other statistics, and increase the difficulty of data analysis. Therefore, data imputation needs to be performed first. Generally, data imputation methods include statistical imputation, regression imputation, [...] Read more.
Missing observations in time series will distort the data characteristics, change the dataset expectations, high-order distances, and other statistics, and increase the difficulty of data analysis. Therefore, data imputation needs to be performed first. Generally, data imputation methods include statistical imputation, regression imputation, multiple imputation, and imputation based on machine learning methods. However, these methods currently have problems such as insufficient utilization of time characteristics, low imputation efficiency, and poor performance under high missing rates. In response to these problems, we propose the informer-WGAN, a network model based on adversarial training and a self-attention mechanism. With the help of the discriminator network and the random missing rate training method, the informer-WGAN can efficiently solve the problem of multidimensional time series imputation. According to the experimental results under different missing rates, the informer-WGAN model achieves better imputation results than the original informer on two datasets. Our model also shows excellent performance on time series imputation of the key parameters of a spacecraft control moment gyroscope (CMG). Full article
Show Figures

Figure 1

12 pages, 437 KiB  
Article
Adaptive IDS for Cooperative Intelligent Transportation Systems Using Deep Belief Networks
by Sultan Ahmed Almalki, Ahmed Abdel-Rahim and Frederick T. Sheldon
Algorithms 2022, 15(7), 251; https://doi.org/10.3390/a15070251 - 20 Jul 2022
Cited by 4 | Viewed by 1690
Abstract
The adoption of cooperative intelligent transportation systems (cITSs) improves road safety and traffic efficiency. Vehicles connected to cITS form vehicular ad hoc networks (VANET) to exchange messages. Like other networks and systems, cITSs are targeted by attackers intent on compromising and disrupting system [...] Read more.
The adoption of cooperative intelligent transportation systems (cITSs) improves road safety and traffic efficiency. Vehicles connected to cITS form vehicular ad hoc networks (VANET) to exchange messages. Like other networks and systems, cITSs are targeted by attackers intent on compromising and disrupting system integrity and availability. They can repeatedly spoof false information causing bottlenecks, traffic jams and even road accidents. The existing security infrastructure assumes that the network topology and/or attack behavior is static. However, the cITS is inherently dynamic in nature. Moreover, attackers may have the ability and resources to change their behavior continuously. Assuming a static IDS security model for VANETs is not suitable and can lead to low detection accuracy and high false alarms. Therefore, this paper proposes an adaptive security solution based on deep learning and contextual references that can cope with the dynamic nature of the cITS topologies and increasingly common attack behaviors. In this study, deep belief networks (DBN) modeling was used to train the detection model. Binary cross entropy was used as a loss function to measure the prediction error. Two activation functions were used, Relu and Softmax, for input–output mapping. The Relu was used in the hidden layers, while the Sigmoid was used in the last layer to map the real vector to output between 0 and 1. The adaptation mechanism was incorporated into the detection model using a moving average that monitors predicted values within a time window. In this way, the model can readjust the classification thresholds on-the-fly as appropriate. The proposed model was evaluated using the Next Generation Simulation (NGSIM) dataset, which is commonly used in such related works. The result is improved accuracy, demonstrating that the adaptation mechanism used in this study was effective. Full article
Show Figures

Figure 1

26 pages, 2207 KiB  
Article
Multi-Fidelity Low-Rank Approximations for Uncertainty Quantification of a Supersonic Aircraft Design
by Sihmehmet Yildiz, Hayriye Pehlivan Solak and Melike Nikbay
Algorithms 2022, 15(7), 250; https://doi.org/10.3390/a15070250 - 19 Jul 2022
Cited by 2 | Viewed by 2220
Abstract
Uncertainty quantification has proven to be an indispensable study for enhancing reliability and robustness of engineering systems in the early design phase. Single and multi-fidelity surrogate modelling methods have been used to replace the expensive high fidelity analyses which must be repeated many [...] Read more.
Uncertainty quantification has proven to be an indispensable study for enhancing reliability and robustness of engineering systems in the early design phase. Single and multi-fidelity surrogate modelling methods have been used to replace the expensive high fidelity analyses which must be repeated many times for uncertainty quantification. However, since the number of analyses required to build an accurate surrogate model increases exponentially with the number of random input variables, most surrogate modelling methods suffer from the curse of dimensionality. As an alternative approach, the Low-Rank Approximation method can be applied to high-dimensional uncertainty quantification studies with a low computational cost, where the number of coefficients for building the surrogate model increases only linearly with the number of random input variables. In this study, the Low-Rank Approximation method is implemented for multi-fidelity applications with additive and multiplicative correction approaches to make the high-dimensional uncertainty quantification analysis more efficient and accurate. The developed uncertainty quantification methodology is tested on supersonic aircraft design problems and its predictions are compared with the results of single- and multi-fidelity Polynomial Chaos Expansion and Monte Carlo methods. For the same computational cost, the Low-Rank Approximation method outperformed both in surrogate modeling and uncertainty quantification cases for all the benchmarks and real-world engineering problems addressed in the present study. Full article
Show Figures

Figure 1

13 pages, 546 KiB  
Article
Intuitionistic and Interval-Valued Fuzzy Set Representations for Data Mining
by Fred Petry and Ronald Yager
Algorithms 2022, 15(7), 249; https://doi.org/10.3390/a15070249 - 19 Jul 2022
Cited by 2 | Viewed by 1288
Abstract
Data mining refers to a variety of techniques in the fields of databases, machine learning and pattern recognition. The intent is to obtain useful patterns and associations from a large collection of data. In this paper we describe extensions to the attribute generalization [...] Read more.
Data mining refers to a variety of techniques in the fields of databases, machine learning and pattern recognition. The intent is to obtain useful patterns and associations from a large collection of data. In this paper we describe extensions to the attribute generalization process to deal with interval and intuitionistic fuzzy information. Specifically, we consider extensions for using interval-valued fuzzy representations in both data and the generalization hierarchy. Moreover, preliminary representations using intuitionistic fuzzy information for attribute generalization are described. Finally, we consider how to use fuzzy hierarchies for the generalization of interval-valued fuzzy representations. Full article
Show Figures

Figure 1

13 pages, 1921 KiB  
Article
Semi-Automatic Multiparametric MR Imaging Classification Using Novel Image Input Sequences and 3D Convolutional Neural Networks
by Bochong Li, Ryo Oka, Ping Xuan, Yuichiro Yoshimura and Toshiya Nakaguchi
Algorithms 2022, 15(7), 248; https://doi.org/10.3390/a15070248 - 18 Jul 2022
Cited by 2 | Viewed by 1728
Abstract
The role of multi-parametric magnetic resonance imaging (mp-MRI) is becoming increasingly important in the diagnosis of the clinical severity of prostate cancer (PCa). However, mp-MRI images usually contain several unaligned 3D sequences, such as DWI image sequences and T2-weighted image sequences, and there [...] Read more.
The role of multi-parametric magnetic resonance imaging (mp-MRI) is becoming increasingly important in the diagnosis of the clinical severity of prostate cancer (PCa). However, mp-MRI images usually contain several unaligned 3D sequences, such as DWI image sequences and T2-weighted image sequences, and there are many images among the entirety of 3D sequence images that do not contain cancerous tissue, which affects the accuracy of large-scale prostate cancer detection. Therefore, there is a great need for a method that uses accurate computer-aided detection of mp-MRI images and minimizes the influence of useless features. Our proposed PCa detection method is divided into three stages: (i) multimodal image alignment, (ii) automatic cropping of the sequence images to the entire prostate region, and, finally, (iii) combining multiple modal images of each patient into novel 3D sequences and using 3D convolutional neural networks to learn the newly composed 3D sequences with different modal alignments. We arrange the different modal methods to make the model fully learn the cancerous tissue features; then, we predict the clinical severity of PCa and generate a 3D cancer response map for the 3D sequence images from the last convolution layer of the network. The prediction results and 3D response map help to understand the features that the model focuses on during the process of 3D-CNN feature learning. We applied our method to Toho hospital prostate cancer patient data; the AUC (=0.85) results were significantly higher than those of other methods. Full article
(This article belongs to the Special Issue Algorithms for Biomedical Image Analysis and Processing)
Show Figures

Figure 1

26 pages, 790 KiB  
Review
Comparative Review of the Intrusion Detection Systems Based on Federated Learning: Advantages and Open Challenges
by Elena Fedorchenko, Evgenia Novikova and Anton Shulepov
Algorithms 2022, 15(7), 247; https://doi.org/10.3390/a15070247 - 15 Jul 2022
Cited by 12 | Viewed by 4397
Abstract
In order to provide an accurate and timely response to different types of the attacks, intrusion and anomaly detection systems collect and analyze a lot of data that may include personal and other sensitive data. These systems could be considered a source of [...] Read more.
In order to provide an accurate and timely response to different types of the attacks, intrusion and anomaly detection systems collect and analyze a lot of data that may include personal and other sensitive data. These systems could be considered a source of privacy-aware risks. Application of the federated learning paradigm for training attack and anomaly detection models may significantly decrease such risks as the data generated locally are not transferred to any party, and training is performed mainly locally on data sources. Another benefit of the usage of federated learning for intrusion detection is its ability to support collaboration between entities that could not share their dataset for confidential or other reasons. While this approach is able to overcome the aforementioned challenges it is rather new and not well-researched. The challenges and research questions appear while using it to implement analytical systems. In this paper, the authors review existing solutions for intrusion and anomaly detection based on the federated learning, and study their advantages as well as open challenges still facing them. The paper analyzes the architecture of the proposed intrusion detection systems and the approaches used to model data partition across the clients. The paper ends with discussion and formulation of the open challenges. Full article
Show Figures

Figure 1

42 pages, 5395 KiB  
Article
A Two-Stage Multi-Objective Genetic Algorithm for a Flexible Job Shop Scheduling Problem with Lot Streaming
by Danial Rooyani and Fantahun Defersha
Algorithms 2022, 15(7), 246; https://doi.org/10.3390/a15070246 - 13 Jul 2022
Cited by 5 | Viewed by 2611
Abstract
The work in this paper is motivated by a recently published article in which the authors developed an efficient two-stage genetic algorithm for a comprehensive model of a flexible job-shop scheduling problem (FJSP). In this paper, we extend the application of the algorithm [...] Read more.
The work in this paper is motivated by a recently published article in which the authors developed an efficient two-stage genetic algorithm for a comprehensive model of a flexible job-shop scheduling problem (FJSP). In this paper, we extend the application of the algorithm to solve a lot streaming problem in FJSP while at the same time expanding the model to incorporate multiple objectives. The objective function terms included in our current work are the minimization of the (1) makespan, (2) maximum sublot flowtime, (3) total sublot flow time, (4) maximum job flowtime, (5) total job flow time, (6) maximum sublot finish-time separation, (7) total sublot finish-time separation, (8) maximum machine load, (9) total machine load, and (10) maximum machine load difference. Numerical examples are presented to illustrate the greater need for multi-objective optimization in larger problems, the interaction of the various objective function terms, and their relevance in providing better solution quality. The ability of the two-stage genetic algorithm to jointly optimize all the objective function terms is also investigated. The results show that the algorithm can generate initial solutions that are highly improved in all of the objective function terms. It also outperforms the regular genetic algorithm in convergence speed and final solution quality in solving the multi-objective FJSP lot streaming. We also demonstrate that high-performance parallel computation can further improve the performance of the two-stage genetic algorithm. Nevertheless, the sequential two-stage genetic algorithm with a single CPU outperforms the parallel regular genetic algorithm that uses many CPUs, asserting the superiority of the two-stage genetic algorithm in solving the proposed multi-objective FJSP lot streaming. Full article
(This article belongs to the Topic Mathematical Modeling in Physical Sciences)
Show Figures

Figure 1

26 pages, 4532 KiB  
Article
Efficient Local Refinement near Parametric Boundaries Using kd-Tree Data Structure and Algebraic Level Sets
by Tao Song, Huanyu Liao and Ganesh Subbarayan
Algorithms 2022, 15(7), 245; https://doi.org/10.3390/a15070245 - 13 Jul 2022
Cited by 1 | Viewed by 1840
Abstract
In analysis of problems with parametric spline boundaries that are immersed or inserted into an underlying domain, the discretization on the underlying domain usually does not conform to the inserted boundaries. While the fixed underlying discretization is of great convenience as the immersed [...] Read more.
In analysis of problems with parametric spline boundaries that are immersed or inserted into an underlying domain, the discretization on the underlying domain usually does not conform to the inserted boundaries. While the fixed underlying discretization is of great convenience as the immersed boundaries evolve, the field approximations near the inserted boundaries require refinement in the underlying domain, as do the quadrature cells. In this paper, a kd-tree data structure together with a sign-based and/or distance-based refinement strategy is proposed for local refinement near the inserted boundaries as well as for adaptive quadrature near the boundaries. The developed algorithms construct and utilize implicit forms of parametric Non-Uniform Rational B-Spline (NURBS) surfaces to algebraically (and non-iteratively) estimate distance as well as sign relative to the inserted boundary. The kd-tree local refinement is demonstrated to produce fewer sub-cells for the same accuracy of solution as compared to the classical quad/oct tree-based subdivision. Consistent with the kd-tree data structure, we describe a new a priori refinement algorithm based on the signed and unsigned distance from the inserted boundary. We first demonstrate the local refinement strategy coupled with the the kd-tree data structure by constructing Truncated Hierarchical B-spline (THB-spline) “meshes”. We next demonstrate the accuracy and efficiency of the developed local refinement strategy through adaptive quadrature near NURBS boundaries inserted within volumetric three-dimensional NURBS discretizations. Full article
Show Figures

Figure 1

18 pages, 1468 KiB  
Article
Inference Acceleration with Adaptive Distributed DNN Partition over Dynamic Video Stream
by Jin Cao, Bo Li, Mengni Fan and Huiyu Liu
Algorithms 2022, 15(7), 244; https://doi.org/10.3390/a15070244 - 13 Jul 2022
Cited by 1 | Viewed by 1743
Abstract
Deep neural network-based computer vision applications have exploded and are widely used in intelligent services for IoT devices. Due to the computationally intensive nature of DNNs, the deployment and execution of intelligent applications in smart scenarios face the challenge of limited device resources. [...] Read more.
Deep neural network-based computer vision applications have exploded and are widely used in intelligent services for IoT devices. Due to the computationally intensive nature of DNNs, the deployment and execution of intelligent applications in smart scenarios face the challenge of limited device resources. Existing job scheduling strategies are single-focused and have limited support for large-scale end-device scenarios. In this paper, we present ADDP, an adaptive distributed DNN partition method that supports video analysis on large-scale smart cameras. ADDP applies to the commonly used DNN models for computer vision and contains a feature-map layer partition module (FLP) supporting edge-to-end collaborative model partition and a feature-map size partition (FSP) module supporting multidevice parallel inference. Based on the inference delay minimization objective, FLP and FSP achieve a tradeoff between the arithmetic and communication resources of different devices. We validate ADDP on heterogeneous devices and show that both the FLP module and the FSP module outperform existing approaches and reduce single-frame response latency by 10–25% compared to the pure on-device processing. Full article
(This article belongs to the Special Issue Deep Learning for Internet of Things)
Show Figures

Figure 1

12 pages, 575 KiB  
Article
A Federated Generalized Linear Model for Privacy-Preserving Analysis
by Matteo Cellamare, Anna J. van Gestel, Hasan Alradhi, Frank Martin and Arturo Moncada-Torres
Algorithms 2022, 15(7), 243; https://doi.org/10.3390/a15070243 - 13 Jul 2022
Cited by 9 | Viewed by 2475
Abstract
In the last few years, federated learning (FL) has emerged as a novel alternative for analyzing data spread across different parties without needing to centralize them. In order to increase the adoption of FL, there is a need to develop more algorithms that [...] Read more.
In the last few years, federated learning (FL) has emerged as a novel alternative for analyzing data spread across different parties without needing to centralize them. In order to increase the adoption of FL, there is a need to develop more algorithms that can be deployed under this novel privacy-preserving paradigm. In this paper, we present our federated generalized linear model (GLM) for horizontally partitioned data. It allows generating models of different families (linear, Poisson, logistic) without disclosing privacy-sensitive individual records. We describe its algorithm (which can be implemented in the user’s platform of choice) and compare the obtained federated models against their centralized counterpart, which were mathematically equivalent. We also validated their execution time with increasing numbers of records and involved parties. We show that our federated GLM is accurate enough to be used for the privacy-preserving analysis of horizontally partitioned data in real-life scenarios. Further development of this type of algorithm has the potential to make FL a much more common practice among researchers. Full article
(This article belongs to the Special Issue Interpretability, Accountability and Robustness in Machine Learning)
Show Figures

Figure 1

22 pages, 474 KiB  
Article
A New String Edit Distance and Applications
by Taylor Petty, Jan Hannig, Tunde I. Huszar and Hari Iyer
Algorithms 2022, 15(7), 242; https://doi.org/10.3390/a15070242 - 12 Jul 2022
Cited by 1 | Viewed by 1938
Abstract
String edit distances have been used for decades in applications ranging from spelling correction and web search suggestions to DNA analysis. Most string edit distances are variations of the Levenshtein distance and consider only single-character edits. In forensic applications polymorphic genetic markers such [...] Read more.
String edit distances have been used for decades in applications ranging from spelling correction and web search suggestions to DNA analysis. Most string edit distances are variations of the Levenshtein distance and consider only single-character edits. In forensic applications polymorphic genetic markers such as short tandem repeats (STRs) are used. At these repetitive motifs the DNA copying errors consist of more than just single base differences. More often the phenomenon of “stutter” is observed, where the number of repeated units differs (by whole units) from the template. To adapt the Levenshtein distance to be suitable for forensic applications where DNA sequence similarity is of interest, a generalized string edit distance is defined that accommodates the addition or deletion of whole motifs in addition to single-nucleotide edits. A dynamic programming implementation is developed for computing this distance between sequences. The novelty of this algorithm is in handling the complex interactions that arise between multiple- and single-character edits. Forensic examples illustrate the purpose and use of the Restricted Forensic Levenshtein (RFL) distance measure, but applications extend to sequence alignment and string similarity in other biological areas, as well as dynamic programming algorithms more broadly. Full article
Show Figures

Figure 1

34 pages, 33546 KiB  
Article
Topology Optimisation under Uncertainties with Neural Networks
by Martin Eigel, Marvin Haase and Johannes Neumann
Algorithms 2022, 15(7), 241; https://doi.org/10.3390/a15070241 - 12 Jul 2022
Cited by 1 | Viewed by 1644
Abstract
Topology optimisation is a mathematical approach relevant to different engineering problems where the distribution of material in a defined domain is distributed in some optimal way, subject to a predefined cost function representing desired (e.g., mechanical) properties and constraints. The computation of such [...] Read more.
Topology optimisation is a mathematical approach relevant to different engineering problems where the distribution of material in a defined domain is distributed in some optimal way, subject to a predefined cost function representing desired (e.g., mechanical) properties and constraints. The computation of such an optimal distribution depends on the numerical solution of some physical model (in our case linear elasticity) and robustness is achieved by introducing uncertainties into the model data, namely the forces acting on the structure and variations of the material stiffness, rendering the task high-dimensional and computationally expensive. To alleviate this computational burden, we develop two neural network architectures (NN) that are capable of predicting the gradient step of the optimisation procedure. Since state-of-the-art methods use adaptive mesh refinement, the neural networks are designed to use a sufficiently fine reference mesh such that only one training phase of the neural network suffices. As a first architecture, a convolutional neural network is adapted to the task. To include sequential information of the optimisation process, a recurrent neural network is constructed as a second architecture. A common 2D bridge benchmark is used to illustrate the performance of the proposed architectures. It is observed that the NN prediction of the gradient step clearly outperforms the classical optimisation method, in particular since larger iteration steps become viable. Full article
(This article belongs to the Special Issue Stochastic Algorithms and Their Applications)
Show Figures

Figure 1

13 pages, 2969 KiB  
Article
Tunnel Traffic Evolution during Capacity Drop Based on High-Resolution Vehicle Trajectory Data
by Lu Yang, Chishe Wang and Zhibin Li
Algorithms 2022, 15(7), 240; https://doi.org/10.3390/a15070240 - 12 Jul 2022
Cited by 3 | Viewed by 1535
Abstract
Capacity drop is the critical phenomenon that triggers traffic congestion, while traffic evolution is very complex during a capacity drop. This study applied the empirical vehicle trajectory data to explore the traffic characteristics during the capacity drop at the tunnel bottleneck section. We [...] Read more.
Capacity drop is the critical phenomenon that triggers traffic congestion, while traffic evolution is very complex during a capacity drop. This study applied the empirical vehicle trajectory data to explore the traffic characteristics during the capacity drop at the tunnel bottleneck section. We first construct a capacity drop analysis model using image processing technology to extract high-precision vehicle trajectories. We then analyze the characteristics of the evolution process of the capacity drop at the bottleneck area. The results show that the capacity drop is a dynamic evolution process from free flow to congested flow where traffic operation is distinct. The capacity drop shows the difference between congested flow and non-congested flow. The driving characteristics of drivers in the two states are also different. The influence of lane-changing behavior on the capacity drop is estimated. In the free flow state, the disturbance caused by lane-changing can be quickly eliminated. With the increase in vehicle numbers in the area, the frequent lane-changing behavior accumulates disturbance. When the disturbance reaches a certain degree, congestion will occur, and the vehicle’s speed will drop sharply, resulting in a capacity drop. Full article
Show Figures

Figure 1

25 pages, 3007 KiB  
Article
IoT Multi-Vector Cyberattack Detection Based on Machine Learning Algorithms: Traffic Features Analysis, Experiments, and Efficiency
by Sergii Lysenko, Kira Bobrovnikova, Vyacheslav Kharchenko and Oleg Savenko
Algorithms 2022, 15(7), 239; https://doi.org/10.3390/a15070239 - 12 Jul 2022
Cited by 5 | Viewed by 2649
Abstract
Cybersecurity is a common Internet of Things security challenge. The lack of security in IoT devices has led to a great number of devices being compromised, with threats from both inside and outside the IoT infrastructure. Attacks on the IoT infrastructure result in [...] Read more.
Cybersecurity is a common Internet of Things security challenge. The lack of security in IoT devices has led to a great number of devices being compromised, with threats from both inside and outside the IoT infrastructure. Attacks on the IoT infrastructure result in device hacking, data theft, financial loss, instability, or even physical damage to devices. This requires the development of new approaches to ensure high-security levels in IoT infrastructure. To solve this problem, we propose a new approach for IoT cyberattack detection based on machine learning algorithms. The core of the method involves network traffic analyses that IoT devices generate during communication. The proposed approach deals with the set of network traffic features that may indicate the presence of cyberattacks in the IoT infrastructure and compromised IoT devices. Based on the obtained features for each IoT device, the feature vectors are formed. To conclude the possible attack presence, machine learning algorithms were employed. We assessed the complexity and time of machine learning algorithm implementation considering multi-vector cyberattacks on IoT infrastructure. Experiments were conducted to approve the method’s efficiency. The results demonstrated that the network traffic feature-based approach allows the detection of multi-vector cyberattacks with high efficiency. Full article
Show Figures

Figure 1

17 pages, 4194 KiB  
Article
Research on Network Attack Traffic Detection HybridAlgorithm Based on UMAP-RF
by Xiaoyu Du, Cheng Cheng, Yujing Wang and Zhijie Han
Algorithms 2022, 15(7), 238; https://doi.org/10.3390/a15070238 - 09 Jul 2022
Cited by 2 | Viewed by 1612
Abstract
Network attack traffic detection plays a crucial role in protecting network operations and services. To accurately detect malicious traffic on the internet, this paper designs a hybrid algorithm UMAP-RF for both binary and multiclassification network attack detection tasks. First, the network traffic data [...] Read more.
Network attack traffic detection plays a crucial role in protecting network operations and services. To accurately detect malicious traffic on the internet, this paper designs a hybrid algorithm UMAP-RF for both binary and multiclassification network attack detection tasks. First, the network traffic data are dimensioned down with UMAP algorithm. The random forest algorithm is improved based on parameter optimization, and the improved random forest algorithm is used to classify the network traffic data, distinguishing normal data from abnormal data and classifying nine different types of network attacks from the abnormal data. Experimental results on the UNSW-NB15 dataset, which are significant improvements compared to traditional machine-learning methods, show that the UMAP-RF hybrid model can perform network attack traffic detection effectively, with accuracy and recall rates of 92.6% and 91%, respectively. Full article
Show Figures

Figure 1

19 pages, 3621 KiB  
Article
An Efficient Parallel Algorithm for Detecting Packet Filter Conflicts
by Chun-Liang Lee, Guan-Yu Lin and Yaw-Chung Chen
Algorithms 2022, 15(7), 237; https://doi.org/10.3390/a15070237 - 07 Jul 2022
Cited by 1 | Viewed by 1610
Abstract
Advanced network services, such as firewalls, policy-based routing, and virtual private networks, must rely on routers to classify packets into different flows based on packet headers and predefined filter tables. When multiple filters are overlapped, conflicts may occur, leading to ambiguity in the [...] Read more.
Advanced network services, such as firewalls, policy-based routing, and virtual private networks, must rely on routers to classify packets into different flows based on packet headers and predefined filter tables. When multiple filters are overlapped, conflicts may occur, leading to ambiguity in the packet classification. Conflict detection ensures the correctness of packet classification and has received considerable attention in recent years. However, most conflict-detection algorithms are implemented on a conventional central processing unit (CPU). Compared with a CPU, a graphics processing unit (GPU) exhibits higher computing power with parallel computing, hence accelerates the execution speed of conflict detection. In this study, we employed a GPU to develop two efficient algorithms for parallel conflict detection: the general parallel conflict-detection algorithm (the GPCDA) and the enhanced parallel conflict-detection algorithm (the EPCDA). In the GPCDA, we demonstrate how to perform conflict detection through parallel execution on GPU cores. While in the EPCDA, we analyze the critical procedure in conflict detection as to reduce the number of matches required for each filter. In addition, the EPCDA adopts a workload balance method to enable load balancing of GPU execution threads, thereby significantly improving performance. The simulation results show that with the 100 K filter database, the GPCDA and the EPCDA execute conflict detection 2.8 to 13.9 and 9.4 to 33.7 times faster, respectively, than the CPU-based algorithm. Full article
(This article belongs to the Special Issue Algorithms for Distributed Computing)
Show Figures

Figure 1

15 pages, 1534 KiB  
Article
Automatic Classification of Foot Thermograms Using Machine Learning Techniques
by Vítor Filipe, Pedro Teixeira and Ana Teixeira
Algorithms 2022, 15(7), 236; https://doi.org/10.3390/a15070236 - 06 Jul 2022
Cited by 7 | Viewed by 2131
Abstract
Diabetic foot is one of the main complications observed in diabetic patients; it is associated with the development of foot ulcers and can lead to amputation. In order to diagnose these complications, specialists have to analyze several factors. To aid their decisions and [...] Read more.
Diabetic foot is one of the main complications observed in diabetic patients; it is associated with the development of foot ulcers and can lead to amputation. In order to diagnose these complications, specialists have to analyze several factors. To aid their decisions and help prevent mistakes, the resort to computer-assisted diagnostic systems using artificial intelligence techniques is gradually increasing. In this paper, two different models for the classification of thermograms of the feet of diabetic and healthy individuals are proposed and compared. In order to detect and classify abnormal changes in the plantar temperature, machine learning algorithms are used in both models. In the first model, the foot thermograms are classified into four classes: healthy and three categories for diabetics. The second model has two stages: in the first stage, the foot is classified as belonging to a diabetic or healthy individual, while, in the second stage, a classification refinement is conducted, classifying diabetic foot into three classes of progressive severity. The results show that both proposed models proved to be efficient, allowing us to classify a foot thermogram as belonging to a healthy or diabetic individual, with the diabetic ones divided into three classes; however, when compared, Model 2 outperforms Model 1 and allows for a better performance classification concerning the healthy category and the first class of diabetic individuals. These results demonstrate that the proposed methodology can be a tool to aid medical diagnosis. Full article
(This article belongs to the Special Issue Algorithms for Machine Learning and Pattern Recognition Tasks)
Show Figures

Figure 1

15 pages, 11309 KiB  
Article
Vector Fitting–Cauchy Method for the Extraction of Complex Natural Resonances in Ground Penetrating Radar Operations
by Andres Gallego, Francisco Roman and Edwin Pineda
Algorithms 2022, 15(7), 235; https://doi.org/10.3390/a15070235 - 03 Jul 2022
Cited by 3 | Viewed by 2110
Abstract
In this paper, we obtain the Complex Natural Resonances of an object from the backscattered response in the frequency domain with a novel rational function approximation method based on both Vector Fitting and Cauchy methods. We determine the system order and an initial [...] Read more.
In this paper, we obtain the Complex Natural Resonances of an object from the backscattered response in the frequency domain with a novel rational function approximation method based on both Vector Fitting and Cauchy methods. We determine the system order and an initial set of poles, which are used as a basis for a rational function approximation. The results from the simulations and experiments show an improvement in the reconstructed signals and the accuracy of the CNRs calculated, with an increased tolerance to the critical Signal-to-Noise Ratio. This is being used in the problem of GPR landmine humanitarian detection in Colombia. Full article
(This article belongs to the Special Issue Computational Methods and Optimization for Numerical Analysis)
Show Figures

Figure 1

17 pages, 621 KiB  
Article
Performance Evaluation of Open-Source Serverless Platforms for Kubernetes
by Jonathan Decker, Piotr Kasprzak and Julian Martin Kunkel
Algorithms 2022, 15(7), 234; https://doi.org/10.3390/a15070234 - 02 Jul 2022
Cited by 1 | Viewed by 3960
Abstract
Serverless computing has grown massively in popularity over the last few years, and has provided developers with a way to deploy function-sized code units without having to take care of the actual servers or deal with logging, monitoring, and scaling of their code. [...] Read more.
Serverless computing has grown massively in popularity over the last few years, and has provided developers with a way to deploy function-sized code units without having to take care of the actual servers or deal with logging, monitoring, and scaling of their code. High-performance computing (HPC) clusters can profit from improved serverless resource sharing capabilities compared to reservation-based systems such as Slurm. However, before running self-hosted serverless platforms in HPC becomes a viable option, serverless platforms must be able to deliver a decent level of performance. Other researchers have already pointed out that there is a distinct lack of studies in the area of comparative benchmarks on serverless platforms, especially for open-source self-hosted platforms. This study takes a step towards filling this gap by systematically benchmarking two promising self-hosted Kubernetes-based serverless platforms in comparison. While the resulting benchmarks signal potential, they demonstrate that many opportunities for performance improvements in serverless computing are being left on the table. Full article
(This article belongs to the Special Issue Performance Optimization and Performance Evaluation)
Show Figures

Figure 1

31 pages, 793 KiB  
Article
ZenoPS: A Distributed Learning System Integrating Communication Efficiency and Security
by Cong Xie, Oluwasanmi Koyejo and Indranil Gupta
Algorithms 2022, 15(7), 233; https://doi.org/10.3390/a15070233 - 01 Jul 2022
Cited by 2 | Viewed by 2058
Abstract
Distributed machine learning is primarily motivated by the promise of increased computation power for accelerating training and mitigating privacy concerns. Unlike machine learning on a single device, distributed machine learning requires collaboration and communication among the devices. This creates several new challenges: (1) [...] Read more.
Distributed machine learning is primarily motivated by the promise of increased computation power for accelerating training and mitigating privacy concerns. Unlike machine learning on a single device, distributed machine learning requires collaboration and communication among the devices. This creates several new challenges: (1) the heavy communication overhead can be a bottleneck that slows down the training, and (2) the unreliable communication and weaker control over the remote entities make the distributed system vulnerable to systematic failures and malicious attacks. This paper presents a variant of stochastic gradient descent (SGD) with improved communication efficiency and security in distributed environments. Our contributions include (1) a new technique called error reset to adapt both infrequent synchronization and message compression for communication reduction in both synchronous and asynchronous training, (2) new score-based approaches for validating the updates, and (3) integration with both error reset and score-based validation. The proposed system provides communication reduction, both synchronous and asynchronous training, Byzantine tolerance, and local privacy preservation. We evaluate our techniques both theoretically and empirically. Full article
(This article belongs to the Special Issue Gradient Methods for Optimization)
Show Figures

Figure 1

12 pages, 584 KiB  
Article
Generating Higher-Fidelity Synthetic Datasets with Privacy Guarantees
by Aleksei Triastcyn and Boi Faltings
Algorithms 2022, 15(7), 232; https://doi.org/10.3390/a15070232 - 01 Jul 2022
Cited by 2 | Viewed by 1535
Abstract
We consider the problem of enhancing user privacy in common data analysis and machine learning development tasks, such as data annotation and inspection, by substituting the real data with samples from a generative adversarial network. We propose employing Bayesian differential privacy as the [...] Read more.
We consider the problem of enhancing user privacy in common data analysis and machine learning development tasks, such as data annotation and inspection, by substituting the real data with samples from a generative adversarial network. We propose employing Bayesian differential privacy as the means to achieve a rigorous theoretical guarantee while providing a better privacy-utility trade-off. We demonstrate experimentally that our approach produces higher-fidelity samples compared to prior work, allowing to (1) detect more subtle data errors and biases, and (2) reduce the need for real data labelling by achieving high accuracy when training directly on artificial samples. Full article
(This article belongs to the Special Issue Privacy Preserving Machine Learning)
Show Figures

Figure 1

21 pages, 1308 KiB  
Review
Automatic Atrial Fibrillation Arrhythmia Detection Using Univariate and Multivariate Data
by Zouhair Haddi, Bouchra Ananou, Miquel Alfaras, Mustapha Ouladsine, Jean-Claude Deharo, Narcís Avellana and Stéphane Delliaux
Algorithms 2022, 15(7), 231; https://doi.org/10.3390/a15070231 - 01 Jul 2022
Cited by 2 | Viewed by 2080
Abstract
Atrial fibrillation (AF) is still a major cause of disease morbidity and mortality, making its early diagnosis desirable and urging researchers to develop efficient methods devoted to automatic AF detection. Till now, the analysis of Holter-ECG recordings remains the gold-standard technique to screen [...] Read more.
Atrial fibrillation (AF) is still a major cause of disease morbidity and mortality, making its early diagnosis desirable and urging researchers to develop efficient methods devoted to automatic AF detection. Till now, the analysis of Holter-ECG recordings remains the gold-standard technique to screen AF. This is usually achieved by studying either RR interval time series analysis, P-wave detection or combinations of both morphological characteristics. After extraction and selection of meaningful features, each of the AF detection methods might be conducted through univariate and multivariate data analysis. Many of these automatic techniques have been proposed over the last years. This work presents an overview of research studies of AF detection based on RR interval time series. The aim of this paper is to provide the scientific community and newcomers to the field of AF screening with a resource that presents introductory concepts, clinical features, and a literature review that describes the techniques that are mostly followed when RR interval time series are used for accurate detection of AF. Full article
Show Figures

Figure 1

13 pages, 900 KiB  
Article
Time Series Analysis of Cryptocurrency Prices Using Long Short-Term Memory
by Jacques Phillipe Fleischer, Gregor von Laszewski, Carlos Theran and Yohn Jairo Parra Bautista
Algorithms 2022, 15(7), 230; https://doi.org/10.3390/a15070230 - 01 Jul 2022
Cited by 7 | Viewed by 5080
Abstract
Digitization is changing our world, creating innovative finance channels and emerging technology such as cryptocurrencies, which are applications of blockchain technology. However, cryptocurrency price volatility is one of this technology’s main trade-offs. In this paper, we explore a time series analysis using deep [...] Read more.
Digitization is changing our world, creating innovative finance channels and emerging technology such as cryptocurrencies, which are applications of blockchain technology. However, cryptocurrency price volatility is one of this technology’s main trade-offs. In this paper, we explore a time series analysis using deep learning to study the volatility and to understand this behavior. We apply a long short-term memory model to learn the patterns within cryptocurrency close prices and to predict future prices. The proposed model learns from the close values. The performance of this model is evaluated using the root-mean-squared error and by comparing it to an ARIMA model. Full article
(This article belongs to the Special Issue Advances in Blockchain Architecture and Consensus)
Show Figures

Figure 1

18 pages, 679 KiB  
Article
Privacy-Preserving Feature Selection with Fully Homomorphic Encryption
by Shinji Ono, Jun Takata, Masaharu Kataoka, Tomohiro I, Kilho Shin and Hiroshi Sakamoto
Algorithms 2022, 15(7), 229; https://doi.org/10.3390/a15070229 - 30 Jun 2022
Viewed by 1851
Abstract
For the feature selection problem, we propose an efficient privacy-preserving algorithm. Let D, F, and C be data, feature, and class sets, respectively, where the feature value x(Fi) and the class label x(C) are [...] Read more.
For the feature selection problem, we propose an efficient privacy-preserving algorithm. Let D, F, and C be data, feature, and class sets, respectively, where the feature value x(Fi) and the class label x(C) are given for each xD and FiF. For a triple (D,F,C), the feature selection problem is to find a consistent and minimal subset FF, where ‘consistent’ means that, for any x,yD, x(C)=y(C) if x(Fi)=y(Fi) for FiF, and ‘minimal’ means that any proper subset of F is no longer consistent. On distributed datasets, we consider feature selection as a privacy-preserving problem: assume that semi-honest parties A and B have their own personal DA and DB. The goal is to solve the feature selection problem for DADB without sacrificing their privacy. In this paper, we propose a secure and efficient algorithm based on fully homomorphic encryption, and we implement our algorithm to show its effectiveness for various practical data. The proposed algorithm is the first one that can directly simulate the CWC (Combination of Weakest Components) algorithm on ciphertext, which is one of the best performers for the feature selection problem on the plaintext. Full article
(This article belongs to the Special Issue Privacy Preserving Machine Learning)
Show Figures

Figure 1

19 pages, 2296 KiB  
Article
On Edge Pruning of Communication Networks under an Age-of-Information Framework
by Abdalaziz Sawwan and Jie Wu
Algorithms 2022, 15(7), 228; https://doi.org/10.3390/a15070228 - 29 Jun 2022
Cited by 1 | Viewed by 1401
Abstract
Effective non-repetitive routing among nodes in a network is an essential function in communication networks. To achieve that, pruning the links of the network is helpful with the trade-off of making the network less robust in transmitting messages while reducing redundancy to increase [...] Read more.
Effective non-repetitive routing among nodes in a network is an essential function in communication networks. To achieve that, pruning the links of the network is helpful with the trade-off of making the network less robust in transmitting messages while reducing redundancy to increase flow with limited network bandwidth, so we enhance the quality of service (QoS). In our paper, we study the case that if a link removal has either no effect or an insignificant effect on the Age of Information (AoI) of the messages delivered in the communication network. The pruning of such links can be applied to the k least significant links in terms of their impact on the AoI of the messages transmitted in the system. The objective of our work is to study the effect of pruning a number of links in a network on the AoI, in order to reduce the redundancy of the messages that may be received at the destination many times while transmitted only once. In our context, the objective of the communication system would be to maintain the information from the source as fresh as possible when it arrives at the destination while reducing the redundancy of messages. In this work, we introduce an efficient reduction method designed for series-parallel networks with links of exponentially distributed wait times. In addition, we consider the deterministic case and present the pruning technique when link removal would not affect the AoI. Lastly, we present a comprehensive simulation to study the effect of pruning the links on the AoI of the network and the redundancy of messages received by the destination. Full article
(This article belongs to the Section Combinatorial Optimization, Graph, and Network Algorithms)
Show Figures

Figure 1

25 pages, 1580 KiB  
Article
Learning-Based Online QoE Optimization in Multi-Agent Video Streaming
by Yimeng Wang, Mridul Agarwal, Tian Lan and Vaneet Aggarwal
Algorithms 2022, 15(7), 227; https://doi.org/10.3390/a15070227 - 28 Jun 2022
Cited by 3 | Viewed by 1964
Abstract
Video streaming has become a major usage scenario for the Internet. The growing popularity of new applications, such as 4K and 360-degree videos, mandates that network resources must be carefully apportioned among different users in order to achieve the optimal Quality of Experience [...] Read more.
Video streaming has become a major usage scenario for the Internet. The growing popularity of new applications, such as 4K and 360-degree videos, mandates that network resources must be carefully apportioned among different users in order to achieve the optimal Quality of Experience (QoE) and fairness objectives. This results in a challenging online optimization problem, as networks grow increasingly complex and the relevant QoE objectives are often nonlinear functions. Recently, data-driven approaches, deep Reinforcement Learning (RL) in particular, have been successfully applied to network optimization problems by modeling them as Markov decision processes. However, existing RL algorithms involving multiple agents fail to address nonlinear objective functions on different agents’ rewards. To this end, we leverage MAPG-finite, a policy gradient algorithm designed for multi-agent learning problems with nonlinear objectives. It allows us to optimize bandwidth distributions among multiple agents and to maximize QoE and fairness objectives on video streaming rewards. Implementing the proposed algorithm, we compare the MAPG-finite strategy with a number of baselines, including static, adaptive, and single-agent learning policies. The numerical results show that MAPG-finite significantly outperforms the baseline strategies with respect to different objective functions and in various settings, including both constant and adaptive bitrate videos. Specifically, our MAPG-finite algorithm maximizes QoE by 15.27% and maximizes fairness by 22.47% compared to the standard SARSA algorithm for a 2000 KB/s link. Full article
(This article belongs to the Special Issue Deep Learning for Internet of Things)
Show Figures

Figure 1

15 pages, 346 KiB  
Article
A Mathematical Model and Two Fuzzy Approaches Based on Credibility and Expected Interval for Project Cost-Quality-Risk Trade-Off Problem in Time-Constrained Conditions
by Mohammad Hossein Haghighi and Seyed Meysam Mousavi
Algorithms 2022, 15(7), 226; https://doi.org/10.3390/a15070226 - 28 Jun 2022
Cited by 4 | Viewed by 1518
Abstract
To successfully finalize projects and attain their determined purposes, it is indispensable to control all success criteria of a project. The time–cost trade-off (TCT) is known as a prevalent and efficient approach applied when the planned finish date of a project is not [...] Read more.
To successfully finalize projects and attain their determined purposes, it is indispensable to control all success criteria of a project. The time–cost trade-off (TCT) is known as a prevalent and efficient approach applied when the planned finish date of a project is not admitted by stakeholders, and consequently, the project duration must be decreased. This paper proposes a new mathematical model under fuzzy uncertainty to deal with the project cost–risk–quality trade-off problem (CRQT) under time constraints. Because of the unique nature of projects and their uncertain circumstances, applying crisp values for some project parameters does not seem appropriate. Hence, this paper employs fuzzy sets to resolve these weaknesses. In this study, two approaches are presented to handle proposed fuzzy multi-objective mathematical model. First, fuzzy credibility theory and then goal attainment method are used. Secondly, the model is solved by a fuzzy method based on expected interval and value and augmented ɛ-constraint method. A project from the literature review is adopted and solved by the presented methodology. The results demonstrate the accuracy and efficiency of the two proposed approaches for the introduced practical problem. Full article
12 pages, 1455 KiB  
Article
Incremental Construction of Motorcycle Graphs
by Franz Aurenhammer, Christoph Ladurner and Michael Steinkogler
Algorithms 2022, 15(7), 225; https://doi.org/10.3390/a15070225 - 27 Jun 2022
Viewed by 1275
Abstract
We show that the so-called motorcycle graph of a planar polygon can be constructed by a randomized incremental algorithm that is simple and experimentally fast. Various test data are given, and a clustering method for speeding up the construction is proposed. Full article
Show Figures

Figure 1

Previous Issue
Next Issue
Back to TopTop