Next Article in Journal
Reconstruction of Velocity Curve in Long Stroke and High Dynamic Range Laser Interferometry
Previous Article in Journal
Effects of Touchscreen Media Use on Toddlers’ Sleep: Insights from Longtime ECG Monitoring
Previous Article in Special Issue
Cost-Effective Vibration Analysis through Data-Backed Pipeline Optimisation
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Review

Deep Learning for the Industrial Internet of Things (IIoT): A Comprehensive Survey of Techniques, Implementation Frameworks, Potential Applications, and Future Directions

1
School of Information Science and Engineering, Fudan University, Shanghai 200433, China
2
Security Engineering Lab, Prince Sultan University, Riyadh 12435, Saudi Arabia
3
RIADI Laboratory, National School of Computer Science, University of Manouba, Manouba 2010, Tunisia
4
Robotics and Internet of Things Lab, Prince Sultan University, Riyadh 12435, Saudi Arabia
5
Department of Electrical Engineering, Institute of Space Technology, Islamabad 44000, Pakistan
6
Department of Mathematics, College of Science, King Khalid University, Abha 61413, Saudi Arabia
7
School of Computing, Edinburgh Napier University, Edinburgh EH10 5DT, UK
*
Author to whom correspondence should be addressed.
Sensors 2021, 21(22), 7518; https://doi.org/10.3390/s21227518
Submission received: 11 October 2021 / Revised: 29 October 2021 / Accepted: 8 November 2021 / Published: 12 November 2021
(This article belongs to the Special Issue Smart Sensors in the Industrial Internet of Things (IIoT))

Abstract

:
The Industrial Internet of Things (IIoT) refers to the use of smart sensors, actuators, fast communication protocols, and efficient cybersecurity mechanisms to improve industrial processes and applications. In large industrial networks, smart devices generate large amounts of data, and thus IIoT frameworks require intelligent, robust techniques for big data analysis. Artificial intelligence (AI) and deep learning (DL) techniques produce promising results in IIoT networks due to their intelligent learning and processing capabilities. This survey article assesses the potential of DL in IIoT applications and presents a brief architecture of IIoT with key enabling technologies. Several well-known DL algorithms are then discussed along with their theoretical backgrounds and several software and hardware frameworks for DL implementations. Potential deployments of DL techniques in IIoT applications are briefly discussed. Finally, this survey highlights significant challenges and future directions for future research endeavors.

1. Introduction

The Internet of Things (IoT) describes a ubiquitous connection between common objects and the Internet. IoT functions by deploying thousands of smart devices in living or industrial environments. These devices collect the information from surrounding environments, perform desired processing activities on the data acquired and transmit that processed data through secure, reliable communication channels. Recent advancements in software, hardware, and communication systems have significantly improved human lifestyles in terms of time, energy, and cost savings [1,2]. The term Industrial Internet of Things (IIoT) refers to the use of conventional IoT concepts in industrial applications. The IIoT improves manufacturing procedures by enabling sustainable and efficient technologies in an industrial environment [3,4]. The IIoT market is currently undergoing rapid growth as well as increasing adaptations in the digital transformations of many industries. Strong alliances and concurrences of interests between IIoT stakeholders and emerging applications has attracted major companies worldwide to invest in this emerging market. According to the latest Statistica report, the size of the IIoT market is expected to increase up to $110.6B USD by 2025 [5]. A comparative analysis of the IIoT market and its increase in size from 2017 to projections for 2025 is presented in Figure 1.
In the fourth industrial revolution, a rapid increase in the number of interconnected devices in the IIoT systems has been recorded. In the IIoT, the performance of smart applications is heavily associated with the intelligent processing of big data. Therefore, IIoT networks require intelligent information processing frameworks for big data analysis. In this context, artificial intelligence (AI), and particularly DL, techniques can be very useful as a means of producing effective, usable results from big data in an IIoT system [4,6,7]. For instance, DL techniques enable the system’s capacities to learn from experience. The efficiency and effectiveness of a DL solution depend upon the characteristics and nature of the data in question as well as the performance of the learning algorithm [8]. The selection of a suitable DL algorithm for a specific application can be a challenging task. Therefore, it is essential to understand the workflow of various DL algorithms and their deployment in real-world applications such as smart homes, smart cities, cybersecurity services, smart healthcare, smart transportation, sustainable agriculture, business enterprises, and many others [9]. The significance of DL in IoT/IIoT applications can be determined by analyzing the emerging and cutting-edge research in this area. The bar chart in Figure 2 presents a comparative analysis of relevant publications records by the top publishers from 2016 to 2020.

1.1. Existing Surveys on DL for IoT and IIoT Applications

In recent literature, multiple DL-based solutions for IoT and IIoT applications have been proposed by both academia and industry. In the following section, some of the latest survey articles are briefly described to demonstrate the depth and scope of existing literature.
Mohammadi et al. [10] have presented a detailed overview of DL techniques and their importance to big data analysis for IoT applications. They also discuss some of the most-used DL algorithms as well as their backgrounds and architectures. They also summarized the deployment of DL techniques in fog and cloud-based IoT environments. They conclude this work by presenting some potential challenges and directions for future research. Meanwhile Ma et al. [11] presented a comprehensive examination of how leveraging DL techniques in IoT applications could be possible. The main goal of this survey is to present the applicability of DL techniques to empower multiple applications in fields including smart healthcare, robotics, home monitoring, traffic control, autonomous driving, and manufacturing. In another survey, Sengupta et al. [12] explored the potential of DL approaches for multiple applications including perspective and predictive analysis, financial time series forecasting, power systems, medical image processing, etc.
Ambika et al. [13] presented a comprehensive survey about ML and DL techniques along with their architectures and impact on IIoT. This survey also discussed some use cases of ML and DL approaches in the IoT and IIoT domains. Saleem et al. [14] then presented a concise discussion on the potential benefits of DL algorithms for IoT applications. This survey formulated DL-based human activity recognition systems and presented a brief comparison of different DL models before discussing advanced research options in the realm of DL-driven IoT applications. As these authors pointed out, with the development of the IoT, several real-time apps collect data about people and their surroundings using IoT sensors and input it into DL models to improve an application’s intelligence and capabilities, allowing it to provide better suggestions and services. In this regard, Deepan and Sudha [15] presented a self-contained review of several DL models for real-time IoT applications. Additionally, the authors provided a better understanding of DL models and their efficiency by adding different DL tools and frameworks. In one of the most recent and best survey articles, Khalil et al. [6] discussed the key potential of DL techniques in IIoT applications. Here the authors reviewed different DL algorithms and their use in different industries. They outlined numerous use cases of DL for IIoT, including smart agriculture, smart manufacturing, smart metering, etc. Moreover, several key challenges and future research directions were discussed.

1.2. Limitations of Existing Surveys

A detailed comparison of the aforementioned surveys is presented in Table 1, demonstrating how researchers have done well in providing a perspective on DL-based IoT and IIoT applications. However, the existing literature has a few notable limitations. First, most of these researchers discussed only the theoretical aspects of DL for IIoT, but not real-time deployment. Second, there is a need for reference architecture of IIoT with key enabling technologies. Third, most of the surveys discussed the mathematical and theoretical background of DL algorithms, but the implementation frameworks and hardware platforms for real-time deployments are not considered or discussed adequately. Finally, these studies used only a limited description of the use cases of DL algorithms in the IIoT sector. Therefore, to overcome these various limitations, we present a comprehensive study on DL technologies for IIoT applications.

1.3. Major Contributions

This study presents a comprehensive deliberation on recent advancements in the integration of DL technologies with IIoT. The major contributions of this survey are as follows.
  • Discussing the great potential of DL schemes for IIoT.
  • Providing a detailed reference architecture of the IIoT with key enabling technologies.
  • Presenting a comprehensive survey on the working principle of well-known DL algorithms, their implementation frameworks, and the relevant hardware platforms.
  • Covering the real-world applications of DL techniques in IIoT systems.
  • Suggesting some potential future research directions.
The remainder of this article is organized as follows. Section 2 describes the detailed reference architecture of the IIoT system. Section 3 briefly discusses the potential of several DL schemes for IIoT and elaborates the working principle of each algorithm. Some well-known software frameworks are discussed in relation to the implementation of DL algorithms in Section 4. Section 5 presents some suitable hardware platforms for the real-time implementation of DL schemes for the IIoT. Section 6 comprises some real-world cases of DL application for IIoT and Section 7 summarizes the challenges faced by DL-based IIoT and presents some potential future research directions. Finally, a brief conclusion is presented in Section 8.

2. Reference Architecture of the Industrial Internet of Things (IIoT)

Over the past decade, IIoT has brought about a great revolution in modern technologies by providing seamless connectivity among devices. In this Internet-enabled era, our mobile applications, devices (smartwatches, tablets/iPads, laptops, etc.), and many more are sending valuable information continuously to IoT networks. These IoT/IIoT systems are built upon specific architectures that include different layers and elements with particular functionalities [16,17]. For instance, a detailed 7-layer architecture for the IIoT is presented in Figure 3.

2.1. Perception Layer

This layer, which is also known as a “physical” layer, collects environmental data and performs the necessary preprocessing. This layer transforms the acquired analog data into digital to make it best compatible with other layers of the system [18]. The main components of this layer are sensors and actuators.
  • Sensors: These are small devices that can detect changes in the surrounding environment and extract useful information from the data acquired. Usually, sensors are considered to be resource-constrained devices that have little processing and computational power combined with limited memory resources. However, modern sensors have great capacity to gather environmental signals with a higher level of accuracy. The most commonly used sensors across multiple industries are used to measure temperature, humidity, air pressure, weight, acceleration, position, and many others.
  • Actuators: These are usually electromechanical devices that convert electrical signals into physical actions. In an industrial environment, linear and rotary are the two general types of actuators most often used. Linear actuators transform electrical signals into linear motions, which are useful in position adjustment applications. Meanwhile, rotary actuators transform electrical energy into rotational energy. These are usually used for the position control of devices such as conveyor belts.

2.2. Connectivity Layer

This layer connects the perception layer and edge layer using advanced communication technologies [19]. Two methods may be adopted for communication in this layer. In the first method, direct communication is performed using a TCP or UDP/IP stack. In another method, smart gateways provide a communication link between the local area network (LAN) and the wide-area network (WAN). Several advanced communication technologies and protocols are used in this layer, as described in the following.
  • WiFi: This is the most versatile and commonly used scheme across communication technologies. WiFi modems are very suitable for both personal and official uses, delivering smooth communications among LAN and WAN.
  • Ethernet: This is an older technology used for connecting devices in a LAN or WAN, which enables them to communicate with each other via a specific communication protocol. Ethernet also enables network communication between different network cables, such as copper to optical fiber and vice versa.
  • Bluetooth: This wireless protocol is widely used for information exchange over short distances by creating personal area networks.
  • NFC: Near Field Communications (NFC) is a wireless technology that enables secure communication between smart devices at short distances. The communication range of NFC is usually considered to be about 10 cm.
  • LPWAN: Low-Power Wide-Area Network (LPWAN) describes a class of radio technologies used for long-distance communication. The top LPWAN technologies are LoRa, Sigfox, and Nwave. As compared to other wireless technologies such as Bluetooth and Wi-Fi, LPWANs are usually used to send smaller amounts of information over longer distances.
  • ZigBee: This is a product from the Zigbee alliance that is designed especially for sensor networks on the IEEE 802.15.4 standard. The mostly used data communication protocols for this communication standard are ISA-100.11.a and WirelessHART. These protocols define Media Access Control (MAC) and physical layers to handle several devices at low-data rates.
  • LTE-M: Long Term Evolution for Machine is a leading LPWA network technology for IoT applications. It is used for interconnecting objects such as IoT sensors and actuators, or other industrial devices via radio modules.
  • NB-IoT: This is a standards-based low-power wide-area (LPWA) technology that enables a wide variety of smart devices and services. NB-IoT improves the power consumption, spectrum efficiency, and system capacity of smart devices.

2.3. Edge Layer

In the early stages, latency often becomes a major challenge to the growth of an IoT network. Edge computing is considered to be a potential solution to this issue, as it accelerates the growth of IoT networks. This layer facilitates the system’s ability to process and analyze data closer to the source. Edge computing technology has now become a new standard for advanced 5G mobile networks that enable broader system connectivity at lower latency. The performance of IoT networks is significantly improved by performing all processes at the edge [20].

2.4. Processing Layer

This layer collects, stores, and analyses information from the edge layer [21]. These operations are all performed by IoT systems and include two main stages.
  • Data Accumulation: Real-time information is acquired through an API and further stored to meet the demands of non-real-time applications. This stage serves as a transient link between query-based data consumption and event-based data generation. This stage also determines the relevance of the data acquired to the stated business requirements.
  • Data Abstraction: Once data accumulation and preparation have been completed, consumer applications may use it to produce insights. Several phases are involved in the end-to-end process, including integrating data from multiple sources, reconciliation of formats, and data aggregation in a single place.
To enhance the compatibility of smart devices, these techniques are usually deployed together. The commonly used protocols for the processing layers are described in the following.
  • Transmission Control Protocol (TCP): It offers host-to-host communication, breaking large sets of data into individual packets and resending and reassembling packets as needed.
  • User Datagram Protocol (UDP): Process-to-process communication is enabled using this protocol, which operates on top of IP. Over TCP, UDP offers faster data transfer speeds, making it the protocol of choice for mission-critical applications.
  • Internet Protocol (IP): Many IoT protocols use IPv4, while more recent executions use IPv6. This recent update to IP routes traffic across the Internet and identifies and locates devices on the network.

2.5. Application Layer

In this layer, software analyses the data to provide promising solutions to critical business problems. Thousands of IIoT applications vary in design complexity and functions, as demonstrated by various versions of this layer. Each one uses various technologies and operating systems (OS). Some prominent applications include device monitoring through software control, business intelligence (BI) services, analytic solutions through AI techniques, and mobile applications for simple interactions. In recent trends, the application layer can be built at the top of IoT/IIoT frameworks that provide software-based architectures with ready-made instruments for data visualization, data mining, and data analytics [22]. Some widely used protocols for the applications layer are discussed in the following.
  • Advanced Message Queuing Protocol (AMQP): It allows messaging middleware to communicate with one another. It enables a variety of systems and applications to communicate with one another, resulting in standardized communications on a large scale.
  • Constrained Application Protocol (CoAP): A constrained-bandwidth and constrained-network protocol designed for machine-to-machine communication between devices with limited capacity. CoAP is a document-transfer protocol that operates on the User Datagram Protocol (UDP).
  • Data Distribution Service (DDS): A flexible peer-to-peer communication protocol capable of running small devices as well as linking high-performance networks. DDS simplifies deployment, boosts dependability, and minimizes complexity.
  • Message Queue Telemetry Transport (MQTT): A messaging protocol developed for low-bandwidth communications to faraway places and mostly used for lightweight machine-to-machine communication. MQTT employs a publisher-subscriber pattern and is suited for tiny devices with limited bandwidth and battery life.

2.6. Business Layer

The IoT-acquired data are considered useful if they are applicable for business planning and strategy. Every business has specific goal-oriented tasks that must be accomplished by extracting useful information from the data gathered. Business enterprises also use past and present data to plan future goals. Modern firms frequently adopt intelligent data analysis techniques to enhance their decision-making capabilities [23]. Today, software and business intelligence techniques have gained great attention from industries as a means of improving their performance and profitability.

2.7. Security Layer

Security has become one of the essential requirements of an IIoT architecture, due to the rapid increase of modern challenges. Hacking, denial of service, malicious software, and data breaches are the main challenges of the current IIoT infrastructures [24]. This layer performs three main functions as, which can be described as follows.
  • Device Security: This is the beginning point of security in the IIoT framework. Many manufactures and companies integrate both software and hardware-based security schemes in IoT devices.
  • Cloud Security: Cloud storage is replacing the traditional data storage servers in modern IoT infrastructures, so in turn new security mechanisms are also adopted to secure that cloud. Cloud security includes encryption schemes and intrusion detection systems, etc., as means of preventing cyberattacks and other malicious activities.
  • Connection Security: In an IIoT network, the data must be encrypted before transmission via any communication channel. In this context, different messaging protocols, such as MQTT, DDS, and AMQP, may be implemented to secure valuable information. In modern trends, the use of TSL cryptographic protocol is recommended for communication in industrial applications.

3. Deep Learning for the IIoT

The use of smart manufacturing in the IIoT has several advantages. For one thing, it can be very helpful to make manufacturing and production processes intelligent [25]. New industrialists are adopting IIoT solutions to enhance the productivity and profitability of their industries. In IoT-enabled industries, the data collected from sensors and smart devices plays an important role to make the production process smarter [25]. Therefore, the deployment of intelligent data analysis techniques has now become an essential requirement of modern industries.
DL is considered to be one of the most powerful techniques in the domain of artificial intelligence (AI). The integration of DL methods in smart industries can upgrade the smart manufacturing process into a highly optimized environment by information processing through its multi-layer architecture. DL approaches are very helpful due to their inherited learning capabilities, underlying patterns identification, and smart decision-making. The biggest advantage of DL over conventional ML techniques is automatic feature learning. With this option, there is no need to implement a separate algorithm for feature learning [25].
The deployment of DL techniques can be very effective to perform the types of aforementioned analysis in smart industries. Some well-known DL-based techniques for the IIoT are discussed in the following section.

3.1. Deep Feedforward Neural Networks

This is the most fundamental type of deep neural network (DNN), in which the connections between nodes move forward. The general architecture of the deep forward neural network (DFNN) is presented in Figure 4. As compared to shallow networks, the multiple hidden layers in DNN can be very helpful to model complex nonlinear relations. This architecture is very popular in all fields of engineering because of its simplicity and robust training process [26].
The most commonly used gradient descent algorithm is preferred for the training of DFNNs. In this algorithm, the weights are first initialized randomly and then the error is minimized using gradient descent. The complete learning process involves multiple executions of forward and backward propagation consecutively [27]. In forward propagation, the input is processed towards the output layer through multiple hidden layers and the calculated output is compared with the desired output of the corresponding input. In the backward process, the rate of change of error is calculated concerning network parameters to adjust the weights. This process will continue until the desired output is achieved by the neural network.
Let x i as the input of neural network and f i as the activation function of layer i. The output of the layer i can be computed as
x i + 1 = f i w i x i + b i
Here, x i + 1 becomes the input for the next layer, w i and b i are the essential parameters that connect the layer i with the previous layer. These parameters are updated during the backward process as shown in the following.
w n e w = w η E W
b n e w = b η E b
Here, w n e w and b n e w are the updated parameters for w and b. E represents the cost function and η is the learning rate. The cost function of the DL model is decided according to the desired task such as classification, regression etc.

3.2. Restricted Boltzmann Machines (RBM)

RBM can also be described as stochastic neural networks. This is one of the well-known and widely used DL approaches because of its capability to learn the input probability distribution in both supervised and unsupervised manners. Paul Smolensky introduced this technique in 1986 with the name Harmonium, and Hinton popularized it in 2002 with the development of new training algorithms [28]. Since then, RBM has been applied widely in a variety of tasks such as dimensionality reduction, representation learning, and prediction problems. Deep belief network training using RMB as a building block was a very famous application. Presently, RBMs are frequently used for collaborative filtering, due to the excellent performance of the Netflix dataset [29]. RBM is an extension of the Boltzmann Machine with a restriction in the intra-layer connection between units. It is an undirected graphical model that contains two layers, visible and hidden, that form a bipartite graph. In recent studies, several variations of RBMs have been introduced in terms of improvised learning algorithms [30]. These variants include conditional RBMs, temporal RBMs, convolutional RBMs, factored conditional RBMs, and Recurrent RBMs. To deal with the data characteristics, different types of nodes may be introduced, such as Gaussian, Bernoulli, etc. In RBM, each node is considered a computational unit that processes the information needed to make stochastic decisions. The general architecture of RBM is presented in Figure 5.
The joint probability distribution of a standard RMB can be described with the Gibbs distribution p ( v , h ) = 1 z e E ( v , h )
Here energy function E ( v , h ) can be described as
E ( v , h ) = i = 1 n j = 1 m w i j h j v i j = 1 m b j v i i = 1 n c i h i
Here m and n represent the number of visible and hidden units. h j and v j are the states of the hidden unit i and visible unit j respectively. b j and c j describes real-valued biases corresponding to the j th and i th units, respectively. w i j are the weights that connect visible units with hidden units. Z indicates the normalization constant that ensures the sub-probability distributions equal to 1. Hinton proposed the Contrastive Divergence algorithm for the training of RBMs. This algorithm trains the RMB to maximize the training sample’s probability. Training stabilizes the model through minimization of its energy through updating the model’s parameters, which can be done using the following Equations (5)–(7).
Δ w i j = ϵ v i h j data v i h j model
Δ b i = ϵ v i data v i model
Δ C i = ϵ h j data h j model
Here, ϵ represents the learning rate, . data , and . model represent the expected value of data and model, respectively.

3.3. Deep Belief Networks (DBN)

DBNs are made up of several layers of latent variables. These variables are typically binary and indicate the hidden characteristics of the input observations. DBN with one layer is the same as an RBM because the connection among the two top layers is undirected [31]. The remaining connections in DBN have directed graphs towards the input layer. A DBN model follows a top-down approach to generate a sample [32]. First, the samples are drawn from the RMB on the top layer using Gibbs sampling. After that, simple execution of ancestral sampling is performed by visible units in a top-down approach. A generic DBN model is presented in Figure 6.
The inference in DBN is an uncontrolled issue due to explaining away effect in the latent variable model. Hin presented an efficient and fast method to train DBN by RBM stacking: at the initial stage, the lowest level of RBM learns the distribution of input data during the training process [33]. At the next stage, RBM computes the higher-order correlation among the hidden units of the previous layer. The entire process is repeated for each hidden layer. At the next stage, RBM computes the higher-order correlation among the hidden units of the previous layer. The entire process is repeated for each hidden layer. A DBN with Z number of hidden layers models the joint distribution among the visible layer v and hidden layer h z , here z = 1 , 2 , 3 , , Z as described in the following.
p v , h 1 , , h z = p v h 1 z = 1 z 2 p h z h z + 1 p h z 1 , h z
Hinton’s training technique led directly to the modern era of DL, as it was the first deep architecture that had been trained efficiently. By adding layers to the NN, the logarithmic probability of the training data can be increased greatly. The DBN has been used as a classifier in several applications, such as computer recognition, phone recognition, etc. In speech recognition, DBN is used for pre-training of the DNN, deep convolutional neural network, and many others.

3.4. Autoencoders (AE)

An autoencoder is an unsupervised learning technique for neural networks that trains the network to ignore noise signals and thus develop more efficient data representations. AE is a 3-layer neural network, as presented in Figure 7. Here the input and output layers have the same number of units, but the hidden layer usually contains a lower number of neurons. The hidden layer encodes the input data into a compact form. AE generally uses deterministic distributions instead of particular distributions, as in the case of RBM [34]. A backpropagation algorithm is usually used to train an AE. This training process contains two stages: encoding and decoding. In the first stage, the model encodes the input into hidden representations through the weight metrics. In the decoding phase, the model reconstructs the same input from an encoded representation using the weight metrics. The encoding and decoding process can be further elaborated mathematically, as presented in Equations (9) and (10).
In the encoding stage.
y = f ( w X + b )
Here, X represents an input vector, f is an activation function, w and b are the parameters that are required to be tuned and y is the hidden representation.
In decoding stage
X = f w y + c
Here X represents the reconstructed input at the output layer, w is the transpose of w and c represent bias value to the output layer.
w n e w = w η E W
b n e w = b η E b
Here, w n e w and b n e w are the updated parameters after the completion of the current iteration. E represents the reconstruction error at the output layer.
A deep autoencoder (DAE) is composed of multiple hidden layers of AE. Because of multiple layers, the training of AE can be a difficult task [35]. This difficulty can be overcome by training each layer of a DAE as simple AE. The DAE has been successfully applied in several applications, such as document encoding for faster subsequent retrieval, speech recognition, and image retrieval, etc. [36]. AEs gained great attention from different researchers because of their non-generative and non-probabilistic nature.

3.5. Convolutional Neural Network (CNN)

CNNs are a type of neural network inspired by the human visual system. CNNs were initially proposed by LeCun in 1998 [37] and gained popularity in the DL frameworks when Krizhevsky et al. [38]. won the ILSVRC-2012 competition with his proposed architecture AlexNet. This remarkable achievement started a new trend in AI, as data scientists observed the great classification capabilities of CNN and its variants. In many applications, CNN algorithms proved their superior performance in human recognition systems.
The fundamental architecture of CNN is presented in Figure 8. It contains multiple convolutional and pooling layers and a fully connected layer at the end. The convolutional layer’s primary role is to extract essential features from the provided input image based on the spatial relationship between pixels. Pooling layers perform dimensionality reduction of feature maps while maintaining feature information. Finally, a fully connected layer connects the neural network with the output layer to obtain the desired output. CNNs are very useful for image descriptor extractions using latent spatial information. A common image contains several characteristics such as color, contours, edge, textures, strokes, gradient, and orientation. A CNN splits an image according to the aforementioned properties and learns them as representation in various layers. CNNs gained great popularity in computer vision applications such as image detection, image classification, image segmentation, and image super-resolution reconstruction. Multiple CNN frameworks have been proposed by considering the real-world application requirements and maintaining the high accuracy threshold. Region-based CNN (R-CNN) and You Only Look Once (YOLO) are popular examples of such modern architectures. The common naive-based CNNs are computationally very expensive because they consider a huge number region proposal to find an object within an image. However, R-CNN-based addresses this issue by the selection of a region of interest (ROI) with a selective search. Redmon et al. [39] initially proposed a YOLO technique in 2016. It has very high speed as compared to R-CNN with little compromise in performance. It understands the generalized representation of an image by looking at the object once with a CNN. However, this technique faces spatial constraints in the detection of smaller objects. Apart from these frameworks, there are several other variants of CNN have been proposed, such as AlexNet, LeNet, VGGNet, ResNet, GoogleNet, ZFNet, and many others [12]. These CNN frameworks have had a great impact on AI-enabled vision research for future applications.

3.6. Recurrent Neural Network (RNN)

An RNN is a class of ANN in which connections between nodes make a directed graph along a temporal sequence [40]. This property allows it to exhibit temporal dynamic behavior. In conventional neural networks (CNNs), all inputs and outputs are independent of one another. In some cases, when it is required to predict the next word or statement, then previous data are required. Therefore, there is a need to remember the previous data. RNN addresses this issue with the help of a hidden layer. The most significant feature of an RNN is the hidden state that remembers the sequence information [40]. RNN has a memory that recollects all the information about what has been calculated during the process. It uses the same parameters for all inputs and the information about what has been calculated during the process. It uses the same parameters for all inputs and performs the same tasks on all the input or hidden layers to generate a suitable output. As compared to other neural networks, this feature reduces the complexity of parameters.
An RNN is a class of ANN in which connections between nodes make a directed graph along a temporal sequence [41]. This property allows it to exhibit temporal dynamic behavior. In conventional neural networks, all the input and outputs are independent of each other. In some cases, when it is required to predict the next word or statement then previous data are required. Therefore, there is a need to remember the previous data. RNN addresses this issue with the help of a hidden layer. The most significant feature of RNN is the hidden state that remembers the sequence information [42]. RNN has a memory that remembers all the information about what has been calculated during the process. It uses the same parameters for all inputs and performs the same tasks on all the input or hidden layers to generate a suitable output. As compared to other neural networks, this feature reduces the complexity of parameters.
The general architecture of RNN is shown in Figure 9. To understand the working mechanism of RNN, we describe a brief example here. Suppose there is a DNN that contains 1 input layer, 3 hidden layers, and 1 output layer. Each layer will have its own set of weights and biases. Let us assume for hidden layers the weights and biases for each layer is described as w 1 , w 2 , w 3 and b 1 , b 2 , b 3 . Each of these layers is independent of the others as they do not memorize the previous output. RNN converts the independent activations into dependent activations by providing the same biases and weights to each layer. It memorizes each of the previous outputs as input to the next hidden layer, which results in complexity reduction. Here, these 3 layers can be converted into a single recurrent layer in such a way that the weights and biases are identical for all the hidden layers. The existing state can then be calculated using the following expression.
h t = f h t 1 , x t
Here h t represents the current state, h t 1 is the previous state and x t is the input state of the neural network.
Now a hyperbolic tangent (tanh) activation function can be applied through the following expression.
h t = tanh w h h h t 1 + w x h x t
Here w h h is weight at current neuron and w x h is weight at input neuron.
Now the output can be calculated using the following expression.
y t = w h y h t
Here y t represents the output and w h y is the weight at the output layer.

3.7. Generative Adversarial Networks (GAN)

GANs are algorithmic frameworks that employ two neural networks to generate new and synthetic data instances that can pass for real data. GANs are used widely in voice, image, and video generation applications [43]. Ian Goodfellow initially introduced GAN in 2014 at the University of Montreal, Canada. Yann LeCun, the AI research director of Facebook, has called adversarial training the most emerging idea in the last 10 years of ML. GANs have great potential in multiple applications because of their ability to mimic any kind of data distribution [44]. These networks can also be trained to create worlds similar to our own choice in several domains, such as speech, music, image, video, etc. In a sense, these networks are considered robot artists that can produce impressive output. GANs can also be used to generate fake media content, a technology that has come to be known as Deepfakes.
GANs are made up of two models. The first model is referred to as a generator, and its primary function is to produce new data that is comparable to the desired data. The generator might be compared to a human art forger, who makes counterfeit pieces of art. The second model is called a discriminator. This model validates the input data, deciding whether it is from the original dataset or was created by a forger. In this context, a discriminator is comparable to an art expert who attempts to determine if artworks are genuine or fraudulent. The working process of GANs are presented in Figure 10. The GAN can be easily understand using universal approximators such as ANN. A generator can be model using a neural network G n , θ 1 . Its main function is to map input noise variables n to the desired data space x. Conversely, the discriminator can be model using a second neural network D x , θ 2 . It generates the output probability of data authenticity in the range of ( 0 , 1 ) . In both cases, θ i describes the weights of each neural network.
Consequently, the discriminator trains to categorize the input data accurately, which means that the weights of neural networks are updated to optimize the probability that any real data input x is categorized as belongs to a real dataset. The generator is trained to generate the data as realistically as possible. This also means that the weights of the generator are tuned to increase the probability that any fake image is classified as part of the original dataset. After multiple executions of the training process, the generator and discriminator will reach the point where further improvement is not possible. At this point, the generator produces realistic data synthetic data, and the discriminator is unable to differentiate between the two types of input. During training, both the generator and discriminator will try to optimize the opposite loss function [45]. The generator tries to maximize its probability of having its output recognized as real, while the discriminator tries to minimize this value.

4. Deep Learning Frameworks

Several ML and DL frameworks are already facilitating the users to use the graphical processing unit (GPU) accelerators and thus expedite the learning process with interactive interfaces. Some of these frameworks allow for the use of optimized libraries, such as OpenCL and CUDA, to further enhance the performance. The most significant feature of multicore accelerators is their significant ability to expedite the computations of matrix-based operations [46]. The software development in ML/DL is highly dynamic and has multiple layers of abstraction. The most commonly known ML/DL frameworks are presented in Figure 11. A short description of each framework, along with its advantages and disadvantages, is described in the following.

4.1. TensorFlow

This is an open-source library for numerical calculations using data flow graphs [47]. This framework was created and maintained by the Google Brain team at Google’s Machine Intelligence research organization for ML and DL, and it is designed especially for large-scale distributed training and inference. The nodes in the network represent mathematical processes, while the graph edges define the multidimensional data arrays transferred between nodes. TensorFlow’s framework is made up of distributed master and slave services with kernel implementations. This framework contains around 200 operations, such as arithmetic, array manipulation, control flow, and state management. TensorFlow was designed for both research and development applications. It can easily execute on CPUs, GPUs, mobile devices, and large-scale systems with hundreds of nodes [48]. Additionally, TensorFlow Lite is a lightweight framework for embedded and mobile devices [49]. It facilitates the on-device ML/DL inference with low latency and small binary size. It supports multiple programming interfaces that include API for C++, Python, GO, Java, R, Haskell, etc. The TensorFlow framework is also supported by Amazon and Google cloud environments.

4.2. Microsoft CNTK

Microsoft Cognitive Toolkit is a commercial-grade distributed DL framework that includes large-scale datasets from Microsoft Research [50]. It facilitates the implementation of DNN for image, speech, text, and handwritten data. Its network is defined as a symbolic graph of vector operations using building blocks. These operations include matrix addition/multiplication or convolution. CNTK supports RNN, CNN, and FFNN models and enables the implementation of stochastic gradient (SGD) algorithms with automatic parallelization and differentiation across multiple servers [51]. CNTK is supported by both Windows and Linux operating systems using C++, C#, Python, and BrainScript API.

4.3. Keras

This is a Python library that enables bindings between several DL frameworks such as Theano, CNTK, TensorFlow, MXNet, and Deeplearning4j [52]. It was released under the MIT license and was developed especially for fast experimentation. Keras runs on Python 2.7 to 3.9 and can execute smoothly on both CPUs and GPUs. Francois Chollet developed and maintained this framework under four guiding principles [53], namely:
  • It follows the best practices by offering simple, consistent APIs to reduce cognitive load.
  • Neural layers, optimizers, cost functions, activation functions, initialization schemes, and regularization methods are all separate modules that can be combined to construct new models.
  • The addition of new modules is easy and existing modules provide sufficient examples that allow for the reduction of expressiveness.
  • It works with Python models that are easy to debug as well as compact and extensible.

4.4. Caffe

This framework was developed by Yangqing Jia and community contributors at the Berkeley Artificial Intelligence Research (BAIR) lab [54]. It provides speed, expression, and modularity. In Caffe, DNN are defined layer by layer. Here, a layer is the fundamental unit of computation essence of a complete model. Accepted data sources include Hierarchical Data Format (HDF5), efficient databases (LMDB or LevelDB), or popular image formats (e.g., JPEG, PNG, TIFF, GIF, PDF). In this framework, common layers provide normalization and vector processing operations. It provides Python support for custom layers but is not very efficient in doing so. Therefore, the new layers must be written in C++ CUDA [55].

4.5. Caffe2

At Facebook Yangqing Jia developed Caffe2, which is a modular, lightweight, and highly scalable DL framework [56]. Its major goal is to provide a fast and simple approach that promotes DL experimentation and exploits research contributions of new models and techniques. Its development is done in PyTorch, and it is used at the production level at Facebook. It has several improvements over Caffe, particularly in terms of new hardware support and mobile deployment. The operator is the fundamental unit of computation in this framework. Currently, Caffe2 has about 400 operators and several other operators are expected to be implemented by the research community. This framework provides Python scripts that facilitate the translation of existing Caffe models into Caffe2. However, this conversion procedure requires a manual verification process of accuracy and loss scores. Caffe also facilitates the conversion of the Torch model to Caffe2 [57].

4.6. MXNet

This DL framework is designed for both flexibility and efficiency. Pedro Domingos developed this framework with colleagues at the University of Washington, USA. This framework is licensed under an Apache 2.0 license. It has wide support for Python, R, Julia, and many other programming languages. In addition, several public cloud providers support this DL framework [58]. MXNet supports imperative programming, which improves productivity and efficiency. At its core, this DL framework contains a dynamic dependency scheduler that parallelized both imperative and symbolic operations automatically. A graph optimization layer expedites the symbolic execution make it memory efficient. It is a lightweight, portable framework that is highly compatible with multiple CPUs, GPUs, and other machines. It also facilitates the efficient deployment of trained DL models on resource-constrained mobile and IoT devices [59].

4.7. Torch

This is a scientific computing platform that supports a wide range of ML/DL algorithms based on the high-level programming language Lua. Many organizations such as Google, Facebook, Twitter, and DeepMind support this framework [60]. This framework is freely available under a BSD license. It uses object-oriented programming and C++ for implementation. Recently, its application interfaces have also been written in Lua, thus enabling optimized C/C++ and CUDA programming. Torch is built around the Tensor library, which is accessible with both CPU and GPU backends. The Tensor library is effectively implemented in C, which enables it to offer a wide range of classical operations. This DL framework supports parallelism on multicore CPUs and GPUs via OpenMP and CUDA, respectively. This framework is mostly used for large-scale learning such as image, voice, and video applications, optimization, supervised and unsupervised learning, image processing, graphical models, and reinforcement learning [61].

4.8. PyTorch

This is a Python library for GPU-accelerated deep learning frameworks. Facebook’s AI research department introduced this framework, which is written in Python, C, and CUDA, in 2016. It incorporates several acceleration libraries, including NVIDIA and Intel MKL. At the core level, it uses CPU and GPU Tensors and NN backend written as independent libraries [62]. Tensor computing is supported in PyTorch, with substantial GPU acceleration. It gained great popularity because of the ease with which it enables users to implement complex architectures. It also employs a reverse mode auto differentiation approach to modify network behavior with minimal effort. Industrial and scientific organizations widely use this framework. Uber created Pyro, a universal probabilistic programming language that leverages PyTorch as a backend, in a popular application. This library is supported by Twitter, Facebook, and NVIDIA and is freely available under the BSD license [63].

4.9. Theano

This is a pioneering DL tool that supports GPU computations and is actively maintained by the Montreal Institute for Learning Algorithms at the University of Montreal, Canada. It is also an open-source framework available under the BSD license [64]. Theano uses the NumPy and BLAS libraries to compile mathematical expressions in Python and thus transform complicated structures into simple and efficient code. These libraries enable speedy executions on CPUs and GPUs platforms. Theano has a distributed framework for training models and support extensions for multi-GPU data parallelism [65].

4.10. Chainer

This is an open-source, Python-based framework for DL models developed by researchers at the University of Tokyo, Japan. It offers a diverse set of DL models, including RNN, CNN, variational encoders, and reinforcement learning [66]. The main intent of Chainer is going beyond invariance. This framework enables automatic differentiation application interfaces based on the Define-by-Run technique. It creates neural networks dynamically, as compared to other frameworks. For high-performance training and inferences, this framework supports CUDA/cuDNN through CuPy. It also uses the Intel MKL library for DNN that accelerates DL schemes for Intel-based architectures. It has several libraries for industrial applications, such as ChainerRL, ChainerCV, and ChainerMN. In a study performed by Akiba in 2017, ChinerMN demonstrated its superior performance in a multi-node setting as compared to the results achieved by CNTK, MXNet, and TensorFlow [67].

5. Hardware Platforms for the Implementation of Deep Learning Algorithms

In the early days of this field, special labs were established for the implementation of AI and DL techniques. These labs were equipped with high-performance computing machines that were not affordable for an individual researcher. However, with ongoing advancements in embedded electronics, single-board computers (SBC) have gained great popularity in both academia and industry for the deployment of real-time DL algorithms in industrial applications. Currently, many development boards with the capabilities of AI use are available on the market. Some well-known hardware development platforms for DL deployments are presented in Figure 12. In the following, we discuss some well-known and powerful SBCs that can or have been used for DL implementations.

5.1. Raspberry Pi 4

This is the latest development board of the popular Raspberry Pi family of SBC. It offers high processing speeds, large memory, efficient multimedia performance, and vast connectivity, particularly as compared to the previous generations of Raspberry Pi modules. This development board is released with a well-documented user manual that can be very helpful for beginners to expedite their learning process. The hardware specifications of Raspberry Pi 4 present the option of 2 GB, 4 GB, and 8 GB DDR4 RAMs that make this SBC an ideal choice for AI and DL-based projects [68]. Due to these characteristics, this SBC has great support from the scientific community. Raspberry Pi 4 is often used in object detection and image processing applications. Several open-source ML/DL libraries provide great support for the speedy implementation of real-world applications. This development board is also complemented by AI-supported third-party accessories. Raspberry Pi used along with Intel Neural Computing Stick 2 makes an ideal combination for the development of AI frameworks. Another alternate third-party accessory is the Coral Edge TPU USB accelerator that enables this SBC for AI-enabled applications [69]. Additional platforms, such as Google’s AIY Vision and Voice kits, can also be paired with Raspberry Pi 4 to make significant contributions in AI and DL projects.

5.2. NVIDIA Jetson XavierTM

This platform provides supercomputer performance at edge in a compact system-on-module. It also has new cloud-native support that accelerates the NVIDIA software stack. The power and efficiency of this module enable multi-modal and accurate AI inference in a small form factor. The development of this SBC opened new doors for innovative edge devices in logistics, manufacturing, retail, agriculture, service, healthcare, smart cities, and many other applications [70]. The Jetson Xavier NX module with cloud-native support enables the design, deployment, and management of AI edge. The NVIDIA Transfer Learning Toolkit with pre-trained AI models from NVIDIA NGC provides a faster interface with optimized AI networks. To improve the performance of existing deployments, this module supports flexible and seamless updates. NVIDIA JetPack™ SDK facilitates the development of applications for Jetson Xavier NX with supporting libraries. It supports several AI frameworks such as graphics, multimedia computer vision, and many others. JetPack ensures reductions in development costs with the integration of the latest NVIDIA tools.

5.3. NVIDIA Jetson NanoTM

This module is an SoM of Jetson Nano Development Kit, and it tends to be used mostly to assemble a development board to deploy in AI-based graphic applications. This powerful AI-enabled SBC enables developers, students, and hobbyists to implement DL algorithms and models for several applications such as object detection, image classification, speech processing, image segmentation, and more [71]. This module is available with the multiple features of Jetson Nano Development Kit, which includes the main CPU, GPU, and several user interfaces such as USB, GPIO, and CSI. Instead of using a MicroSD card, the Jetson Nano module has 16 GB eMMC storage. All functionalities of this SBC can be accessed via an M.2 E connector.

5.4. NVIDIA Jetson AGX XavierTM

Jetson AGX Xavier development board provides great support for the development of end-to-end AI-enabled robotics applications for delivery, agriculture, retail, manufacturing, and many other fields. This development board facilitates the industrial or business circumstances related to automobile upgrades, horticulture, production, manufacturing, and retailing, among others. NVIDIA designed this SBC especially for advanced- level developers and engineers to explore the great potential of AI deployments in industrial, scientific, and business applications. The advanced capabilities of NVIDIA Jetson AGX Xavier make it an ideal choice for companies, businesses, and industries seeking to create versatile applications [72].

5.5. Google Coral

When a deployment of rapid ML/DL inferences in a compact form factor is required, this development board is an appropriate choice. This SBC can be used to prototype the embedded system and subsequently scale to production using the onboard Coral SoM combined with custom hardware. Google Coral provides a fully integrated system that includes NXP’s iMX 8M, LPDDR4 RAM, eMMC memory, Bluetooth, and WiFi. Its distinct processing capability is provided by Google’s Edge TPU co-processor. Google created the Edge TPU, a tiny ASIC architecture that offers high-performance ML/DL inferences while consuming relatively little power. It can execute advanced mobile vision models such as MobileNet v2 at almost 400 FPS, all in an energy-efficient manner [73].

5.6. Google Coral Dev Board Mini

This is a low-cost SBC with a built-in Edge TPU and real-time inference module for powerful ML/DL algorithms. This development board integrates MediaTek 8167 SoC with the Edge TPU, making it a standalone hardware platform capable of executing AI applications smoothly. TensorFlow Lite can be used with Edge TPU for quick and energy-efficient inferences. This integration provides adequate data protection according to the General Data Protection Regulation (GDPR) [74]. Google uses ML/DL-based solutions in its services at a large scale. Google introduced specialized processors TPU that can execute ML/DL algorithms in a faster and energy-efficient manner using TensorFlow. The best example is Google Maps Street view, which is analyzed by a TensorFlow-based neural network. Edge TPU supports TensorFlow Lite, which is highly compatible with embedded and mobile devices for efficient ML/DL implementations [74].

5.7. Rock Pi N10

This development board is a new member of the Rock Pi family. It contains a powerful SoC RK3399Pro that is integrated into GPU, CPU, and NPU. RK3399Pro’s CPU is a six-core that includes Dual Cortex-A72 and quad Cortex-A53. Its GPU supports Open CL 1.1 1.2, Vulkan, OpenGL ES 1.1/2.0/3.0/3.1/3.2, and DXI. Its NPU supports 8/16bit computing. The NPU has very good performance for complex calculations, which can be very helpful in DL applications [75]. This development board has multiple resources for storage. A 64 GB eMMC 5.1 and 64 bits dual-channel 8 GB LPDDR are embedded on the mainboard. Rock Pi N10 also supports multiple interfaces for camera, audio, USB, Ethernet, display, and I/O pins. This development board uses Android 8.1 and Debian as its firmware [75].

5.8. HiKey 970

This is a Super-Edge AI computing platform that is supported by Kirin 970 SoC with 4 × Cortex A53 and 4 × A73 processors. It has 6 GB UFS storage, 6 GB LPDDR4 RAM, and Gigabit Ethernet onboard. It is considered the world’s first dedicated NPU AI platform, which integrates popular neural network frameworks and Huawei HiAI computing architecture [76]. It supports CPU, GPU, and NPU all dedicated to AI acceleration. This SBC is mostly used to implement DL algorithms in automobiles, robots, and smart city applications [76].

5.9. BeagleBone AI

This development board provides great support for the use of AI in everyday applications. It has embedded vision engine (EVE) cores and the T1 C66x digital signal processor cores supported by OpenCL API with pre-installed tools. It is widely used in home automation as well as commercial and industrial applications [77]. This SBC is considered an ideal candidate for AI and DL implementations. It contains a rich collection of features and mimics the Linux approach, which makes it as an open-source platform. Some powerful specifications of this SBC make it more powerful than other development boards, such as its a Dual ARM Cortex A-15 processor, 2 C66x floating-point VLIW DSPs, dual-core PowerVR SGX544 3D GPU, and the dual-core programmable real-time unit, which consists of four embedded vision engines [77].

5.10. BeagleV

This development board offers open-source RISC-V computing on Linux distributions such as Fedora and Debian. This is a powerful SBC that contains a SiFive U74 processor, 8 GB RAM, and multiple I/O ports. BeagleV is comparatively affordable, as contrasted with other SBCs. The SiFive U74 processor contains two cores and an L2 cache that can all compete with the performance of other processors, such as ARM Cortex-A55 [78]. The SiFive U74 also includes a Vision DSP Tensilica-VP6, an NVDLA Engine, a Neural Network Engine, and an audio processing DSP. Currently, this development board is only available without GPU, but according to CNX software, it will also be available with GPU from September 2021. This SBC facilitates developers’ ability to extend their projects more strongly and flexibly.

6. Applications of Deep Learning in the IIoT

DL techniques have shown promising results in several applications. This section discusses some potential use cases of DL in IIoT applications.

6.1. Agriculture

Agriculture is a major industry and foundation of the economy in several countries. However, multiple factors such as population growth, climate change, and food safety concerns have propelled researchers to seek more innovative techniques for the protection and improvement of crops. In developed economies, AI and DL techniques are frequently used for improving the productivity and quality of agricultural products. Some popular research contributions in the context of DL techniques for the agricultural industry can be found here [79,80,81,82,83,84,85,86,87,88]. A comparison of some prominent studies is presented in Table 2. In the following, we discuss several prominent applications of DL in the agricultural sector as shown in Figure 13.

6.1.1. Weed Detection

When a vast number of data are produced by certain applications, then DL techniques are considered to be an ideal choice of solution. In weed detection problems, DL models can be trained using thousands of images of weeds. After optimized training, the DL model can successfully distinguish weeds from crops.

6.1.2. Smart Greenhouse

Greenhouses enable independent climate inside a specific area. It is quite complex to manage the micro-climate for a greenhouse. In this context, several environmental factors must be managed, such as temperature, humidity, wind direction, the intensity of light, heating, ventilation, etc. The use of DL methods in the greenhouse can be worthwhile to predict these outcomes and control these variables. DL can efficiently manage the inside climate by considering all the internal and external environmental factors.

6.1.3. Hydroponics

This system enables the growth of plants in a nutrient-rich environment instead of soil. Several factors affect the performance of the hydroponic system, such as temperature, sunlight, pH balance, and nutrition density. DL models can be trained according to these environmental parameters to launch an appropriate control action.

6.1.4. Soil Nutrient Monitoring

Soil fertility or nutrient level is determined by the organic matter in the soil, and it determines the yield and quality of an agricultural product. DL frameworks can be used to predict the presence of organic matter for the calculation of soil fertility. As a result, farmers can more easily predict the productivity and health of the soil on their farms.

6.1.5. Smart Irrigation

A DL-based framework for a plant recognition system can easily determine the water requirement for a particular plant type. It can also control the amount of water flow according to the plant’s requirements. In this particular application, the convolutional neural network (CNN) can be an ideal choice, since it can recognize the plant type after training by on datasets of plant images.

6.1.6. Fruit and Vegetable Plucking

Picking ripe crops requires labor and the ability to distinguish targets accurately. A DL model can be trained using image datasets of fruit and vegetables and then deployed in fruit and vegetable-picking tools such as robots. This robot can easily distinguish the specific fruit or vegetable from other objects and then pluck it. For this particular application, CNN can be an ideal choice.

6.1.7. Early Detection of Plant Diseases

DL frameworks can also be used for the early-stage detection of plant diseases. In this particular application, an image dataset will be required that contains the images of plant leaves with a particular disease. A deep convolutional neural network can then be trained to identify this disease and perform this task.

6.2. Education

DL is a type of individualized learning in the education industry. It can be used to provide an individualized educational experience to each student. Here, the students are guided for their own learning, can follow the pace they want, and make their own decisions about what to learn. Some state-of-the-art research contributions in the context of DL techniques for the education industry can be found here [89,90,91,92,93,94]. A comparison of some prominent studies is presented in Table 3. In the following, we discuss some top applications of DL in the education sector as shown in Figure 14.

6.2.1. Adaptive Learning

It helps the students to analyze their real-time performance and modifies teaching methodologies and curriculum according to the collected information. DL models can assist to have a customized engagement and tries to adapt to the individual for a better education. DL frameworks can suggest better learning paths to students using different materials and methodologies.

6.2.2. Increasing Efficiency

DL has a great capability to organize and manage better content and curriculum. It helps to understand the potential of everyone and bifurcate the work accordingly. It can efficiently analyze that which work is best suitable for teacher and student. DL can make instructors’ and students’ jobs simpler, making them more comfortable with their educational commitments. These modern techniques can motivate the students to increase their involvement and towards participation and learning. The DL frameworks also have the great potential to make education more efficient by classroom management and scheduling etc. in summary the DL methods can significantly increase the efficiency of the education sector.

6.2.3. Learning Analytics

Several times teachers can also become stuck while teaching. Because of this, the insights and gist are not properly understood by the students. Learning analytics can help the teacher to perform deep dives into the data. The instructor can go through hundreds of pieces of material, interpret them, and then draw appropriate connections and conclusions. It can make a very positive impact on the learning and teaching process. DL methods can help the students to gain benefits by learning advanced methodologies through learning analytics.

6.2.4. Predictive Analytics

Predictive analytics is helpful to understand the student’s needs and their behaviors in the education sector. It helps in drawing inferences about what could happen in the future. With the use of class assessments and half-year results, it is possible to predict which students will show the best performance in the exams and which students will face difficulties. It also assists teachers and parents in being aware and taking appropriate measures. In short, predictive analysis can help the students in a better way and can also work on their weak points in education.

6.2.5. Personalized Learning

This is one of the best applications that deep learning provides for education. Students can define their own learning ways through this educational model. They can study at their own pace and choose what and how they want to learn. They can select the subjects they want to study, the teacher, standards, curriculum, and pattern they want to follow.

6.2.6. Evaluating Assessments

DL frameworks can be used to grade student’s assignments and examinations more accurately as compared to the human. Although some inputs are necessarily required from humans. However, the best results will have more reliability and validity when a machine performs the task with more efficiency, higher accuracy, and low chances of error.

6.3. Healthcare

Healthcare is an important sector that provides value treatment to millions of people while also being a top income producer for many countries. Technology is helping healthcare experts to establish alternate staffing models, IP capitalization, deliver smart healthcare, and reduce administrative and supply expenses, in addition to playing a vital role in patient care, billing, and medical records. Deep learning is gaining great attention from healthcare industries. DL can help to analyze millions of data points, precise resource allocation, provide timely risk sources, and many other applications. Some latest research contributions in the context of DL techniques for the healthcare industry can be found here [95,96,97,98,99,100,101,102,103,104,105,106,107]. A comparison of some prominent studies is presented in Table 4. In the following, we discuss some top applications of DL in the healthcare industry as shown in Figure 15.

6.3.1. Diseases Identification and Diagnosis

DL can play a significant role in a patient’s disease identification, monitoring his real-time health condition, and suggest necessary measures to prevent it. It can efficiently identify the minor diseases as well as the major ones such as cancer detection which is difficult to identify at early stages. It also provides suitable information that enables more accurate diagnosis, improved outcomes, and individualized treatments.

6.3.2. Drug Discovery and Manufacturing

Next-generation sequencing and precision medicine are two research and development technologies that can help in the treatment of a wide range of health problems. DL algorithms can identify the patterns in medical data without providing any predictions. Discovering or producing a novel medication may be a costly and time-consuming procedure because many chemicals are tested and only one outcome might be beneficial. In this context, DL techniques can expedite this process.

6.3.3. Medical Imaging

DL algorithms have made it feasible to detect microscopic abnormalities in scanned images of patients, allowing physicians to make an accurate diagnosis. Traditional procedures such as X-ray and CT scans were sufficient for inspecting small abnormalities, but with the increasing diseases, there is a need to check them thoroughly. DL made advanced contributions in the field of computer vision. This technique was deployed in Microsoft’s Inner-Eye project, which develops image analysis technologies. This research project uses DL algorithms to provide novel tools for the automated, quantitative interpretation of 3D radiological images. This project also helps the experts in the field of surgical planning and radiotherapy.

6.3.4. Personalized Medicine/Treatment

Doctors can now provide personalized treatment to individual patients based on their specific needs according to the explosion of patient data in the form of electronic health records. They aim to extract insights from enormous volumes of datasets and use them to make patients healthier at individual levels. DL technologies can help to suggest personalized combinations and predict the risks of disease. Watson healthcare creates powerful resources using ML and DL for the patient’s health improvement. This platform decreased the doctor’s time that was spent on making treatment decisions based on research analysis, clinical practices, and trials. This DL-based platform now offers blood cancer treatment and many other diseases.

6.3.5. Smart Health Records

There are still some processes exist that take a lot of time for data entry. Maintaining every day’s health records is a time-consuming activity. In this context, DL can save time, effort, and money to maintain health records. Google’s Cloud Vision API and MATLAB’s DL-based handwriting recognition technology are frequently used for document classification. It facilitates the recording of clinical data, modernizes the workflow, and improvement of health information accuracy.

6.3.6. Diseases Prediction

Multiple DL technologies are being used to monitor and predict epidemics all around the world. Scientists have access to vast amounts of data collected from satellites, websites, and social media platforms. DL assists in the collaboration of this information and the prediction of everything from small diseases to serious chronic infectious diseases.

6.4. Intelligent Transportation Systems

Intelligent transportation systems (ITS) have been influenced by the rapid growth of deep learning. DL gained great popularity in ITS with the proliferation of data and advancements in computational technologies. In ITS, the DL models are being used to provides powerful and viable solutions to address the multiple challenges in the traditional transport system. In the context of DL approaches, some latest research contributions related to the ITS can be found in [108,109,110,111,112,113,114,115,116,117]. A comparison of prominent studies is presented in Table 5. In the following, we discuss potential applications of DL in the ITS as shown in Figure 16.

6.4.1. Traffic Characteristics Prediction

This is one of the most important applications of DL in intelligent transportation. It can help the drivers to select the most feasible routes and enable the traffic control agencies for the efficient management of transport. The main characteristics are traveling time, traffic speed, and traffic flow. As these characteristics are not mutually exclusive, however, the techniques used to predict one of them can also be used to predict the other features

6.4.2. Traffic Incidents Inference

The main goal of traffic incident risk prediction for a specific location helps the traffic management authorities in reducing incident risks in a hazardous region and traffic congestion in incident locations. Several key factors can be helpful for traffic incidents prediction. These factors include traffic flow, human mobility, weather, geographical position, period, and day of the week. All these features can be evaluated as indicators of a traffic incident. DL models can be effectively used to predict and detect traffic incidents.

6.4.3. Vehicle Detection

Vehicle detection in highway monitoring videos is a challenging task and very important in terms of intelligent traffic management and control. Traffic surveillance cameras are installed on highways that generate a vast amount of data in the form of traffic videos. These data can be further analyzed to identify the vehicle identification. DL algorithms provide fast and robust solutions for vehicle detection that can be helpful for traffic management authorities to ensure traffic safety on highways.

6.4.4. Traffic Signal Timing

One of the major responsibilities of ITS management is to control the traffic flow using traffic signals. The optimum selection of signal timing plays a significant role in traffic flow management. This is one of the biggest challenges in the field of modern transportation. DL studies providing new paths by modeling the dynamics of traffic to obtain the best performance on the roads. In several studies, the DL algorithms proved their superior performance for the optimum selection of traffic signal timing.

6.4.5. Visual Recognition Tasks

The use of non-intrusive recognition and detection systems, such as camera-image-based systems, is one of the most significant uses of DL. These applications might range from providing appropriate roadway infrastructure for driving cars to provide autonomous vehicles with a safe and dependable driving strategy.

6.5. Manufacturing Industry

DL-based solutions are transforming manufacturing industries into highly efficient organizations by improving productivity, increasing capacity use, decreasing production defects, and reducing maintenance costs. In the context of DL approaches some latest research contributions for the manufacturing industry can be found here [118,119,120,121,122,123,124,125,126]. A comparison of some prominent studies is presented in Table 6. In the following, we discuss some potential applications of DL in the manufacturing industry as shown in Figure 17.

6.5.1. Maintenance

DL algorithms are widely used for predictive maintenance in an industry to prevent failures of the machine by identifying the upcoming potential challenges more accurately. DL frameworks analyze the real-time images, sounds, and other information collected by industrial sensors to reduce the downtime of systems.

6.5.2. Predictive Analytics

DL models can accurately predict operational outcomes. It enables the companies to optimize their manufacturing processes. DL algorithms use real-time sensor data monitoring production lines, inventory, waiting for the time of machines, technical and physical conditions of machines, and behavior of workers. Manufacturing companies can also examine the effectiveness of their processes by analyzing the information about quality issues, raw materials, maintenance procedures, and environmental factors such as temperature and humidity. The predictive analysis can be very helpful to identify, unprofitable lines, non-value-added activities, and bottlenecks in industrial operations.

6.5.3. Product Development

DL is increasingly used in product designing. DL-based software enables the manufactures to input the important aspects of product design such as material, cost, and durability, etc. Based on the user input, this software proposes a highly suitable product design and facilitates the user for more realistic testing without having to build extremely expensive prototypes. Engineers of the automotive industry believe that the DL will be the most promising approach to design high-speed racing cars and can also be helpful to create more realistic platforms to achieve optimal performance.

6.5.4. Quality Assurance

DL-enabled computer vision techniques are used for the automatic detection of defective products. Different manufacturers claim that such quality testing solutions can achieve up to 90% accuracy in defect identification for specific applications. It also enables the production teams to address the quality problems at an earlier stage. As an example, Audi uses DL-based image recognition system for the successful identification of fine cracks on metal sheets.

6.5.5. Robotics

Several manufacturers use industrial robots to handle dangerous and complicated processes. Now DL frameworks enable the robots to learn on their own. A DL-enabled robot can train itself for performing new tasks using object and pattern recognition capabilities.

6.5.6. Supply Chain Management

DL approaches are considered highly accurate analytical techniques that can add value in complex supply chain management operations. Using DL models companies can predict real-time demands, optimize their production schedules and supply chain operations, and can efficiently manage the inventory to reduce purchasing costs of raw materials. These capabilities of DL models enable the companies to quickly respond to the change in market demand.

6.5.7. Logistics

DL models can significantly improve fuel efficiency and delivery time by analyzing the real-time information about drivers and vehicles in logistic operations.

6.6. Aviation Industry

AI and DL can streamline and automate customer services, analytics, machinery maintenance, and many other internal procedures and operations in the aviation industry. The most recent and relevant research contributions related to the use of DL techniques in the aviation industry field can be found in [127,128,129,130,131,132,133,134,135]. A comparison of prominent studies is presented in Table 7. In the following, we discuss the most important applications of DL in the aviation industry as shown in Figure 18.

6.6.1. Revenue Management

It is a popular application of data and analytics that determines the sale of an appropriate product to relevant customers at a reasonable price using the right channel. Revenue management specialists in the aviation industry use DL techniques to define attractive destinations and adjust prices, find efficient and convenient distribution channels, and seat management to keep the airline competitive and customer-friendly.

6.6.2. Air Safety and Airplane Maintenance

Airlines usually bear the huge cost because of flights delays and cancellations that include expenditures on maintenance and compensations to passengers stuck in airports. According to a study, about 30% of total delay time is caused by unplanned maintenance. To address this challenge, a DL-based predictive analysis can be applied to fleet technical support. Airline carriers deploy intelligent predictive maintenance frameworks for better data management from aircraft health monitoring sensors. These systems enable the technician’s access to real-time and historical information from any location. The information about the current technical condition of an aircraft through notifications, alerts, and reports can be helpful for employees to identify the possible malfunction and parts replacement proactively.

6.6.3. Feedback Analysis

AI and DL-enabled systems can swiftly assess if there is a chance to favorably intervene in the customer journey and transform a bad experience into a pleasurable one. It also enables the companies to respond rapidly in an aligned and synchronized way that is on board with the business’s values. Using DL for market research and feedback analysis enables airlines to make intelligent decisions and meet customer’s expectations.

6.6.4. Messaging Automation

Travelers usually become nervous when a flight delay or baggage loss occurs. If passengers do not receive a timely response from an airline representative regarding their problem, then they will not choose this airline for their upcoming trips. The speed of response to customers matters a lot. DL-based solutions simplify and expedite the customer service processes using advanced algorithms of natural language processing. These solutions can automate several routine processes and can create ease for passengers.

6.6.5. Crew Management

One of the major responsibilities of the scheduling department is to assign crews to each of thousands of flights every day. Crew management is a complex task that includes several factors such as flight route, aircraft type, crew member license, work regulations, vaccination of staff, and days off to approve conflict-free schedules for pilots and flight attendants. Several aviation companies are using AI and DL-based software to make optimal scheduling in terms of crew qualification, working hours, aircraft expenses, and use. Such software integrates predictive models with an airline operation management system.

6.6.6. Fuel Efficiency Optimization

Airlines use AI frameworks with built-in DL algorithms to gather and analyze flight data regarding aircraft type, weather, weight, route distance, and altitudes, etc. DL models can efficiently estimate the optimal amount of fuel required for a specific flight.

6.6.7. In-Flight Food Service Management

DL techniques can be helpful for supply management specialists to determine the number of snacks and drinks on board without being wasteful. The effective use of DL methods can predict the amount of food required for any specific flight. The optimization in foodservice management can reduce the cost and improve the service quality.

6.7. Defense

ML and DL have become an essential part of modern warfare. Modern military systems integrated with ML/DL can process enormous amounts of data more effectively compared to traditional systems. AI techniques improve self-control, self-regulation, and self-actuation of combat systems because of their inherent decision-making capabilities. AI and DL are being deployed in almost every military application. Military research organizations have increased the research funding to drive the adoption of ML/DL-driven applications for military applications. Recent research contributions related to the use of DL techniques in defense and military applications can be found in [136,137,138,139,140,141]. A comparison of prominent studies is presented in Table 8. In the following, we discuss the most relevant potential applications of DL in the military domain as shown in Figure 19.

6.7.1. Warfare Platforms

Armed forces from several countries across the globe are incorporating AI techniques into weapons and other military equipment used on ground, airborne, naval, and space platforms. The deployment of DL methods on these platforms has enabled the development of highly efficient autonomous warfare systems. DL enhanced the performance of warfare platforms with less maintenance.

6.7.2. Cybersecurity

Military platforms are frequently vulnerable to cyber-attacks, which can result in massive data loss and damage to defense systems. However, the deployment of ML/DL techniques can autonomously protect computers, networks, data, and programs from any intrusion. Additionally, DL-enabled cybersecurity systems can record the pattern of new attacks and develop firewalls to tackle them.

6.7.3. Logistics and Transportation

The efficient transportation of ammunition, goods, troops and armaments is an integral component of successful military operations. The integration of DL techniques with military transportation can reduce human operational efforts and lower transportation costs. It also enables the military fleets to quickly predict the component failures and easily detect anomalies.

6.7.4. Target Recognition

ML/DL frameworks are being developed to improve the accuracy of target identification in complex combat environments. These techniques allow the armed forced to gain an in-depth understanding of critical operation areas by analyzing documents, reports, and news feeds, etc. The capabilities of DL-based target identification systems include aggregation of weather conditions, probability-based forecasts of enemy behavior, assessments of mission approaches, and suggested mitigation strategies. In addition, the integration of DL in target recognition systems improves the ability of these systems to identify the position of targets.

6.7.5. Battlefield Healthcare

ML/DL can be integrated with Robotic Ground Platforms (RGPs) and Robotic Surgical Systems (RSS) to provide evacuation activities and surgical support in war zones. Under critical conditions, the health care systems equipped with ML/DL can search the soldier’s medical records and efficiently assist in complex diagnoses.

6.7.6. Combat Simulation and Training

Simulation and Training is a broad field that combines software engineering, system engineering, and computer science to develop computerized models that acquaint soldiers with the various combat systems during military operations.

6.8. Sports Industry

ML and DL techniques are being adopted to better analyze, organize, and optimize every aspect of the sports industry. Management and athletes are required to gather every information about the individual and team performance to gain an edge over their opponent. AI and virtual reality techniques are increasingly used to provide real-time analytics in sports. The availability of large data in sports can be very helpful to develop predictive DL models to make better management decisions. In the context of DL approaches, the latest and relevant research contributions related to the sports industry field can be found in [142,143,144,145,146,147,148,149]. A comparison of prominent studies is presented in Table 9. In the following, we discuss relevant applications of DL in the sports industry as shown in Figure 20.

6.8.1. Sports Coaching

The presence of experienced coaches is an essential factor behind every winning team. Presently, augmented AI with wearable sensors and high-speed cameras are helping a lot to improve training. The weakest point of conventional coaching is that it takes years to hone the skills. Now AI-enabled assistants are helping the coaches to improve the game strategy and optimize the team’s lineup according to modern game requirements.

6.8.2. Analyzing Player Behavior

DL evaluates the multiple parameters of sports side by side. The individual player’s data can be tracked using wearable sensors, high-speed cameras, and DL throughout the match. This technology can identify minuscule differences and help us understand the player’s performance under stress. DL can communicate better real-time game plan change to coaches and players and can also identify the minuscule differences.

6.8.3. Refereeing

It is one of the earliest applications of ML and DL in sports. These techniques are helpful in precise judgments to make the game fair and law-abiding. The best example is the use of hawk-eye technology in cricket to determine if the batter is out or not in the cases of LBW.

6.8.4. Health and Fitness Improvement

ML and DL are transforming the healthcare industry in several ways. The predictive and diagnostic capabilities of DL can also be helpful in the sports industry. The physical health and fitness of the players is an extremely important element in sports. Team managements spend a lot of money to maintain the mental and physical wellbeing of the players. Each sport measures the mental and physical skills of a player differently. The adaptation of DL methods in sports can be very helpful to analyze these skills accurately.

6.8.5. Streaming and Broadcasting

AI and DL frameworks are used in match recording, where broadcasters can selectively capture the highlights. DL can automatically generate the subtitles for live sports events in multiple languages based on the viewer’s location and language preferences. DL is also used in sports marketing to determine the best camera angles during the match. In addition, these methods are very helpful to identify the top moments of the game for multiple brands to obtain better advertising opportunities.

7. Potential Challenges and Future Research Directions

In the previous sections, we discussed the importance and potential applications of DL techniques in various sectors. However, the successful implementation of these techniques and obtaining desired results in an IIoT is challenging. This section discusses some key issues and future research directions for DL-based IIoT.

7.1. Key Challenges

The implementation of DL techniques in the IIoT faces several challenges. In the following, we discuss some key challenges faced by several IIoT applications.

7.1.1. Complexity

This is one of the biggest challenges faced by DL models. Some extra efforts are required to address this problem [150]. The first main issue is the computation requirement and time consumption of the training process because of the complexity of the DL algorithm and large IIoT datasets. The second major problem is the scarcity of large numbers of training data in industrial situations, which can impair the efficiency and accuracy of DL models through overfitting. The complexity problem can be addressed by employing a tensor-train deep compression (TTDC) model for effective feature learning from industrial data [151]. This approach increases model speed by condensing many components in the DL algorithm.

7.1.2. Algorithm Selection

There are several well-known DL algorithms are available for IIoT applications. Most of these algorithms work fine for any general scenario but for a specific IIoT application, the selection of the DL algorithm is a challenge [152]. For implementation in a specific application, it is very important to determine which algorithm is best suitable for this particular application. The inaccurate selection of the DL algorithm can result in several issues, including a waste of money, time, and efforts.

7.1.3. Data Selection

The selection of training data is directly connected to the success of a DL method. In every DL model, the right type and amount of data are important [153]. As part of an industrial process, it is essential to prevent such data that might lead to selective bias.

7.1.4. Data Preprocessing

In DL models, data preprocessing is another mandatory process. It transforms the selected data to make it best compatible with desired DL model [154,155]. This process includes cleaning, data conversion, feature scaling, and removal or replacement of missing entries.

7.1.5. Data Labeling

In terms of implementation, training, and deployment in various IIoT applications, supervised DL techniques are simplest and highly suitable [156]. On the other hand, unsupervised techniques are difficult to implement and sometimes require a lengthy training process. Additionally, data labeling is challenging for both supervised and unsupervised DL models and cannot be outsourced for intensive tasks. DL algorithms are continuously developing, and their feature learning capabilities are changing. The incorporation of new features in the DL framework can be sometimes a nightmare if the earlier datasets, features, and models are not properly documented.

7.2. Future Directions

There are multiple aspects of DL frameworks in the IIoT environments that should be rectified for future implementations. This feature will necessitate various improvements, such as the selection of intelligent algorithms with increased efficiency and compatibility with different hardware platforms. In the following, we discuss some potential future research directions in the context of DL-based IIoT.

7.2.1. DL-Enabled Edge/Cloud Computing

In smart industrial environments, fast and efficient computing is considered to be the main feature that affects reliability, latency, and many other important performance parameters [157,158,159]. IIoT applications require powerful machines to compute the large amounts of data generated by diverse operations. To conduct effective computation and provide understandable computing infrastructure for IIoT, a hybrid cloud-edge computing environment is now required. However, this is still not recognized as a realistic method of coping with complex learning issues. There are several reasons behind this problem, such as low processing power, the resource-constrained nature of IoT devices, and the complexity of DL algorithms [160,161]. Therefore, it is deemed appropriate to employ edge-based computing frameworks, simply because of their potential to minimize latency and improve the learning process. However, the integration of DL techniques with an edge-enabled framework for IIoT scenarios remains an open research problem. To achieve self-organization, greater productivity, and reduced runtime, the combined implementation of parallel and distributed learning for edge-based designs needs further optimization [162].

7.2.2. Distributed Deep Learning

Distributed DL is a specialized approach for large-scale DL operations with massive training datasets and long training durations. It functions by assigning several computational resources to collaborate on a single task [163]. Distributed DL can perform several tasks, such as data collection, data mining, and testing, via multiple distributed nodes that work on it and address the initial problem simultaneously and promptly. Therefore, distributed DL is considered one of the most appropriate techniques for implementation in the IIoT. However, its implementation in the smart industrial environment is still a challenging task. The main issue is to determine an effective way of managing the overall distributed computation resources.

7.2.3. Low Latency and Improved Reliability

Smart industrial frameworks require multiple synchronized processes that require lower latency and higher reliability to obtain the desired performance [164,165]. Furthermore, the DL techniques used in IIoT should be capable of dealing with these challenges as well as other aspects such as resource management and network deployment [166]. However, the competency of DL-based IIoT scenarios is still in the early stages with low latency and high-reliability requirements in IIoT. Therefore, research efforts in this area are necessary to build a theoretical and practical foundation for DL-based IIoT to provide low-latency and ultra-reliable communication.

7.2.4. Intelligent Sensing and Decision-Making

Control challenges in DL-based IIoT include both sensing and assessment methods that involve a huge number of sensors and actuators. This does enables smart sensing capabilities though, such as classification, prediction, decision-making, and direct management of the complete IIoT system [167]. Intelligent sensing and useful decision-making abilities are strictly enforced in a smart manufacturing environment, with no room for economic loss or safety issues caused by failure. Therefore, the successful deployment of DL-based IIoT has stringent criteria for accurate prediction, categorization, and decision-making, which can only be met using intelligent sensing and data-driven decisions [168,169].

7.2.5. Lightweight Learning Frameworks

The IIoT framework contains multiple intelligent and connected devices to establish a smart industrial setup [170]. These devices include sensors, actuators, and controllers that use different DL algorithms for the learning process. These devices are considered resource-constrained devices that require lightweight learning algorithms [171]. This will enhance the learning process for diverse industrial devices, resulting in an intelligent IIoT network with reduced computing complexity and increased network lifetime. The best example is the employment of a hardware-in-the-loop (HIL) simulation platform [172,173,174,175,176]. The HIL combines computations for numerous processes with the internal hardware of sensors and actuators. The essential capability of the HIL platform is the use of real-time data generated by installed hardware testbeds.

8. Conclusions

This paper presented a comprehensive survey of deep learning (DL) techniques and applications for the Industrial Internet of Things (IIoT). A brief introduction is presented in the context of DL deployments for industrial applications along with the latest state-of-the-art contributions from survey articles. The survey highlighted the major drawbacks of existing studies and overcame these shortcomings by including additional information. Most of the existing surveys lack a detailed description of standard IIoT architecture. This study described the detailed seven-layer architecture of the IIoT along with key enabling technology and protocols. This survey discussed the theories of well-known DL algorithms along with mathematical backgrounds and reference architectures. One of the major shortcomings of the existing studies is the non-consideration of software and hardware implementation platform. To address this issue, a detailed description of software and hardware deployment frameworks is presented in the context of DL and IIoT. To evaluate the effectiveness of DL for IIoT, several potential use cases of DL technologies for the IIoT are discussed. Finally, this survey is concluded by highlighting the key challenges in existing DL-based IIoT systems and presented potential research directions for future endeavors.

Author Contributions

Conceptualisation, S.L., M.D., W.B., Z.e.H. and S.S.J.; Writing original draft preparation, S.L., Z.e.H., J.A., S.S.J. and Z.I.; Review and editing, M.D., Z.I.; Supervision, W.B. and J.A. All authors have read and agreed to the published version of the manuscript.

Funding

This research received no external funding.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Acknowledgments

The authors would like to acknowledge the support of Prince Sultan University for paying the Article Processing Charges (APC) of this publication.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Sisinni, E.; Saifullah, A.; Han, S.; Jennehag, U.; Gidlund, M. Industrial internet of things: Challenges, opportunities, and directions. IEEE Trans. Ind. Inform. 2018, 14, 4724–4734. [Google Scholar] [CrossRef]
  2. Saeed, N.; Alouini, M.S.; Al-Naffouri, T.Y. Toward the internet of underground things: A systematic survey. IEEE Commun. Surv. Tutor. 2019, 21, 3443–3466. [Google Scholar] [CrossRef] [Green Version]
  3. Al-Fuqaha, A.; Guizani, M.; Mohammadi, M.; Aledhari, M.; Ayyash, M. Internet of things: A survey on enabling technologies, protocols, and applications. IEEE Commun. Surv. Tutor. 2015, 17, 2347–2376. [Google Scholar] [CrossRef]
  4. Liang, F.; Yu, W.; Liu, X.; Griffith, D.; Golmie, N. Toward edge-based deep learning in industrial Internet of Things. IEEE Internet Things J. 2020, 7, 4329–4341. [Google Scholar] [CrossRef]
  5. Statista. Industrial Internet of Things Market Size Worldwide from 2017 to 2025. 2020. Available online: https://www.statista.com/statistics/611004/global-industrial-internet-of-things-market-size/ (accessed on 1 August 2021).
  6. Khalil, R.A.; Saeed, N.; Masood, M.; Fard, Y.M.; Alouini, M.S.; Al-Naffouri, T.Y. Deep Learning in the Industrial Internet of Things: Potentials, Challenges, and Emerging Applications. IEEE Internet Things J. 2021, 8, 11016–11040. [Google Scholar] [CrossRef]
  7. Zhu, S.; Ota, K.; Dong, M. Green AI for IIoT: Energy Efficient Intelligent Edge Computing for Industrial Internet of Things. IEEE Trans. Green Commun. Netw. 2021. [Google Scholar] [CrossRef]
  8. Ullah, F.; Naeem, H.; Jabbar, S.; Khalid, S.; Latif, M.A.; Al-Turjman, F.; Mostarda, L. Cyber security threats detection in internet of things using deep learning approach. IEEE Access 2019, 7, 124379–124389. [Google Scholar] [CrossRef]
  9. Zhang, Q.; Yang, L.T.; Chen, Z.; Li, P.; Bu, F. An adaptive dropout deep computation model for industrial IoT big data learning with crowdsourcing to cloud computing. IEEE Trans. Ind. Inform. 2018, 15, 2330–2337. [Google Scholar] [CrossRef]
  10. Mohammadi, M.; Al-Fuqaha, A.; Sorour, S.; Guizani, M. Deep learning for IoT big data and streaming analytics: A survey. IEEE Commun. Surv. Tutor. 2018, 20, 2923–2960. [Google Scholar] [CrossRef] [Green Version]
  11. Ma, X.; Yao, T.; Hu, M.; Dong, Y.; Liu, W.; Wang, F.; Liu, J. A survey on deep learning empowered IoT applications. IEEE Access 2019, 7, 181721–181732. [Google Scholar] [CrossRef]
  12. Boulila, W.; Sellami, M.; Driss, M.; Al-Sarem, M.; Safaei, M.; Ghaleb, F.A. RS-DCNN: A novel distributed convolutional-neural-networks based-approach for big remote-sensing image classification. Comput. Electron. Agric. 2021, 182, 106014. [Google Scholar] [CrossRef]
  13. Ambika, P. Machine learning and deep learning algorithms on the Industrial Internet of Things (IIoT). Adv. Comput. 2020, 117, 321–338. [Google Scholar]
  14. Saleem, T.J.; Chishti, M.A. Deep learning for the internet of things: Potential benefits and use-cases. Digit. Commun. Netw. 2020. [Google Scholar] [CrossRef]
  15. Deepan, P.; Sudha, L. Deep Learning Algorithm and Its Applications to IoT and Computer Vision. Artif. Intell. IOT Smart Converg. Eco-Friendly Topogr. 2021, 85, 223. [Google Scholar]
  16. Boyes, H.; Hallaq, B.; Cunningham, J.; Watson, T. The industrial internet of things (IIoT): An analysis framework. Comput. Ind. 2018, 101, 1–12. [Google Scholar] [CrossRef]
  17. Lin, S.W.; Miller, B.; Durand, J.; Joshi, R.; Didier, P.; Chigani, A.; Torenbeek, R.; Duggal, D.; Martin, R.; Bleakley, G.; et al. Industrial Internet Reference Architecture; Technical Reports; Industrial Internet Consortium (IIC): Boston, MA, USA, 2015. [Google Scholar]
  18. Cheng, C.F.; Chen, Y.C.; Lin, J.C.W. A Carrier-Based Sensor Deployment Algorithm for Perception Layer in the IoT Architecture. IEEE Sens. J. 2020, 20, 10295–10305. [Google Scholar] [CrossRef]
  19. Kaur, H.; Kumar, R. A survey on Internet of Things (IoT): Layer-specific, domain-specific and industry-defined architectures. In Advances in Computational Intelligence and Communication Technology; Springer: Berlin/Heidelberg, Germany, 2021; pp. 265–275. [Google Scholar]
  20. HaddadPajouh, H.; Khayami, R.; Dehghantanha, A.; Choo, K.K.R.; Parizi, R.M. AI4SAFE-IoT: An AI-powered secure architecture for edge layer of Internet of things. Neural Comput. Appl. 2020, 32, 16119–16133. [Google Scholar] [CrossRef]
  21. Abdullah, A.; Kaur, H.; Biswas, R. Universal Layers of IoT Architecture and Its Security Analysis. In New Paradigm in Decision Science and Management; Springer: Berlin/Heidelberg, Germany, 2020; pp. 293–302. [Google Scholar]
  22. Mrabet, H.; Belguith, S.; Alhomoud, A.; Jemai, A. A survey of IoT security based on a layered architecture of sensing and data analysis. Sensors 2020, 20, 3625. [Google Scholar] [CrossRef]
  23. Alidoosti, M.; Nowroozi, A.; Nickabadi, A. Evaluating the web-application resiliency to business-layer DoS attacks. ETRI J. 2020, 42, 433–445. [Google Scholar] [CrossRef]
  24. Patnaik, R.; Padhy, N.; Raju, K.S. A systematic survey on IoT security issues, vulnerability and open challenges. In Intelligent System Design; Springer: Berlin/Heidelberg, Germany, 2021; pp. 723–730. [Google Scholar]
  25. Chen, B.; Wan, J. Emerging trends of ml-based intelligent services for industrial internet of things (iiot). In Proceedings of the 2019 Computing, Communications and IoT Applications (ComComAp), Shenzhen, China, 26–28 October 2019; pp. 135–139. [Google Scholar]
  26. Ozanich, E.; Gerstoft, P.; Niu, H. A feedforward neural network for direction-of-arrival estimation. J. Acoust. Soc. Am. 2020, 147, 2035–2048. [Google Scholar] [CrossRef]
  27. Orimoloye, L.O.; Sung, M.C.; Ma, T.; Johnson, J.E. Comparing the effectiveness of deep feedforward neural networks and shallow architectures for predicting stock price indices. Expert Syst. Appl. 2020, 139, 112828. [Google Scholar] [CrossRef]
  28. Zhang, N.; Sun, S. Multiview Graph Restricted Boltzmann Machines. IEEE Trans. Cybern. 2021. [Google Scholar] [CrossRef] [PubMed]
  29. Gu, L.; Zhou, F.; Yang, L. Towards the representational power of restricted Boltzmann machines. Neurocomputing 2020, 415, 358–367. [Google Scholar] [CrossRef]
  30. Deshwal, D.; Sangwan, P. A Comprehensive Study of Deep Neural Networks for Unsupervised Deep Learning. In Artificial Intelligence for Sustainable Development: Theory, Practice and Future Applications; Springer: Berlin/Heidelberg, Germany, 2021; pp. 101–126. [Google Scholar]
  31. Chen, C.; Ma, Y.; Ren, G. Hyperspectral classification using deep belief networks based on conjugate gradient update and pixel-centric spectral block features. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2020, 13, 4060–4069. [Google Scholar] [CrossRef]
  32. Hong, C.; Zeng, Z.Y.; Fu, Y.Z.; Guo, M.F. Deep-Belief-Networks Based Fault Classification in Power Distribution Networks. IEEJ Trans. Electr. Electron. Eng. 2020, 15, 1428–1435. [Google Scholar] [CrossRef]
  33. Chu, H.; Wei, J.; Wu, W.; Jiang, Y.; Chu, Q.; Meng, X. A classification-based deep belief networks model framework for daily streamflow forecasting. J. Hydrol. 2021, 595, 125967. [Google Scholar] [CrossRef]
  34. Zhuang, X.; Guo, H.; Alajlan, N.; Zhu, H.; Rabczuk, T. Deep autoencoder based energy method for the bending, vibration, and buckling analysis of Kirchhoff plates with transfer learning. Eur. J. Mech. -A/Solids 2021, 87, 104225. [Google Scholar] [CrossRef]
  35. Pawar, K.; Attar, V.Z. Assessment of autoencoder architectures for data representation. In Deep Learning: Concepts and Architectures; Springer: Berlin/Heidelberg, Germany, 2020; pp. 101–132. [Google Scholar]
  36. Deng, Z.; Li, Y.; Zhu, H.; Huang, K.; Tang, Z.; Wang, Z. Sparse stacked autoencoder network for complex system monitoring with industrial applications. Chaos Solitons Fractals 2020, 137, 109838. [Google Scholar] [CrossRef]
  37. LeCun, Y.; Bottou, L.; Bengio, Y.; Haffner, P. Gradient-based learning applied to document recognition. Proc. IEEE 1998, 86, 2278–2324. [Google Scholar] [CrossRef] [Green Version]
  38. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Adv. Neural Inf. Process. Syst. 2012, 25, 1097–1105. [Google Scholar] [CrossRef]
  39. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Las Vegas, NV, USA, 27–30 June 2016; pp. 779–788. [Google Scholar]
  40. Zhang, Y.; Wang, Y.; Luo, G. A new optimization algorithm for non-stationary time series prediction based on recurrent neural networks. Future Gener. Comput. Syst. 2020, 102, 738–745. [Google Scholar] [CrossRef]
  41. Boulila, W.; Ghandorh, H.; Khan, M.A.; Ahmed, F.; Ahmad, J. A novel CNN-LSTM-based approach to predict urban expansion. Ecol. Inform. 2021, 101325. [Google Scholar] [CrossRef]
  42. Kousik, N.; Natarajan, Y.; Raja, R.A.; Kallam, S.; Patan, R.; Gandomi, A.H. Improved salient object detection using hybrid Convolution Recurrent Neural Network. Expert Syst. Appl. 2021, 166, 114064. [Google Scholar] [CrossRef]
  43. Aggarwal, A.; Mittal, M.; Battineni, G. Generative adversarial network: An overview of theory and applications. Int. J. Inf. Manag. Data Insights 2021, 100004. [Google Scholar] [CrossRef]
  44. Yoon, J.; Drumright, L.N.; Van Der Schaar, M. Anonymization through data synthesis using generative adversarial networks (ads-gan). IEEE J. Biomed. Health Inform. 2020, 24, 2378–2388. [Google Scholar] [CrossRef]
  45. Fekri, M.N.; Ghosh, A.M.; Grolinger, K. Generating energy data for machine learning with recurrent generative adversarial networks. Energies 2020, 13, 130. [Google Scholar] [CrossRef] [Green Version]
  46. Nguyen, G.; Dlugolinsky, S.; Bobák, M.; Tran, V.; García, Á.L.; Heredia, I.; Malík, P.; Hluchỳ, L. Machine learning and deep learning frameworks and libraries for large-scale data mining: A survey. Artif. Intell. Rev. 2019, 52, 77–124. [Google Scholar] [CrossRef] [Green Version]
  47. Pang, B.; Nijkamp, E.; Wu, Y.N. Deep learning with tensorflow: A review. J. Educ. Behav. Stat. 2020, 45, 227–248. [Google Scholar] [CrossRef]
  48. TensorFlow: An End-to-End Open Source Machine Learning Platform. Available online: https://www.tensorflow.org/ (accessed on 4 August 2021).
  49. TensorFlow Lite. Available online: https://www.tensorflow.org/lite/guide (accessed on 4 August 2021).
  50. Yapici, M.M.; Topaloğlu, N. Performance comparison of deep learning frameworks. Comput. Inform. 2021, 1, 1–11. [Google Scholar]
  51. Microsoft. The Microsoft Cognitive Toolkit. Available online: https://docs.microsoft.com/en-us/cognitive-toolkit/ (accessed on 4 August 2021).
  52. Gulli, A.; Pal, S. Deep Learning with Keras; Packt Publishing Ltd.: Birmingham, UK, 2017. [Google Scholar]
  53. Keras: The Python Deep Learning API. Available online: https://keras.io/about/ (accessed on 4 August 2021).
  54. Bahrampour, S.; Ramakrishnan, N.; Schott, L.; Shah, M. Comparative Study of Caffe, Neon, Theano, and Torch for Deep Learning 2016. Available online: https://openreview.net/pdf/q7kEN7WoXU8LEkD3t7BQ.pdf (accessed on 6 August 2021).
  55. Caffe: Deep Learning Framework. Available online: https://caffe.berkeleyvision.org/ (accessed on 8 August 2021).
  56. Wang, Z.; Liu, K.; Li, J.; Zhu, Y.; Zhang, Y. Various frameworks and libraries of machine learning and deep learning: A survey. Arch. Comput. Methods Eng. 2019, 1–24. [Google Scholar] [CrossRef]
  57. Caffe2: A New Lightweight, Modular, and Scalable Deep Learning Framework. Available online: https://caffe2.ai/docs/caffe-migration.html (accessed on 8 August 2021).
  58. Hodnett, M.; Wiley, J.F. R Deep Learning Essentials: A Step-by-Step Guide to Building Deep Learning Models Using Tensorflow, Keras, and MXNet; Packt Publishing Ltd.: Birmingham, UK, 2018. [Google Scholar]
  59. Apache MXNet: A Flexible and Efficient Library for Deep Learning. Available online: https://mxnet.apache.org/versions/1.8.0/ (accessed on 8 August 2021).
  60. Kabakuş, A.T. A Comparison of the State-of-the-Art Deep Learning Platforms: An Experimental Study. Sak. Univ. J. Comput. Inf. Sci. 2020, 3, 169–182. [Google Scholar] [CrossRef]
  61. Torch: A Scientific Computing Framewotk for LuaJIT. Available online: http://torch.ch/ (accessed on 8 August 2021).
  62. Paszke, A.; Gross, S.; Massa, F.; Lerer, A.; Bradbury, J.; Chanan, G.; Killeen, T.; Lin, Z.; Gimelshein, N.; Antiga, L.; et al. Pytorch: An imperative style, high-performance deep learning library. Adv. Neural Inf. Process. Syst. 2019, 32, 8026–8037. [Google Scholar]
  63. Stevens, E.; Antiga, L.; Viehmann, T. Deep learning with PyTorch; Manning Publications Company: Shelter Island, NY, USA, 2020. [Google Scholar]
  64. Bourez, C. Deep Learning with Theano; Packt Publishing Ltd.: Birmingham, UK, 2017. [Google Scholar]
  65. Brownlee, J. Introduction to the Python Deep Learning Library Theano. 2016. Available online: https://machinelearningmastery.com/introduction-python-deep-learning-library-theano/ (accessed on 8 August 2021).
  66. Tokui, S.; Okuta, R.; Akiba, T.; Niitani, Y.; Ogawa, T.; Saito, S.; Suzuki, S.; Uenishi, K.; Vogel, B.; Yamazaki Vincent, H. Chainer: A deep learning framework for accelerating the research cycle. In Proceedings of the 25th ACM SIGKDD International Conference on Knowledge Discovery & Data Mining, Anchorage, AK, USA, 4–8 August 2019; pp. 2002–2011. [Google Scholar]
  67. Chainer: A Powerful, Flexible, and Intuitive Framework for Neural Networks. Available online: https://chainer.org/ (accessed on 8 August 2021).
  68. Hattersley, L. Learn Artificial Intelligence with Raspberry Pi. 2019. Available online: https://magpi.raspberrypi.org/articles/learn-artificial-intelligence-with-raspberry-pi (accessed on 12 August 2021).
  69. Hattersley, L. Raspberry Pi 4. Available online: https://www.raspberrypi.org/products/raspberry-pi-4-model-b/ (accessed on 12 August 2021).
  70. Developer, N. Jetson Xavier NX Developer Kit. Available online: https://developer.nvidia.com/embedded/jetson-xavier-nx-devkit (accessed on 12 August 2021).
  71. Developer, N. Jetson Nano: Deep Learning Inference Benchmarks. Available online: https://developer.nvidia.com/embedded/jetson-nano-dl-inference-benchmarks (accessed on 12 August 2021).
  72. Developer, N. NVIDIA Jetson AGX Xavier: The AI Platform for Autonomous Machines. Available online: https://www.nvidia.com/en-us/autonomous-machines/jetson-agx-xavier/ (accessed on 12 August 2021).
  73. Dev Board: Coral. Available online: https://coral.ai/products/dev-board/ (accessed on 12 August 2021).
  74. Dev Board Mini: Coral. Available online: https://coral.ai/products/dev-board-mini/ (accessed on 12 August 2021).
  75. Yida. Introducing the Rock Pi N10 RK3399Pro—SBC for AI and Deep Learning. 2019. Available online: https://www.seeedstudio.com/blog/2019/12/04/introducing-the-rock-pi-n10-rk3399pro-sbc-for-ai-and-deep-learning/ (accessed on 13 August 2021).
  76. Synced. Huawei Introduces AI Development Board HiKey 970. 2018. Available online: https://medium.com/syncedreview/huawei-introduces-ai-development-board-hikey-970-763ac996b29a (accessed on 13 August 2021).
  77. BeagleBone AI: Fast Track to Embedded Artificial Intelligence. Available online: https://beagleboard.org/ai (accessed on 13 August 2021).
  78. Cloudware, O. BeagleV Development Board Features RISC-V Architecture. 2021. Available online: https://opencloudware.com/beaglev-development-board-features-risc-v-architecture/ (accessed on 13 August 2021).
  79. Sehgal, G.; Gupta, B.; Paneri, K.; Singh, K.; Sharma, G.; Shroff, G. Crop planning using stochastic visual optimization. In Proceedings of the 2017 IEEE Visualization in Data Science (VDS), Phoenix, AZ, USA, 1 October 2017; pp. 47–51. [Google Scholar]
  80. Song, X.; Zhang, G.; Liu, F.; Li, D.; Zhao, Y.; Yang, J. Modeling spatio-temporal distribution of soil moisture by deep learning-based cellular automata model. J. Arid Land 2016, 8, 734–748. [Google Scholar] [CrossRef] [Green Version]
  81. Douarre, C.; Schielein, R.; Frindel, C.; Gerth, S.; Rousseau, D. Deep learning based root-soil segmentation from X-ray tomography images. bioRxiv 2016, 071662. [Google Scholar] [CrossRef]
  82. Aliev, K.; Pasero, E.; Jawaid, M.M.; Narejo, S.; Pulatov, A. Internet of plants application for smart agriculture. Int. J. Adv. Comput. Sci. Appl 2018, 9, 421–429. [Google Scholar] [CrossRef]
  83. Huang, H.; Deng, J.; Lan, Y.; Yang, A.; Deng, X.; Zhang, L. A fully convolutional network for weed mapping of unmanned aerial vehicle (UAV) imagery. PLoS ONE 2018, 13, e0196302. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  84. Rahnemoonfar, M.; Sheppard, C. Deep count: Fruit counting based on deep simulated learning. Sensors 2017, 17, 905. [Google Scholar] [CrossRef] [Green Version]
  85. Jiang, Z.; Liu, C.; Hendricks, N.P.; Ganapathysubramanian, B.; Hayes, D.J.; Sarkar, S. Predicting county level corn yields using deep long short term memory models. arXiv 2018, arXiv:1805.12044. [Google Scholar]
  86. Ferentinos, K.P. Deep learning models for plant disease detection and diagnosis. Comput. Electron. Agric. 2018, 145, 311–318. [Google Scholar] [CrossRef]
  87. Toda, Y.; Okura, F. How convolutional neural networks diagnose plant disease. Plant Phenomics 2019, 2019, 9237136. [Google Scholar] [CrossRef]
  88. Grinblat, G.L.; Uzal, L.C.; Larese, M.G.; Granitto, P.M. Deep learning for plant identification using vein morphological patterns. Comput. Electron. Agric. 2016, 127, 418–424. [Google Scholar] [CrossRef] [Green Version]
  89. Bhardwaj, P.; Gupta, P.; Panwar, H.; Siddiqui, M.K.; Morales-Menendez, R.; Bhaik, A. Application of Deep Learning on Student Engagement in e-learning environments. Comput. Electr. Eng. 2021, 93, 107277. [Google Scholar] [CrossRef]
  90. Han, Z.; Xu, A. Ecological evolution path of smart education platform based on deep learning and image detection. Microprocess. Microsyst. 2021, 80, 103343. [Google Scholar] [CrossRef]
  91. Tsai, S.C.; Chen, C.H.; Shiao, Y.T.; Ciou, J.S.; Wu, T.N. Precision education with statistical learning and deep learning: A case study in Taiwan. Int. J. Educ. Technol. High. Educ. 2020, 17, 1–13. [Google Scholar] [CrossRef]
  92. Fok, W.W.; He, Y.; Yeung, H.A.; Law, K.; Cheung, K.; Ai, Y.; Ho, P. Prediction model for students’ future development by deep learning and tensorflow artificial intelligence engine. In Proceedings of the 2018 4th International Conference on Information Management (ICIM), Oxford, UK, 25–27 May 2018; pp. 103–106. [Google Scholar]
  93. Nandal, P. Deep Learning in diverse Computing and Network Applications Student Admission Predictor using Deep Learning. In Proceedings of the International Conference on Innovative Computing & Communications (ICICC), Delhi, India, 20–22 February 2020. [Google Scholar]
  94. Khaleel, M.F.; Sharkh, M.A.; Kalil, M. A Cloud-based Architecture for Automated Grading of Computer-Aided Design Student Work Using Deep Learning. In Proceedings of the 2020 IEEE Canadian Conference on Electrical and Computer Engineering (CCECE), London, ON, Canada, 30 August–2 September 2020; pp. 1–5. [Google Scholar]
  95. Choi, E.; Biswal, S.; Malin, B.; Duke, J.; Stewart, W.F.; Sun, J. Generating multi-label discrete patient records using generative adversarial networks. In Proceedings of the Machine Learning for Healthcare Conference, PMLR, Boston, MA, USA, 18–19 August 2017; pp. 286–305. [Google Scholar]
  96. Nie, D.; Trullo, R.; Lian, J.; Wang, L.; Petitjean, C.; Ruan, S.; Wang, Q.; Shen, D. Medical image synthesis with deep convolutional adversarial networks. IEEE Trans. Biomed. Eng. 2018, 65, 2720–2730. [Google Scholar] [CrossRef]
  97. Sha, Y.; Wang, M.D. Interpretable predictions of clinical outcomes with an attention-based recurrent neural network. In Proceedings of the 8th ACM International Conference on Bioinformatics, Computational Biology, and Health Informatics, Boston, MA, USA, 20–23 August 2017; pp. 233–240. [Google Scholar]
  98. Verma, H.; Kumar, S. An accurate missing data prediction method using LSTM based deep learning for health care. In Proceedings of the 20th International Conference on Distributed Computing and Networking, Bangalore, India, 4–7 January 2019; pp. 371–376. [Google Scholar]
  99. Sun, M.; Min, T.; Zang, T.; Wang, Y. CDL4CDRP: A collaborative deep learning approach for clinical decision and risk prediction. Processes 2019, 7, 265. [Google Scholar] [CrossRef] [Green Version]
  100. Najdi, S.; Gharbali, A.A.; Fonseca, J.M. Feature transformation based on stacked sparse autoencoders for sleep stage classification. In Proceedings of the Doctoral Conference on Computing, Electrical and Industrial Systems, Costa de Caparica, Portugal, 3–5 May 2017; Springer: Berlin/Heidelberg, Germany, 2017; pp. 191–200. [Google Scholar]
  101. Nguyen, M.; Sun, N.; Alexander, D.C.; Feng, J.; Yeo, B.T. Modeling Alzheimer’s disease progression using deep recurrent neural networks. In Proceedings of the 2018 International Workshop on Pattern Recognition in Neuroimaging (PRNI), Singapore, 12–14 June 2018; pp. 1–4. [Google Scholar]
  102. Xue, Q.; Wang, X.; Meehan, S.; Kuang, J.; Gao, J.A.; Chuah, M.C. Recurrent neural networks based obesity status prediction using activity data. In Proceedings of the 2018 17th IEEE International Conference on Machine Learning and Applications (ICMLA), Orlando, FL, USA, 17–20 December 2018; pp. 865–870. [Google Scholar]
  103. Amin, S.U.; Hossain, M.S.; Muhammad, G.; Alhussein, M.; Rahman, M.A. Cognitive smart healthcare for pathology detection and monitoring. IEEE Access 2019, 7, 10745–10753. [Google Scholar] [CrossRef]
  104. Wang, L.; Zhou, X. Detection of congestive heart failure based on LSTM-based deep network via short-term RR intervals. Sensors 2019, 19, 1502. [Google Scholar] [CrossRef] [Green Version]
  105. Alhussein, M.; Muhammad, G. Voice pathology detection using deep learning on mobile healthcare framework. IEEE Access 2018, 6, 41034–41041. [Google Scholar] [CrossRef]
  106. Maragatham, G.; Devi, S. LSTM model for prediction of heart failure in big data. J. Med. Syst. 2019, 43, 111. [Google Scholar] [CrossRef]
  107. Kim, J.; Kang, U.; Lee, Y. Statistics and deep belief network-based cardiovascular risk prediction. Healthc. Inform. Res. 2017, 23, 169–175. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  108. Xu, W.; Feng, X.; Wang, J.; Luo, C.; Li, J.; Ming, Z. Energy harvesting-based smart transportation mode detection system via attention-based lstm. IEEE Access 2019, 7, 66423–66434. [Google Scholar] [CrossRef]
  109. Song, X.; Kanasugi, H.; Shibasaki, R. Deeptransport: Prediction and simulation of human mobility and transportation mode at a citywide level. In Proceedings of the Twenty-Fifth International Joint Conference on Artificial Intelligence, New York, NY, USA, 9–15 July 2016; pp. 2618–2624. [Google Scholar]
  110. Mohammadi, M.; Al-Fuqaha, A.; Oh, J.S. Path planning in support of smart mobility applications using generative adversarial networks. In Proceedings of the 2018 IEEE International Conference on Internet of Things (iThings) and IEEE Green Computing and Communications (GreenCom) and IEEE Cyber, Physical and Social Computing (CPSCom) and IEEE Smart Data (SmartData), Halifax, NS, Canada, 30 July–3 August 2018; pp. 878–885. [Google Scholar]
  111. Camero, A.; Toutouh, J.; Stolfi, D.H.; Alba, E. Evolutionary deep learning for car park occupancy prediction in smart cities. In Proceedings of the International Conference on Learning and Intelligent Optimization, Kalamata, Greece, 10–15 June 2018; Springer: Berlin/Heidelberg, Germany, 2018; pp. 386–401. [Google Scholar]
  112. Singh, D.; Mohan, C.K. Deep spatio-temporal representation for detection of road accidents using stacked autoencoder. IEEE Trans. Intell. Transp. Syst. 2018, 20, 879–887. [Google Scholar] [CrossRef]
  113. Lv, Z.; Xu, J.; Zheng, K.; Yin, H.; Zhao, P.; Zhou, X. Lc-rnn: A deep learning model for traffic speed prediction. In Proceedings of the Twenty-Seventh International Joint Conference on Artificial Intelligence (IJCAI-18), Stockholm, Sweden, 13–19 July 2018; pp. 3470–3476. [Google Scholar]
  114. Ma, X.; Yu, H.; Wang, Y.; Wang, Y. Large-scale transportation network congestion evolution prediction using deep learning theory. PLoS ONE 2015, 10, e0119044. [Google Scholar] [CrossRef]
  115. Pérez, J.L.; Gutierrez-Torre, A.; Berral, J.L.; Carrera, D. A resilient and distributed near real-time traffic forecasting application for Fog computing environments. Future Gener. Comput. Syst. 2018, 87, 198–212. [Google Scholar] [CrossRef]
  116. Xiangxue, W.; Lunhui, X.; Kaixun, C. Data-driven short-term forecasting for urban road network traffic based on data processing and LSTM-RNN. Arab. J. Sci. Eng. 2019, 44, 3043–3060. [Google Scholar] [CrossRef]
  117. Goudarzi, S.; Kama, M.N.; Anisi, M.H.; Soleymani, S.A.; Doctor, F. Self-organizing traffic flow prediction with an optimized deep belief network for internet of vehicles. Sensors 2018, 18, 3459. [Google Scholar] [CrossRef] [Green Version]
  118. Park, S.T.; Li, G.; Hong, J.C. A study on smart factory-based ambient intelligence context-aware intrusion detection system using machine learning. J. Ambient Intell. Humaniz. Comput. 2020, 11, 1405–1412. [Google Scholar] [CrossRef]
  119. Tao, W.; Lai, Z.H.; Leu, M.C.; Yin, Z. Worker activity recognition in smart manufacturing using IMU and sEMG signals with convolutional neural networks. Procedia Manuf. 2018, 26, 1159–1166. [Google Scholar] [CrossRef]
  120. Ren, L.; Sun, Y.; Cui, J.; Zhang, L. Bearing remaining useful life prediction based on deep autoencoder and deep neural networks. J. Manuf. Syst. 2018, 48, 71–77. [Google Scholar] [CrossRef]
  121. Yan, H.; Wan, J.; Zhang, C.; Tang, S.; Hua, Q.; Wang, Z. Industrial big data analytics for prediction of remaining useful life based on deep learning. IEEE Access 2018, 6, 17190–17197. [Google Scholar] [CrossRef]
  122. Jiang, L.; Ge, Z.; Song, Z. Semi-supervised fault classification based on dynamic Sparse Stacked auto-encoders model. Chemom. Intell. Lab. Syst. 2017, 168, 72–83. [Google Scholar] [CrossRef]
  123. Yuan, Y.; Ma, G.; Cheng, C.; Zhou, B.; Zhao, H.; Zhang, H.T.; Ding, H. Artificial intelligent diagnosis and monitoring in manufacturing. arXiv 2018, arXiv:1901.02057. [Google Scholar]
  124. Li, L.; Ota, K.; Dong, M. Deep learning for smart industry: Efficient manufacture inspection system with fog computing. IEEE Trans. Ind. Inform. 2018, 14, 4665–4673. [Google Scholar] [CrossRef] [Green Version]
  125. Wang, J.; Wang, K.; Wang, Y.; Huang, Z.; Xue, R. Deep Boltzmann machine based condition prediction for smart manufacturing. J. Ambient Intell. Humaniz. Comput. 2019, 10, 851–861. [Google Scholar] [CrossRef]
  126. Zhang, W.; Guo, W.; Liu, X.; Liu, Y.; Zhou, J.; Li, B.; Lu, Q.; Yang, S. LSTM-based analysis of industrial IoT equipment. IEEE Access 2018, 6, 23551–23560. [Google Scholar] [CrossRef]
  127. Alkhamisi, A.O.; Mehmood, R. An ensemble machine and deep learning model for risk prediction in aviation systems. In Proceedings of the 2020 6th Conference on Data Science and Machine Learning Applications (CDMA), Riyadh, Saudi Arabia, 4–5 March 2020; pp. 54–59. [Google Scholar]
  128. Acuna-Agost, R.; Thomas, E.; Lhéritier, A. Price elasticity estimation for deep learning-based choice models: An application to air itinerary choices. J. Revenue Pricing Manag. 2021, 20, 213–226. [Google Scholar] [CrossRef]
  129. Barakat, H.; Yeniterzi, R.; Martín-Domingo, L. Applying deep learning models to twitter data to detect airport service quality. J. Air Transp. Manag. 2021, 91, 102003. [Google Scholar] [CrossRef]
  130. Wu, E.Q.; Deng, P.Y.; Qu, X.Y.; Tang, Z.; Zhang, W.M.; Zhu, L.M.; Ren, H.; Zhou, G.R.; Sheng, R.S. Detecting fatigue status of pilots based on deep learning network using EEG signals. IEEE Trans. Cogn. Dev. Syst. 2020, 13, 575–585. [Google Scholar] [CrossRef]
  131. Dong, T.; Yang, Q.; Ebadi, N.; Luo, X.R.; Rad, P. Identifying Incident Causal Factors to Improve Aviation Transportation Safety: Proposing a Deep Learning Approach. J. Adv. Transp. 2021, 2021, 5540046. [Google Scholar] [CrossRef]
  132. Yazdi, M.F.; Kamel, S.R.; Chabok, S.J.M.; Kheirabadi, M. Flight delay prediction based on deep learning and Levenberg-Marquart algorithm. J. Big Data 2020, 7, 106. [Google Scholar] [CrossRef]
  133. Wang, L.; Mykityshyn, A.; Johnson, C.M.; Marple, B.D. Deep Learning for Flight Demand and Delays Forecasting. In Proceedings of the AIAA Aviation 2021 Forum, Virtual, 2–6 August 2021; p. 2399. [Google Scholar]
  134. Corrado, S.J.; Puranik, T.G.; Pinon-Fischer, O.J.; Mavris, D.; Rose, R.; Williams, J.; Heidary, R. Deep Autoencoder for Anomaly Detection in Terminal Airspace Operations. In Proceedings of the AIAA Aviation 2021 Forum, Virtual, 2–6 August 2021; p. 2405. [Google Scholar]
  135. Hasib, K.M.; Habib, M.A.; Towhid, N.A.; Showrov, M.I.H. A Novel Deep Learning based Sentiment Analysis of Twitter Data for US Airline Service. In Proceedings of the 2021 International Conference on Information and Communication Technology for Sustainable Development (ICICT4SD), Dhaka, Bangladesh, 27–28 February 2021; pp. 450–455. [Google Scholar]
  136. Das, L.B.; Lijiya, A.; Jagadanand, G.; Aadith, A.; Gautham, S.; Mohan, V.; Reuben, S.; George, G. Human Target Search and Detection using Autonomous UAV and Deep learning. In Proceedings of the 2020 IEEE International Conference on Industry 4.0, Artificial Intelligence, and Communications Technology (IAICT), Bali, Indonesia, 7–8 July 2020; pp. 55–61. [Google Scholar]
  137. Calderón, M.; Aguilar, W.G.; Merizalde, D. Visual-Based Real-Time Detection Using Neural Networks and Micro-UAVs for Military Operations. In Developments and Advances in Defense and Security; Springer: Berlin/Heidelberg, Germany, 2020; p. 55. [Google Scholar]
  138. Krishnaveni, P.; Sutha, J. Novel deep learning framework for broadcasting abnormal events obtained from surveillance applications. J. Ambient Intell. Humaniz. Comput. 2020, 1–15. [Google Scholar] [CrossRef]
  139. Sharma, D.Y.K.; Pradeep, S. Deep Learning based Real Time Object Recognition for Security in Air Defense. In Proceedings of the 13th INDIACom, New Delhi, India, 13–15 March 2019; pp. 64–67. [Google Scholar]
  140. Shi, Y.; Sagduyu, Y.E.; Erpek, T.; Davaslioglu, K.; Lu, Z.; Li, J.H. Adversarial deep learning for cognitive radio security: Jamming attack and defense strategies. In Proceedings of the 2018 IEEE International Conference on Communications Workshops (ICC Workshops), Kansas City, MO, USA, 20–24 May 2018; pp. 1–6. [Google Scholar]
  141. Wang, F.; Zhong, C.; Gursoy, M.C.; Velipasalar, S. Defense strategies against adversarial jamming attacks via deep reinforcement learning. In Proceedings of the 2020 54th Annual Conference on Information Sciences and Systems (CISS), Princeton, NJ, USA, 18–20 March 2020; pp. 1–6. [Google Scholar]
  142. Chen, C.Y.; Lai, W.; Hsieh, H.Y.; Zheng, W.H.; Wang, Y.S.; Chuang, J.H. Generating defensive plays in basketball games. In Proceedings of the 26th ACM international conference on Multimedia, Seoul, Korea, 22–26 October 2018; pp. 1580–1588. [Google Scholar]
  143. Chung, J.; Kastner, K.; Dinh, L.; Goel, K.; Courville, A.C.; Bengio, Y. A recurrent latent variable model for sequential data. Adv. Neural Inf. Process. Syst. 2015, 28, 2980–2988. [Google Scholar]
  144. Baccouche, M.; Mamalet, F.; Wolf, C.; Garcia, C.; Baskurt, A. Action classification in soccer videos with long short-term memory recurrent neural networks. In Proceedings of the International Conference on Artificial Neural Networks, Thessaloniki, Greece, 15–18 September 2010; Springer: Berlin/Heidelberg, Germany, 2010; pp. 154–159. [Google Scholar]
  145. Theagarajan, R.; Pala, F.; Zhang, X.; Bhanu, B. Soccer: Who has the ball? Generating visual analytics and player statistics. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, Salt Lake City, UT, USA, 18–22 June 2018; pp. 1749–1757. [Google Scholar]
  146. Le, H.M.; Carr, P.; Yue, Y.; Lucey, P. Data-driven ghosting using deep imitation learning. In Proceedings of the MIT Sloan Sports Analytics Conference, Boston, MA, USA, 3–4 March 2017. [Google Scholar]
  147. Kautz, T.; Groh, B.H.; Hannink, J.; Jensen, U.; Strubberg, H.; Eskofier, B.M. Activity recognition in beach volleyball using a deep convolutional neural network. Data Min. Knowl. Discov. 2017, 31, 1678–1705. [Google Scholar] [CrossRef]
  148. Qiao, F. Application of deep learning in automatic detection of technical and tactical indicators of table tennis. PLoS ONE 2021, 16, e0245259. [Google Scholar] [CrossRef]
  149. Cao, Z.; Liao, T.; Song, W.; Chen, Z.; Li, C. Detecting the shuttlecock for a badminton robot: A YOLO based approach. Expert Syst. Appl. 2021, 164, 113833. [Google Scholar] [CrossRef]
  150. Li, H.; Ota, K.; Dong, M. Learning IoT in edge: Deep learning for the Internet of Things with edge computing. IEEE Netw. 2018, 32, 96–101. [Google Scholar] [CrossRef] [Green Version]
  151. Roopaei, M.; Rad, P.; Jamshidi, M. Deep learning control for complex and large scale cloud systems. Intell. Autom. Soft Comput. 2017, 23, 389–391. [Google Scholar] [CrossRef]
  152. Rymarczyk, T.; Kłosowski, G.; Kozłowski, E.; Tchórzewski, P. Comparison of selected machine learning algorithms for industrial electrical tomography. Sensors 2019, 19, 1521. [Google Scholar] [CrossRef] [Green Version]
  153. Wang, Y.; Pan, Z.; Yuan, X.; Yang, C.; Gui, W. A novel deep learning based fault diagnosis approach for chemical process with extended deep belief network. ISA Trans. 2020, 96, 457–467. [Google Scholar] [CrossRef]
  154. Latif, S.; Zou, Z.; Idrees, Z.; Ahmad, J. A novel attack detection scheme for the industrial internet of things using a lightweight random neural network. IEEE Access 2020, 8, 89337–89350. [Google Scholar] [CrossRef]
  155. Khan, M.A.; Khan, M.A.; Latif, S.; Shah, A.A.; Rehman, M.U.; Boulila, W.; Driss, M.; Ahmad, J. Voting Classifier-based Intrusion Detection for IoT Networks. arXiv 2021, arXiv:2104.10015. [Google Scholar]
  156. Zhou, L.; Wu, D.; Chen, J.; Dong, Z. When computation hugs intelligence: Content-aware data processing for industrial IoT. IEEE Internet Things J. 2017, 5, 1657–1666. [Google Scholar] [CrossRef]
  157. Ai, Y.; Peng, M.; Zhang, K. Edge computing technologies for Internet of Things: A primer. Digit. Commun. Netw. 2018, 4, 77–86. [Google Scholar] [CrossRef]
  158. Latif, S.; Idrees, Z.; Zou, Z.; Ahmad, J. DRaNN: A deep random neural network model for intrusion detection in industrial IoT. In Proceedings of the 2020 International Conference on UK-China Emerging Technologies (UCET), Glasgow, UK, 20–21 August 2020; pp. 1–4. [Google Scholar]
  159. Savaglio, C.; Gerace, P.; Di Fatta, G.; Fortino, G. Data mining at the IoT edge. In Proceedings of the 2019 28th International Conference on Computer Communication and Networks (ICCCN), Valencia, Spain, 29 July–1 August 2019; pp. 1–6. [Google Scholar]
  160. Hatcher, W.G.; Yu, W. A survey of deep learning: Platforms, applications and emerging research trends. IEEE Access 2018, 6, 24411–24432. [Google Scholar] [CrossRef]
  161. Chen, J.; Li, K.; Deng, Q.; Li, K.; Philip, S.Y. Distributed deep learning model for intelligent video surveillance systems with edge computing. IEEE Trans. Ind. Inform. 2019. [Google Scholar] [CrossRef] [Green Version]
  162. Latif, S.; Idrees, Z.; Ahmad, J.; Zheng, L.; Zou, Z. A blockchain-based architecture for secure and trustworthy operations in the industrial Internet of Things. J. Ind. Inf. Integr. 2021, 21, 100190. [Google Scholar] [CrossRef]
  163. Luckow, A.; Cook, M.; Ashcraft, N.; Weill, E.; Djerekarov, E.; Vorster, B. Deep learning in the automotive industry: Applications and tools. In Proceedings of the 2016 IEEE International Conference on Big Data (Big Data), Washington, DC, USA, 5–8 December 2016; pp. 3759–3768. [Google Scholar]
  164. Yang, H.; Alphones, A.; Zhong, W.D.; Chen, C.; Xie, X. Learning-based energy-efficient resource management by heterogeneous RF/VLC for ultra-reliable low-latency industrial IoT networks. IEEE Trans. Ind. Inform. 2019, 16, 5565–5576. [Google Scholar] [CrossRef]
  165. Huma, Z.E.; Latif, S.; Ahmad, J.; Idrees, Z.; Ibrar, A.; Zou, Z.; Alqahtani, F.; Baothman, F. A Hybrid Deep Random Neural Network for Cyberattack Detection in the Industrial Internet of Things. IEEE Access 2021, 9, 55595–55605. [Google Scholar] [CrossRef]
  166. Bennis, M.; Debbah, M.; Poor, H.V. Ultrareliable and low-latency wireless communication: Tail, risk, and scale. Proc. IEEE 2018, 106, 1834–1853. [Google Scholar] [CrossRef] [Green Version]
  167. Gungor, V.C.; Hancke, G.P. Industrial wireless sensor networks: Challenges, design principles, and technical approaches. IEEE Trans. Ind. Electron. 2009, 56, 4258–4265. [Google Scholar] [CrossRef] [Green Version]
  168. Popescu, G.H. The economic value of the industrial internet of things. J. Self-Gov. Manag. Econ. 2015, 3, 86–91. [Google Scholar]
  169. Banaie, F.; Hashemzadeh, M. Complementing IIoT Services through AI: Feasibility and Suitability. In AI-Enabled Threat Detection and Security Analysis for Industrial IoT; Springer: Berlin/Heidelberg, Germany, 2021; pp. 7–19. [Google Scholar]
  170. Latif, S.; Idrees, Z.; e Huma, Z.; Ahmad, J. Blockchain technology for the industrial Internet of Things: A comprehensive survey on security challenges, architectures, applications, and future research directions. Trans. Emerg. Telecommun. Technol. 2021, e4337. [Google Scholar] [CrossRef]
  171. Cantor, B.; Grant, P.; Johnston, C. Automotive Engineering: Lightweight, Functional, and Novel Materials; CRC Press: Boca Raton, FL, USA, 2008. [Google Scholar]
  172. Alvarez-Gonzalez, F.; Griffo, A.; Sen, B.; Wang, J. Real-time hardware-in-the-loop simulation of permanent-magnet synchronous motor drives under stator faults. IEEE Trans. Ind. Electron. 2017, 64, 6960–6969. [Google Scholar] [CrossRef] [Green Version]
  173. Idrees, Z.; Granados, J.; Sun, Y.; Latif, S.; Gong, L.; Zou, Z.; Zheng, L. IEEE 1588 for Clock Synchronization in Industrial IoT and Related Applications: A Review on Contributing Technologies, Protocols and Enhancement Methodologies. IEEE Access 2020, 8, 155660–155678. [Google Scholar] [CrossRef]
  174. Abbas, A.; Khan, M.A.; Latif, S.; Ajaz, M.; Shah, A.A.; Ahmad, J. A New Ensemble-Based Intrusion Detection System for Internet of Things. Arab. J. Sci. Eng. 2021, 1–15. [Google Scholar] [CrossRef]
  175. Huerta, F.; Gruber, J.K.; Prodanovic, M.; Matatagui, P. Power-hardware-in-the-loop test beds: Evaluation tools for grid integration of distributed energy resources. IEEE Ind. Appl. Mag. 2016, 22, 18–26. [Google Scholar] [CrossRef]
  176. Mai, X.H.; Kwak, S.K.; Jung, J.H.; Kim, K.A. Comprehensive electric-thermal photovoltaic modeling for power-hardware-in-the-loop simulation (PHILS) applications. IEEE Trans. Ind. Electron. 2017, 64, 6255–6264. [Google Scholar] [CrossRef]
Figure 1. The worldwide market size of the IIoT from 2017 to 2025 [5].
Figure 1. The worldwide market size of the IIoT from 2017 to 2025 [5].
Sensors 21 07518 g001
Figure 2. A comparison of publication records for DL-based IoT/IIoT applications.
Figure 2. A comparison of publication records for DL-based IoT/IIoT applications.
Sensors 21 07518 g002
Figure 3. Reference architecture of the Industrial Internet of Things.
Figure 3. Reference architecture of the Industrial Internet of Things.
Sensors 21 07518 g003
Figure 4. A general architecture of the deep feedforward neural network (DFNN).
Figure 4. A general architecture of the deep feedforward neural network (DFNN).
Sensors 21 07518 g004
Figure 5. A general architecture of the Restricted Boltzmann Machines (RBM).
Figure 5. A general architecture of the Restricted Boltzmann Machines (RBM).
Sensors 21 07518 g005
Figure 6. A general architecture of the Deep Belief Networks (DBN).
Figure 6. A general architecture of the Deep Belief Networks (DBN).
Sensors 21 07518 g006
Figure 7. A general architecture of an Autoencoder (AE).
Figure 7. A general architecture of an Autoencoder (AE).
Sensors 21 07518 g007
Figure 8. A general architecture of the Convolutional Neural Network.
Figure 8. A general architecture of the Convolutional Neural Network.
Sensors 21 07518 g008
Figure 9. A general architecture of the Recurrent Neural Network (RNN).
Figure 9. A general architecture of the Recurrent Neural Network (RNN).
Sensors 21 07518 g009
Figure 10. A general architecture of the Generative Adversarial Networks (GAN).
Figure 10. A general architecture of the Generative Adversarial Networks (GAN).
Sensors 21 07518 g010
Figure 11. Some well-known ML/DL frameworks.
Figure 11. Some well-known ML/DL frameworks.
Sensors 21 07518 g011
Figure 12. Some well-known Development Boards for DL Implementations.
Figure 12. Some well-known Development Boards for DL Implementations.
Sensors 21 07518 g012
Figure 13. Applications of DL in agriculture.
Figure 13. Applications of DL in agriculture.
Sensors 21 07518 g013
Figure 14. Applications of DL in education.
Figure 14. Applications of DL in education.
Sensors 21 07518 g014
Figure 15. Applications of DL in healthcare.
Figure 15. Applications of DL in healthcare.
Sensors 21 07518 g015
Figure 16. Applications of DL in intelligent transport system.
Figure 16. Applications of DL in intelligent transport system.
Sensors 21 07518 g016
Figure 17. Applications of DL in manufacturing industry.
Figure 17. Applications of DL in manufacturing industry.
Sensors 21 07518 g017
Figure 18. Applications of DL in aviation industry.
Figure 18. Applications of DL in aviation industry.
Sensors 21 07518 g018
Figure 19. Applications of DL in defense.
Figure 19. Applications of DL in defense.
Sensors 21 07518 g019
Figure 20. Applications of DL in sports industry.
Figure 20. Applications of DL in sports industry.
Sensors 21 07518 g020
Table 1. A comparison of latest survey articles.
Table 1. A comparison of latest survey articles.
AuthorsYearContributions of Survey Articles
IIoT ArchitectureAlgorithmsFrameworksHardwareApplicationsFuture Directions
Mohammadi et al. [10]2018×
Ma et al. [11]2019×××
Sengupta et al. [12]2020×××
Ambika et al. [13]2020×××
Saleem et al. [14]2020××××
Deepan et al. [15]2021×××
Khalil et al. [6]2021××
Our Study2021
Table 2. State-of-the-art contributions in DL-based agriculture.
Table 2. State-of-the-art contributions in DL-based agriculture.
AuthorDL AlgorithmDatasetApplicationPurpose of DL Technique
Sehgal et al. [79]LSTMDataset from Syngeta crop challenge (2016)Weather predictionWeather prediction according to the conditions of preceding year
Song et al. [80]DBNData gathered from corn field (irrigated) in ChinaSoil moisture content predictionPrediction of moisture content in the soil
Douarre et al. [81]CNNX-ray tomographic images of soilRoot and soil segmentationImage categorization into two classes root and soil
Aliev et al. [82]RNNSensory dataInternet of plants-based systemTo envisage the minimum and temperature records for ten days
Huang et al. [83]CNNData collected using multirotor UAVWeed mapping in smart agricultureClassification of input images into three categories: weed, rice and others
Rahnemoonfar et al. [84]CNNDataset consisting of 24,000 imagesTomato countingPrediction of tomatoes quantity
Jiang et al. [85]LSTMData obtained from the National Agricultural Statistics Service (NASS) Quick StatsCrop yield predictionCorn yield prediction
Ferentinos et al. [86]CNNLeaf images of plantsPlant disease detectionImage classification into health and diseased categories
Toda et al. [87]CNNPlant Village datasetPlant disease diagnosisLeaf image classification into healthy and diseased categories and diagnosis of disease type
Grinblat et al. [88]CNNDataset consisting of vein leaf images of soybean, red beans, and white beansPlant identificationLegume’s classification into three categories: red beans, soybean, and white beans
Table 3. State-of-the-art contributions in DL-based educational sector.
Table 3. State-of-the-art contributions in DL-based educational sector.
AuthorDL AlgorithmDatasetApplicationPurpose of DL Technique
Bhardwaj et al. [89]CNNFER-2013, MES datasetStudent engagementMonitoring the student’s emotions in real time such as anger, fear, disgust, sadness, happiness, and surprise
Han et al. [90]DNNAmazonSmart education platformDesigning an intelligent educational environment
Tsai et al. [91]MLPUniversity’s institutional research databasePrecision educationTo help universities to more precisely understand student backgrounds
Fok et al. [92]CNNSelf-generated datasetPrediction model for students’ future developmentAnalyzing students’ performance and prediction of their future program of studies
Nandal et al. [93]DNNSelf-generated datasetStudent admission predictorDevelopment of student admission predictor program for students to find the chances of gaining admission to a university
Khaleel et al. [94]DCNNSelf-generated datasetAutomated gradingAutomatic grade prediction system for the students of computer-aided drawing
Table 4. State-of-the-art contributions in DL-based smart healthcare.
Table 4. State-of-the-art contributions in DL-based smart healthcare.
AuthorDL AlgorithmDatasetApplicationPurpose of DL Technique
Choi et al. [95]AE + GANSutter PAMF, IMIC-III, Sutter Heart FailureGenerating patient recordsPatient’s, record synthesis
Nie et al. [96]GANBrain data from ADNI dataset, Pelvic datasetMedical image synthesisSynthetization of CT image from MRI
Sha et al. [97]RNNMedical Information Mart for Intensive Care (MIMIC) datasetClinical outcome predictionMortality prediction
Verma et al. [98]LSTMMIT-BIH datasetMissing data prediction in healthcarePrediction of missing data in healthcare scenarios
Sun et al. [99]RBMChronic kidney disease (CKD) and dermatology datasetsClinical decision and risk predictionCapturing high-level features from the clinical data and predict missing values
Najdi et al. [100]AEISRUC-Sleep datasetSleep stage classificationDimensionality reduction, feature extraction, and classification
Nguyen et al. [101]RNNAlzheimer’s Disease Neuroimaging Initiative (ADNI) datasetAlzheimer’s disease recognitionModeling the succession of Alzheimer’s disease for seven years
Xue et al. [102]RNNElectronic medical records, Sensory data from wearablesObesity status predictionPrediction of improvement in obesity status based on blood demographics, pressure, and step count
Amin et al. [103]CNNTemple University Hospital datasetPathology detection and monitoringClassification of EEG signals into two categories, normal and pathological
Wang et al. [104]LSTMNormal Sinus Rhythm (NSR), Fantasia Database (FD),Congestive heart failureDetection of congestive heart failure
Alhussein et al. [105]CNNSVD database, MEEI databaseVoice pathology detectionClassification of voice signals into normal and pathological categories
Maragatham et al. [106]LSTMElectronic Health RecordsHeart failure predictionModeling the risk prediction of heart failure
Kim et al. [107]DBNSixth Korea National Health and Nutrition Examination Survey (KNHANES-VI) 2013 datasetCardiovascular risk predictionDevelopment of cardiovascular risk prediction model
Table 5. State-of-the-art contributions in DL-based intelligent transportation system.
Table 5. State-of-the-art contributions in DL-based intelligent transportation system.
AuthorDL AlgorithmDatasetApplicationPurpose of DL Technique
Su et al. [108]LSTM38.6 h of transportation dataMode detection systemIdentification of mode of transport based on kinetic energy harvester
Song et al. [109]LSTMGPS data and transportation network dataHuman mobility and transportation mode predictionPrediction of human movements
Mohammadi et al. [110]GANLocalization dataset, Path planning datasetPath planningSafe and reliable paths generation
Camero et al. [111]RNNData from 29 Parking slots in BirminghamCar Park occupancy predictionPrediction of occupancies rate of car parks
Singh et al. [112]AETraffic videosRoad Accident detectionExtraction of Spatio-temporal features from the surveillance video
Lv et al. [113]RNN + CNNTrajectory data from Beijing and ShanghaiTraffic speed predictionTraffic speed prediction
Ma et al. [114]RBM + RNNGPS dataCongestion Evolution Prediction in the transportation networkTraffic congestion evolution from GPS data
Pérez et al. [115]RBMFloating car data gathered in BarcelonaReal-time traffic forecastingTraffic prediction in real time
Xiangxue et al. [116]LSTMFloating Car DataShort-term traffic predictionModeling of traffic flow in urban road networks
Goudarzi et al. [117]DBNData containing historical road traffic flowTraffic flow predictionTraffic flows prediction
Table 6. State-of-the-art contributions in DL-based manufacturing industries.
Table 6. State-of-the-art contributions in DL-based manufacturing industries.
AuthorDL AlgorithmDatasetApplicationPurpose of DL Technique
Park et al. [118]AESensory data, network traffic dataIntrusion detection systemDevelopment of IDS
Tao et al. [119]CNNSensory dataWorker activity recognitionClassification of worker’s activities into 6 groups: screwdriver, used power, grab tool, hammer, rest arm, turn a screwdriver, and wrench usage
Ren et al. [120]AEIEEE PHM2012 data provided by the FEMTO-ST Institute in FranceRemaining useful life prediction of bearingsFeatures extraction that is important for the remaining bearings’ life prediction
Yan et al. [121]AEData collected from CNC machining centersRemaining useful life prediction in machinesFeatures extraction that is important for the remaining machine’s life prediction
Jiang et al. [122]AEProcess data samplesFault classificationFeature learning from a wide variety of faults
Yuan et al. [123]CNNBearing data offered by Case Western Reserve University (CWRU)Diagnosis and monitoring in manufacturingIdentification and prediction of machine faults
Li et al. [124]CNNSensory dataManufacture inspection systemClassification of production items into two categories: defected and non-defected.
Wang et al. [125]DBNSensory data gathered from a centrifugal compressorCondition predictionPrediction of machine’s condition in manufacturing systems
Zhang et al. [126]LSTMSensory data obtained from 33 sensors deployed on a pump in power stationIndustrial IoT equipment analysisPrediction of the working condition of industrial equipment to enhance operation quality
Table 7. State-of-the-art contributions in DL-based aviation industries.
Table 7. State-of-the-art contributions in DL-based aviation industries.
AuthorDL AlgorithmDatasetApplicationPurpose of DL Technique
Alkhamisi et al. [127]RNNAviation Safety Reporting System (ASRS) datasetRisk prediction in Aviation SystemsImprovements of risks prediction in aviation systems
Rodrigo et al. [128]PCMC-NetData extracted from a global distribution system (GDS)Price elasticity estimationDifferentiate the price elasticity between business and leisure trips
Barakat et al. [129]CNN + LSTMTwitter US Airline Sentiment datasetAirport service qualityMeasurement of airport service quality using passengers’ tweets about airports
Wu et al. [130]CSAESelf-generated EEG datasetDetecting fatigue status of pilotsDevelopment of fatigue recognition system based on EEG signals and DL algorithms
Dong et al. [131]LSTMAviation Safety Reporting System (ASRS)Aviation transportation safetyIdentification of incident causal factors for aviation transportation safety improvement.
Yazdi et al. [132]SAE-LM + SDAU.S flight datasetFlight delay predictionDevelopment of flight delays prediction system.
Wang et al. [133]LSTMASPM datasetsFlight demand and delays forecastingPrediction of flight departure demand in a multiple-stage time horizon
Corrado et al. [134]DAEFlight data collected from San Francisco International AirportAnomaly detectionDevelopment of an anomaly detection system to identify deviated trajectories
Hasib et al. [135]DNN + CNNUS airline service datasetSentiment analysisEvaluation of six major US airlines and multi-class sentiment analysis
Table 8. State-of-the-art contributions in DL-based defense systems.
Table 8. State-of-the-art contributions in DL-based defense systems.
AuthorDL AlgorithmDatasetApplicationPurpose of DL Technique
Das et al. [136]R-CNNSelf-generated datasetTarget detectionDevelopment of a new search algorithm for object detection through UAV
Calderón et al. [137]CNNSelf-generated datasetReal-time object detectionDevelopment of a vision-based object detection system for a micro-UAV
Krishnaveni et al. [138]DCNNData collected from wildlife television.Surveillance applicationsIdentification of abnormal events and the data streaming by creating a multipath routing in WSN
Pradeep et al. [139]CNNSelf-generated datasetReal-time object recognition in air defense systemsAccurate identification of definite target with DL algorithm and real-time camera of FWN aircraft
Shi et al. [140]FNNSelf-generatedCognitive radio securityLaunching of jamming attacks on wireless communications and development of a defense strategy
Wang et al. [141]DRLSelf-generated datasetDefense strategies against adversarial jamming attacksDesign and development of defense strategies against DRL-based jamming attackers on a multichannel access agent
Table 9. State-of-the-art contributions in DL-based sports industry.
Table 9. State-of-the-art contributions in DL-based sports industry.
AuthorDL AlgorithmDatasetApplicationPurpose of DL Technique
Chen et al. [142]GANNBA SportVuBasketballDevelopment of realistic defensive plays conditioned on the ball and offensive term movements
Chung et al. [143]GANSTATS SportVuBasketballSimulation of offensive tactic sketched by coaches
Baccouche et al. [144]LSTM + RNNMICC-Soccer-Actions-4 datasetFootballClassifying four football actions
Theagarajan et al. [145]CNN3 different soccer matchesFootballGeneration of sports highlights
Le et al. [146]RNNSTATSFootballGhost modeling in football.
Kautz et al. [147]DCNNVideo Recordings from GoPro Hero 3 action cameraVolleyballActivity recognition in volleyball
Qiao et al. [148]DCNN + LSTMSelf-built video datasetTable TennisRecognition and tracking of table tennis’s real-time trajectories in complex environments
Cao et al. [149]Tiny YOLOv2Self-generated shuttlecock detection datasetBadmintonPrecise and detection of the shuttlecock with badminton robot
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Latif, S.; Driss, M.; Boulila, W.; Huma, Z.e.; Jamal, S.S.; Idrees, Z.; Ahmad, J. Deep Learning for the Industrial Internet of Things (IIoT): A Comprehensive Survey of Techniques, Implementation Frameworks, Potential Applications, and Future Directions. Sensors 2021, 21, 7518. https://doi.org/10.3390/s21227518

AMA Style

Latif S, Driss M, Boulila W, Huma Ze, Jamal SS, Idrees Z, Ahmad J. Deep Learning for the Industrial Internet of Things (IIoT): A Comprehensive Survey of Techniques, Implementation Frameworks, Potential Applications, and Future Directions. Sensors. 2021; 21(22):7518. https://doi.org/10.3390/s21227518

Chicago/Turabian Style

Latif, Shahid, Maha Driss, Wadii Boulila, Zil e Huma, Sajjad Shaukat Jamal, Zeba Idrees, and Jawad Ahmad. 2021. "Deep Learning for the Industrial Internet of Things (IIoT): A Comprehensive Survey of Techniques, Implementation Frameworks, Potential Applications, and Future Directions" Sensors 21, no. 22: 7518. https://doi.org/10.3390/s21227518

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop