Next Article in Journal
Effect of Mother’s Age and Pathology on Functional Behavior of Amniotic Mesenchymal Stromal Cells—Hints for Bone Regeneration
Previous Article in Journal
Identification of Knee Cartilage Changing Pattern
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Bidirectional Long Short-Term Memory Neural Networks for Linear Sum Assignment Problems

Department of Electronic Engineering, Myongji University, Yongin 17058, Korea
*
Author to whom correspondence should be addressed.
Appl. Sci. 2019, 9(17), 3470; https://doi.org/10.3390/app9173470
Submission received: 5 June 2019 / Revised: 31 July 2019 / Accepted: 19 August 2019 / Published: 22 August 2019
(This article belongs to the Section Electrical, Electronics and Communications Engineering)

Abstract

:
Many resource allocation problems can be modeled as a linear sum assignment problem (LSAP) in wireless communications. Deep learning techniques such as the fully-connected neural network and convolutional neural network have been used to solve the LSAP. We herein propose a new deep learning model based on the bidirectional long short-term memory (BDLSTM) structure for the LSAP. In the proposed method, the LSAP is divided into sequential sub-assignment problems, and BDLSTM extracts the features from sequential data. Simulation results indicate that the proposed BDLSTM is more memory efficient and achieves a higher accuracy than conventional techniques.

1. Introduction

The linear sum assignment problem (LSAP) is a special case of the linear programming problem that aims to minimize the cost of assigning multiple tasks to a number of agents on a one-to-one basis. In wireless communications, many resource allocation tasks can be modeled as the LSAP, such as joint relay selection and resource allocation for device-to-device communications [1], joint resource allocation with multi-cell cooperation in virtual multiple-input multiple-output (MIMO) systems [2], and unlicensed channel allocation for LTE systems [3].
For LSAPs, the Hungarian algorithm has been proposed to obtain the optimal solution without an exhaustive search [4]. Furthermore, heuristic algorithms such as deep greedy switching [5], the interior point [6], or the dual forest algorithm [7] have been investigated to estimate near-optimal solutions in real-world problems with time constraints.
In recent years, deep neural networks (DNNs) have recently achieved state-of-the-art performance in different fields such as audio processing [8,9], visual object recognition [10], and other domains [11,12,13]. Neural networks (NNs) can extract features from input implicitly and approximate an arbitrary math function. Several approaches have applied DNNs to solve mathematical optimization problems [14,15]. In [2], a DNN was used to approximate an optimization algorithm in interference management. In [16], an LSAP was decomposed into several sub-assignment problems and two types of DNNs, i.e., the feed-forward neural network (FNN) and convolutional neural network (CNN) [17] were applied to address sub-assignment problems. However, the DNNs discussed in [16] have not achieved much success owing to the increasing size of the problem.
Hence, a bidirectional long short-term memory neural network (BDLSTM) is proposed to solve LSAPs. In the proposed method, an LSAP is decomposed into a sequence of smaller sub-assignment tasks that become the model input, in which each layer consists of two separated smaller LSTM layers. The intermediate outputs of these layers are combined to process data in both the forward and backward directions. The BDLSTM structure is better at remembering the LSAP than other structures by combining the hidden layer outputs of each sub-assignment problem together. Finally, a collision-avoidance algorithm is applied to the output to obtain the final result. Simulation results indicate that the proposed BDLSTM outperforms the FNN and CNN models when the size of the problem increases, while requiring less parameters to model the patterns.

2. Preliminary

2.1. Problem Formulation

In LSAPs [4], given a set of n jobs I = { 1 , , n } and n agents J = { 1 , , n } , we wish to minimize the cost of assigning each job to each agent. Assigning job i to agent j results in a cost value of U i j , and the problem can be formulated as follows:
min i = 1 n j = 1 n u i j x i j ,
subject to:
i = 1 n x i j = 1 , j = 1 , , n ,
j = 1 n x i j = 1 , i = 1 , , n ,
x i j { 0 , 1 } , i , j = 1 , , n ,
where x i j is the decision indicator value. x i j = 1 when job i is assigned to agent j; otherwise, x i j = 0 . We can rewrite the cost values as an n by n matrix U = [ u 1 , , u n ] T and decision indicator values as an n by n permutation matrix X = [ x 1 , , x n ] T , where u i = [ u i 1 , , u i n ] T and x i = [ x i 1 , , x i n ] T . In the communication field, many optimization problems are similar or can be formulated as LSAP [1,2,3].
In [16], the LSAP problem was firstly decomposed into multiple sub-assignment problems to implement DNNs. However, the DNNs in [16] did not address the sequential information in the cost matrix, as multiple FNN and CNN models must be applied to solve all sub-assignment problems simultaneously. Here, a single neural network model is considered to extract the necessary features to solve LSAPs.

2.2. Bidirectional LSTM

The structure of a conventional LSTM model is illustrated in Figure 1. Here, we provide a short introduction to the LSTM model. For further details, we refer the interested reader to [18]. LSTM [18] was designed with special memory cells to store temporal information. This structure allows LSTM to remember long-range features better than conventional recurrent neural networks. With a multilayer model, components of a cell at time step i at layer l in the forward direction can be implemented by the following functions:
f i l = σ W ( f ) l h i l 1 + V ( f ) l h i 1 l + b ( f ) l ,
i i l = σ W ( i ) l h i l 1 + V ( i ) l h i 1 l + b ( i ) l ,
o i l = σ W ( o ) l h i l 1 + V ( o ) l h i 1 l + b ( o ) l ,
g i l = tanh W ( g ) l h i l 1 + V ( g ) l h i 1 l + b ( g ) l ,
C i l = f i l C i 1 l + i i l g i l ,
h i l = o i l tanh C i l ,
where i i l , f i l , o i l , g i l , C i l , and h i l are the input gate, forget gate, output gate, candidate gate, cell state, and hidden state, respectively, all of which are of the size of an N l -dimensional vector. In (6)–(9), the W l terms are the weight matrices between cells of layer ( l 1 ) l, the V l terms are the weight matrices between consecutive cells of layer l, and the b l terms are the bias vector at each layer. The weight matrices and bias values in a cell are shared along the length of the sequence, thus reducing the total number of weights and hidden neurons in the network. The sigmoid function σ and hyperbolic tangent function are used as activation functions, and ⊙ denote element-wise multiplication.
A bidirectional LSTM can process the data in both the forward and backward directions using two separate LSTM layers. The forward hidden state, h i l , calculated by the formulas above, and the backward state, h i l , calculated similarly, are concatenated and then fed forward to the next layer, as demonstrated in Figure 2:
h i l = h i l h i l ,
where l = 0 is the input layer. BDLSTM is better at obtaining the relations among elements in a whole sequence by utilizing information in both directions, instead of remembering the features in only one direction such as that of the conventional LSTM. Furthermore, by using the parameter-sharing technique in a layer, the BDLSTM model requires less memory to solve the problem compared to the conventional FNN and CNN.

3. Proposed Method

This section presents the proposed BDLSTM model for LSAPs. The architecture and details of the proposed model are elaborated in Section 3.1. In Section 3.2, the training process is explained.

3.1. System Model

Figure 3 shows the proposed system architecture for solving an LSAP. A deep-learning-based BDLSTM model is first applied to the cost matrix of the LSAP. After that, a collision-avoidance algorithm is applied to solve any interference in the outputs of the proposed BDLSTM model. In the proposed BDLSTM model, the constraint in (3) is not taken into consideration, which guarantees that one job can be assigned to only one agent. This is because the proposed BDLSTM model solves each sub-assignment task as a separate classification problem. The collision-avoidance algorithm prevents one job from being assigned to multiple agents simultaneously.
As described in Figure 4, in the proposed BDLSTM model, the cost matrix U is divided into n sub-assignment tasks u i , where i = 1 , , n and u i is the cost vector for job i. The u i are used as the input sequence of the BDLSTM model as h i 0 = h i 0 = u i . After L consecutive BDLSTM layers to extract the features, all hidden states at the last layer h i L will be applied in an identical fully-connected layer with the softmax activation function to obtain the output matrix Y = [ y 1 , , y n ] T :
y i = softmax ( W y h i l + b y ) ,
where W y and b y are the weight matrix and bias of the output layer, respectively. After the BDLSTM model, a decision matrix can be obtained by Y ¯ = y ¯ 1 , , y ¯ n T , where y ¯ i is obtained by one-hot encoding of y i . In collision avoidance, the collision cases in Y ¯ are corrected using the Hungarian algorithm by the following steps:
  • Step 1: Find jobs that are assigned to multiple or no agents.
  • Step 2: Extract the cost values of jobs from Step 1 as a small LSAP.
  • Step 3: Solve the LSAP from Step 2 using the Hungarian algorithm.
  • Step 4: Y ¯ is modified into a decision matrix Z using Step 3, where Z is an n by n permutation matrix.
For sequential data, FNNs, CNNs, and especially recurrent structures such as LSTM and BDLSTM are widely used and have achieved many successes. FNNs can be applied to every problem as the baseline architectures. CNNs are more complicated and have good performance in computer vision and some natural language processing tasks. On the other hand, LSTMs and BDLSTMs are recurrent structures built specifically for problems relating to sequential data. In the proposed method, BDLSTM is used to extract the features from the cost matrix C , and each classification is based on the information of the whole cost matrix. Compared to the FNN and CNN in [16], BDLSTM has memory cells, which can remember the relationship between the cost vectors of consecutive sub-assignment tasks.

3.2. Training Stage

The training dataset consists of M data entries, each of which is a pair of cost and decision matrices { U , X } . First, we generate the cost matrix U , where each cost value u i j is a real value and is generated from a uniform distribution on (0, 1]. Then, the optimal decision matrix X according to U is obtained for the training dataset using the Hungarian algorithm [4]. After repeating this step M times, we can generate the whole dataset. For this experiment, we used M = 500 , 000 samples. Among these, 10% of this dataset was sampled randomly for validation at the end of each training epoch to select the hyperparameters while preventing the overfitting problem. After completing the training step, 50,000 unseen samples were used as the test dataset to verify the performance of the trained model.
In the training process, the mini-batch gradient descent procedure was applied to optimize the parameters of the models to minimize the following total loss:
L o s s = 1 | B | m B L ( m ) ,
where B is a minibatch sampled from the dataset with size | B | and L ( m ) is the loss computed from samples m B . For this classification task, we chose the cross-entropy as the loss function, which can be formulated as follows:
L ( m ) = i = 1 n x i ( m ) log y i ( m ) .
The weights were updated based on the gradient information of the loss function. We chose the Adam algorithm [19] as the gradient descent method as it required only the first-order gradient to be computed, thus reducing the calculation complexity.

4. Performance Evaluation

In this section, we evaluate the simulation results of the proposed scheme using a BDLSTM model with different configurations for the LSAP. To obtain the optimized hyper-parameters, such as batch size, number of epochs, and learning rate, the proposed model was trained using several combinations of parameters using grid search, and the best combination was selected. The BDLSTM model with L = 4 layers was chosen, and each layer contained N l = 16 hidden units at each gate.
Figure 5 shows the convergence of the BDLSTM over each training epoch on the training and validation set with n = 16 and a batch size | B | = 1024 . The loss of the model on both datasets decreased steadily after each epoch, and the gap was minimal, suggesting no overfitting during training.
Table 1 shows the accuracy performance comparisons between the proposed BDLSTM model and other neural network structures, where the FNN and CNN were based on [16], and a uni-directional LSTM model was used with the same number of layers and hidden units for the proposed model. The FNN consisted of n smaller models, each having four layers with 32, 64, 256, and n hidden neurons. The CNN consisted of five convolutional layers, each containing 32, 32, 32, 32, and n kernels, and n output layers, each with n neurons. The accuracy was defined as the number of jobs correctly assigned to their optimal agents divided by the total number of jobs according to decision matrices X in the dataset. We can see that the proposed BDLSTM performed better than LSTM, FNN, and CNN for all values of n. Because of the bidirectional structure, the proposed BDLSTM exhibited performance improvements between 18% ( n = 16 ) and 27% ( n = 8 ) compared with the conventional LSTM. The performance gap became more significant as n, the size of the LSAP, increased.
Figure 6 shows the complexity comparison of deep neural networks with a size n for the LSAP. Here, we focused on the number of trainable parameters to determine the complexity of the network, where the trainable parameters were the weight and bias values. The proposed BDLSTM was robust with respect to the complexity owing to the change in n compared to FNN and CNN. This was because the FNN model required n different fully-connected layers for n sub-assignment problems, and the sizes of the final layer and output layers of CNN scaled according to n. Because the weights in a layer of the BDLSTM were shared among all sub-assignment problems, the number of trainable parameters was stable to the change in n, resulting in low memory requirements. When n = 16 , the BDLSTM achieved the best result while requiring 18- and seven-times fewer parameters than FNN and CNN, respectively.
Following the proposed BDLSTM model, the collision-avoidance algorithm was applied to obtain the final output. Table 2 shows the performance and operating time after collision avoidance with n = 8 , where all the tests were conducted on the same system with CPU E5-2620 v4, having eight cores at 2.10 GHz, and 32 GB of RAM. Table 2 shows the accuracy performances and operating time after collision avoidance with n = 8 , where the accuracy after the collision avoidance could be obtained using X and Z . As shown in Table 2, by achieving the highest classification accuracy, the proposed method reduced the percentage of collisions that occurred in the prediction result, hence limiting the usage of the Hungarian algorithm to solve the collision cases.

5. Conclusions

We herein proposed a BDLSTM technique to solve LSAPs. In the proposed method, the LSAP was decomposed into sub-assignment problems for classification. The BDLSTM addressed a series of sub-assignment problems, and collision avoidance was performed. The proposed BDLSTM was trained by connecting the information among the sub-assignment problems. Simulation results indicated that the proposed method achieved a higher accuracy compared to the FNN and CNN while requiring less memory for trainable parameters. The proposed method exhibited great potential for solving the LSAP as the model scaled better with the complexity of the problem, requiring less memory for the parameters compared to conventional DNN structures.

Author Contributions

Y.-H.K. conceived of the presented idea. N.M.-T. developed the model and performed the computation. All authors discussed the results and contributed to the final manuscript.

Funding

This research was supported in part by Korea Electric Power Corporation (Grant Number R18XA01) and in part by the Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Science and ICT (NRF-2017R1C1B1012259).

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Kim, T.; Dong, M. An Iterative Hungarian Method to Joint Relay Selection and Resource Allocation for D2D Communications. IEEE Wirel. Commun. Lett. 2014, 3, 625–628. [Google Scholar] [CrossRef]
  2. Sun, H.; Chen, X.; Shi, Q.; Hong, M.; Fu, X.; Sidiropoulos, N.D. Learning to Optimize: Training Deep Neural Networks for Interference Management. IEEE Trans. Signal Process. 2018, 66, 5438–5453. [Google Scholar] [CrossRef]
  3. Song, H.; Fang, X.; Fang, Y. Unlicensed Spectra Fusion and Interference Coordination for LTE Systems. IEEE Trans. Mob. Comput. 2016, 15, 3171–3184. [Google Scholar] [CrossRef]
  4. Kuhn, H.W. The Hungarian method for the assignment problem. Naval Res. Logist. Q. 1955, 2, 83–97. [Google Scholar] [CrossRef] [Green Version]
  5. Naiem, A.; El-Beltagy, M.; Ab, P. Deep greedy switching: A fast and simple approach for linear assignment problems. In Proceedings of the 7th International Conference of Numerical Analysis and Applied Mathematics, Thessaloniki, Greece, 25–30 September 2017; pp. 1–6. [Google Scholar]
  6. Karmarkar, N.K.; Ramakrishnan, K.G. Computational results of an interior point algorithm for large scale linear programming. Math. Programm. 1991, 52, 555–586. [Google Scholar] [CrossRef]
  7. Akgül, M.; Ekin, O. A dual feasible forest algorithm for the linear assignment problem. RAIRO Oper. Res. 1991, 25, 403–411. [Google Scholar] [CrossRef] [Green Version]
  8. Lee, J.; Kim, K.; Shabestary, T.; Kang, H. Deep bi-directional long short-term memory based speech enhancement for wind noise reduction. In Proceedings of the 2017 Hands-free Speech Communications and Microphone Arrays (HSCMA), San Francisco, CA, USA, 1–3 March 2017; pp. 41–45. [Google Scholar] [CrossRef]
  9. Graves, A.; Jaitly, N.; Mohamed, A. Hybrid speech recognition with Deep Bidirectional LSTM. In Proceedings of the 2013 IEEE Workshop on Automatic Speech Recognition and Understanding, Olomouc, Czech Republic, 8–12 December 2013; pp. 273–278. [Google Scholar] [CrossRef]
  10. Zhao, Z.Q.; Zheng, P.; Xu, S.T.; Wu, X. Object Detection with Deep Learning: A Review. arXiv 2018, arXiv:1807.05511. [Google Scholar] [CrossRef] [PubMed]
  11. Nguyen, M.T.; Nguyen, V.H.; Yun, S.J.; Kim, Y.H. Recurrent Neural Network for Partial Discharge Diagnosis in Gas-Insulated Switchgear. Energies 2018, 11, 1202. [Google Scholar] [CrossRef]
  12. Nguyen, V.H.; Nguyen, M.T.; Choi, J.; Kim, Y.H. NLOS Identification in WLANs Using Deep LSTM with CNN Features. Sensors 2018, 18, 4057. [Google Scholar] [CrossRef] [PubMed]
  13. Boutaba, R.; Salahuddin, M.A.; Limam, N.; Ayoubi, S.; Shahriar, N.; Estrada-Solano, F.; Caicedo, O.M. A comprehensive survey on machine learning for networking: evolution, applications and research opportunities. J. Internet Serv. Appl. 2018, 9, 16. [Google Scholar] [CrossRef]
  14. Bennett, K.P.; Parrado-Hernández, E. The Interplay of Optimization and Machine Learning Research. J. Mach. Learn. Res. 2006, 7, 1265–1281. [Google Scholar]
  15. Gambella, C.; Ghaddar, B.; Naoum-Sawaya, J. Optimization Models for Machine Learning: A Survey. arXiv 2019, arXiv:1901.05331. [Google Scholar]
  16. Lee, M.; Xiong, Y.; Yu, G.; Li, G.Y. Deep Neural Networks for Linear Sum Assignment Problems. IEEE Wirel. Commun. Lett. 2018, 7, 962–965. [Google Scholar] [CrossRef]
  17. LeCun, Y.; Bengio, Y.; Hinton, G. Deep learning. Nature 2015, 521, 436–444. [Google Scholar] [CrossRef] [PubMed]
  18. Hochreiter, S.; Schmidhuber, J. Long Short-Term Memory. Neural Comput. 1997, 9, 1735–1780. [Google Scholar] [CrossRef] [PubMed]
  19. Kingma, D.P.; Ba, J. Adam: A Method for Stochastic Optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
Figure 1. Detailed structure within an LSTM cell in the forward direction.
Figure 1. Detailed structure within an LSTM cell in the forward direction.
Applsci 09 03470 g001
Figure 2. BDLSTM layer with both forward and backward LSTM layers. LSAP, linear sum assignment problem.
Figure 2. BDLSTM layer with both forward and backward LSTM layers. LSAP, linear sum assignment problem.
Applsci 09 03470 g002
Figure 3. Proposed system model.
Figure 3. Proposed system model.
Applsci 09 03470 g003
Figure 4. Proposed BDLSTM model.
Figure 4. Proposed BDLSTM model.
Applsci 09 03470 g004
Figure 5. Learning curve of the BDLSTM model with n = 16 .
Figure 5. Learning curve of the BDLSTM model with n = 16 .
Applsci 09 03470 g005
Figure 6. Complexity comparison.
Figure 6. Complexity comparison.
Applsci 09 03470 g006
Table 1. Accuracy performance comparison.
Table 1. Accuracy performance comparison.
n = 4 n = 8 n = 12 n = 16
FNN [16]97.6370.1959.1856.14
CNN [16]98.2982.9566.0562.74
LSTM77.1567.8264.2060.35
BDLSTM98.8594.4786.5478.82
Table 2. Accuracy performance before and after collision avoidance with n = 8 .
Table 2. Accuracy performance before and after collision avoidance with n = 8 .
BeforeAfterOperating Time
FNN [16]70.19%82.11%23.06 s
CNN [16]82.95%86.65%14.85 s
BDLSTM94.47%96.08%6.5 s

Share and Cite

MDPI and ACS Style

Minh-Tuan, N.; Kim, Y.-H. Bidirectional Long Short-Term Memory Neural Networks for Linear Sum Assignment Problems. Appl. Sci. 2019, 9, 3470. https://doi.org/10.3390/app9173470

AMA Style

Minh-Tuan N, Kim Y-H. Bidirectional Long Short-Term Memory Neural Networks for Linear Sum Assignment Problems. Applied Sciences. 2019; 9(17):3470. https://doi.org/10.3390/app9173470

Chicago/Turabian Style

Minh-Tuan, Nguyen, and Yong-Hwa Kim. 2019. "Bidirectional Long Short-Term Memory Neural Networks for Linear Sum Assignment Problems" Applied Sciences 9, no. 17: 3470. https://doi.org/10.3390/app9173470

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop