The Study of Intelligent Image Classification Systems: An Exploration of Generative Adversarial Networks with Texture Information on Coastal Driftwood
Abstract
:1. Introduction
2. Data Collection and Methods
2.1. Data
2.2. Methods
- Generative: It learns a generative model to describes how data are generated considering a probabilistic model.
- Adversarial: The training data of a model are made by an adversarial setting.
- Networks: Use deep neural networks as artificial intelligence (AI) algorithms for training progress.
- (a)
- The Generator Model
- (b)
- The Discriminator Model
- (c)
- Generation of training samples and testing samples
- (d)
- Generate the confusion matrix and thematic map for verification
3. Results
3.1. The Post-Progress for GAN
- (1)
- Dropout in discriminator: During training, we applied dropout to the discriminator’s hidden layers. For example, you can set a dropout rate of 0.5, meaning 50% of the neurons are randomly deactivated during each forward and backward pass.
- (2)
- Training the discriminator: We sampled a batch of real data from the dataset and generated a batch of fake data using the generator. Then, we trained the discriminator with the real and fake data batches: We computed the discriminator loss based on how well it distinguishes real from fake data. Then, we updated the discriminator’s weights and biases using backpropagation.
- (3)
- Dropout in generator: During training, we applied dropout to the generator’s hidden layers (if necessary). See Figure 5.
- (4)
- Training the generator: We generated a new batch of fake data using the generator. Then, we trained the generator to “fool” the discriminator. We computed the generator loss based on the discriminator’s response to the generated data. Afterwards, we updated the generator’s weights and biases using backpropagation.
- (i)
- Alternating training: Step 2 and Step 4 are used to exchange information between each other.
- (ii)
- Loss function: We used appropriate loss functions, such as binary cross-entropy, to guide the training of both the discriminator and the generator.
- (iii)
- Hyperparameter tuning: Usually, we experiment with different learning rates, batch sizes, dropout rates, and architectures to optimize GAN training.
- (iv)
- Stopping criteria: We decided on a stopping criterion based on the performance of the discriminator and generator. The latter will discuss the outcomes of accuracy.
3.2. Epoch for Accuracy
3.3. Confusion Matrix
3.4. Prediction for Thematic Map
4. Conclusions
- (a)
- Original RGB + IR image has about 70% overall classification outcomes. All the driftwood accuracy is misclassified. But a big amount of rock was misclassified, reducing the producer accuracy.
- (b)
- The texture gray-based image has about 78% overall classification outcomes. But a large amount of grass has been misclassified to the tree, reducing the producer accuracy. The driftwood has about 79.5% producer accuracy.
Author Contributions
Funding
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Wan, S.; Lei, T.C. A development of a robust machine for removing Irregular noise with the intelligent system of auto-encoder for image classification of coastal waste. Environments 2022, 9, 114. [Google Scholar] [CrossRef]
- Kwak, G.H.; Park, N.W. Impact of texture information on crop classification with machine learning and UAV images. Appl. Sci. 2019, 9, 643. [Google Scholar] [CrossRef]
- Goodfellow, I.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Bengio, Y. Generative adversarial networks. Commun. ACM 2020, 63, 139–144. [Google Scholar] [CrossRef]
- Gui, J.; Sun, Z.; Wen, Y.; Tao, D.; Ye, J. A Review on Generative Adversarial Networks: Algorithms, Theory, and Applications. IEEE Trans. Knowl. Data Eng. 2023, 35, 3313–3332. [Google Scholar] [CrossRef]
- Radford, A.; Metz, L.; Chintala, S. Unsupervised representation learning with deep convolutional generative adversarial networks. arXiv 2015, arXiv:1511.06434. [Google Scholar]
- Luc, P.; Couprie, C.; Chintala, S.; Verbeek, J. Semantic segmentation using adversarial networks. arXiv 2016, arXiv:1611.08408. [Google Scholar]
- Chen, Y.; Lin, Z.; Zhao, X.; Wang, G.; Gu, Y. Deep learning-based classification of hyperspectral data. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2014, 7, 2094–2107. [Google Scholar] [CrossRef]
- Wan, S. A spatial decision support system for extracting the core factors and thresholds for landslide susceptibility map. Eng. Geol. 2009, 108, 237–251. [Google Scholar] [CrossRef]
- Lu, X.; Matsuda, S.; Hori, C.; Kashioka, H. Speech restoration based on deep learning Auto-encoder with layer-wised pretraining. In Proceedings of the Thirteenth Annual Conference of the International Speech Communication Association, Portland, OR, USA, 9–13 September 2012. [Google Scholar]
- Kudo, Y.; Aoki, Y. Dilated convolutions for image classification and object localization. In Proceedings of the 2017 Fifteenth IAPR International Conference on Machine Vision Applications (MVA), Nagoya, Japan, 8–12 May 2017; pp. 452–455. [Google Scholar]
- Chang, S.H.; Wan, S. A novel study on ant-based clustering for paddy rice image classification. Arab. J. Geosci. 2015, 8, 6305–6316. [Google Scholar] [CrossRef]
- Pham, D.T.; Sagiroglu, S. Training multilayered perceptrons for pattern recognition: A comparative study of four training algorithms. Int. J. Mach. Tools Manuf. 2001, 41, 419–430. [Google Scholar] [CrossRef]
- Pang, Y.; Sun, M.; Jiang, X. Convolution in convolution for network in network. IEEE Trans. Neural Netw. Learn. Syst. 2018, 29, 1587–1597. [Google Scholar] [CrossRef]
- Odena, A.; Olah, C.; Shlens, J. Conditional image synthesis with auxiliary classifier gans. In Proceedings of the 34th International Conference on Machine Learning, Sydney, Australia, 6–11 August 2017; Volume 70, pp. 2642–2651. [Google Scholar]
- Available online: https://docs.opencv.org/3.4/de/d25/imgproc_color_conversions.html (accessed on 11 June 2021).
- Armanious, K.; Hepp, T.; Küstner, T.; Dittmann, H.; Nikolaou, K.; La Fougère, C.; Yang, B.; Gatidis, S. Independent attenuation correction of whole body [18F] FDG-PET using a deep learning approach with Generative Adversarial Networks. EJNMMI Res. 2020, 10, 53. [Google Scholar] [CrossRef]
- Mao, X.; Li, Q.; Xie, H. Aligngan: Learning to align cross-domain images with conditional generative adversarial networks. arXiv 2017, arXiv:1707.01400. [Google Scholar]
- Schlegl, T.; Seeböck, P.; Waldstein, S.M.; Schmidt-Erfurth, U.; Langs, G. Unsupervised anomaly detection with generative adversarial networks to guide marker discovery. In Proceedings of the International Conference on Information Processing in Medical Imaging, Boone, NC, USA, 25–30 June 2017; Springer: Berlin/Heidelberg, Germany, 2017; pp. 146–157. [Google Scholar]
- Hsu, C.C.; Lee, C.Y.; Zhuang, Y.X. Learning to Detect Fake Face Images in the Wild. In Proceedings of the 2018 International Symposium on Computer, Consumer and Control (IS3C), Taichung, Taiwan, 6–8 December 2018; IEEE: New York City, NY, USA, 2018; pp. 388–391. [Google Scholar]
- Wu, Q.; Wang, P.; Shen, C.; Reid, I.; van den Hengel, A. Are you talking to me? Reasoned visual dialog generation through adversarial learning. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 6106–6115. [Google Scholar]
- Joo, D.; Kim, D.; Kim, J. Generating a Fusion Image: One’s Identity and Another’s Shape. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA, 18–23 June 2018; pp. 1635–1643. [Google Scholar]
- Reed, S.E.; Akata, Z.; Mohan, S.; Tenka, S.; Schiele, B.; Lee, H. Learning what and where to draw. In Advances in Neural Information Processing Systems; MIT Press: Cambridge, MA, USA, 2016; pp. 217–225. [Google Scholar]
- Karpathy Atoderici, G. Large-scale video classification with convolutional neural networks. In Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, Nanjing, China, 28–30 July 2014; pp. 1725–1732. [Google Scholar] [CrossRef]
- Krizhevsky Ahinton, G. Neural ImageNet Classification with Deep Convolutional Neural Networks. Inf. Process. Syst. 2012, 25, 1–9. [Google Scholar] [CrossRef]
- Gu, J.; Wang, Z.; Kuen, J.; Ma, L.; Shahroudy, A.; Shuai, B.; Liu, T.; Wang, X.; Wang, G.; Cai, J.; et al. Recent advances in convolutional neural networks. Pattern Recognit. 2018, 77, 354–377. [Google Scholar] [CrossRef]
- Badrinarayanan, V.; Kendall, A.; Cipolla, R. SegNet: A Deep Convolutional Encoder-Decoder Architecture for Image Segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 2017, 39, 2481–2495. [Google Scholar] [CrossRef] [PubMed]
- Lange, S.; Riedmiller, M. Deep auto-encoder neural networks in reinforcement learning. In Proceedings of the International Joint Conference on Neural Networks, Barcelona, Spain, 18–23 July 2010. [Google Scholar] [CrossRef]
- Espindola, G.M.; Camara, G.; Reis, I.A.; Bins, L.S.; Monteiro, A.M. Parameter selection for region-growing image segmentation algorithms using spatial autocorrelation. Int. J. Remote Sens. 2006, 27, 3035–3040. [Google Scholar] [CrossRef]
- Du, P.; Tan, K.; Xing, X. A Novel Binary Tree Support Vector Machine for Hyperspectral Remote Sensing Image Classification. Opt. Commun. 2012, 285, 3054–3060. [Google Scholar] [CrossRef]
- Chapelle, O.; Haffner, P.; Vapnik, V. Support vector machines for histogram-based image classification. IEEE Trans. Neural Netw. 1999, 10, 1055–1064. [Google Scholar] [CrossRef] [PubMed]
- Furey, T.; Cristianini, N.; Duffy, N.; Bednarski, D.; Schummer, M.; Haussler, D. Support vector machine classification and validation of cancer tissue samples using microarray expression data. Bioinformatics 2000, 16, 906–914. [Google Scholar] [CrossRef] [PubMed]
Texture Indices | Formula |
---|---|
Homogeneity | |
Contrast | |
Dissimilarity | |
Entropy |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Yeh, M.-L.; Wan, S.; Ma, H.-L.; Chou, T.-Y. The Study of Intelligent Image Classification Systems: An Exploration of Generative Adversarial Networks with Texture Information on Coastal Driftwood. Environments 2023, 10, 167. https://doi.org/10.3390/environments10100167
Yeh M-L, Wan S, Ma H-L, Chou T-Y. The Study of Intelligent Image Classification Systems: An Exploration of Generative Adversarial Networks with Texture Information on Coastal Driftwood. Environments. 2023; 10(10):167. https://doi.org/10.3390/environments10100167
Chicago/Turabian StyleYeh, Mei-Ling, Shiuan Wan, Hong-Lin Ma, and Tien-Yin Chou. 2023. "The Study of Intelligent Image Classification Systems: An Exploration of Generative Adversarial Networks with Texture Information on Coastal Driftwood" Environments 10, no. 10: 167. https://doi.org/10.3390/environments10100167
APA StyleYeh, M. -L., Wan, S., Ma, H. -L., & Chou, T. -Y. (2023). The Study of Intelligent Image Classification Systems: An Exploration of Generative Adversarial Networks with Texture Information on Coastal Driftwood. Environments, 10(10), 167. https://doi.org/10.3390/environments10100167