Research on Augmentation of Wood Microscopic Image Dataset Based on Generative Adversarial Networks
Abstract
1. Introduction
- (1)
- We provide the first systematic, cross-architecture evaluation of multiple GAN models for microscopic wood imagery under extreme limited-data conditions, with a specific focus on softwood species. This distinguishes our work from prior studies like Lopes et al. [11], which focused on hardwood species, and addresses the distinct anatomical characteristics and associated generation challenges of softwoods.
- (2)
- We introduce a domain-adapted evaluation framework that incorporates woody anatomical characteristics via KID using DenseNet wood embeddings, LPIPS diversity, and Density–Coverage metrics—evaluation dimensions not addressed in Lopes et al. [11].
- (3)
- We analyze model robustness, distributional coverage, and failure modes specific to microscopic textures, offering practical insights for real-world small-sample wood identification augmentation.
2. Materials and Methods
2.1. Wood Microscopic Image Dataset
2.2. Dataset Augmentation Based on Basic Image Processing Methods
2.3. Dataset Augmentation Based on Generative Adversarial Networks
2.3.1. BGAN
2.3.2. DCGAN
2.3.3. WGAN-GP
2.3.4. LSGAN
2.3.5. StyleGAN2
2.4. Quality Assessment Indicators
2.4.1. KID
2.4.2. IS
2.4.3. SSIM
3. Results and Discussion
3.1. Data Augmentation Using Basic Image Processing Methods
3.2. Ablation Study on Data Augmentation Strategies
3.3. Data Augmentation Based on GAN
3.3.1. Overall Training Dynamics

3.3.2. Model-Specific Performance Analysis
3.3.3. Qualitative Evaluation
3.3.4. Metric Behavior and Limitations
3.3.5. Impact of Different Dataset Sizes on Model Training
4. Conclusions
Author Contributions
Funding
Institutional Review Board Statement
Informed Consent Statement
Data Availability Statement
Conflicts of Interest
References
- FAO Global Forest Products Facts and Figures 2023; FAO: Rome, Italy, 2024; pp. 1–2.
- Ravindran, P.; Costa, A.; Soares, R.; Wiedenhoeft, A. Classification of CITES-Listed and Other Neotropical Meliaceae Wood Images Using Convolutional Neural Networks. Plant Methods 2018, 14, 25. [Google Scholar] [CrossRef]
- He, T.; Mu, S.; Zhou, H.; Hu, J. Wood Species Identification Based on an Ensemble of Deep Convolution Neural Networks. Wood Res. 2021, 66, 1–13. [Google Scholar] [CrossRef] [PubMed]
- Chen, Y.; Sun, C.; Ren, Z.; Na, B. Review of the Current State of Application of Wood Defect Recognition Technology. Bioresources 2023, 18, 49. [Google Scholar] [CrossRef]
- Hwang, S.-W.; Sugiyama, J. Computer Vision-Based Wood Identification and Its Expansion and Contribution Potentials in Wood Science: A Review. Plant Methods 2021, 17, 47. [Google Scholar] [CrossRef]
- Dhanya, V.; Subeesh, A.; Kushwaha, N.; Vishwakarma, D.; Kumar, T.; Ritika, G.; Singh, A. Deep Learning Based Computer Vision Approaches for Smart Agricultural Applications. Artif. Intell. Agric. 2022, 6, 211–229. [Google Scholar] [CrossRef]
- Wang, Y.; Zhang, W.; Gao, R.; Jin, Z.; Wang, X. Recent Advances in the Application of Deep Learning Methods to Forestry. Wood Sci. Technol. 2021, 55, 1171–1202. [Google Scholar] [CrossRef]
- Sun, Y.; Lin, Q.; He, X.; Zhao, Y.; Dai, F.; Qiu, J.; Cao, Y. Wood Species Recognition with Small Data: A Deep Learning Approach. Int. J. Comput. Intell. Syst. 2021, 14, 1451–1460. [Google Scholar] [CrossRef]
- Wu, F.; Gazo, R.; Haviarova, E.; Benes, B. Wood Identification Based on Longitudinal Section Images by Using Deep Learning. Wood Sci. Technol. 2021, 55, 553–563. [Google Scholar] [CrossRef]
- Pan, X.; Qiu, J.; Yang, Z. Identification of Softwood Species Using Convolutional Neural Networks and Raw Near-Infrared Spectroscopy. Wood Mater. Sci. Eng. 2023, 18, 1338–1348. [Google Scholar] [CrossRef]
- Lopes, D.J.V.; Monti, G.F.; Burgreen, G.W.; Moulin, J.C.; Dos Santos Bobadilha, G.; Entsminger, E.D.; Oliveira, R.F. Creating High-Resolution Microscopic Cross-Section Images of Hardwood Species Using Generative Adversarial Networks. Front. Plant Sci. 2021, 12, 760139. [Google Scholar] [CrossRef]
- Martins, J.; Oliveira, L.S.; Nisgoski, S.; Sabourin, R. A Database for Automatic Classification of Forest Species. Mach. Vis. Appl. 2013, 24, 567–578. [Google Scholar] [CrossRef]
- Wiedenhoeft, A. Atlas of Macroscopic Wood Identification with a Special Focus on Timbers Used in Europe and CITES-Listed Species. Wood Sci. Technol. 2020, 54, 1401–1403. [Google Scholar] [CrossRef]
- Goodfellow, I.J.; Pouget-Abadie, J.; Mirza, M.; Xu, B.; Warde-Farley, D.; Ozair, S.; Courville, A.; Bengio, Y. Generative Adversarial Networks. arXiv 2014, arXiv:1406.2661. [Google Scholar] [CrossRef]
- Cheng, Y.; Fang, X.; Tang, Z.; Yu, Z.; Sun, L.; Zhu, L. SDR2Tr-GAN: A Novel Medical Image Fusion Pipeline Based on GAN With SDR2 Module and Transformer Optimization Strategy. Int. J. Imaging Syst. Technol. 2024, 34, e23208. [Google Scholar] [CrossRef]
- Bakht, A.B.; Jia, Z.; Din, M.U.; Akram, W.; Saoud, L.S.; Seneviratne, L.; Lin, D.; He, S.; Hussain, I. MuLA-GAN: Multi-Level Attention GAN for Enhanced Underwater Visibility. Ecol. Inform. 2024, 81, 102631. [Google Scholar] [CrossRef]
- Xu, J.; Lu, K.; Shi, X.; Qin, S.; Wang, H.; Ma, J. A DenseUnet Generative Adversarial Network for Near-Infrared Face Image Colorization. Signal Process. 2021, 183, 108007. [Google Scholar] [CrossRef]
- Qi, H.; Huang, Z.; Jin, B.; Tang, Q.; Jia, L.; Zhao, G.; Cao, D.; Sun, Z.; Zhang, C. SAM-GAN: An Improved DCGAN for Rice Seed Viability Determination Using near-Infrared Hyperspectral Imaging. Comput. Electron. Agric. 2024, 216, 108473. [Google Scholar] [CrossRef]
- Zhang, Z.; Gao, Q.; Liu, L.; He, Y. A High-Quality Rice Leaf Disease Image Data Augmentation Method Based on a Dual GAN. IEEE Access 2023, 11, 21176–21191. [Google Scholar] [CrossRef]
- Fathallah, M.; Sakr, M.; Eletriby, S. Stabilizing and Improving Training of Generative Adversarial Networks Through Identity Blocks and Modified Loss Function. IEEE Access 2023, 11, 43276–43285. [Google Scholar] [CrossRef]
- He, X.; Luo, Z.; Li, Q.; Chen, H.; Li, F. DG-GAN: A High Quality Defect Image Generation Method for Defect Detection. Sensors 2023, 23, 5922. [Google Scholar] [CrossRef]
- Israr, S.M.; Saeed, R.; Zhao, F. Few-Shot Adaptation of GANs Using Self-Supervised Consistency Regularization. Knowl.-Based Syst. 2024, 302, 112256. [Google Scholar] [CrossRef]
- Yildiz, E.; Yuksel, M.E.; Sevgen, S. A Single-Image GAN Model Using Self-Attention Mechanism and DenseNets. Neurocomputing 2024, 596, 127873. [Google Scholar] [CrossRef]
- Sun, S.; Li, G.; Ding, R.; Zhao, L.; Zhang, Y.; Zhao, S.; Zhang, J.; Ye, J. Seismic Random Noise Suppression Using Improved CycleGAN. Front. Earth Sci. 2023, 11, 1102656. [Google Scholar] [CrossRef]
- Chen, Z.; Zhang, Y. CA-GAN: The Synthesis of Chinese Art Paintings Using Generative Adversarial Networks. Vis. Comput. 2023, 40, 5451–5463. [Google Scholar] [CrossRef]
- Hjelm, R.D.; Jacob, A.P.; Che, T.; Trischler, A.; Cho, K.; Bengio, Y. Boundary-Seeking Generative Adversarial Networks. arXiv 2018, arXiv:1702.08431. [Google Scholar] [CrossRef]
- Radford, A.; Metz, L.; Chintala, S. Unsupervised Representation Learning with Deep Convolutional Generative Adversarial Networks. arXiv 2016, arXiv:1511.06434. [Google Scholar] [CrossRef]
- Gulrajani, I.; Ahmed, F.; Arjovsky, M.; Dumoulin, V.; Courville, A. Improved Training of Wasserstein GANs. arXiv 2017, arXiv:1704.00028. [Google Scholar] [CrossRef]
- Mao, X.; Li, Q.; Xie, H.; Lau, R.Y.K.; Wang, Z.; Smolley, S.P. Least Squares Generative Adversarial Networks. arXiv 2017, arXiv:1611.04076. [Google Scholar] [CrossRef]
- Karras, T.; Laine, S.; Aittala, M.; Hellsten, J.; Lehtinen, J.; Aila, T. Analyzing and Improving the Image Quality of StyleGAN. arXiv 2020, arXiv:1912.04958. [Google Scholar] [CrossRef]
- Yang, R.; Chen, J.; Wang, J.; Liu, S. Toward Field Soil Surveys: Identifying and Delineating Soil Diagnostic Horizons Based on Deep Learning and RGB Image. Agronomy 2022, 12, 2664. [Google Scholar] [CrossRef]
- Appiahene, P.; Asare, J.W.; Donkoh, E.T.; Dimauro, G.; Maglietta, R. Detection of Iron Deficiency Anemia by Medical Images: A Comparative Study of Machine Learning Algorithms. BioData Min. 2023, 16, 2. [Google Scholar] [CrossRef]
- Bińkowski, M.; Sutherland, D.J.; Arbel, M.; Gretton, A. Demystifying MMD GANs. arXiv 2021, arXiv:1801.01401. [Google Scholar] [CrossRef]
- Salimans, T.; Goodfellow, I.; Zaremba, W.; Cheung, V.; Radford, A.; Chen, X. Improved Techniques for Training GANs. arXiv 2016, arXiv:1606.03498. [Google Scholar] [CrossRef]
- Horé, A.; Ziou, D. Image Quality Metrics: PSNR vs. SSIM. In Proceedings of the 2010 20th International Conference on Pattern Recognition, Istanbul, Turkey, 23–26 August 2010; pp. 2366–2369. [Google Scholar] [CrossRef]







| Model | Core Contributions | Main Limitations & Criticisms |
|---|---|---|
| BGAN [26] | Proposed a boundary-seeking objective function, particularly suitable for discrete data. It generates samples by having the generator approximate the “boundary” of the real data distribution, aiding training stability and mitigating mode collapse. | Performance Dependency: Its performance heavily relies on the effective use of RL techniques like policy gradient, increasing implementation complexity. |
| DCGAN [27] | Established the foundation for CNNs in GAN architectures, proposing a set of architectural guidelines (e.g., using strided convs, batch norm, ReLU/LeakyReLU), which laid the groundwork for subsequent CNN-based GAN research. | Training Instability: Training can still be unstable and suffer from mode collapse despite improvements. Resolution Limitation: The original architecture struggles to generate high-resolution, highly realistic images. |
| WGAN-GP [28] | Introduced the Wasserstein distance as the loss function, providing theoretically smoother gradients and significantly improving training stability. Proposed Gradient Penalty to enforce the Lipschitz constraint, avoiding the optimization issues caused by weight clipping. | Computational Overhead: The gradient penalty term adds extra cost as it requires computing gradient norms on interpolated points between real and fake data. |
| LSGAN [29] | Replaced the traditional cross-entropy loss with a least squares loss. This provides the discriminator with smoother and non-saturating gradients, helping to stabilize training and generate higher-quality images. | Mode Collapse: While alleviated, LSGAN can still suffer from mode collapse, especially on complex datasets. |
| StyleGAN2 [30] | Redesigned the generator architecture to eliminate “water droplet” artifacts present in StyleGAN1. Introduced weight demodulation and path length regularization, enabling more precise control over styles and generating exceptionally high-quality and diverse images. | Extreme Complexity: The model is very complex and requires substantial computational resources and time to train. Data Hungry: Heavily relies on large-scale, high-quality datasets to reach its potential; prone to overfitting or underperformance on small datasets. Latent Space Interpretability: While offering control, its advanced latent space structure remains complex and is not fully interpretable. |
| Parameter | Value | Description |
|---|---|---|
| GPU | NVIDIA GeForce RTX 4060Ti 8 GB | Graphics processing unit for accelerated model training |
| CPU | 13th Gen Intel(R) Core (TM) i5-13490F 2.50 GHz | Central processing unit for data preprocessing operations |
| Python | 3.9 | Programming language environment for implementation |
| PyTorch | 1.12 | Deep learning framework for neural network operations |
| Total Epochs | 4000 | Maximum training epochs |
| Batch Size | 64 | Samples per batch |
| Learning Rate | 0.0002 | Adam optimizer learning rate |
| Optimizer | Adam | β1 = 0.5, β2 = 0.999 |
| Image Size | 256 × 256 | Output image resolution |
| Gradient Penalty | λ = 10 | WGAN-GP gradient penalty coefficient |
| Augmentation Configuration | Number of Images | KID (↓) | IS (↑) | SSIM (↑) |
|---|---|---|---|---|
| Original Only | 20 | 25.25 | 2.35 | 0.373 |
| Flips | 60 | 29.97 | 2.55 | 0.414 |
| Flips + Translation + Brightness | 140 | 29.83 | 2.64 | 0.423 |
| All 13 Methods | 280 | 28.58 | 2.83 | 0.427 |
| Models | Training Time (s) |
|---|---|
| BGAN | 32,187 |
| DCGAN | 152,289 |
| WGAN-GP | 37,382 |
| LSGAN | 72,079 |
| StyleGAN2 | 465,796 |
| Models | KID | IS | SSIM | Density | Coverage | LPIPS |
|---|---|---|---|---|---|---|
| BGAN | 20.42 | 1.74 | 0.39 | 0.52 | 0.90 | 0.32 |
| DCGAN | 45.97 | 1.67 | 0.37 | 0 | 0 | 0.50 |
| WGAN-GP | 28.58 | 3.04 | 0.39 | 0.26 | 0.95 | 0.49 |
| LSGAN | 30.51 | 1.83 | 0.37 | 0.03 | 0.25 | 0.46 |
| StyleGAN2 | 25.59 | 1.33 | 0.36 | 0.44 | 0.50 | 0.33 |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2025 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Xu, S.; Su, H.; Zhao, L. Research on Augmentation of Wood Microscopic Image Dataset Based on Generative Adversarial Networks. J. Imaging 2025, 11, 445. https://doi.org/10.3390/jimaging11120445
Xu S, Su H, Zhao L. Research on Augmentation of Wood Microscopic Image Dataset Based on Generative Adversarial Networks. Journal of Imaging. 2025; 11(12):445. https://doi.org/10.3390/jimaging11120445
Chicago/Turabian StyleXu, Shuo, Hang Su, and Lei Zhao. 2025. "Research on Augmentation of Wood Microscopic Image Dataset Based on Generative Adversarial Networks" Journal of Imaging 11, no. 12: 445. https://doi.org/10.3390/jimaging11120445
APA StyleXu, S., Su, H., & Zhao, L. (2025). Research on Augmentation of Wood Microscopic Image Dataset Based on Generative Adversarial Networks. Journal of Imaging, 11(12), 445. https://doi.org/10.3390/jimaging11120445

