Parallel Communication Optimization Based on Graph Partition for Hexagonal Neutron Transport Simulation Using MOC Method
Abstract
:1. Introduction
2. Methodologies
2.1. Hexagonal Lattice Processing
2.2. Correspondence Relationship of MPI Communication
2.3. MPI Communication Analysis of Hexagonal Core
2.4. Graph Partitioning Program METIS
3. Results
3.1. Test Environment and the Benchmark
3.2. Calculation
4. Discussion
Author Contributions
Funding
Data Availability Statement
Acknowledgments
Conflicts of Interest
References
- Cho, J.Y.; Kim, K.S.; Shim, H.J.; Song, J.S.; Lee, C.C.; Joo, H.G. Whole Core Transport Calculation Employing Hexagonal Modular Ray Tracing and CMFD Formulation. J. Nucl. Sci. Technol. 2008, 45, 740–751. [Google Scholar] [CrossRef]
- Chen, Q.; Si, S.; Zhao, J.; Bei, H. SONG-Development of transport modules. Hedongli Gongcheng/Nucl. Power Eng. 2014, 35, 127–130. [Google Scholar]
- Liu, Z.; Chen, J.; Li, S.; Wu, H.; Cao, L. Development and Verification of the Hexagonal Core Simulation Capability of the NECP-X Code. Ann. Nucl. Energy 2022, 179, 109388. [Google Scholar] [CrossRef]
- Yang, W.; Hu, C.; Liu, T.; Wang, A.; Wu, M. Research Progress of China Virtual Reactor (CVR1.0). Yuanzineng Kexue Jishu/At. Energy Sci. Technol. 2019, 53, 1821–1832. [Google Scholar]
- Boyd, W.; Shaner, S.; Li, L.; Forget, B.; Smith, K. The OpenMOC Method of Characteristics Neutral Particle Transport Code. Ann. Nucl. Energy 2014, 68, 43–52. [Google Scholar] [CrossRef] [Green Version]
- Wu, W.; Wang, Z.; Zheng, J.; Xie, Z.; Zhao, C.; Peng, X. Hexagonal Method of Characteristics with area decomposition in parallel. J. Harbin Eng. Univ 2022, 43, 1–5. [Google Scholar]
- Farhat, C. A simple and efficient automatic fem domain decomposer. Comput. Struct. 1988, 28, 579–602. [Google Scholar] [CrossRef]
- Elsner, U. Graph Partitioning—A Survey. Encycl. Parallel Comput. 1999, 97, 27. [Google Scholar] [CrossRef]
- Yao, Y.F.; Richards, B.E. Parallel CFD Computation on Unstructured Grids. In Parallel Computational Fluid Dynamics 1997; Emerson, D.R., Periaux, J., Ecer, A., Satofuka, N., Fox, P., Eds.; North-Holland: Amsterdam, The Netherlands, 1998; pp. 289–296. ISBN 978-0-444-82849-1. [Google Scholar]
- Fitzgerald, A.P. Spatial Decomposition of Structured Grids for Nuclear Reactor Simulations. Ann. Nucl. Energy 2019, 132, 686–701. [Google Scholar] [CrossRef]
- Zhao, C.; Peng, X.; Zhang, H.; Zhao, W.; Li, Q.; Chen, Z. Analysis and Comparison of the 2D/1D and Quasi-3D Methods with the Direct Transport Code SHARK. Nucl. Eng. Technol. 2022, 54, 19–29. [Google Scholar] [CrossRef]
- Zhao, C.; Peng, X.; Zhao, W.; Feng, J.; Zhao, Y.; Zhang, H.; Wang, B.; Chen, Z.; Gong, Z.; Li, Q. Verification of the Direct Transport Code SHARK with the JRR-3M Macro Benchmark. Ann. Nucl. Energy 2022, 177, 109294. [Google Scholar] [CrossRef]
- Karypis, G.; Kumar, V. METIS: A Software Package for Partitioning Unstructured Graphs, Partitioning Meshes, and Computing Fill-Reducing Orderings of Sparse Matrices. Comput. Sci. Eng. 1997. Available online: https://hdl.handle.net/11299/215346 (accessed on 3 February 2023).
- Intel® MPI Library Developer Reference for Linux* OS. 26 August 2022. Available online: https://www.intel.com/content/www/us/en/content-details/740630/intel-mpi-library-developer-reference-for-linux-os.html (accessed on 3 February 2023).
- Karypis, G.; Kumar, V. Multilevelk-Way Partitioning Scheme for Irregular Graphs. J. Parallel Distrib. Comput. 1998, 48, 96–129. [Google Scholar] [CrossRef] [Green Version]
- Kan, W.; Li, Z.; Ding, S.; Liu, Y.; Yu, G. Progress on RMC: A Monte Carlo neutron transport code for reactor analysis. In Proceedings of the International Conference on Mathematics and Computational Methods Applied to Nuclear Science and Engineering, Rio de Janeiro, Brazil, 8–12 May 2011. [Google Scholar]
Rank # | Edge Data | Graph Data |
---|---|---|
19 42 | ||
0 | [12,−1,11,−1,1,−1] | 13 12 2 |
1 | [13,−1,12,−1,2,0] | 14 13 3 1 |
2 | [3,−1,13,−1,−1,1] | 4 14 2 |
3 | [4,2,14,−1,−1,13] | 5 3 15 14 |
4 | [−1,3,5,−1,−1,14] | 4 6 15 |
5 | [−1,14,6,4,−1,15] | 15 7 5 16 |
6 | [−1,15,−1,5,−1,7] | 16 6 8 |
7 | [−1,16,−1,15,6,8] | 17 16 7 9 |
8 | [−1,9,−1,16,7,−1] | 10 17 8 |
9 | [8,10,−1,17,16,−1] | 9 11 18 17 |
10 | [9,−1,−1,11,17,−1] | 10 12 18 |
11 | [17,−1,10,0,12,−1] | 18 11 1 13 |
12 | [18,0,17,1,13,11] | 19 1 18 2 14 12 |
13 | [14,1,18,2,3,12] | 15 2 19 3 4 13 |
14 | [5,13,15,3,4,18] | 6 14 15 4 5 19 |
15 | [6,18,7,14,5,16] | 7 19 8 15 6 17 |
16 | [7,17,8,18,15,9] | 8 18 9 19 16 10 |
17 | [16,11,9,12,18,10] | 17 12 10 13 19 11 |
18 | [15,12,16,13,14,17] | 16 13 17 14 15 18 |
Test Environment | Platform 1 (v6_384 Queue of T-Partition of BSCC) | Platform 2 (t2 Queue of T-Partition of BSCC) |
---|---|---|
CPU | Intel Xeon Platinum 9242 CPU @ 2.30 GHz (96 cores per node) | Genuine Intel CPU 0000% @ 2.20 GHz (48 cores per node) |
Memory | 376 GB | 187 GB |
System | Linux Kernel: 3.10.0-1127.18.2.el7.x86_64 | Linux Kernel: 3.10.0-1127.18.2.el7.x86_64 |
Compiler | ICC 17.0.5 | ICC 17.0.5 |
MPI version | Intel MPI 2017 | Intel MPI 2017 |
Before Optimization | After Optimization | ||||
---|---|---|---|---|---|
Node Name | IB Send (MB) | IB Recv (MB) | Node Name | IB Send (MB) | IB Recv (MB) |
cb1205 | 758.46 | 763.06 | cb1203 | 112.09 | 115.96 |
cb1203 | 758.58 | 763.65 | cb1108 | 107.10 | 112.56 |
cb1108 | 753.89 | 757.90 | cb1106 | 122.12 | 128.02 |
cb1106 | 754.49 | 757.80 | cb1103 | 79.07 | 81.98 |
cb1103 | 756.74 | 761.94 | cb1101 | 68.41 | 71.47 |
cb1101 | 758.32 | 769.29 | cb1007 | 78.69 | 81.10 |
cb1007 | 757.23 | 758.80 | cb1003 | 70.19 | 73.22 |
cb1003 | 766.00 | 770.49 | cb1001 | 77.59 | 82.68 |
cb1001 | 769.05 | 774.12 | cb0907 | 124.80 | 128.97 |
cb0807 | 770.95 | 775.51 | cb0807 | 65.76 | 70.15 |
Before Optimization | After Optimization | ||||
---|---|---|---|---|---|
Node Name | IB Send (MB) | IB Recv (MB) | Node Name | IB Send (MB) | IB Recv (MB) |
cd0511 | 1121.69 | 1132.32 | cd0511 | 259.36 | 274.02 |
cd0510 | 1160.05 | 1176.01 | cd0510 | 210.98 | 232.36 |
cd0509 | 1160.61 | 1172.60 | cd0509 | 200.61 | 219.88 |
cd0508 | 1148.82 | 1158.95 | cd0508 | 238.68 | 255.61 |
cd0507 | 1157.81 | 1168.56 | cd0507 | 186.19 | 201.42 |
cd0506 | 1149.84 | 1164.02 | cd0506 | 193.93 | 207.14 |
cd0505 | 1147.87 | 1155.77 | cd0505 | 231.25 | 243.40 |
cd0415 | 1140.68 | 1150.67 | cd0415 | 169.61 | 190.58 |
cd0414 | 1145.60 | 1156.05 | cd0414 | 128.65 | 151.51 |
cd0413 | 1154.64 | 1164.28 | cd0413 | 165.99 | 187.12 |
cd0412 | 1148.46 | 1158.73 | cd0412 | 151.08 | 163.02 |
cd0411 | 1146.83 | 1160.91 | cd0411 | 161.74 | 180.15 |
cd0410 | 1148.74 | 1159.48 | cd0410 | 138.20 | 149.41 |
cd0408 | 1153.17 | 1167.11 | cd0408 | 152.14 | 168.90 |
cd0407 | 1150.50 | 1166.40 | cd0407 | 117.27 | 139.07 |
cd0405 | 1150.15 | 1164.01 | cd0405 | 187.36 | 201.21 |
cd0404 | 1153.88 | 1166.70 | cd0404 | 118.19 | 132.46 |
cd0403 | 1146.97 | 1154.43 | cd0403 | 177.83 | 188.47 |
cd0402 | 1139.48 | 1150.77 | cd0402 | 244.91 | 256.90 |
cd0401 | 1137.00 | 1156.12 | cd0401 | 129.05 | 152.29 |
Rings # | MPI # | Nodes # | Before Optimization | After Optimization | Comm Speedup | ||
---|---|---|---|---|---|---|---|
Cal Time (s) | Comm Time (s) | Cal Time (s) | Comm Time (s) | ||||
6 | 91 | 1 | 15.002 | 0.1442 | / | / | / |
7 | 127 | 2 | 14.264 | 0.5215 | 13.691 | 0.1098 | 4.75 |
8 | 169 | 2 | 15.746 | 0.6687 | 15.52 | 0.1329 | 5.03 |
9 | 217 | 3 | 15.021 | 0.6778 | 15.008 | 0.1183 | 5.73 |
10 | 271 | 3 | 16.836 | 0.8623 | 15.329 | 0.1391 | 6.2 |
11 | 331 | 4 | 16.165 | 0.8924 | 15.777 | 0.1275 | 7 |
12 | 397 | 5 | 16.343 | 0.8901 | 15.024 | 0.1302 | 6.84 |
13 | 469 | 5 | 16.851 | 1.0596 | 16.268 | 0.1408 | 7.53 |
14 | 547 | 6 | 16.335 | 1.0579 | 15.547 | 0.1484 | 7.13 |
15 | 631 | 7 | 17.594 | 1.1152 | 16.076 | 0.1505 | 7.41 |
16 | 721 | 8 | 17.134 | 1.7117 | 15.807 | 0.1609 | 10.64 |
17 | 817 | 9 | 17.036 | 1.7211 | 16.669 | 0.1783 | 9.65 |
18 | 919 | 10 | 17.849 | 1.7989 | 16.609 | 0.1629 | 11.04 |
19 | 1027 | 11 | 18.237 | 1.8685 | 16.407 | 0.1808 | 10.34 |
Parallel efficiency of 19-ring core example | 0.823 | / | 0.914 | / | / |
Rings # | MPI # | Nodes # | Before Optimization | After Optimization | Comm Speedup | ||
---|---|---|---|---|---|---|---|
Cal Time (s) | Comm Time (s) | Cal Time (s) | Comm Time (s) | ||||
4 | 37 | 1 | 13.339 | 0.0847 | / | / | / |
5 | 61 | 2 | 13.617 | 0.1975 | 12.691 | 0.0718 | 2.75 |
6 | 91 | 2 | 13.685 | 0.3603 | 13.388 | 0.1272 | 2.83 |
7 | 127 | 3 | 13.706 | 0.3862 | 13.389 | 0.0905 | 4.27 |
8 | 169 | 4 | 14.37 | 0.4498 | 13.395 | 0.0946 | 4.75 |
9 | 217 | 5 | 13.726 | 0.4847 | 13.495 | 0.094 | 5.16 |
10 | 271 | 6 | 14.233 | 0.524 | 13.795 | 0.1496 | 3.5 |
11 | 331 | 7 | 14.459 | 0.5713 | 13.667 | 0.1105 | 5.17 |
12 | 397 | 9 | 13.798 | 0.5385 | 13.641 | 0.162 | 3.32 |
13 | 469 | 10 | 14.772 | 0.7997 | 14.001 | 0.2334 | 3.43 |
14 | 547 | 12 | 14.371 | 0.753 | 13.848 | 0.2078 | 3.62 |
15 | 631 | 14 | 14.188 | 0.6245 | 13.692 | 0.1499 | 4.17 |
16 | 721 | 16 | 14.598 | 0.6917 | 14.236 | 0.1896 | 3.65 |
17 | 817 | 18 | 14.142 | 0.687 | 13.894 | 0.2084 | 3.3 |
18 | 919 | 20 | 14.473 | 0.846 | 13.746 | 0.1804 | 4.69 |
Parallel efficiency of 18-ring core example | 0.922 | / | 0.97 | / | / |
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content. |
© 2023 by the authors. Licensee MDPI, Basel, Switzerland. This article is an open access article distributed under the terms and conditions of the Creative Commons Attribution (CC BY) license (https://creativecommons.org/licenses/by/4.0/).
Share and Cite
Zheng, J.; Wang, Z.; Xie, Z.; Peng, X.; Zhao, C.; Wu, W. Parallel Communication Optimization Based on Graph Partition for Hexagonal Neutron Transport Simulation Using MOC Method. Energies 2023, 16, 2823. https://doi.org/10.3390/en16062823
Zheng J, Wang Z, Xie Z, Peng X, Zhao C, Wu W. Parallel Communication Optimization Based on Graph Partition for Hexagonal Neutron Transport Simulation Using MOC Method. Energies. 2023; 16(6):2823. https://doi.org/10.3390/en16062823
Chicago/Turabian StyleZheng, Jingchao, Zhiqiang Wang, Zeyi Xie, Xingjie Peng, Chen Zhao, and Wenbin Wu. 2023. "Parallel Communication Optimization Based on Graph Partition for Hexagonal Neutron Transport Simulation Using MOC Method" Energies 16, no. 6: 2823. https://doi.org/10.3390/en16062823
APA StyleZheng, J., Wang, Z., Xie, Z., Peng, X., Zhao, C., & Wu, W. (2023). Parallel Communication Optimization Based on Graph Partition for Hexagonal Neutron Transport Simulation Using MOC Method. Energies, 16(6), 2823. https://doi.org/10.3390/en16062823