SLIDE 7 XXVIth Telemac & Mascaret User Club Toulouse, FR, 16-17 October, 2019 reasonable to stay with the default implementation of PARTEL for a moderate large mesh, in case of an extremely large mesh
- ne should consider using the implementation V of PARTEL.
Finally, implementation VI is the combination of V with the parallelization of PARTEL. In the benchmark we only tested for the parallelization of two MPI processes and the result is
- bvious telling that the implementation is the most efficient
- ne. It is about 4 times faster in case 1 and case 2. Although,
we were not able to carry out the result for case 3 because of the memory constraints on our workstation, but from the embarrassing characteristic of the implementation of the parallelization we can actually get a significant better speed up as we increase the number of processes for the parallelized PARTEL. IV. CONCLUSION AND FUTURE WORK In this study, we provide several parallel implementations
- f the domain decomposer, PARTEL, in TELEMAC. The
- riginal serial version of PARTEL has been parallelized and it
is successful to maintain low memory consumption while
- btaining a good speedup. As a result, we can effortlessly run
simulations with significantly larger meshes, for example 100 million element mesh. The parallel implementations, however, still has a limitation running on a single computational machine node, since the total memory used by all MPI process
- f PARTEL is high. This leads to a consequence that, for a
significantly large mesh (hundreds of millions of elements) the code must run on multiple computational nodes as it is no longer suitale for a workstation. In this case, we suggest a new scheme for future development of PARTEL that can be either implemented with OpenMP for
MPI communication, with which we enjoy the benefit of parallelization and memory efficiecy
domain decomposition of PARTEL on a single computational node. This implementation should be made optional so that user can select on their suitable scenarios. ACKNOWLEDGMENT The authors gratefully acknowledge the support of the Leibnitz Supercomputing Center, Garching, Germany. We would like to thank the KONWIHR initiative (Bavarian Competence Network for Technical and Scientific HPC) for funding this study. REFERENCES
[1] Adouin Yoann, Moulinec Charles, Barber Robert, Sunderland Andrew (2011), "Preparing TELEMAC-2D for extremely large simulations". In: Violeau, Damien; Hervouet, Jean-Michel;Razafindrakoto, Emile; Denis, Christophe (Hrsg.): XVIIIth TELEMAC & MASCARET User Club 2011. Chatou: Bundesanstalt für Wasserbau (BAW). S. 41-48 [2] Moulinec, Charles, et al. "TELEMAC: An efficient hydrodynamics suite for massively parallel architectures." Computers & Fluids 51.1 (2011): 30-34 [3] Judicael Grasset, Yoann Audouin, Jacques Fontaine, Charles Moulinec, David R. Emerson, “Improving TELEMAC system pre-processing and IO stages”. TELEMAC User Conference 2018 [4] Reisenbüchler M, Liepert T, Nguyen ND, Bui MD, Rutschmann P (2018), “Preliminary study on a Bavaria-wide coupled hydrological and hydromorphological model,” Environmental Informatics: Techniques and Trends. Adjunct Proceedings of the 32nd edition of the EnviroInfo – the long-standing and established international and interdisciplinary conference series on leading environmental information and communication technologies. (eds H-J Bungartz, D Kranzlmüller, V Weinberg, J Weismüller, V Wohlgemuth. ISBN: 978-3-8440-6138-3, DOI: http://doi.org/10.2370/9783844061383, Part IV Disaster Management for Resilience and Public Safety, pp. 145-148, Herzogenrath, Germany [5] Reisenbüchler, M; Nguyen N.D.; Liepert, T.; Bui, M.D.; Rutschmann,
- P. (2018) TELEMAC - A hydrodynamic solver for HPC?. InSiDE
Magazine Spring 2018, Vol. 16, p 95:97 [6] Leibniz-Rechenzentrum (LRZ), https://www.lrz.de/ [7] METIS, http://glaros.dtc.umn.edu/gkhome/metis/metis/overview [8] Open TELEMAC-MASCARET, http://www.opentelemac.org/ [9] SCOTCH, https://www.labri.fr/perso/pelegrin/scotch/ [10] https://github.com/abseil/abseil-cpp [11] https://github.com/martinus/robin-hood-hashing [12] https://github.com/skarupke/flat_hash_map