Se você prefere baixar um arquivo único com todas as referências do LARCC, você pode encontrá-lo neste link. Você também pode acompanhar novas publicações via RSS.
Adicionalmente, você também pode encontrar as publicações no perfil do LARCC no Google Scholar .
2024 |
Tremarin, Gabriel Dineck; Marciano, Anna Victoria Gonc¸alves; Schepke, Claudio; Vogel, Adriano Fortran DO CONCURRENT Evaluation in Multi-core for NAS-PB Conjugate Gradient and a Porous Media Application Inproceedings Tremarin, Gabriel Dineck; Marciano, Anna Victoria Gonc¸alves; Schepke, Claudio; Vogel, Adriano (Ed.): pp. 133–143, 2024. Abstract | Links | BibTeX | Tags: multicore, Parallel programming, parallelism @inproceedings{Tremarin2024, title = {Fortran DO CONCURRENT Evaluation in Multi-core for NAS-PB Conjugate Gradient and a Porous Media Application}, author = {Gabriel Dineck Tremarin and Anna Victoria Gonc¸alves Marciano and Claudio Schepke and Adriano Vogel}, editor = {Gabriel Dineck Tremarin and Anna Victoria Gonc¸alves Marciano and Claudio Schepke and Adriano Vogel}, url = {https://larcc.setrem.com.br/wp-content/uploads/2025/03/244796_1.pdf}, year = {2024}, date = {2024-10-23}, journal = {Simpósio em Sistemas Computacionais de Alto Desempenho (SSCAD)}, pages = {133--143}, abstract = {High-performance computing exploits the hardware resources available to accelerate the applications’ executions, whereas achieving such an exploitation of hardware resources demands software programming. Hence, several parallel programming interfaces (PPIs) are used for sequential programs to call thread resources and parallelism routines. There are explicit PPIs (eg, Pthreads and TBB) or implicit (eg, OpenMP and OpenACC). Another approach is parallel programming languages like the Fortran 2008 specification, which natively provides the DO CONCURRENT resource. However, DO CONCURRENT’s evaluation is still limited. In this paper, we explore and compare the native parallelism of FORTRAN with the directives provided by the OpenMP and OpenACC PPIs in the NAS-PB CG benchmark and a porous media application. The results show that the DO CONCURRENT provides parallel CPU code with numerical compatibility for scientific applications. Moreover, DO CONCURRENT achieves in multi-cores a performance comparable to and even slightly better than other PPIs, such as OpenMP. Our work also contributes with a method to use DO CONCURRENT.}, keywords = {multicore, Parallel programming, parallelism}, pubstate = {published}, tppubtype = {inproceedings} } High-performance computing exploits the hardware resources available to accelerate the applications’ executions, whereas achieving such an exploitation of hardware resources demands software programming. Hence, several parallel programming interfaces (PPIs) are used for sequential programs to call thread resources and parallelism routines. There are explicit PPIs (eg, Pthreads and TBB) or implicit (eg, OpenMP and OpenACC). Another approach is parallel programming languages like the Fortran 2008 specification, which natively provides the DO CONCURRENT resource. However, DO CONCURRENT’s evaluation is still limited. In this paper, we explore and compare the native parallelism of FORTRAN with the directives provided by the OpenMP and OpenACC PPIs in the NAS-PB CG benchmark and a porous media application. The results show that the DO CONCURRENT provides parallel CPU code with numerical compatibility for scientific applications. Moreover, DO CONCURRENT achieves in multi-cores a performance comparable to and even slightly better than other PPIs, such as OpenMP. Our work also contributes with a method to use DO CONCURRENT. |