On using an hybrid MPI-Thread programming for the implementation of a parallel sparse direct solver on a network of SMP nodes
HÉNON, Pascal
Algorithms and high performance computing for grand challenge applications [SCALAPPLIX]
Algorithms and high performance computing for grand challenge applications [SCALAPPLIX]
RAMET, Pierre
Algorithms and high performance computing for grand challenge applications [SCALAPPLIX]
Laboratoire Bordelais de Recherche en Informatique [LaBRI]
Algorithms and high performance computing for grand challenge applications [SCALAPPLIX]
Laboratoire Bordelais de Recherche en Informatique [LaBRI]
ROMAN, Jean
Algorithms and high performance computing for grand challenge applications [SCALAPPLIX]
Laboratoire Bordelais de Recherche en Informatique [LaBRI]
Algorithms and high performance computing for grand challenge applications [SCALAPPLIX]
Laboratoire Bordelais de Recherche en Informatique [LaBRI]
HÉNON, Pascal
Algorithms and high performance computing for grand challenge applications [SCALAPPLIX]
Algorithms and high performance computing for grand challenge applications [SCALAPPLIX]
RAMET, Pierre
Algorithms and high performance computing for grand challenge applications [SCALAPPLIX]
Laboratoire Bordelais de Recherche en Informatique [LaBRI]
Algorithms and high performance computing for grand challenge applications [SCALAPPLIX]
Laboratoire Bordelais de Recherche en Informatique [LaBRI]
ROMAN, Jean
Algorithms and high performance computing for grand challenge applications [SCALAPPLIX]
Laboratoire Bordelais de Recherche en Informatique [LaBRI]
< Réduire
Algorithms and high performance computing for grand challenge applications [SCALAPPLIX]
Laboratoire Bordelais de Recherche en Informatique [LaBRI]
Langue
en
Communication dans un congrès
Ce document a été publié dans
Sixth International Conference on Parallel Processing and Applied Mathematics, Workshop HPC Linear Algebra, 2005, Poznan. 2005, vol. 3911, p. 1050--1057
Springer Verlag
Résumé en anglais
Since the last decade, most of the supercomputer architectures are based on cluster of SMP nodes. In those architectures the exchanges between processors are made through shared memory when the processor are located on a ...Lire la suite >
Since the last decade, most of the supercomputer architectures are based on cluster of SMP nodes. In those architectures the exchanges between processors are made through shared memory when the processor are located on a same SMP node and through the network otherwise. Generally, the MPI implementations provided by the constructor on those machines are adapted to this situation and take advantage of the share memory to treat messages between processors in a same SMP node. Nevertheless, this transparent approach to exploit shared memory do not avoid the storage of buffers needed in asynchronous communications. In the parallel direct solvers the storage of these buffers can become a bottleneck. In this paper, we propose an hybrid thread-MPI implementation of a direct solver and analyse the benefits of this approach in term of memory and run-time performances.< Réduire
Origine
Importé de halUnités de recherche