Mostrar o rexistro simple do ítem

dc.contributor.authorMallón, Damián A.
dc.contributor.authorTaboada, Guillermo L.
dc.contributor.authorKoesterke, Lars
dc.date.accessioned2018-11-07T15:27:24Z
dc.date.available2018-11-07T15:27:24Z
dc.date.issued2016-05-06
dc.identifier.citationMallón, D. A., Taboada, G. L., & Koesterke, L. (2016). MPI and UPC broadcast, scatter and gather algorithms in Xeon Phi. Concurrency and Computation: Practice and Experience, 28(8), 2322-2340.es_ES
dc.identifier.issn1532-0626
dc.identifier.issn1532-0634
dc.identifier.urihttp://hdl.handle.net/2183/21246
dc.descriptionThis is the peer reviewed version of the following article: Mallón, D. A., Taboada, G. L., & Koesterke, L. (2016). MPI and UPC broadcast, scatter and gather algorithms in Xeon Phi. Concurrency and Computation: Practice and Experience, 28(8), 2322-2340, which has been published in final form at https://doi.org/10.1002/cpe.3552. This article may be used for non-commercial purposes in accordance with Wiley Terms and Conditions for Use of Self-Archived Versions.es_ES
dc.description.abstract[Abstract] Accelerators have revolutionised the high performance computing (HPC) community. Despite their advantages, their very specific programming models and limited communication capabilities have kept them in a supporting role of the main processors. With the introduction of Xeon Phi, this is no longer true, as it can be programmed as the main processor and has direct access to the InfiniBand network adapter. Collective operations play a key role in many HPC applications. Therefore, studying its behaviour in the context of manycore coprocessors has great importance. This work analyses the performance of different algorithms for broadcast, scatter and gather, in a large‐scale Xeon Phi supercomputer. The algorithms evaluated are those available in the reference message passing interface (MPI) implementation for Xeon Phi (Intel MPI), the default algorithm in an optimised MPI implementation (MVAPICH2‐MIC), and a new set of algorithms, developed by the authors of this work, designed with modern processors and new communication features in mind. The latter are implemented in Unified Parallel C (UPC), a partitioned global address space language, leveraging one‐sided communications, hierarchical trees and message pipelining. This study scales the experiments to 15360 cores in the Stampede supercomputer and compares the results to Xeon and hybrid Xeon + Xeon Phi experiments, with up to 19456 cores.es_ES
dc.description.sponsorshipNational Science Foundation; OCI-1134872es_ES
dc.language.isoenges_ES
dc.publisherJohn Wiley & Sons Ltd.es_ES
dc.relation.urihttps://doi.org/10.1002/cpe.3552es_ES
dc.subjectCollective operationses_ES
dc.subjectXeon Phies_ES
dc.subjectManycorees_ES
dc.subjectUPCes_ES
dc.subjectMPIes_ES
dc.subjectInfiniBandes_ES
dc.titleMPI and UPC broadcast, scatter and gather algorithms in Xeon Phies_ES
dc.typeinfo:eu-repo/semantics/articlees_ES
dc.rights.accessinfo:eu-repo/semantics/openAccesses_ES
UDC.journalTitleConcurrency and Computation: Practice & Experiencees_ES
UDC.volume28es_ES
UDC.issue8es_ES
UDC.startPage2322es_ES
UDC.endPage2340es_ES
dc.identifier.doi10.1002/cpe.3552


Ficheiros no ítem

Thumbnail

Este ítem aparece na(s) seguinte(s) colección(s)

Mostrar o rexistro simple do ítem