Spinal-Fusion 发表于 2025-3-23 10:40:22
Characteristics of the Unexpected Message Queue of MPI Applicationses. We find that for the particular inputs used, these applications have widely varying characteristics with regard to UMQ length and show patterns for specific applications which persist over various scales.transdermal 发表于 2025-3-23 15:11:54
http://reply.papertrans.cn/83/8231/823051/823051_12.pngPetechiae 发表于 2025-3-23 19:42:29
An HDF5 MPI Virtual File Driver for Parallel In-situ Post-processingcess data as efficiently as possible with minimal disruption to the simulation itself, we have developed a parallel virtual file driver for the HDF5 library which acts as an MPI-IO virtual file layer, allowing the simulation to write in parallel to remotely located distributed shared memory instead of writing to disk.Exuberance 发表于 2025-3-24 00:35:39
http://reply.papertrans.cn/83/8231/823051/823051_14.pngHemodialysis 发表于 2025-3-24 05:50:12
Design of Kernel-Level Asynchronous Collective Communicationd KACC is proposed to provide fast asynchronous collective communications. KACC is implemented in the OS kernel interrupt context to perform non-blocking asynchronous collective operations without an extra thread. The experimental results show that the CPU time cost of this method is sufficiently small.BUMP 发表于 2025-3-24 08:51:06
An In-Place Algorithm for Irregular All-to-All Communication with Limited Memorynd on the message sizes. Additional memory of arbitrary size can be used to improve its performance. Performance results for a Blue Gene/P system are shown to demonstrate the performance of the approach.掺假 发表于 2025-3-24 12:11:13
Parallel Zero-Copy Algorithms for Fast Fourier Transform and Conjugate Gradient Using MPI Datatypesgnificant speedups up to a factor of 3.8 and 18%, respectively, in both cases. Our work can be used as a template to utilize datatypes for application developers. For MPI implementers, we show two practically relevant access patterns that deserve special optimization.follicle 发表于 2025-3-24 15:16:21
http://reply.papertrans.cn/83/8231/823051/823051_18.png得体 发表于 2025-3-24 22:58:28
Enabling Concurrent Multithreaded MPI Communication on Multicore Petascale Systemsode, combined with MPI across nodes. Achieving high performance when a large number of concurrent threads make MPI calls is a challenging task for an MPI implementation. We describe the design and implementation of our solution in MPICH2 to achieve high-performance multithreaded communication on thecipher 发表于 2025-3-25 02:13:42
http://reply.papertrans.cn/83/8231/823051/823051_20.png