oalib
Search Results: 1 - 10 of 100 matches for " "
All listed articles are free for downloading (OA Articles)
Page 1 /100
Display every page Item
МИНИМИЗАЦИЯ ВРЕМЕНИ ВЫПОЛНЕНИЯ MPI-ПРОГРАММ С УЧЕТОМ КОНКУРЕНЦИИ ЗА КАНАЛЫ ПЕРЕДАЧИ ДАННЫХ КОММУНИКАЦИОННОЙ СРЕДЫ КЛАСТЕРНОЙ СИСТЕМЫ  [cached]
А. В. Юлдашев
Vestnik UGATU , 2011,
Abstract: В данной работе исследуется влияние конкуренции за каналы передачи данных коммуникационной среды на время выполнения MPI-программ на кластерных системах, узлы которых построены на основе многоядерных процессоров. Предлагается модель конкурентного использования каналов передачи данных. Описывается разработанный метод назначения задач (MPI-программ) на узлы кластерной системы, позволяющий сократить время выполнения программ за счет минимизации задержек, возникающих при конкурентном использовании каналов передачи данных. Представлена апробация разработанного метода назначения задач на кластерной системе УГАТУ.
МИНИМИЗАЦИЯ ВРЕМЕНИ ВЫПОЛНЕНИЯ MPI-ПРОГРАММ С УЧЕТОМ КОНКУРЕНЦИИ ЗА КАНАЛЫ ПЕРЕДАЧИ ДАННЫХ КОММУНИКАЦИОННОЙ СРЕДЫ КЛАСТЕРНОЙ СИСТЕМЫ  [cached]
А. В. Юлдашев
Vestnik UGATU , 2011,
Abstract: В данной работе исследуется влияние конкуренции за каналы передачи данных коммуникационной среды на время выполнения MPI-программ на кластерных системах, узлы которых построены на основе многоядерных процессоров. Предлагается модель конкурентного использования каналов передачи данных. Описывается разработанный метод назначения задач (MPI-программ) на узлы кластерной системы, позволяющий сократить время выполнения программ за счет минимизации задержек, возникающих при конкурентном использовании каналов передачи данных. Представлена апробация разработанного метода назначения задач на кластерной системе УГАТУ.
DART-MPI: An MPI-based Implementation of a PGAS Runtime System  [PDF]
Huan Zhou,Yousri Mhedheb,Kamran Idrees,Colin W. Glass,José Gracia,Karl Fürlinger,Jie Tao
Computer Science , 2015, DOI: 10.1145/2676870.2676875
Abstract: A Partitioned Global Address Space (PGAS) approach treats a distributed system as if the memory were shared on a global level. Given such a global view on memory, the user may program applications very much like shared memory systems. This greatly simplifies the tasks of developing parallel applications, because no explicit communication has to be specified in the program for data exchange between different computing nodes. In this paper we present DART, a runtime environment, which implements the PGAS paradigm on large-scale high-performance computing clusters. A specific feature of our implementation is the use of one-sided communication of the Message Passing Interface (MPI) version 3 (i.e. MPI-3) as the underlying communication substrate. We evaluated the performance of the implementation with several low-level kernels in order to determine overheads and limitations in comparison to the underlying MPI-3.
MPISE: Symbolic Execution of MPI Programs  [PDF]
Xianjin Fu,Zhenbang Chen,Yufeng Zhang,Chun Huang,Wei Dong,Ji Wang
Computer Science , 2014,
Abstract: Message Passing Interfaces (MPI) plays an important role in parallel computing. Many parallel applications are implemented as MPI programs. The existing methods of bug detection for MPI programs have the shortage of providing both input and non-determinism coverage, leading to missed bugs. In this paper, we employ symbolic execution to ensure the input coverage, and propose an on-the-fly schedule algorithm to reduce the interleaving explorations for non-determinism coverage, while ensuring the soundness and completeness. We have implemented our approach as a tool, called MPISE, which can automatically detect the deadlock and runtime bugs in MPI programs. The results of the experiments on benchmark programs and real world MPI programs indicate that MPISE finds bugs effectively and efficiently. In addition, our tool also provides diagnostic information and replay mechanism to help understanding bugs.
Research of MPI Programs Optimization Technology on Multi-core Clusters
多核机群下MPI程序优化技术的研究

WANG Jie,ZHONG Lu-jie,ZENG Yu,
王洁
,衷璐洁,曾宇

计算机科学 , 2011,
Abstract: The new features of multi-core make the memory hierarchy of multi-core clusters more complex, and also add the optimization space for MPI programs. We tested the communication performance of three different multi-core clus- ters, and evaluated some general optimization technologies, such as hybrid MPI/OpenMP, tuning MPI runtime parame- ters and optimization of MPI process placement in Intel and AMD multi core cluster. I}he experiments result and opti- mization performances were also analyzed.
Implementing Parareal - OpenMP or MPI?  [PDF]
Daniel Ruprecht
Computer Science , 2015,
Abstract: The paper presents a comparison between MPI and OpenMP implementations of the parallel-in-time integration method Parareal. A special-purpose, lightweight FORTRAN code is described, which serves as a benchmark. To allow for a fair comparison, an OpenMP implementation of Parareal with pipelining is introduced, which relies on manual control of locks for synchronisation. Performance is compared in terms of runtime, speedup, memory footprint and energy-to-solution. The pipelined shared memory implementation is found to be the most efficient, particularly with respect to memory footprint. Its higher implementation complexity, however, might make it difficult to use in legacy codes.
Deadlock Detection in Basic Models of MPI Synchronization Communication Programs  [PDF]
Ming-xue Liao,Zhi-hua Fan
Computer Science , 2007,
Abstract: A model of MPI synchronization communication programs is presented and its three basic simplified models are also defined. A series of theorems and methods for deciding whether deadlocks will occur among the three models are given and proved strictly. These theories and methods for simple models' deadlock detection are the necessary base for real MPI program deadlock detection. The methods are based on a static analysis through programs and with runtime detection in necessary cases and they are able to determine before compiling whether it will be deadlocked for two of the three basic models. For another model, some deadlock cases can be found before compiling and others at runtime. Our theorems can be used to prove the correctness of currently popular MPI program deadlock detection algorithms. Our methods may decrease codes that those algorithms need to change to MPI source or profiling interface and may detects deadlocks ahead of program execution, thus the overheads can be reduced greatly.
Runtime系统综述
张宏莉,胡铭曾,方滨兴
计算机科学 , 1999,
Abstract: Runtime systems play an important role in parallel programming and parallel compilation. In this paper,goals and key techniques of runtime systems are presented. And some experiences and its trend are given in the end.
Asynchronous MPI for the Masses  [PDF]
Markus Wittmann,Georg Hager,Thomas Zeiser,Gerhard Wellein
Computer Science , 2013,
Abstract: We present a simple library which equips MPI implementations with truly asynchronous non-blocking point-to-point operations, and which is independent of the underlying communication infrastructure. It utilizes the MPI profiling interface (PMPI) and the MPI_THREAD_MULTIPLE thread compatibility level, and works with current versions of Intel MPI, Open MPI, MPICH2, MVAPICH2, Cray MPI, and IBM MPI. We show performance comparisons on a commodity InfiniBand cluster and two tier-1 systems in Germany, using low-level and application benchmarks. Issues of thread/process placement and the peculiarities of different MPI implementations are discussed in detail. We also identify the MPI libraries that already support asynchronous operations. Finally we show how our ideas can be extended to MPI-IO.
MPI Language Bindings:MPI-Delphi,MPI-Java and MPI-Ruby
MPI语言绑定:MPI-Delphi,MPI-Java与MPI-Ruby

魏兵海
计算机科学 , 2004,
Abstract: MPI(消息传递接口)是最重要的主流并行计算模式之一,它既能应用于当今的分布式环境,也可用于未来的网格环境。本文对以下三种API语言绑定作了综合性分析:MPI-Delphi,MPI-Java和MPI-Ruby,并探讨了其体系架构、实现机制及相关的技术特征。MPI-Delphi基于DLL(动态语言连接)模式实现Delphi语言到MPI库的绑定。在MPI-Java绑定实现时,JVM(Java虚拟机)、JNI(Java本地接接口)、对象串行化范型和Java新I/O库Java.nio都是用作MPI性能优化的关键技术。MPI-Ruby能够提供给用户最易使用和最强大的接口。本文还对一些相关的绑定机制进行了介绍和分析。
Page 1 /100
Display every page Item


Home
Copyright © 2008-2017 Open Access Library. All rights reserved.