Results 6926–6950 of 9,998
Search results
Jump to search filters
Lofstead, Gerald F.
Littlewood, David J.; Tikare, Veena T.
Knupp, Patrick K.; Day, David M.
Barrett, Richard F.; Hammond, Simon D.; Vaughan, Courtenay T.; Doerfler, Douglas W.; Heroux, Michael A.
Hammond, Simon D.
Silling, Stewart A.
Olivier, Stephen L.
Olivier, Stephen L.
Doerfler, Douglas W.
Littlewood, David J.; Mish, Kyran D.; Pierson, Kendall H.
Rajan, Mahesh R.; Doerfler, Douglas W.; Lin, Paul L.; Hammond, Simon D.; Barrett, Richard F.; Vaughan, Courtenay T.
Proposed for publication in Human Factors: The Journal of Human Factors and Ergonomics Society.
Abbott, Robert G.; Haass, Michael J.
Moreland, Kenneth D.
Moreland, Kenneth D.
Littlewood, David J.; Bignell, John B.; Tikare, Veena T.
Barrett, Brian B.; Brightwell, Ronald B.; Hemmert, Karl S.
Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
Hoefler, Torsten; Dinan, James; Buntinas, Darius; Balaji, Pavan; Barrett, Brian W.; Brightwell, Ronald B.; Gropp, William; Kale, Vivek; Thakur, Rajeev
Hybrid parallel programming with MPI for internode communication in conjunction with a shared-memory programming model to manage intranode parallelism has become a dominant approach to scalable parallel programming. While this model provides a great deal of flexibility and performance potential, it saddles programmers with the complexity of utilizing two parallel programming systems in the same application. We introduce an MPI-integrated shared-memory programming model that is incorporated into MPI through a small extension to the one-sided communication interface. We discuss the integration of this interface with the upcoming MPI 3.0 one-sided semantics and describe solutions for providing portable and efficient data sharing, atomic operations, and memory consistency. We describe an implementation of the new interface in the MPICH2 and Open MPI implementations and demonstrate an average performance improvement of 40% to the communication component of a five-point stencil solver. © 2012 Springer-Verlag.
Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)
Barrett, Brian W.; Brightwell, Ronald B.; Underwood, Keith D.
Message passing paradigms provide for many to one messaging patterns that result in receive side resource exhaustion. Traditionally, MPI implementations layered over the Portals network programming interface provided a large default unexpected receive buffer space, the user was expected to configure the buffer size to the application demand, and the application was aborted when the buffer space was overrun. The Portals 4 design provides a set of primitives for implementing scalable resource exhaustion recovery without negatively impacting normal operation. A resource exhaustion recovery protocol for MPI implementations is presented, as well as performance results for an Open MPI implementation of the protocol. © 2012 Springer-Verlag.
Proceedings of the 2012 IEEE 26th International Parallel and Distributed Processing Symposium, IPDPS 2012
Rajamanickam, Sivasankaran R.; Boman, Erik G.; Heroux, Michael A.
Proceedings of the 2012 IEEE 26th International Parallel and Distributed Processing Symposium, IPDPS 2012
Azad, Ariful; Halappanavar, Mahantesh; Rajamanickam, Sivasankaran R.; Boman, Erik G.; Khan, Arif; Pothen, Alex
We design, implement, and evaluate algorithms for computing a matching of maximum cardinality in a bipartite graph on multicore and massively multithreaded computers. As computers with larger numbers of slower cores dominate the commodity processor market, the design of multithreaded algorithms to solve large matching problems becomes a necessity. Recent work on serial algorithms for the matching problem has shown that their performance is sensitive to the order in which the vertices are processed for matching. In a multithreaded environment, imposing a serial order in which vertices are considered for matching would lead to loss of concurrency and performance. But this raises the question: Would parallel matching algorithms on multithreaded machines improve performance over a serial algorithm? We answer this question in the affirmative. We report efficient multithreaded implementations of three classes of algorithms based on their manner of searching for augmenting paths: breadth-first-search, depth-first-search, and a combination of both. The Karp-Sipser initialization algorithm is used to make the parallel algorithms practical. We report extensive results and insights using three shared-memory platforms (a 48-core AMD Opteron, a 32-coreIntel Nehalem, and a 128-processor Cray XMT) on a representative set of real-world and synthetic graphs. To the best of our knowledge, this is the first study of augmentation-based parallel algorithms for bipartite cardinality matching that demonstrates good speedups on multithreaded shared memory multiprocessors. © 2012 IEEE.
Journal of Computational Physics
Scovazzi, Guglielmo S.
Lofstead, Gerald F.
Proposed for publication in IEEE Computer Magazine.
Moreland, Kenneth D.
Lofstead, Gerald F.
Lofstead, Gerald F.
Results 6926–6950 of 9,998