Thomas Sterling – Panel/Talk

Keynote: Coordinated Computing: Defining the Third Domain

T
homas Sterling
Center for Computation and Technology
Louisiana State University

Center for Advanced Computing Research
California Institute of Technology

Computer Science and Mathematics Division
Oak Ridge National Laboratory

A vast amount of scientific and technical computing may be characterized as “capacity” computing, also referred to as “throughput” computing. Conventionally, “capability” computing is reserved for more tightly couple supercomputing. Yet, this second domain is blurry with few vendors willing to admit that their system offerings are largely ensembles of commodity parts interconnected by medium bandwidth system area networks. Custom capability architectures if optimized for the task of near-fine grain parallel processing incorporate mechanisms to lower overheads, additional mechanisms to compensate for latency, embody the semantics of parallel execution, and employ a high bandwidth low latency communication fabric to minimize wasted cycles. Most systems today do not provide these functions but nonetheless are referred to as capability machines, even as they deliver single digit floating point efficiencies on important mainstream computational problems. Capacity machines do not need any of these global attributes when optimized for throughput workloads. Yet many systems like commodity clusters or MPPs of COTS microprocessors and DRAMs are applied to single problems, usually by means of message passing libraries. And, this strategy has proven very successful. It is clear, then, that the distinction of capability versus capacity computing is inadequate to describe what we really do. Something between the two that recognizes the reality rather than the hype of conventional practices is required. This talk will present such an alternative: “Coordinated Computing” that provides a third domain, positioned between these two that describes the use of communicating sequential processes on ensembles of commodity components, thereby reserving the former terms, “capacity” and “capability”, for the operational ranges to which they are optimally suited.

A Petaflops Post-Mortem and the Computing Imperative

Thomas Sterling
Department of Computer Science
Center for Computation and Technology
Louisiana State University

Center for Advanced Computing Research
California Institute of Technology

and

Computer Science and Mathematics Division
Oak Ridge National Laboratory

All technological advances ultimately are driven by need (recognized or latent) and enabled by an opportunity gap. Together these are represented by the classic S-Curve where the gap is the vertical distance between the two asymptotes and the investment barrier, prior to realization, is the length of the bottom start-up segment. But S-curves are soft-sloped entities and at any point in time, they feel like straight lines often misconstrued as implying a perpetual future direction. However, the ultimate emergent behavior is saturation and the resulting stagnation due to their insidious property of one and only one point of inflection; that is S-curves bend twice. This talk will describe the realm of the S-curve in technology development with past examples and apply it to the current domain of high performance computing and its implication for future strategic goals.