Algorithms for higher order accuracy modeling of kinematic behavior within the ALEGRA framework are presented. These techniques improve the behavior of the code when kinematic errors are found, ensure orthonormality of the rotation tensor at each time step, and increase the accuracy of the Lagrangian stretch and rotation tensor update algorithm. The implementation of these improvements in ALEGRA is described. A short discussion of issues related to improving the accuracy of the stress update procedures is also included.
Solidification and blood flow seemingly have little in common, but each involves a fluid in contact with a deformable solid. In these systems, the solid-fluid interface moves as the solid advects and deforms, often traversing the entire domain of interest. Currently, these problems cannot be simulated without innumerable expensive remeshing steps, mesh manipulations or decoupling the solid and fluid motion. Despite the wealth of progress recently made in mechanics modeling, this glaring inadequacy persists. We propose a new technique that tracks the interface implicitly and circumvents the need for remeshing and remapping the solution onto the new mesh. The solid-fluid boundary is tracked with a level set algorithm that changes the equation type dynamically depending on the phases present. This novel approach to coupled mechanics problems promises to give accurate stresses, displacements and velocities in both phases, simultaneously.
High throughput instruments and analysis techniques are required in order to make good use of the genomic sequences that have recently become available for many species, including humans. These instruments and methods must work with tens of thousands of genes simultaneously, and must be able to identify the small subsets of those genes that are implicated in the observed phenotypes, or, for instance, in responses to therapies. Microarrays represent one such high throughput method, which continue to find increasingly broad application. This project has improved microarray technology in several important areas. First, we developed the hyperspectral scanner, which has discovered and diagnosed numerous flaws in techniques broadly employed by microarray researchers. Second, we used a series of statistically designed experiments to identify and correct errors in our microarray data to dramatically improve the accuracy, precision, and repeatability of the microarray gene expression data. Third, our research developed new informatics techniques to identify genes with significantly different expression levels. Finally, natural language processing techniques were applied to improve our ability to make use of online literature annotating the important genes. In combination, this research has improved the reliability and precision of laboratory methods and instruments, while also enabling substantially faster analysis and discovery.
ALEGRA is an arbitrary Lagrangian-Eulerian finite element code that emphasizes large distortion and shock propagation in inviscid fluids and solids. This document describes user options for modeling magnetohydrodynamic, thermal conduction, and radiation emission effects.
Wireless communication networks are highly resource-constrained; thus many security protocols which work in other settings may not be efficient enough for use in wireless environments. This report considers a variety of cryptographic techniques which enable secure, authenticated communication when resources such as processor speed, battery power, memory, and bandwidth are tightly limited.
We present here the details of the implementation of the parallel tempering Monte Carlo technique into a LAMMPS, a heavily used massively parallel molecular dynamics code at Sandia. This technique allows for many replicas of a system to be run at different simulation temperatures. At various points in the simulation, configurations can be swapped between different temperature environments and then continued. This allows for large regions of energy space to be sampled very quickly, and allows for minimum energy configurations to emerge in very complex systems, such as large biomolecular systems. By including this algorithm into an existing code, we immediately gain all of the previous work that had been put into LAMMPS, and allow this technique to quickly be available to the entire Sandia and international LAMMPS community. Finally, we present an example of this code applied to folding a small protein.
Genetic programming is a powerful methodology for automatically producing solutions to problems in a variety of domains. It has been used successfully to develop behaviors for RoboCup soccer players and simple combat agents. We will attempt to use genetic programming to solve a problem in the domain of strategic combat, keeping in mind the end goal of developing sophisticated behaviors for compound defense and infiltration. The simplified problem at hand is that of two armed agents in a small room, containing obstacles, fighting against each other for survival. The base case and three changes are considered: a memory of positions using stacks, context-dependent genetic programming, and strongly typed genetic programming. Our work demonstrates slight improvements from the first two techniques, and no significant improvement from the last.
This report is based on the Statement of Work (SOW) describing the various requirements for delivering 3 new supercomputer system to Sandia National Laboratories (Sandia) as part of the Department of Energy's (DOE) Accelerated Strategic Computing Initiative (ASCI) program. This system is named Red Storm and will be a distributed memory, massively parallel processor (MPP) machine built primarily out of commodity parts. The requirements presented here distill extensive architectural and design experience accumulated over a decade and a half of research, development and production operation of similar machines at Sandia. Red Storm will have an unusually high bandwidth, low latency interconnect, specially designed hardware and software reliability features, a light weight kernel compute node operating system and the ability to rapidly switch major sections of the machine between classified and unclassified computing environments. Particular attention has been paid to architectural balance in the design of Red Storm, and it is therefore expected to achieve an atypically high fraction of its peak speed of 41 TeraOPS on real scientific computing applications. In addition, Red Storm is designed to be upgradeable to many times this initial peak capability while still retaining appropriate balance in key design dimensions. Installation of the Red Storm computer system at Sandia's New Mexico site is planned for 2004, and it is expected that the system will be operated for a minimum of five years following installation.
The Sandia Petaflops Planner is a tool for projecting the design and performance of parallel supercomputers into the future. The mathematical basis of these projections is the International Technology Roadmap for Semiconductors (ITRS, or a detailed version of Moore's Law) and DOE balance factors for supercomputer procurements. The planner is capable of various forms of scenario analysis, cost estimation, and technology analysis. The tool is described along with technology conclusions regarding PFLOPS-level supercomputers in the upcoming decade.
The Computational Plant or Cplant is a commodity-based distributed-memory supercomputer under development at Sandia National Laboratories. Distributed-memory supercomputers run many parallel programs simultaneously. Users submit their programs to a job queue. When a job is scheduled to run, it is assigned to a set of available processors. Job runtime depends not only on the number of processors but also on the particular set of processors assigned to it. Jobs should be allocated to localized clusters of processors to minimize communication costs and to avoid bandwidth contention caused by overlapping jobs. This report introduces new allocation strategies and performance metrics based on space-filling curves and one dimensional allocation strategies. These algorithms are general and simple. Preliminary simulations and Cplant experiments indicate that both space-filling curves and one-dimensional packing improve processor locality compared to the sorted free list strategy previously used on Cplant. These new allocation strategies are implemented in Release 2.0 of the Cplant System Software that was phased into the Cplant systems at Sandia by May 2002. Experimental results then demonstrated that the average number of communication hops between the processors allocated to a job strongly correlates with the job's completion time. This report also gives processor-allocation algorithms for minimizing the average number of communication hops between the assigned processors for grid architectures. The associated clustering problem is as follows: Given n points in {Re}d, find k points that minimize their average pairwise L{sub 1} distance. Exact and approximate algorithms are given for these optimization problems. One of these algorithms has been implemented on Cplant and will be included in Cplant System Software, Version 2.1, to be released. In more preliminary work, we suggest improvements to the scheduler separate from the allocator.
This report summarizes the results of a three-year LDRD project on prognostics and health management. System failure over some future time interval (an alternative definition is the capability to predict the remaining useful life of a system). Prognostics are integrated with health monitoring (through inspections, sensors, etc.) to provide an overall PHM capability that optimizes maintenance actions and results in higher availability at a lower cost. Our goal in this research was to develop PHM tools that could be applied to a wide variety of equipment (repairable, non-repairable, manufacturing, weapons, battlefield equipment, etc.) and require minimal customization to move from one system to the next. Thus, our approach was to develop a toolkit of reusable software objects/components and architecture for their use. We have developed two software tools: an Evidence Engine and a Consequence Engine. The Evidence Engine integrates information from a variety of sources in order to take into account all the evidence that impacts a prognosis for system health. The Evidence Engine has the capability for feature extraction, trend detection, information fusion through Bayesian Belief Networks (BBN), and estimation of remaining useful life. The Consequence Engine involves algorithms to analyze the consequences of various maintenance actions. The Consequence Engine takes as input a maintenance and use schedule, spares information, and time-to-failure data on components, then generates maintenance and failure events, and evaluates performance measures such as equipment availability, mission capable rate, time to failure, and cost. This report summarizes the capabilities we have developed, describes the approach and architecture of the two engines, and provides examples of their use. 'Prognostics' refers to the capability to predict the probability of
A fundamental challenge for all communication systems, engineered or living, is the problem of achieving efficient, secure, and error-free communication over noisy channels. Information theoretic principals have been used to develop effective coding theory algorithms to successfully transmit information in engineering systems. Living systems also successfully transmit biological information through genetic processes such as replication, transcription, and translation, where the genome of an organism is the contents of the transmission. Decoding of received bit streams is fairly straightforward when the channel encoding algorithms are efficient and known. If the encoding scheme is unknown or part of the data is missing or intercepted, how would one design a viable decoder for the received transmission? For such systems blind reconstruction of the encoding/decoding system would be a vital step in recovering the original message. Communication engineers may not frequently encounter this situation, but for computational biologists and biotechnologist this is an immediate challenge. The goal of this work is to develop methods for detecting and reconstructing the encoder/decoder system for engineered and biological data. Building on Sandia's strengths in discrete mathematics, algorithms, and communication theory, we use linear programming and will use evolutionary computing techniques to construct efficient algorithms for modeling the coding system for minimally errored engineered data stream and genomic regulatory DNA and RNA sequences. The objective for the initial phase of this project is to construct solid parallels between biological literature and fundamental elements of communication theory. In this light, the milestones for FY2003 were focused on defining genetic channel characteristics and providing an initial approximation for key parameters, including coding rate, memory length, and minimum distance values. A secondary objective addressed the question of determining similar parameters for a received, noisy, error-control encoded data set. In addition to these goals, we initiated exploration of algorithmic approaches to determine if a data set could be approximated with an error-control code and performed initial investigations into optimization based methodologies for extracting the encoding algorithm given the coding rate of an encoded noise-free and noisy data stream.
Evidence for the existence of discrete sub-movements underlying continuous human movement has motivated many attempts to "extract" them. Although they produce visually convincing results, all of the methodologies that have been employed are prone to produce spurious decompositions. Examples of potential failures are given. A branch-and-bound algorithm for submovement extraction, capable of global nonlinear minimization (and hence capable of avoiding spurious decompositions), is developed and demonstrated.
If we are to build a supercomputer with a speed of 10{sup 15} floating operations per second (1 PetaFLOPS), interconnect technology will need to be improved considerably over what it is today. In this report, we explore one possible interconnect design for such a network. The guiding principle in this design is the optimization of all components for the finiteness of the speed of light. To achieve a linear speedup in time over well-tested supercomputers of todays' designs will require scaling up of processor power and bandwidth and scaling down of latency. Latency scaling is the most challenging: it requires a 100 ns user-to-user latency for messages traveling the full diameter of the machine. To meet this constraint requires simultaneously minimizing wire length through 3D packaging, new low-latency electrical signaling mechanisms, extremely fast routers, and new network interfaces. In this report, we outline approaches and implementations that will meet the requirements when implemented as a system. No technology breakthroughs are required.
High-resolution finite volume methods for solving systems of conservation laws have been widely embraced in research areas ranging from astrophysics to geophysics and aero-thermodynamics. These methods are typically at least second-order accurate in space and time, deliver non-oscillatory solutions in the presence of near discontinuities, e.g., shocks, and introduce minimal dispersive and diffusive effects. High-resolution methods promise to provide greatly enhanced solution methods for Sandia's mainstream shock hydrodynamics and compressible flow applications, and they admit the possibility of a generalized framework for treating multi-physics problems such as the coupled hydrodynamics, electro-magnetics and radiative transport found in Z pinch physics. In this work, we describe initial efforts to develop a generalized 'black-box' conservation law framework based on modern high-resolution methods and implemented in an object-oriented software framework. The framework is based on the solution of systems of general non-linear hyperbolic conservation laws using Godunov-type central schemes. In our initial efforts, we have focused on central or central-upwind schemes that can be implemented with only a knowledge of the physical flux function and the minimal/maximal eigenvalues of the Jacobian of the flux functions, i.e., they do not rely on extensive Riemann decompositions. Initial experimentation with high-resolution central schemes suggests that contact discontinuities with the concomitant linearly degenerate eigenvalues of the flux Jacobian do not pose algorithmic difficulties. However, central schemes can produce significant smearing of contact discontinuities and excessive dissipation for rotational flows. Comparisons between 'black-box' central schemes and the piecewise parabolic method (PPM), which relies heavily on a Riemann decomposition, shows that roughly equivalent accuracy can be achieved for the same computational cost with both methods. However, PPM clearly outperforms the central schemes in terms of accuracy at a given grid resolution and the cost of additional complexity in the numerical flux functions. Overall we have observed that the finite volume schemes, implemented within a well-designed framework, are extremely efficient with (potentially) very low memory storage. Finally, we have found by computational experiment that second and third-order strong-stability preserving (SSP) time integration methods with the number of stages greater than the order provide a useful enhanced stability region. However, we observe that non-SSP and non-optimal SSP schemes with SSP factors less than one can still be very useful if used with time-steps below the standard CFL limit. The 'well-designed' integration schemes that we have examined appear to perform well in all instances where the time step is maintained below the standard physical CFL limit.
An understanding of the dynamics of z-pinch wire array explosion and collapse is of critical interest to the development and future of pulsed power inertial confinement fusion experiments. Experimental results clearly show the extreme three-dimensional nature of the wire explosion and collapse process. The physics of this process can be approximated by the resistive magnetohydrodynamic (MHD) equations augmented by thermal and radiative transport modeling. Z-pinch MHD physics is dominated by material regions whose conductivity properties vary drastically as material passes from solid through melt into plasma regimes. At the same time void regions between the wires are modeled as regions of very low conductivity. This challenging physical situation requires a sophisticated three-dimensional modeling approach matched by sufficient computational resources to make progress in predictive modeling and improved physical understanding.
This paper describes a methodology for implementing disk-less cluster systems using the Network File System (NFS) that scales to thousands of nodes. This method has been successfully deployed and is currently in use on several production systems at Sandia National Labs. This paper will outline our methodology and implementation, discuss hardware and software considerations in detail and present cluster configurations with performance numbers for various management operations like booting.
A new capability for modeling thin-shell structures within the coupled Euler-Lagrange code, Zapotec, is under development. The new algorithm creates an artificial material interface for the Eulerian portion of the problem by expanding a Lagrangian shell element such that it has an effective thickness that spans one or more Eulerian cells. The algorithm implementation is discussed along with several examples involving blast loading on plates.
Implicit time integration coupled with SUPG discretization in space leads to additional terms that provide consistency and improve the phase accuracy for convection dominated flows. Recently, it has been suggested that for small Courant numbers these terms may dominate the streamline diffusion term, ostensibly causing destabilization of the SUPG method. While consistent with a straightforward finite element stability analysis, this contention is not supported by computational experiments and contradicts earlier Von-Neumann stability analyses of the semidiscrete SUPG equations. This prompts us to re-examine finite element stability of the fully discrete SUPG equations. A careful analysis of the additional terms reveals that, regardless of the time step size, they are always dominated by the consistent mass matrix. Consequently, SUPG cannot be destabilized for small Courant numbers. Numerical results that illustrate our conclusions are reported.
A recently developed Centroidal Voronoi Tessellation (CVT) unstructured sampling method is investigated here to assess its suitability for use in statistical sampling and function integration. CVT efficiently generates a highly uniform distribution of sample points over arbitrarily shaped M-Dimensional parameter spaces. It has recently been shown on several 2-D test problems to provide superior point distributions for generating locally conforming response surfaces. In this paper, its performance as a statistical sampling and function integration method is compared to that of Latin-Hypercube Sampling (LHS) and Simple Random Sampling (SRS) Monte Carlo methods, and Halton and Hammersley quasi-Monte-Carlo sequence methods. Specifically, sampling efficiencies are compared for function integration and for resolving various statistics of response in a 2-D test problem. It is found that on balance CVT performs best of all these sampling methods on our test problems.
Broadcasting messages through the earth is a daunting task. Indeed, broadcasting a normal telephone conversion through the earth by wireless means is impossible with todays technology. Most of us don't care, but some do. Industries that drill into the earth need wireless communication to broadcast navigation parameters. This allows them to steer their drill bits. They also need information about the natural formation that they are drilling. Measurements of parameters such as pressure, temperature, and gamma radiation levels can tell them if they have found a valuable resource such as a geothermal reservoir or a stratum bearing natural gas. Wireless communication methods are available to the drilling industry. Information is broadcast via either pressure waves in the drilling fluid or electromagnetic waves in the earth and well tubing. Data transmission can only travel one way at rates around a few baud. Given that normal Internet telephone modems operate near 20,000 baud, these data rates are truly very slow. Moreover, communication is often interrupted or permanently blocked by drilling conditions or natural formation properties. Here we describe a tool that communicates with stress waves traveling through the steel drill pipe and production tubing in the well. It's based on an old idea called Acoustic Telemetry. But what we present here is more than an idea. This tool exists, it's drilled several wells, and it works. Currently, it's the first and only acoustic telemetry tool that can withstand the drilling environment. It broadcasts one way over a limited range at much faster rates than existing methods, but we also know how build a system that can communicate both up and down wells of indefinite length.
This report presents a perspective on the role of code comparison activities in verification and validation. We formally define the act of code comparison as the Code Comparison Principle (CCP) and investigate its application in both verification and validation. One of our primary conclusions is that the use of code comparisons for validation is improper and dangerous. We also conclude that while code comparisons may be argued to provide a beneficial component in code verification activities, there are higher quality code verification tasks that should take precedence. Finally, we provide a process for application of the CCP that we believe is minimal for achieving benefit in verification processes.
We seek to understand which supercomputer architecture will be best for supercomputers at the Petaflops scale and beyond. The process we use is to predict the cost and performance of several leading architectures at various years in the future. The basis for predicting the future is an expanded version of Moore's Law called the International Technology Roadmap for Semiconductors (ITRS). We abstract leading supercomputer architectures into chips connected by wires, where the chips and wires have electrical parameters predicted by the ITRS. We then compute the cost of a supercomputer system and the run time on a key problem of interest to the DOE (radiation transport). These calculations are parameterized by the time into the future and the technology expected to be available at that point. We find the new advanced architectures have substantial performance advantages but conventional designs are likely to be less expensive (due to economies of scale). We do not find a universal ''winner'', but instead the right architectural choice is likely to involve non-technical factors such as the availability of capital and how long people are willing to wait for results.
The Trilinos Project is an effort to facilitate the design, development, integration and ongoing support of mathematical software libraries. In particular, our goal is to develop parallel solver algorithms and libraries within an object-oriented software framework for the solution of large-scale, complex multi-physics engineering and scientific applications. Our emphasis is on developing robust, scalable algorithms in a software framework, using abstract interfaces for flexible interoperability of components while providing a full-featured set of concrete classes that implement all abstract interfaces. Trilinos uses a two-level software structure designed around collections of packages. A Trilinos package is an integral unit usually developed by a small team of experts in a particular algorithms area such as algebraic preconditioners, nonlinear solvers, etc. Packages exist underneath the Trilinos top level, which provides a common look-and-feel, including configuration, documentation, licensing, and bug-tracking. Trilinos packages are primarily written in C++, but provide some C and Fortran user interface support. We provide an open architecture that allows easy integration with other solver packages and we deliver our software to the outside community via the Gnu Lesser General Public License (LGPL). This report provides an overview of Trilinos, discussing the objectives, history, current development and future plans of the project.
The Trilinos Project is an effort to facilitate the design, development, integration and ongoing support of mathematical software libraries. A new software capability is introduced into Trilinos as a package. A Trilinos package is an integral unit usually developed by a small team of experts in a particular algorithms area such as algebraic preconditioners, nonlinear solvers, etc. The Trilinos Users Guide is a resource for new and existing Trilinos users. Topics covered include how to configure and build Trilinos, what is required to integrate an existing package into Trilinos and examples of how those requirements can be met, as well as what tools and services are available to Trilinos packages. Also discussed are some common practices that are followed by many Trilinos package developers. Finally, a snapshot of current Trilinos packages and their interoperability status is provided, along with a list of supported computer platforms.
Protein microtubules (MTs) 25 nm in diameter and tens of micrometers long have been used as templates for the biomimetic mineralization of FeOOH. Exposure of MTs to anaerobic aqueous solutions of Fe{sup 2+} buffered to neutral pH followed by aerial oxidation leads to the formation of iron oxide coated MTs. The iron oxide layer was found to grow via a two-step process: initially formed 10-30 nm thick coatings were found to be amorphous in structure and comprised of several iron-containing species. Further growth resulted in MTs coated with highly crystalline layers of lepidocrocite with a controllable thickness of up to 125 nm. On the micrometer size scale, these coated MTs were observed to form large, irregular bundles containing hundreds of individually coated MTs. Iron oxide grew selectively on the MT surface, a result of the highly charged MT surface that provided an interface favorable for iron oxide nucleation. This result illustrates that MTs can be used as scaffolds for the in-situ production of high-aspect-ratio inorganic nanowires.
Currently, the Egyptian Atomic Energy Authority is designing a shallow-land disposal facility for low-level radioactive waste. To insure containment and prevent migration of radionuclides from the site, the use of a reactive backfill material is being considered. One material under consideration is hydroxyapatite, Ca{sub 10}(PO{sub 4}){sub 6}(OH){sub 2}, which has a high affinity for the sorption of many radionuclides. Hydroxyapatite has many properties that make it an ideal material for use as a backfill including low water solubility (K{sub sp} > 10{sup -40}), high stability under reducing and oxidizing conditions over a wide temperature range, availability, and low cost. However, there is often considerable variation in the properties of apatites depending on source and method of preparation. In this work, we characterized and compared a synthetic hydroxyapatite with hydroxyapatites prepared from cattle bone calcined at 500 C, 700 C, 900 C and 1100 C. The analysis indicated the synthetic hydroxyapatite was similar in morphology to 500 C prepared cattle hydroxyapatite. With increasing calcination temperature the crystallinity and crystal size of the hydroxyapatites increased and the BET surface area and carbonate concentration decreased. Batch sorption experiments were performed to determine the effectiveness of each material to sorb uranium. Sorption of U was strong regardless of apatite type indicating all apatite materials evaluated. Sixty day desorption experiments indicated desorption of uranium for each hydroxyapatite was negligible.
The maximum contact map overlap (MAX-CMO) between a pair of protein structures can be used as a measure of protein similarity. It is a purely topological measure and does not depend on the sequence of the pairs involved in the comparison. More importantly, the MAX-CMO present a very favorable mathematical structure which allows the formulation of integer, linear and Lagrangian models that can be used to obtain guarantees of optimality. It is not the intention of this paper to discuss the mathematical properties of MAX-CMO in detail as this has been dealt elsewhere. In this paper we compare three algorithms that can be used to obtain maximum contact map overlaps between protein structures. We will point to the weaknesses and strengths of each one. It is our hope that this paper will encourage researchers to develop new and improve methods for protein comparison based on MAX-CMO.
We consider the convergence properties of a non-elitist self-adaptive evolutionary strategy (ES) on multi-dimensional problems. In particular, we apply our recent convergence theory for a discretized (1,{lambda})-ES to design a related (1,{lambda})-ES that converges on a class of seperable, unimodal multi-dimensional problems. The distinguishing feature of self-adaptive evolutionary algorithms (EAs) is that the control parameters (like mutation step lengths) are evolved by the evolutionary algorithm. Thus the control parameters are adapted in an implicit manner that relies on the evolutionary dynamics to ensure that more effective control parameters are propagated during the search. Self-adaptation is a central feature of EAs like evolutionary stategies (ES) and evolutionary programming (EP), which are applied to continuous design spaces. Rudolph summarizes theoretical results concerning self-adaptive EAs and notes that the theoretical underpinnings for these methods are essentially unexplored. In particular, convergence theories that ensure convergence to a limit point on continuous spaces have only been developed by Rudolph, Hart, DeLaurentis and Ferguson, and Auger et al. In this paper, we illustrate how our analysis of a (1,{lambda})-ES for one-dimensional unimodal functions can be used to ensure convergence of a related ES on multidimensional functions. This (1,{lambda})-ES randomly selects a search dimension in each iteration, along which points generated. For a general class of separable functions, our analysis shows that the ES searches along each dimension independently, and thus this ES converges to the (global) minimum.
Blastwalls are often assumed to be the answer for facility protection from malevolent explosive assault, particularly from large vehicle bombs (LVB's). The assumption is made that the blastwall, if it is built strong enough to survive, will provide substantial protection to facilities and people on the side opposite the LVB. This paper will demonstrate through computer simulations and experimental data the behavior of explosively induced air blasts during interaction with blastwalls. It will be shown that air blasts can effectively wrap around and over blastwalls. Significant pressure reduction can be expected on the downstream side of the blastwall but substantial pressure will continue to propagate. The effectiveness of the blastwall to reduce blast overpressure depends on the geometry of the blastwall and the location of the explosive relative to the blastwall.
The purpose of this study was to investigate the impact of instructions on aircraft visual inspection performance and strategy. Forty-two inspectors from industry were asked to perform inspections of six areas of a Boeing 737. Six different instruction versions were developed for each inspection task, varying in the number and type of directed inspections. The amount of time spent inspecting, the number of calls made, and the number of the feedback calls detected all varied widely across the inspectors. However, inspectors who used instructions with a higher number of directed inspections referred to the instructions more often during and after the task, and found a higher percentage of a selected set of feedback cracks than inspectors using other instruction versions. This suggests that specific instructions can help overall inspection performance, not just performance on the defects specified. Further, instructions were shown to change the way an inspector approaches a task.
Memory may be the only system component that is more commoditized than a microprocessor. To simultaneously exploit this and address the impending memory wall, processing in memory (PIM) research efforts are considering ways to move processing into memory without significantly increasing the cost of the memory. As such, PIM devices may become the basis for future commodity clusters. Although these PIM devices may leverage new computational paradigms such as hardware support for multi-threading and traveling threads, they must provide support for legacy programming models if they are to supplant commodity clusters. This paper presents a prototype implementation of MPI over a traveling thread mechanism called parcels. A performance analysis indicates that the direct hardware support of a traveling thread model can lead to an efficient, lightweight MPI implementation.
Mechanisms for enhanced low-dose-rate sensitivity are described. In these mechanisms, bimolecular reactions dominate the kinetics at high dose rates thereby causing a sub-linear dependence on total dose, and this leads to a dose-rate dependence. These bimolecular mechanisms include electron-hole recombination, hydrogen recapture at hydrogen source sites, and hydrogen dimerization to form hydrogen molecules. The essence of each of these mechanisms is the dominance of the bimolecular reactions over the radiolysis reaction at high dose rates. However, at low dose rates, the radiolysis reaction dominates leading to a maximum effect of the radiation.
Properties of relevance for the equation of state for a high-density glass are discussed. We review the effects of failure waves, comminuted phase, and compaction on the validity of the Mie-Grueneisen EOS. The specific heat and the Grueneisen parameter at standard conditions for a {rho}{sub 0} = 5.085 g/cm{sup 3} glass ('Glass A') is then estimated to be 522 mJ/g/K and 0.1-0.3, respectively. The latter value is substantially smaller than the value of 2.1751 given in the SESAME tables for a high-density glass with {rho}{sub 0} = 5.46 g/cm{sup 3}. The present unusual value of the Grueneisen parameter is confirmed from the volume dependence determined from fitting the Mie-Grueneisen EOS to shock data in Ref. [2].
Submovements are hypothesized building blocks of human movement, discrete ballistic movements of which more complex movements are composed. Using a novel algorithm, submovements were extracted from the point-to-point movements of 41 persons recovering from stroke. Analysis of the extracted submovements showed that, over the course of therapy, patients' submovements tended to increase in peak speed and duration. The number of submovements employed to produce a given movement decreased. The time between the peaks of adjacent submovements decreased for inpatients (those less than 1 month post-stroke), but not for outpatients (those greater than 12 months post-stroke) as a group. Submovements became more overlapped for all patients, but more markedly for inpatients. The strength and consistency with which it quantified patients' recovery indicates that analysis of submovement overlap might be a useful tool for measuring learning or other changes in motor behavior in future human movement studies.
Proposed for publication in Coordinated & Multiple Views in Exploratory Visualization, Special Issue of Information Visualization Journal, Vol 2 No. 4, Palgrave/Macmillan.
PIM (Processor in Memory) architectures are being proposed for future supercomputers, because they reduce the problems that SMP MMPs have with latency. However, they do not meet the SMP MPP balance factors. Being relatively processor rich and memory starved, it is unclear whether an ASCI application could run on them, either as-is or with recoding. The KBA (Koch-Baker-Alcouffe) algorithm (Koch, 1992) for particle transport (radiation transport) is shown not to fit on PIMs as written. When redesigned with a 3-D allocation of cells to PIMs, the resulting algorithm is projected to execute an order of magnitude faster and more cost-effectively than the KBA algorithm, albeit with high initial hardware costs.
We present a series of electronic structure calculations that demonstrate a mechanism for spontaneous ionization of hydrogen at the Si-SiO{sub 2} interface. Specifically, we show that an isolated neutral hydrogen atom will spontaneously give up its charge and bond to a threefold coordinated oxygen atom. We refer to this entity as a proton. We have calculated the potential surface and found it to be entirely attractive. In contrast, hydrogen molecules will not undergo an analogous reaction. We relate these calculations both to proton generation experiments and to hydrogen plasma experiments.