Publications

Results 9951–9975 of 9,998

Search results

Jump to search filters

Welding Behavior of Free Machining Stainless Steel

Welding Journal Research Supplement

Brooks, John A.; Robino, Charles V.; Headley, Thomas J.; Michael, Joseph R.

The weld solidification and cracking behavior of sulfur bearing free machining austenitic stainless steel was investigated for both gas-tungsten arc (GTA) and pulsed laser beam weld processes. The GTA weld solidification was consistent with those predicted with existing solidification diagrams and the cracking response was controlled primarily by solidification mode. The solidification behavior of the pulsed laser welds was complex, and often contained regions of primary ferrite and primary austenite solidification, although in all cases the welds were found to be completely austenite at room temperature. Electron backscattered diffraction (EBSD) pattern analysis indicated that the nature of the base metal at the time of solidification plays a primary role in initial solidification. The solid state transformation of austenite to ferrite at the fusion zone boundary, and ferrite to austenite on cooling may both be massive in nature. A range of alloy compositions that exhibited good resistance to solidification cracking and was compatible with both welding processes was identified. The compositional range is bounded by laser weldability at lower Cr{sub eq}/Ni{sub eq} ratios and by the GTA weldability at higher ratios. It was found with both processes that the limiting ratios were somewhat dependent upon sulfur content.

More Details

What Makes a Beam Shaping Problem Difficult?

Romero, L.A.; Dickey, Fred M.

The authors have discussed the three factors that they believe are the most important in determining the difficulty of a beam shaping problem: scaling, smoothness, and coherence. The arguments have been almost completely based on considering how these factors influence beam shaping lenses that were designed using geometrical optics. However, they believe that these factors control the difficulty of beam shaping problems even if one does not base ones design strategy on geometrical optics. For example, they have shown that a lens designed using geometrical optics will not work well unless {beta} is large. However, they have also shown that if {beta} is small the uncertainty principle shows that it is impossible to do a good job of beam shaping no matter how one designs ones lens.

More Details

Unconstrained and Constrained Minimization, Linear Scaling, and the Grassmann Manifold: Theory and Applications

Physical Review B

Lippert, Ross A.; Schultz, Peter A.

An unconstrained minimization algorithm for electronic structure calculations using density functional for systems with a gap is developed to solve for nonorthogonal Wannier-like orbitals in the spirit of E. B. Stechel, A. R. Williams, and P. J. Feibelman, Phys. Rev. B 49, 10,008 (1994). The search for the occupied sub-space is a Grassmann conjugate gradient algorithm generalized from the algorithm of A. Edelman, T.A. Arias, and S. T. Smith, SIAM J. on Matrix Anal. Appl. 20, 303 (1998). The gradient takes into account the nonorthogonality of a local atom-centered basis, gaussian in their implementation. With a localization constraint on the Wannier-like orbitals, well-constructed sparse matrix multiplies lead to O(N) scaling of the computationally intensive parts of the algorithm. Using silicon carbide as a test system, the accuracy, convergence, and implementation of this algorithm as a quantitative alternative to diagonalization are investigated. Results up to 1,458 atoms on a single processor are presented.

More Details

Dynamics of Exchange at Gas-Zeolite Interfaces 1: Pure Component n-Butane and Isobutane

Jornal of Physical Chemistry B

Chandross, M.; Webb, Edmund B.; Grest, Gary S.; Martin, Marcus G.; Thompson, Aidan P.

The authors present the results of molecular dynamics simulations of n-butane and isobutane in silicalite. They begin with a comparison of the bulk adsorption and diffusion properties for two different parameterizations of the interaction potential between the hydrocarbon species, both of which have been shown to reproduce experimental gas-liquid coexistence curves. They examine diffusion as a function of the loading of the zeolite, as well as the temperature dependence of the diffusion constant at loading and for infinite dilution. They continue with simulations in which interfaces are formed between single component gases and the zeolite. After reaching equilibrium, they examine the dynamics of exchange between the bulk gas and the zeolite. Finally, they calculate the permeability of the zeolite for n-butane and isobutane as a function of pressure. Their simulations are performed for a number of different gas temperatures and pressures, covering a wide range of state points.

More Details

Cooperative sentry vehicles and differential GPS leapfrog

Feddema, John T.; Lewis, Christopher L.; Lafarge, Robert A.

As part of a project for the Defense Advanced Research Projects Agency, Sandia National Laboratories Intelligent Systems and Robotics Center is developing and testing the feasibility of using a cooperative team of robotic sentry vehicles to guard a perimeter, perform a surround task, and travel extended distances. This paper describes the authors most recent activities. In particular, this paper highlights the development of a Differential Global Positioning System (DGPS) leapfrog capability that allows two or more vehicles to alternate sending DGPS corrections. Using this leapfrog technique, this paper shows that a group of autonomous vehicles can travel 22.68 kilometers with a root mean square positioning error of only 5 meters.

More Details

Evolutionary pattern search algorithms for unconstrained and linearly constrained optimization

IEEE Transactions on Evolutionary Computation

Hart, William E.

The authors describe a convergence theory for evolutionary pattern search algorithms (EPSAs) on a broad class of unconstrained and linearly constrained problems. EPSAs adaptively modify the step size of the mutation operator in response to the success of previous optimization steps. The design of EPSAs is inspired by recent analyses of pattern search methods. The analysis significantly extends the previous convergence theory for EPSAs. The analysis applies to a broader class of EPSAs,and it applies to problems that are nonsmooth, have unbounded objective functions, and which are linearly constrained. Further, they describe a modest change to the algorithmic framework of EPSAs for which a non-probabilistic convergence theory applies. These analyses are also noteworthy because they are considerably simpler than previous analyses of EPSAs.

More Details

Code Verification by the Method of Manufactured Solutions

Salari, Kambiz S.; Knupp, Patrick K.

A procedure for code Verification by the Method of Manufactured Solutions (MMS) is presented. Although the procedure requires a certain amount of creativity and skill, we show that MMS can be applied to a variety of engineering codes which numerically solve partial differential equations. This is illustrated by detailed examples from computational fluid dynamics. The strength of the MMS procedure is that it can identify any coding mistake that affects the order-of-accuracy of the numerical method. A set of examples which use a blind-test protocol demonstrates the kinds of coding mistakes that can (and cannot) be exposed via the MMS code Verification procedure. The principle advantage of the MMS procedure over traditional methods of code Verification is that code capabilities are tested in full generality. The procedure thus results in a high degree of confidence that all coding mistakes which prevent the equations from being solved correctly have been identified.

More Details

Load balancing fictions, falsehoods and fallacies

Applied Mathematical Modeling

Hendrickson, Bruce A.

Effective use of a parallel computer requires that a calculation be carefully divided among the processors. This load balancing problem appears in many guises and has been a fervent area of research for the past decade or more. Although great progress has been made, and useful software tools developed, a number of challenges remain. It is the conviction of the author that these challenges will be easier to address if programmers first come to terms with some significant shortcomings in their current perspectives. This paper tries to identify several areas in which the prevailing point of view is either mistaken or insufficient. The goal is to motivate new ideas and directions for this important field.

More Details

Interprocessor communication with memory constraints

Hendrickson, Bruce A.; Hendrickson, Bruce A.

Many parallel applications require periodic redistribution of workloads and associated data. In a distributed memory computer, this redistribution can be difficult if limited memory is available for receiving messages. The authors propose a model for optimizing the exchange of messages under such circumstances which they call the minimum phase remapping problem. They first show that the problem is NP-Complete, and then analyze several methodologies for addressing it. First, they show how the problem can be phrased as an instance of multi-commodity flow. Next, they study a continuous approximation to the problem. They show that this continuous approximation has a solution which requires at most two more phases than the optimal discrete solution, but the question of how to consistently obtain a good discrete solution from the continuous problem remains open. Finally, they devise a simple and practical approximation algorithm for the problem with a bound of 1.5 times the optimal number of phases.

More Details

Solving complex-valued linear systems via equivalent real formulations

SIAM Journal of Scientific Computing

Day, David M.; Heroux, Michael A.

Most algorithms used in preconditioned iterative methods are generally applicable to complex valued linear systems, with real valued linear systems simply being a special case. However, most iterative solver packages available today focus exclusively on real valued systems, or deal with complex valued systems as an afterthought. One obvious approach to addressing this problem is to recast the complex problem into one of a several equivalent real forms and then use a real valued solver to solve the related system. However, well-known theoretical results showing unfavorable spectral properties for the equivalent real forms have diminished enthusiasm for this approach. At the same time, experience has shown that there are situations where using an equivalent real form can be very effective. In this paper, the authors explore this approach, giving both theoretical and experimental evidence that an equivalent real form can be useful for a number of practical situations. Furthermore, they show that by making good use of some of the advance features of modem solver packages, they can easily generate equivalent real form preconditioners that are computationally efficient and mathematically identical to their complex counterparts. Using their techniques, they are able to solve very ill-conditioned complex valued linear systems for a variety of large scale applications. However, more importantly, they shed more light on the effectiveness of equivalent real forms and more clearly delineate how and when they should be used.

More Details

Direct simulation of particle-laden fluids

Cook, Benjamin K.; Noble, David R.; Preece, Dale S.

Processes that involve particle-laden fluids are common in geomechanics and especially in the petroleum industry. Understanding the physics of these processes and the ability to predict their behavior requires the development of coupled fluid-flow and particle-motion computational methods. This paper outlines an accurate and robust coupled computational scheme using the lattice-Boltzmann method for fluid flow and the discrete-element method for solid particle motion. Results from several two-dimensional validation simulations are presented. Simulations reported include the sedimentation of an ellipse, a disc and two interacting discs in a closed column of fluid. The recently discovered phenomenon of drafting, kissing, and tumbling is fully reproduced in the two-disc simulation.

More Details

A naturalistic decision making model for simulated human combatants

Hunter, Keith O.; Hart, William E.; Forsythe, James C.

The authors describe a naturalistic behavioral model for the simulation of small unit combat. This model, Klein's recognition-primed decision making (RPD) model, is driven by situational awareness rather than a rational process of selecting from a set of action options. They argue that simulated combatants modeled with RPD will have more flexible and realistic responses to a broad range of small-scale combat scenarios. Furthermore, they note that the predictability of a simulation using an RPD framework can be easily controlled to provide multiple evaluations of a given combat scenario. Finally, they discuss computational issues for building an RPD-based behavior engine for fully automated combatants in small conflict scenarios, which are being investigated within Sandia's Next Generation Site Security project.

More Details

Materials Issues for Micromachines Development - ASCI Program Plan

Fang, H.E.; Miller, Samuel L.; Dugger, Michael T.; Prasad, Somuri V.; Reedy, Earl D.; Thompson, Aidan P.; Wong, Chungnin C.; Yang, Pin Y.; Battaile, Corbett C.; Benavides, Gilbert L.; Ensz, M.T.; Buchheit, Thomas E.; Lavan, David A.; Chen, Er-Ping C.; Christenson, Todd R.; De Boer, Maarten P.

This report summarizes materials issues associated with advanced micromachines development at Sandia. The intent of this report is to provide a perspective on the scope of the issues and suggest future technical directions, with a focus on computational materials science. Materials issues in surface micromachining (SMM), Lithographic-Galvanoformung-Abformung (LIGA: lithography, electrodeposition, and molding), and meso-machining technologies were identified. Each individual issue was assessed in four categories: degree of basic understanding; amount of existing experimental data capability of existing models; and, based on the perspective of component developers, the importance of the issue to be resolved. Three broad requirements for micromachines emerged from this process. They are: (1) tribological behavior, including stiction, friction, wear, and the use of surface treatments to control these, (2) mechanical behavior at microscale, including elasticity, plasticity, and the effect of microstructural features on mechanical strength, and (3) degradation of tribological and mechanical properties in normal (including aging), abnormal and hostile environments. Resolving all the identified critical issues requires a significant cooperative and complementary effort between computational and experimental programs. The breadth of this work is greater than any single program is likely to support. This report should serve as a guide to plan micromachines development at Sandia.

More Details

Application of finite element, global polynomial, and kriging response surfaces in Progressive Lattice Sampling designs

Romero, Vicente J.; Swiler, Laura P.; Giunta, Anthony A.

This paper examines the modeling accuracy of finite element interpolation, kriging, and polynomial regression used in conjunction with the Progressive Lattice Sampling (PLS) incremental design-of-experiments approach. PLS is a paradigm for sampling a deterministic hypercubic parameter space by placing and incrementally adding samples in a manner intended to maximally reduce lack of knowledge in the parameter space. When combined with suitable interpolation methods, PLS is a formulation for progressive construction of response surface approximations (RSA) in which the RSA are efficiently upgradable, and upon upgrading, offer convergence information essential in estimating error introduced by the use of RSA in the problem. The three interpolation methods tried here are examined for performance in replicating an analytic test function as measured by several different indicators. The process described here provides a framework for future studies using other interpolation schemes, test functions, and measures of approximation quality.

More Details

Scalable rendering on PC clusters

Wylie, Brian N.; Lewis, Vasily L.; Shirley, David N.; Pavlakos, Constantine P.

This case study presents initial results from research targeted at the development of cost-effective scalable visualization and rendering technologies. The implementations of two 3D graphics libraries based on the popular sort-last and sort-middle parallel rendering techniques are discussed. An important goal of these implementations is to provide scalable rendering capability for extremely large datasets (>> 5 million polygons). Applications can use these libraries for either run-time visualization, by linking to an existing parallel simulation, or for traditional post-processing by linking to an interactive display program. The use of parallel, hardware-accelerated rendering on commodity hardware is leveraged to achieve high performance. Current performance results show that, using current hardware (a small 16-node cluster), they can utilize up to 85% of the aggregate graphics performance and achieve rendering rates in excess of 20 million polygons/second using OpenGL{reg_sign} with lighting, Gouraud shading, and individually specified triangles (not t-stripped).

More Details

Algebraic mesh quality metrics

SIAM Journal of Scientific Computing

Knupp, Patrick K.

Quality metrics for structured and unstructured mesh generation are placed within an algebraic framework to form a mathematical theory of mesh quality metrics. The theory, based on the Jacobian and related matrices, provides a means of constructing, classifying, and evaluating mesh quality metrics. The Jacobian matrix is factored into geometrically meaningful parts. A nodally-invariant Jacobian matrix can be defined for simplicial elements using a weight matrix derived from the Jacobian matrix of an ideal reference element. Scale and orientation-invariant algebraic mesh quality metrics are defined. the singular value decomposition is used to study relationships between metrics. Equivalence of the element condition number and mean ratio metrics is proved. Condition number is shown to measure the distance of an element to the set of degenerate elements. Algebraic measures for skew, length ratio, shape, volume, and orientation are defined abstractly, with specific examples given. Combined metrics for shape and volume, shape-volume-orientation are algebraically defined and examples of such metrics are given. Algebraic mesh quality metrics are extended to non-simplical elements. A series of numerical tests verify the theoretical properties of the metrics defined.

More Details

Scalability limitations of VIA-based technologies in supporting MPI

Brightwell, Ronald B.; Maccabe, Arthur B.

This paper analyzes the scalability limitations of networking technologies based on the Virtual Interface Architecture (VIA) in supporting the runtime environment needed for an implementation of the Message Passing Interface. The authors present an overview of the important characteristics of VIA and an overview of the runtime system being developed as part of the Computational Plant (Cplant) project at Sandia National Laboratories. They discuss the characteristics of VIA that prevent implementations based on this system to meet the scalability and performance requirements of Cplant.

More Details

Salinas - An implicit finite element structural dynamics code developed for massively parallel platforms

Reese, Garth M.; Driessen, Brian D.; Alvin, Kenneth F.; Day, David M.

As computational needs for structural finite element analysis increase, a robust implicit structural dynamics code is needed which can handle millions of degrees of freedom in the model and produce results with quick turn around time. A parallel code is needed to avoid limitations of serial platforms. Salinas is an implicit structural dynamics code specifically designed for massively parallel platforms. It computes the structural response of very large complex structures and provides solutions faster than any existing serial machine. This paper gives a current status of Salinas and uses demonstration problems to show Salinas' performance.

More Details

Computational methods for coupling microstructural and micromechanical materials response simulations

Holm, Elizabeth A.; Wellman, Gerald W.; Battaile, Corbett C.; Buchheit, Thomas E.; Fang, H.E.; Rintoul, Mark D.; Vedula, Venkata R.; Glass, Sarah J.; Knorovsky, Gerald A.; Neilsen, Michael K.

Computational materials simulations have traditionally focused on individual phenomena: grain growth, crack propagation, plastic flow, etc. However, real materials behavior results from a complex interplay between phenomena. In this project, the authors explored methods for coupling mesoscale simulations of microstructural evolution and micromechanical response. In one case, massively parallel (MP) simulations for grain evolution and microcracking in alumina stronglink materials were dynamically coupled. In the other, codes for domain coarsening and plastic deformation in CuSi braze alloys were iteratively linked. this program provided the first comparison of two promising ways to integrate mesoscale computer codes. Coupled microstructural/micromechanical codes were applied to experimentally observed microstructures for the first time. In addition to the coupled codes, this project developed a suite of new computational capabilities (PARGRAIN, GLAD, OOF, MPM, polycrystal plasticity, front tracking). The problem of plasticity length scale in continuum calculations was recognized and a solution strategy was developed. The simulations were experimentally validated on stockpile materials.

More Details

A case study in working with cell-centered data

Crossno, Patricia J.

This case study provides examples of how some simple decisions the authors made in structuring their algorithms for handling cell-centered data can dramatically influence the results. Although they all know that these decisions produce variations in results, they think that they underestimate the potential magnitude of the differences. More importantly, the users of the codes may not be aware that these choices have been made or what they mean to the resulting visualizations of their data. This raises the question of whether or not these decisions are inadvertently distorting user interpretations of data sets.

More Details

An agent-based microsimulation of critical infrastructure systems

Barton, Dianne C.; Stamber, Kevin L.

US infrastructures provide essential services that support the economic prosperity and quality of life. Today, the latest threat to these infrastructures is the increasing complexity and interconnectedness of the system. On balance, added connectivity will improve economic efficiency; however, increased coupling could also result in situations where a disturbance in an isolated infrastructure unexpectedly cascades across diverse infrastructures. An understanding of the behavior of complex systems can be critical to understanding and predicting infrastructure responses to unexpected perturbation. Sandia National Laboratories has developed an agent-based model of critical US infrastructures using time-dependent Monte Carlo methods and a genetic algorithm learning classifier system to control decision making. The model is currently under development and contains agents that represent the several areas within the interconnected infrastructures, including electric power and fuel supply. Previous work shows that agent-based simulations models have the potential to improve the accuracy of complex system forecasting and to provide new insights into the factors that are the primary drivers of emergent behaviors in interdependent systems. Simulation results can be examined both computationally and analytically, offering new ways of theorizing about the impact of perturbations to an infrastructure network.

More Details

Salvo: Seismic imaging software for complex geologies

Ober, Curtis C.; Womble, David E.

This report describes Salvo, a three-dimensional seismic-imaging software for complex geologies. Regions of complex geology, such as overthrusts and salt structures, can cause difficulties for many seismic-imaging algorithms used in production today. The paraxial wave equation and finite-difference methods used within Salvo can produce high-quality seismic images in these difficult regions. However this approach comes with higher computational costs which have been too expensive for standard production. Salvo uses improved numerical algorithms and methods, along with parallel computing, to produce high-quality images and to reduce the computational and the data input/output (I/O) costs. This report documents the numerical algorithms implemented for the paraxial wave equation, including absorbing boundary conditions, phase corrections, imaging conditions, phase encoding, and reduced-source migration. This report also describes I/O algorithms for large seismic data sets and images and parallelization methods used to obtain high efficiencies for both the computations and the I/O of seismic data sets. Finally, this report describes the required steps to compile, port and optimize the Salvo software, and describes the validation data sets used to help verify a working copy of Salvo.

More Details
Results 9951–9975 of 9,998
Results 9951–9975 of 9,998