Publications

Results 7401–7450 of 9,998

Search results

Jump to search filters

The potential, limitations, and challenges of divide and conquer quantum electronic structure calculations on energetic materials

Tucker, Jon R.; Magyar, Rudolph J.

High explosives are an important class of energetic materials used in many weapons applications. Even with modern computers, the simulation of the dynamic chemical reactions and energy release is exceedingly challenging. While the scale of the detonation process may be macroscopic, the dynamic bond breaking responsible for the explosive release of energy is fundamentally quantum mechanical. Thus, any method that does not adequately describe bonding is destined to lack predictive capability on some level. Performing quantum mechanics calculations on systems with more than dozens of atoms is a gargantuan task, and severe approximation schemes must be employed in practical calculations. We have developed and tested a divide and conquer (DnC) scheme to obtain total energies, forces, and harmonic frequencies within semi-empirical quantum mechanics. The method is intended as an approximate but faster solution to the full problem and is possible due to the sparsity of the density matrix in many applications. The resulting total energy calculation scales linearly as the number of subsystems, and the method provides a path-forward to quantum mechanical simulations of millions of atoms.

More Details

Formulating and analyzing multi-stage sensor placement problems

Water Distribution Systems Analysis 2010 - Proceedings of the 12th International Conference, WDSA 2010

Watson, Jean-Paul W.; Hart, William E.; Woodruff, David L.; Murray, Regan

The optimization of sensor placements is a key aspect of the design of contaminant warning systems for automatically detecting contaminants in water distribution systems. Although researchers have generally assumed that all sensors are placed at the same time, in practice sensor networks will likely grow and evolve over time. For example, limitations for a water utility's budget may dictate an staged, incremental deployment of sensors over many years. We describe optimization formulations of multi-stage sensor placement problems. The objective of these formulations includes an explicit trade-off between the value of the initially deployed and final sensor networks. This trade-off motivates the deployment of sensors in initial stages of the deployment schedule, even though these choices typically lead to a solution that is suboptimal when compared to placing all sensors at once. These multi-stage sensor placement problems can be represented as mixed-integer programs, and we illustrate the impact of this trade-off using standard commercial solvers. We also describe a multi-stage formulation that models budget uncertainty, expressed as a tree of potential budget scenarios through time. Budget uncertainty is used to assess and hedge against risks due to a potentially incomplete deployment of a planned sensor network. This formulation is a multi-stage stochastic mixed-integer program, which are notoriously difficult to solve. We apply standard commercial solvers to small-scale test problems, enabling us to effectively analyze multi-stage sensor placement problems subject to budget uncertainties, and assess the impact of accounting for such uncertainty relative to a deterministic multi-stage model. © 2012 ASCE.

More Details

Optimal determination of grab sample locations and source inversion in large-scale water distribution systems

Water Distribution Systems Analysis 2010 - Proceedings of the 12th International Conference, WDSA 2010

Wong, Angelica; Young, James; Laird, Carl D.; Hart, William E.; Mckenna, Sean A.

We present a mixed-integer linear programming formulation to determine optimal locations for manual grab sampling after the detection of contaminants in a water distribution system. The formulation selects optimal manual grab sample locations that maximize the total pair-wise distinguishability of candidate contamination events. Given an initial contaminant detection location, a source inversion is performed that will eliminate unlikely events resulting in a much smaller set of candidate contamination events. We then propose a cyclical process where optimal grab samples locations are determined and manual grab samples taken. Relying only on YES/NO indicators of the presence of contaminant, source inversion is performed to reduce the set of candidate contamination events. The process is repeated until the number of candidate events is sufficiently small. Case studies testing this process are presented using water network models ranging from 4 to approximately 13000 nodes. The results demonstrate that the contamination event can be identified within a remarkably small number of sampling cycles using very few sampling teams. Furthermore, solution times were reasonable making this formulation suitable for real-time settings. © 2012 ASCE.

More Details

Introducing the target-matrix paradigm for mesh optimization via node-movement

Engineering with Computers

Knupp, Patrick K.

A general-purpose algorithm for mesh optimization via node-movement, known as the Target-Matrix Paradigm, is introduced. The algorithm is general purpose in that it can be applied to a wide variety of mesh and element types, and to various commonly recurring mesh optimization problems such as shape improvement, and to more unusual problems like boundary-layer preservation with sliver removal, high-order mesh improvement, and edge-length equalization. The algorithm can be considered to be a direct optimization method in which weights are automatically constructed to enable definitions of application- specific mesh quality. The high-level concepts of the paradigm have been implemented in the Mesquite mesh improvement library, along with a number of concrete algorithms that address mesh quality issues such as those shown in the examples of the present paper. © Springer-Verlag (outside the USA) 2011.

More Details

A tunable, software-based DRAM error detection and correction library for HPC

Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics)

Fiala, David; Ferreira, Kurt; Mueller, Frank; Engelmann, Christian

Proposed exascale systems will present a number of considerable resiliency challenges. In particular, DRAM soft-errors, or bit-flips, are expected to greatly increase due to the increased memory density of these systems. Current hardware-based fault-tolerance methods will be unsuitable for addressing the expected soft error frequency rate. As a result, additional software will be needed to address this challenge. In this paper we introduce LIBSDC, a tunable, transparent silent data corruption detection and correction library for HPC applications. LIBSDC provides comprehensive SDC protection for program memory by implementing on-demand page integrity verification. Experimental benchmarks with Mantevo HPCCG show that once tuned, LIBSDC is able to achieve SDC protection with 50% overhead of resources, less than the 100% needed for double modular redundancy. © 2012 Springer-Verlag Berlin Heidelberg.

More Details

Stabilization and scalable block preconditioning for the Navier-Stokes equations

Journal of Computational Physics

Cyr, Eric C.; Shadid, John N.; Tuminaro, Raymond S.

This study compares several block-oriented preconditioners for the stabilized finite element discretization of the incompressible Navier-Stokes equations. This includes standard additive Schwarz domain decomposition methods, aggressive coarsening multigrid, and three preconditioners based on an approximate block LU factorization, specifically SIMPLEC, LSC, and PCD. Robustness is considered with a particular focus on the impact that different stabilization methods have on preconditioner performance. Additionally, parallel scaling studies are undertaken. The numerical results indicate that aggressive coarsening multigrid, LSC and PCD all have good algorithmic scalability. Coupling this with the fact that block methods can be applied to systems arising from stable mixed discretizations implies that these techniques are a promising direction for developing scalable methods for Navier-Stokes. © 2011.

More Details

New density functional theory approaches for enabling prediction of chemical and physical properties of plutonium and other actinides

Wills, Ann E.

Density Functional Theory (DFT) based Equation of State (EOS) construction is a prominent part of Sandia's capabilities to support engineering sciences. This capability is based on amending experimental data with information gained from computational investigations, in parts of the phase space where experimental data is hard, dangerous, or expensive to obtain. A prominent materials area where such computational investigations are hard to perform today because of limited accuracy is actinide and lanthanide materials. The Science of Extreme Environment Lab Directed Research and Development project described in this Report has had the aim to cure this accuracy problem. We have focused on the two major factors which would allow for accurate computational investigations of actinide and lanthanide materials: (1) The fully relativistic treatment needed for materials containing heavy atoms, and (2) the needed improved performance of DFT exchange-correlation functionals. We have implemented a fully relativistic treatment based on the Dirac Equation into the LANL code RSPt and we have shown that such a treatment is imperative when calculating properties of materials containing actinides and/or lanthanides. The present standard treatment that only includes some of the relativistic terms is not accurate enough and can even give misleading results. Compared to calculations previously considered state of the art, the Dirac treatment gives a substantial change in equilibrium volume predictions for materials with large spin-orbit coupling. For actinide and lanthanide materials, a Dirac treatment is thus a fundamental requirement in any computational investigation, including those for DFT-based EOS construction. For a full capability, a DFT functional capable of describing strongly correlated systems such as actinide materials need to be developed. Using the previously successful subsystem functional scheme developed by Mattsson et.al., we have created such a functional. In this functional the Harmonic Oscillator Gas is providing the necessary reference system for the strong correlation and localization occurring in actinides. Preliminary testing shows that the new Hao-Armiento-Mattsson (HAM) functional gives a trend towards improved results for the crystalline copper oxide test system we have chosen. This test system exhibits the same exchange-correlation physics as the actinide systems do, but without the relativistic effects, giving access to a pure testing ground for functionals. During the work important insights have been gained. An example is that currently available functionals, contrary to common belief, make large errors in so called hybridization regions where electrons from different ions interact and form new states. Together with the new understanding of functional issues, the Dirac implementation into the RSPt code will permit us to gain more fundamental understanding, both quantitatively and qualitatively, of materials of importance for Sandia and the rest of the Nuclear Weapons complex.

More Details

Modeling and Optimization of Superstructure-based Stochastic Programs for Risk-aware Decision Support

Computer Aided Chemical Engineering

Siirola, John D.; Watson, Jean-Paul W.

This manuscript presents a unified software framework for modeling and optimizing large-scale engineered systems with uncertainty. We propose a Python-based " block- oriented" modeling approach for representing the discrete components within the system. Through the use of a modeling components library, the block-oriented approach facilitates a clean separation of system superstructure from the details of individual components. This approach also lends itself naturally to expressing design and operational decisions as disjunctive expressions over the component blocks. We then apply a Python- based risk and uncertainty analysis library that leverages the explicit representation of the mathematical program in Python to automatically expand the deterministic system model into a multi-stage stochastic program, which can then be solved either directly or via decomposition-based solution strategies. This manuscript demonstrates the application of this modeling approach for risk-aware analysis of an electric distribution system. © 2012 Elsevier B.V.

More Details

Modeling reactive transport in deformable porous media using the theory of interacting continua

Turner, Daniel Z.

This report gives an overview of the work done as part of an Early Career LDRD aimed at modeling flow induced damage of materials involving chemical reactions, deformation of the porous matrix, and complex flow phenomena. The numerical formulation is motivated by a mixture theory or theory of interacting continua type approach to coupling the behavior of the fluid and the porous matrix. Results for the proposed method are presented for several engineering problems of interest including carbon dioxide sequestration, hydraulic fracturing, and energetic materials applications. This work is intended to create a general framework for flow induced damage that can be further developed in each of the particular areas addressed below. The results show both convincing proof of the methodologies potential and the need for further validation of the models developed.

More Details

Optimization of large-scale heterogeneous system-of-systems models

Gray, Genetha A.; Hart, William E.; Hough, Patricia D.; Parekh, Ojas D.; Phillips, Cynthia A.; Siirola, John D.; Swiler, Laura P.; Watson, Jean-Paul W.

Decision makers increasingly rely on large-scale computational models to simulate and analyze complex man-made systems. For example, computational models of national infrastructures are being used to inform government policy, assess economic and national security risks, evaluate infrastructure interdependencies, and plan for the growth and evolution of infrastructure capabilities. A major challenge for decision makers is the analysis of national-scale models that are composed of interacting systems: effective integration of system models is difficult, there are many parameters to analyze in these systems, and fundamental modeling uncertainties complicate analysis. This project is developing optimization methods to effectively represent and analyze large-scale heterogeneous system of systems (HSoS) models, which have emerged as a promising approach for describing such complex man-made systems. These optimization methods enable decision makers to predict future system behavior, manage system risk, assess tradeoffs between system criteria, and identify critical modeling uncertainties.

More Details

TriBITS lifecycle model. Version 1.0, a lean/agile software lifecycle model for research-based computational science and engineering and applied mathematical software

Willenbring, James M.; Heroux, Michael A.

Software lifecycles are becoming an increasingly important issue for computational science and engineering (CSE) software. The process by which a piece of CSE software begins life as a set of research requirements and then matures into a trusted high-quality capability is both commonplace and extremely challenging. Although an implicit lifecycle is obviously being used in any effort, the challenges of this process - respecting the competing needs of research vs. production - cannot be overstated. Here we describe a proposal for a well-defined software lifecycle process based on modern Lean/Agile software engineering principles. What we propose is appropriate for many CSE software projects that are initially heavily focused on research but also are expected to eventually produce usable high-quality capabilities. The model is related to TriBITS, a build, integration and testing system, which serves as a strong foundation for this lifecycle model, and aspects of this lifecycle model are ingrained in the TriBITS system. Here, we advocate three to four phases or maturity levels that address the appropriate handling of many issues associated with the transition from research to production software. The goals of this lifecycle model are to better communicate maturity levels with customers and to help to identify and promote Software Engineering (SE) practices that will help to improve productivity and produce better software. An important collection of software in this domain is Trilinos, which is used as the motivation and the initial target for this lifecycle model. However, many other related and similar CSE (and non-CSE) software projects can also make good use of this lifecycle model, especially those that use the TriBITS system. Indeed this lifecycle process, if followed, will enable large-scale sustainable integration of many complex CSE software efforts across several institutions.

More Details

A Stochastic Programming Formulation for Disinfectant Booster Station Placement to Protect Large-Scale Water Distribution Systems

Computer Aided Chemical Engineering

Hackebeil, Gabriel A.; Mann, Angelica V.; Hart, William E.; Klise, Katherine A.; Laird, Carl D.

We present a methodology for optimally locating disinfectant booster stations for response to contamination events in water distribution systems. A stochastic programming problem considering uncertainty in both the location and time of the contamination event is formulated resulting in an extensive form that is equivalent to the weighted maximum coverage problem. Although the original full-space problem is intractably large, we show a series of reductions that reduce the size of the problem by five orders of magnitude and allow solutions of the optimal placement problem for realistically sized water network models. © 2012 Elsevier B.V.

More Details
Results 7401–7450 of 9,998
Results 7401–7450 of 9,998