Publications

Results 5401–5600 of 9,998

Search results

Jump to search filters

Motivation and Design of the Sirocco Storage System Version 1.0

Curry, Matthew L.; Ward, Harry L.; Danielson, Geoffrey C.

Sirocco is a massively parallel, high performance storage system for the exascale era. It emphasizes client-to-client coordination, low server-side coupling, and free data movement to improve resilience and performance. Its architecture is inspired by peer-to-peer and victim- cache architectures. By leveraging these ideas, Sirocco natively supports several media types, including RAM, flash, disk, and archival storage, with automatic migration between levels. Sirocco also includes storage interfaces and support that are more advanced than typical block storage. Sirocco enables clients to efficiently use key-value storage or block-based storage with the same interface. It also provides several levels of transactional data updates within a single storage command, including full ACID-compliant updates. This transaction support extends to updating several objects within a single transaction. Further support is provided for con- currency control, enabling greater performance for workloads while providing safe concurrent modification. By pioneering these and other technologies and techniques in the storage system, Sirocco is poised to fulfill a need for a massively scalable, write-optimized storage system for exascale systems. This is version 1.0 of a document reflecting the current and planned state of Sirocco. Further versions of this document will be accessible at http://www.cs.sandia.gov/Scalable_IO/ sirocco .

More Details

Adiabatic release measurements in aluminum between 400 and 1200 GPa: Characterization of aluminum as a shock standard in the multimegabar regime

Physical Review B - Condensed Matter and Materials Physics

Knudson, Marcus D.; Desjarlais, Michael P.; Pribram-Jones, Aurora P.

Aluminum has been used prolifically as an impedance matching standard in the multimegabar regime (1 Mbar = 100 GPa), particularly in nuclear driven, early laser driven, and early magnetically driven flyer plate experiments. The accuracy of these impedance matching measurements depends upon the knowledge of both the Hugoniot and release or reshock response of aluminum. Here, we present the results of several adiabatic release measurements of aluminum from ∼400-1200 GPa states along the principal Hugoniot using full density polymethylpentene (commonly known as TPX), and both ∼190 and ∼110 mg/cc silica aerogel standards. These data were analyzed within the framework of a simple, analytical model that was motivated by a first-principles molecular dynamics investigation into the release response of aluminum, as well as by a survey of the release response determined from several tabular equations of state for aluminum. Combined, this theoretical and experimental study provides a method to perform impedance matching calculations without the need to appeal to any tabular equation of state for aluminum. As an analytical model, this method allows for propagation of all uncertainty, including the random measurement uncertainties and the systematic uncertainties of the Hugoniot and release response of aluminum. This work establishes aluminum for use as a high-precision standard for impedance matching in the multimegabar regime.

More Details

Achieving performance isolation with lightweight co-kernels

HPDC 2015 - Proceedings of the 24th International Symposium on High-Performance Parallel and Distributed Computing

Ouyang, Jiannan; Kocoloski, Brian; Lange, John; Laros, James H.

Performance isolation is emerging as a requirement for High Performance Computing (HPC) applications, particularly as HPC architectures turn to in situ data processing and application composition techniques to increase system throughput. These approaches require the co-location of disparate workloads on the same compute node, each with different resource and runtime requirements. In this paper we claim that these workloads cannot be effectively managed by a single Operating System/Runtime (OS/R). Therefore, we present Pisces, a system software architecture that enables the co-existence of multiple independent and fully isolated OS/Rs, or enclaves, that can be customized to address the disparate requirements of next generation HPC workloads. Each enclave consists of a specialized lightweight OS cokernel and runtime, which is capable of independently managing partitions of dynamically assigned hardware resources. Contrary to other co-kernel approaches, in this work we consider performance isolation to be a primary requirement and present a novel co-kernel architecture to achieve this goal. We further present a set of design requirements necessary to ensure performance isolation, including: (1) elimination of cross OS dependencies, (2) internalized management of I/O, (3) limiting cross enclave communication to explicit shared memory channels, and (4) using virtualization techniques to provide missing OS features. The implementation of the Pisces co-kernel architecture is based on the Kitten Lightweight Kernel and Palacios Virtual Machine Monitor, two system software architectures designed specifically for HPC systems. Finally we will show that lightweight isolated co-kernels can provide better performance for HPC applications, and that isolated virtual machines are even capable of outperforming native environments in the presence of competing workloads.

More Details

Exploring failure recovery for stencil-based applications at extreme scales

HPDC 2015 - Proceedings of the 24th International Symposium on High-Performance Parallel and Distributed Computing

Gamell Balmana, Marc G.; Teranishi, Keita T.; Heroux, Michael A.; Mayo, Jackson M.; Kolla, Hemanth K.; Chen, Jacqueline H.; Parashar, Manish

Application resilience is a key challenge that must be ad-dressed in order to realize the exascale vision. Previous work has shown that online recovery, even when done in a global manner (i.e., involving all processes), can dramatically re-duce the overhead of failures when compared to the more traditional approach of terminating the job and restarting it from the last stored checkpoint. In this paper we suggest going one step further, and explore how local recovery can be used for certain classes of applications to reduce the over-heads due to failures. Specifically we study the feasibility of local recovery for stencil-based parallel applications and we show how multiple independent failures can be masked to effectively reduce the impact on the total time to solution.

More Details

PaCMap: Topology mapping of unstructured communication patterns onto non-contiguous allocations

Proceedings of the International Conference on Supercomputing

Tuncer, Ozan; Leung, Vitus J.; Coskun, Ayse K.

In high performance computing (HPC), applications usually have many parallel tasks running on multiple machine nodes. As these tasks intensively communicate with each other, the communication overhead has a significant impact on an application's execution time. This overhead is determined by the application's communication pattern as well as the network distances between communicating tasks. By mapping the tasks to the available machine nodes in a communication-aware manner, the network distances and the execution times can be significantly reduced. Existing techniques first allocate available nodes to an application, and then map the tasks onto the allocated nodes. In this paper, we discuss the potential benefits of simultaneous allocation and mapping for applications with irregular communication patterns. We also propose a novel graphbased allocation and mapping technique to reduce the execution time in HPC machines that use non-contiguous allocation, such as Cray XK series. Simulations calibrated with real-life experiments show that our technique reduces hop-bytes up to 30% compared to the state-of-the-art.

More Details

ACM TOMS replicated computational results initiative

ACM Transactions on Mathematical Software

Heroux, Michael A.

In this study, the scientific community relies on the peer review process for assuring the quality of published material, the goal of which is to build a body of work we can trust. Computational journals such as The ACM Transactions on Mathematical Software (TOMS) use this process for rigorously promoting the clarity and completeness of content, and citation of prior work. At the same time, it is unusual to independently confirm computational results.

More Details

Decreasing the temporal complexity for nonlinear, implicit reduced-order models by forecasting

Computer Methods in Applied Mechanics and Engineering

Carlberg, Kevin T.; Ray, Jaideep R.; van Bloemen Waanders, Bart G.

Implicit numerical integration of nonlinear ODEs requires solving a system of nonlinear algebraic equations at each time step. Each of these systems is often solved by a Newton-like method, which incurs a sequence of linear-system solves. Most model-reduction techniques for nonlinear ODEs exploit knowledge of a system's spatial behavior to reduce the computational complexity of each linear-system solve. However, the number of linear-system solves for the reduced-order simulation often remains roughly the same as that for the full-order simulation.We propose exploiting knowledge of the model's temporal behavior to (1) forecast the unknown variable of the reduced-order system of nonlinear equations at future time steps, and (2) use this forecast as an initial guess for the Newton-like solver during the reduced-order-model simulation. To compute the forecast, we propose using the Gappy POD technique. The goal is to generate an accurate initial guess so that the Newton solver requires many fewer iterations to converge, thereby decreasing the number of linear-system solves in the reduced-order-model simulation.

More Details

Simulation and optimization of HPC job allocation for jointly reducing communication and cooling costs

Sustainable Computing: Informatics and Systems

Meng, Jie; McCauley, Samuel; Kaplan, Fulya; Leung, Vitus J.; Coskun, Ayse K.

Performance and energy are critical aspects in high performance computing (HPC) data centers. Highly parallel HPC applications that require multiple nodes usually run for long durations in the range of minutes, hours or days. As the threads of parallel applications communicate with each other intensively, the communication cost of these applications has a significant impact on data center performance. Energy consumption has also become a first-order constraint of HPC data centers. Nearly half of the energy in the computing clusters today is consumed by the cooling infrastructure. Existing job allocation policies either target improving the system performance or reducing the cooling energy cost of the server nodes. How to optimize the system performance while minimizing the cooling energy consumption is still an open question. This paper proposes a job allocation methodology aimed at jointly reducing the communication cost and the cooling energy of HPC data centers. In order to evaluate and validate our optimization algorithm, we implement our joint job allocation methodology in the structural simulation toolkit (SST) - a simulation framework for large-scale data centers. We evaluate our joint optimization algorithm using traces extracted from real-world workloads. Experimental results show that, in comparison to performance-aware job allocation algorithms, our algorithm achieves comparable running times and reduces the cooling power by up to 42.21% across all the jobs.

More Details

Optimizing Your Options: Extracting the Full Economic Value of Transmission When Planning Under Uncertainty

Electricity Journal

Watson, Jean-Paul W.; Munoz-Espinoza, Francisco D.; Hobbs, Benjamin F.

The anticipated magnitude of needed investments in new transmission infrastructure in the U.S. requires that these be allocated in a way that maximizes the likelihood of achieving society's goals for power system operation. The use of state-of-the-art optimization tools can identify cost-effective investment alternatives, extract more benefits out of transmission expansion portfolios, and account for the huge economic, technology, and policy uncertainties that the power sector faces over the next several decades.

More Details

Evaluating the Potential of a Laplacian Linear Solver

Informatica

Boman, Erik G.; Deweese, Kevin; Gilbert, John R.

A new approach for solving Laplacian linear systems proposed by Kelner et al. involves the random sampling and update of fundamental cycles in a graph. We evaluate the performance of this approach on a variety of real world graphs. We examine di erent ways to choose the set of cycles and their sequence of updates with the goal of providing more exibility and potential parallelism. We propose a parallel model of the Kelner et al. method for evaluating potential parallelism concerned with minimizing the span of edges updated at every iteration. We provide experimental results comparing the potential parallelism of the fundamental cycle basis and the extended basis. Our preliminary experiments show that choosing a non-fundamental set of cycles can save signi cant work compared to a fundamental cycle basis.

More Details

ASCR Workshop on Quantum Computing for Science

Aspuru-Guzik, Alan; Van Dam, Wim; Farhi, Edward; Gaitan, Frank; Humble, Travis; Jordan, Stephen; Landahl, Andrew J.; Love, Peter; Lucas, Robert; Preskill, John; Muller, Richard P.; Svore, Krysta; Wiebe, Nathan; Williams, Carl

This report details the findings of the DOE ASCR Workshop on Quantum Computing for Science that was organized to assess the viability of quantum computing technologies to meet the computational requirements of the DOE’s science and energy mission, and to identify the potential impact of quantum technologies. The workshop was held on February 17-18, 2015, in Bethesda, MD, to solicit input from members of the quantum computing community. The workshop considered models of quantum computation and programming environments, physical science applications relevant to DOE's science mission as well as quantum simulation, and applied mathematics topics including potential quantum algorithms for linear algebra, graph theory, and machine learning. This report summarizes these perspectives into an outlook on the opportunities for quantum computing to impact problems relevant to the DOE’s mission as well as the additional research required to bring quantum computing to the point where it can have such impact.

More Details

Delta: Data Reduction for Integrated Application Workflows

Lofstead, Gerald F.; Jean-Baptiste, Gregory; Oldfield, Ron A.

Integrated Application Workflows (IAWs) run multiple simulation workflow components con- currently on an HPC resource connecting these components using compute area resources and compensating for any performance or data processing rate mismatches. These IAWs require high frequency and high volume data transfers between compute nodes and staging area nodes during the lifetime of a large parallel computation. The available network band- width between the two areas may not be enough to efficiently support the data movement. As the processing power available to compute resources increases, the requirements for this data transfer will become more difficult to satisfy and perhaps will not be satisfiable at all since network capabilities are not expanding at a comparable rate. Furthermore, energy consumption in HPC environments is expected to grow by an order of magnitude as exas- cale systems become a reality. The energy cost of moving large amounts of data frequently will contribute to this issue. It is necessary to reduce the volume of data without reducing the quality of data when it is being processed and analyzed. Delta resolves the issue by addressing the lifetime data transfer operations. Delta removes subsequent identical copies of already transmitted data during transfers and restores those copies once the data has reached the destination. Delta is able to identify duplicated information and determine the most space efficient way to represent it. Initial tests show about 50% reduction in data movement while maintaining the same data quality and transmission frequency.

More Details

Trends in Microfabrication Capabilities & Device Architectures

Bauer, Todd B.; Jones, Adam J.; Lentine, Anthony L.; Mudrick, John M.; Okandan, Murat; Rodrigues, Arun

More Details

The fundamental downscaling limit of field effect transistors

Applied Physics Letters

Mamaluy, Denis M.; Gao, Xujiao G.

We predict that within next 15 years a fundamental down-scaling limit for CMOS technology and other Field-Effect Transistors (FETs) will be reached. Specifically, we show that at room temperatures all FETs, irrespective of their channel material, will start experiencing unacceptable level of thermally induced errors around 5-nm gate lengths. These findings were confirmed by performing quantum mechanical transport simulations for a variety of 6-, 5-, and 4-nm gate length Si devices, optimized to satisfy high-performance logic specifications by ITRS. Different channel materials and wafer/channel orientations have also been studied; it is found that altering channel-source-drain materials achieves only insignificant increase in switching energy, which overall cannot sufficiently delay the approaching downscaling limit. Alternative possibilities are discussed to continue the increase of logic element densities for room temperature operation below the said limit.

More Details

Towards efficient backward-in-time adjoint computations using data compression techniques

Computer Methods in Applied Mechanics and Engineering

Cyr, E.C.; Shadid, John N.; Wildey, T.

In the context of a posteriori error estimation for nonlinear time-dependent partial differential equations, the state-of-the-practice is to use adjoint approaches which require the solution of a backward-in-time problem defined by a linearization of the forward problem. One of the major obstacles in the practical application of these approaches is the need to store, or recompute, the forward solution to define the adjoint problem and to evaluate the error representation. This study considers the use of data compression techniques to approximate forward solutions employed in the backward-in-time integration. The development derives an error representation that accounts for the difference between the standard-approach and the compressed approximation of the forward solution. This representation is algorithmically similar to the standard representation and only requires the computation of the quantity of interest for the forward solution and the data-compressed reconstructed solution (i.e.scalar quantities that can be evaluated as the forward problem is integrated). This approach is then compared with existing techniques, such as checkpointing and time-averaged adjoints. Finally, we provide numerical results indicating the potential efficiency of our approach on a transient diffusion-reaction equation and on the Navier-Stokes equations. These results demonstrate memory compression ratios up to 450× while maintaining reasonable accuracy in the error-estimates.

More Details

Defect reaction network in Si-doped InAs. Numerical predictions

Schultz, Peter A.

This Report characterizes the defects in the def ect reaction network in silicon - doped, n - type InAs predicted with first principles density functional theory. The reaction network is deduced by following exothermic defect reactions starting with the initially mobile interstitial defects reacting with common displacement damage defects in Si - doped InAs , until culminating in immobile reaction p roducts. The defect reactions and reaction energies are tabulated, along with the properties of all the silicon - related defects in the reaction network. This Report serves to extend the results for the properties of intrinsic defects in bulk InAs as colla ted in SAND 2013 - 2477 : Simple intrinsic defects in InAs : Numerical predictions to include Si - containing simple defects likely to be present in a radiation - induced defect reaction sequence . This page intentionally left blank

More Details

Next-generation Algorithms for Assessing Infrastructure Vulnerability and Optimizing System Resilience

Burchett, Deon L.; Chen, Richard L.; Phillips, Cynthia A.; Richard, Jean-Philippe

This report summarizes the work performed under the project project Next-Generation Algo- rithms for Assessing Infrastructure Vulnerability and Optimizing System Resilience. The goal of the project was to improve mathematical programming-based optimization technology for in- frastructure protection. In general, the owner of a network wishes to design a network a network that can perform well when certain transportation channels are inhibited (e.g. destroyed) by an adversary. These are typically bi-level problems where the owner designs a system, an adversary optimally attacks it, and then the owner can recover by optimally using the remaining network. This project funded three years of Deon Burchett's graduate research. Deon's graduate advisor, Professor Jean-Philippe Richard, and his Sandia advisors, Richard Chen and Cynthia Phillips, supported Deon on other funds or volunteer time. This report is, therefore. essentially a replication of the Ph.D. dissertation it funded [12] in a format required for project documentation. The thesis had some general polyhedral research. This is the study of the structure of the feasi- ble region of mathematical programs, such as integer programs. For example, an integer program optimizes a linear objective function subject to linear constraints, and (nonlinear) integrality con- straints on the variables. The feasible region without the integrality constraints is a convex polygon. Careful study of additional valid constraints can significantly improve computational performance. Here is the abstract from the dissertation: We perform a polyhedral study of a multi-commodity generalization of variable upper bound flow models. In particular, we establish some relations between facets of single- and multi- commodity models. We then introduce a new family of inequalities, which generalizes traditional flow cover inequalities to the multi-commodity context. We present encouraging numerical results. We also consider the directed edge-failure resilient network design problem (DRNDP). This problem entails the design of a directed multi-commodity flow network that is capable of fulfilling a specified percentage of demands in the event that any G arcs are destroyed, where G is a constant parameter. We present a formulation of DRNDP and solve it in a branch-column-cut framework. We present computational results.

More Details

Cyber Graph Queries for Geographically Distributed Data Centers

Berry, Jonathan W.; Collins, Michael; Kearns, Aaron; Phillips, Cynthia A.; Saia, Jared

We present new algorithms for a distributed model for graph computations motivated by limited information sharing we first discussed in [20]. Two or more independent entities have collected large social graphs. They wish to compute the result of running graph algorithms on the entire set of relationships. Because the information is sensitive or economically valuable, they do not wish to simply combine the information in a single location. We consider two models for computing the solution to graph algorithms in this setting: 1) limited-sharing: the two entities can share only a polylogarithmic size subgraph; 2) low-trust: the entities must not reveal any information beyond the query answer, assuming they are all honest but curious. We believe this model captures realistic constraints on cooperating autonomous data centers. We have algorithms in both setting for s - t connectivity in both models. We also give an algorithm in the low-communication model for finding a planted clique. This is an anomaly- detection problem, finding a subgraph that is larger and denser than expected. For both the low- communication algorithms, we exploit structural properties of social networks to prove perfor- mance bounds better than what is possible for general graphs. For s - t connectivity, we use known properties. For planted clique, we propose a new property: bounded number of triangles per node. This property is based upon evidence from the social science literature. We found that classic examples of social networks do not have the bounded-triangles property. This is because many social networks contain elements that are non-human, such as accounts for a business, or other automated accounts. We describe some initial attempts to distinguish human nodes from automated nodes in social networks based only on topological properties.

More Details

Storage Systems and Input/Output to Support Extreme Scale Science

Ross, Robert; Grider, Gary; Felix, Evan; Gary, Mark; Klasky, Scott; Oldfield, Ron A.; Shipman, Galen; Wu, John

Storage systems are a foundational component of computational, experimental, and observational science today. The success of Department of Energy (DOE) activities in these areas is inextricably tied to the usability, performance, and reliability of storage and input/output (I/O) technologies.

More Details

Replicated computational results (RCR) report for "BLIS: A framework for rapidly instantiating BLAS functionality"

ACM Transactions on Mathematical Software

Willenbring, James M.

"BLIS: A Framework for Rapidly Instantiating BLAS Functionality" includes single-platform BLIS performance results for both level-2 and level-3 operations that is competitive with OpenBLAS, ATLAS, and Intel MKL. A detailed description of the configuration used to generate the performance results was provided to the reviewer by the authors. All the software components used in the comparison were reinstalled and new performance results were generated and compared to the original results. After completing this process, the published results are deemed replicable by the reviewer.

More Details

A sparse reconstruction method for the estimation of multi-resolution emission fields via atmospheric inversion

Geoscientific Model Development

Ray, Jaideep R.; Lee, Jina L.; Yadav, V.; Lefantzi, Sophia L.; Michalak, A.M.; van Bloemen Waanders, Bart G.

Atmospheric inversions are frequently used to estimate fluxes of atmospheric greenhouse gases (e.g., biospheric CO2 flux fields) at Earth's surface. These inversions typically assume that flux departures from a prior model are spatially smoothly varying, which are then modeled using a multi-variate Gaussian. When the field being estimated is spatially rough, multi-variate Gaussian models are difficult to construct and a wavelet-based field model may be more suitable. Unfortunately, such models are very high dimensional and are most conveniently used when the estimation method can simultaneously perform data-driven model simplification (removal of model parameters that cannot be reliably estimated) and fitting. Such sparse reconstruction methods are typically not used in atmospheric inversions. In this work, we devise a sparse reconstruction method, and illustrate it in an idealized atmospheric inversion problem for the estimation of fossil fuel CO2 (ffCO2) emissions in the lower 48 states of the USA. Our new method is based on stagewise orthogonal matching pursuit (StOMP), a method used to reconstruct compressively sensed images. Our adaptations bestow three properties to the sparse reconstruction procedure which are useful in atmospheric inversions. We have modified StOMP to incorporate prior information on the emission field being estimated and to enforce non-negativity on the estimated field. Finally, though based on wavelets, our method allows for the estimation of fields in non-rectangular geometries, e.g., emission fields inside geographical and political boundaries. Our idealized inversions use a recently developed multi-resolution (i.e., wavelet-based) random field model developed for ffCO2 emissions and synthetic observations of ffCO2 concentrations from a limited set of measurement sites. We find that our method for limiting the estimated field within an irregularly shaped region is about a factor of 10 faster than conventional approaches. It also reduces the overall computational cost by a factor of 2. Further, the sparse reconstruction scheme imposes non-negativity without introducing strong nonlinearities, such as those introduced by employing log-transformed fields, and thus reaps the benefits of simplicity and computational speed that are characteristic of linear inverse problems.

More Details

Ethane-xenon mixtures under shock conditions

Physical Review B - Condensed Matter and Materials Physics

Magyar, Rudolph J.; Root, Seth R.; Mattsson, Thomas M.; Cochrane, Kyle C.; Flicker, Dawn G.

Mixtures of light elements with heavy elements are important in inertial confinement fusion. We explore the physics of molecular scale mixing through a validation study of equation of state (EOS) properties. Density functional theory molecular dynamics (DFT-MD) at elevated temperature and pressure is used to obtain the thermodynamic state properties of pure xenon, ethane, and various compressed mixture compositions along their principal Hugoniots. To validate these simulations, we have performed shock compression experiments using the Sandia Z-Machine. A bond tracking analysis correlates the sharp rise in the Hugoniot curve with the completion of dissociation in ethane. The DFT-based simulation results compare well with the experimental data along the principal Hugoniots and are used to provide insight into the dissociation and temperature along the Hugoniots as a function of mixture composition. Interestingly, we find that the compression ratio for complete dissociation is similar for several compositions suggesting a limiting compression for C-C bonded systems.

More Details

Integral approximations to classical diffusion and smoothed particle hydrodynamics

Computer Methods in Applied Mechanics and Engineering

Du, Qiang; Lehoucq, Richard B.; Tartakovsky, A.M.

The contribution of the paper is the approximation of a classical diffusion operator by an integral equation with a volume constraint. A particular focus is on classical diffusion problems associated with Neumann boundary conditions. By exploiting this approximation, we can also approximate other quantities such as the flux out of a domain. Our analysis of the model equation on the continuum level is closely related to the recent work on nonlocal diffusion and peridynamic mechanics. In particular, we elucidate the role of a volumetric constraint as an approximation to a classical Neumann boundary condition in the presence of physical boundary. The volume-constrained integral equation then provides the basis for accurate and robust discretization methods. An immediate application is to the understanding and improvement of the Smoothed Particle Hydrodynamics (SPH) method.

More Details

Benchmarking Adiabatic Quantum Optimization for Complex Network Analysis

Parekh, Ojas D.; Wendt, Jeremy D.; Shulenburger, Luke N.; Landahl, Andrew J.; Moussa, Jonathan E.; Aidun, John B.

We lay the foundation for a benchmarking methodology for assessing current and future quantum computers. We pose and begin addressing fundamental questions about how to fairly compare computational devices at vastly different stages of technological maturity. We critically evaluate and offer our own contributions to current quantum benchmarking efforts, in particular those involving adiabatic quantum computation and the Adiabatic Quantum Optimizers produced by D-Wave Systems, Inc. We find that the performance of D-Wave's Adiabatic Quantum Optimizers scales roughly on par with classical approaches for some hard combinatorial optimization problems; however, architectural limitations of D-Wave devices present a significant hurdle in evaluating real-world applications. In addition to identifying and isolating such limitations, we develop algorithmic tools for circumventing these limitations on future D-Wave devices, assuming they continue to grow and mature at an exponential rate for the next several years.

More Details

A variational flux recovery approach for elastodynamics problems with interfaces

PANACM 2015 - 1st Pan-American Congress on Computational Mechanics, in conjunction with the 11th Argentine Congress on Computational Mechanics, MECOM 2015

Bochev, Pavel B.; Kuberry, Paul A.

We present a new explicit algorithm for linear elastodynamic problems with material interfaces. The method discretizes the governing equations independently on each material subdomain and then connects them by exchanging forces and masses across the material interface. Variational flux recovery techniques provide the force and mass approximations. The new algorithm has attractive computational properties. It allows different discretizations on each material subdomain and enables partitioned solution of the discretized equations. The method passes a linear patch test and recovers the solution of a monolithic discretization of the governing equations when interface grids match.

More Details

A conservative, optimization-based semi-lagrangian spectral element method for passive tracer transport

COUPLED PROBLEMS 2015 - Proceedings of the 6th International Conference on Coupled Problems in Science and Engineering

Bochev, Pavel B.; Moe, Scott A.; Peterson, Kara J.; Ridzal, Denis R.

We present a new optimization-based, conservative, and quasi-monotone method for passive tracer transport. The scheme combines high-order spectral element discretization in space with semi-Lagrangian time stepping. Solution of a singly linearly constrained quadratic program with simple bounds enforces conservation and physically motivated solution bounds. The scheme can handle efficiently a large number of passive tracers because the semi-Lagrangian time stepping only needs to evolve the grid points where the primitive variables are stored and allows for larger time steps than a conventional explicit spectral element method. Numerical examples show that the use of optimization to enforce physical properties does not affect significantly the spectral accuracy for smooth solutions. Performance studies reveal the benefits of high-order approximations, including for discontinuous solutions.

More Details
Results 5401–5600 of 9,998
Results 5401–5600 of 9,998