The use of oxidized metal powders in mechanical shock or crush safety enhancers in nuclear weapons has been investigated. The functioning of these devices is based on the remarkable electrical behavior of compacts of certain oxidized metal powders when subjected to compressive stress. For example, the low voltage resistivity of a compact of oxidized tantalum powder was found to decrease by over six orders of magnitude during compaction between 1 MPa, where the thin, insulating oxide coatings on the particles are intact, to 10 MPa, where the oxide coatings have broken down along a chain of particles spanning the electrodes. In this work, the behavior of tantalum and aluminum powders was investigated. The low voltage resistivity during compaction of powders oxidized under various conditions was measured and compared. In addition, the resistivity at higher voltages and the dielectric breakdown strength during compaction were also measured. A key finding was that significant changes in the electrical properties persist after the removal of the stress so that a mechanical shock enhancer is feasible. This was verified by preliminary shock experiments. Finally, conceptual designs for both types of enhancers are presented.
Preliminary thermal decomposition experiments with Ablefoam and EF-AR20 foam (Ablefoam replacement) were done to determine the important chemical and associated physical phenomena that should be investigated to develop the foam decomposition chemistry sub-models that are required in numerical simulations of the fire-induced response of foam-filled engineered systems for nuclear safety applications. Although the two epoxy foams are physically and chemically similar, the thermal decomposition of each foam involves different chemical mechanisms, and the associated physical behavior of the foams, particularly ''foaming'' and ''liquefaction,'' have significant implications for modeling. A simplified decomposition chemistry sub-model is suggested that, subject to certain caveats, may be appropriate for ''scoping-type'' calculations.
This report is a presentation of modeling and simulation work for analyzing three designs of Micro Electro Mechanical (MEM) Compound Pivot Mirrors (CPM). These CPMs were made at Sandia National Laboratories using the SUMMiT{trademark} process. At 75 volts and above, initial experimental analysis of fabricated mirrors showed tilt angles of up to 7.5 degrees for one design, and 5 degrees for the other two. Nevertheless, geometric design models predicted higher tilt angles. Therefore, a detailed study was conducted to explain why lower tilt angles occurred and if design modifications could be made to produce higher tilt angles at lower voltages. This study showed that the spring stiffnesses of the CPMs were too great to allow for desired levels of rotation at lower levels of voltage. To produce these lower stiffnesses, a redesign is needed.
The semiconductor bridge (SCB) is an electroexplosive device used to initiate detonators. A C cable is commonly used to connect the SCB to a firing set. A series of tests were performed to identify smaller, lighter cables for firing single and multiple SCBs. This report provides a description of these tests and their results. It was demonstrated that lower threshold voltages and faster firing times can be achieved by increasing the wire size, which reduces ohmic losses. The RF 100 appears to be a reasonable substitute for C cable when firing single SCBs. This would reduce the cable volume by 68% and the weight by 67% while increasing the threshold voltage by only 22%. In general, RG 58 outperforms twisted pair when firing multiple SCBs in parallel. The RG 58's superior performance is attributed to its larger conductor size.
Chemometric analysis of nuclear magnetic resonance (NMR) spectroscopy has increased dramatically in recent years. Various chemometric techniques have been applied to a wide range of problems in food, agricultural, medical, process, and industrial system. This article gives a brief review of chemometric analysis of NMR spectral data, including a summary of the types of mixtures and experiments analyzed with chemometric techniques. Common experiment problems encountered during the chemometric analysis of NMR data are also discussed.
Stiction and friction in micromachines is commonly inhibited through the use of silane coupling agents such as 1H-, 1H-, 2H-, 2H-perfluorodecyltrichlorosilane (FDTS). FDTS coatings have allowed micromachine parts processed in water to be released without debilitating capillary adhesion occurring. These coatings are frequently considered as densely-packed monolayers, well-bonded to the substrate. In this paper, it is demonstrated that FDTS coatings can exhibit complex nanoscale structures, which control whether micromachine parts release or not. Surface images obtained via atomic force microscopy reveal that FDTS coating solutions can generate micellar aggregates that deposit on substrate surfaces. Interferometric imaging of model beam structures shows that stiction is high when the droplets are present and low when only monolayers are deposited. As the aggregate thickness (tens of nanometers) is insufficient to bridge the 2 μm gap under the beams, the aggregates appear to promote beam-substrate adhesion by changing the wetting characteristics of coated surfaces. Contact angle measurements and condensation figure experiments have been performed on surfaces and under coated beams to quantify the changes in interfacial properties that accompany different coating structures. These results may explain the irreproducibility that is often observed with these films.
A DOE/Sandia project termed the Blade Manufacturing Program was established at Sandia to develop means of advancing manufacturing processes in ways that lower costs and improve the reliability of turbine blades. Through industry contracts, manufacturers are improving processes such as resin infusion, resin transfer molding, and thermoplastic casting. Testing and modeling research at universities and national labs are adding to the knowledge of how composite materials perform in substructures and sub-scale blades as a function of their fabrication process.
Optimal estimation theory has been applied to the problem of estimating process variables during vacuum arc remelting (VAR), a process widely used in the specialty metals industry to cast large ingots of segregation sensitive and/or reactive metal alloys. Four state variables were used to develop a simple state-space model of the VAR process: electrode gap (G), electrode mass (M), electrode position (X) and electrode melting rate (R). The optimal estimator consists of a Kalman filter that incorporates the model and uses electrode feed rate and measurement based estimates of G, M and X to produce optimal estimates of all four state variables. Simulations show that the filter provides estimates that have error variances between one and three orders-of-magnitude less than estimates based solely on measurements. Examples are presented that verify this for electrode gap, an extremely important control parameter for the process.
Direct Simulation Monte Carlo (DSMC) and Navier-Stokes calculations are performed for a Mach 11 25 deg.-55 deg. spherically blunted biconic. The conditions are such that flow is laminar, with separation occurring at the cone-cone juncture. The simulations account for thermochemical nonequilibrium based on standard Arrhenius chemical rates for nitrogen dissociation and Millikan and White vibrational relaxation. The simulation error for the Navier-Stokes (NS) code is estimated to be 2% for the surface pressure and 10% for the surface heat flux. The grid spacing for the DSMC simulations was adjusted to be less than the local mean-freepath (mfp) and the time step less than the cell transient time of a computational particle. There was overall good agreement between the two simulations; however, the recirculation zone was computed to be larger for the NS simulation. A sensitivity study is performed to examine the effects of experimental uncertainty in the freestream properties on the surface pressure and heat flux distributions. The surface quantities are found to be extremely sensitive to the vibrational excitation state of the gas at the test section, with differences of 25% found in the surface pressure and 25%-35% for the surface heat flux. These calculations are part of a blind validation comparison and thus the experimental data has not yet been released.
Simulations of a turbulent methanol pool fire are conducted using both Reynolds-Averaged Navier-Stokes (RANS) and Large Eddy Simulation (LES) modeling methodologies. Two simple conserved scalar flameletbased combustion models with assumed PDF are developed and implemented. The first model assumes statistical independence between mixture fraction and its variance and results in poor predictions of time-averaged temperature and velocity. The second combustion model makes use of the PDF transport equation for mixture fraction and does not employ the statistical independence assumption. Results using this model show good agreement with experimental data for both the 2D and 3D LES, indicating that the use of statistical independence between mixture fraction and its dissipation is not valid for pool fire simulations. Lastly, "finger-like" flow structures near the base of the plume, generated from stream-wise vorticity, are shown to be important mixing mechanisms for accurate prediction of time-averaged temperature and velocity.
The concept of genetic divisors can be given a quantitative measure with a non-Archimedean p-adic metric that is both computationally convenient and physically motivated. For two particles possessing distinct mass parameters x and y, the metric distance D(x, y) is expressed on the field of rational numbers Q as the inverse of the greatest common divisor [gcd (x , y)]. As a measure of genetic similarity, this metric can be applied to (1) the mass numbers of particle states and (2) the corresponding subgroup orders of these systems. The use of the Bezout identity in the form of a congruence for the expression of the gcd (x , y) corresponding to the v{sub e} and {sub {mu}} neutrinos (a) connects the genetic divisor concept to the cosmic seesaw congruence, (b) provides support for the {delta}-conjecture concerning the subgroup structure of particle states, and (c) quantitatively strengthens the interlocking relationships joining the values of the prospectively derived (i) electron neutrino (v{sub e}) mass (0.808 meV), (ii) muon neutrino (v{sub {mu}}) mass (27.68 meV), and (iii) unified strong-electroweak coupling constant ({alpha}*{sup -1} = 34.26).
Alkylation reactions of benzene with propylene using zeolites were studied for their affinity for cumene production. The current process for the production of cumene involves heating corrosive acid catalysts, cooling, transporting, and distillation. This study focused on the reaction of products in a static one-pot vessel using non-corrosive zeolite catalysts, working towards a more efficient one-step process with a potentially large energy savings. A series of experiments were conducted to find the best reaction conditions yielding the highest production of cumene. The experiments looked at cumene formation amounts in two different reaction vessels that had different physical traits. Different zeolites, temperatures, mixing speeds, and amounts of reactants were also investigated to find their affects on the amount of cumene produced. Quantitative analysis of product mixture was performed by gas chromatography. Mass spectroscopy was also utilized to observe the gas phase components during the alkylation process.
The ultimate goal of many environmental measurements is to determine the risk posed to humans or ecosystems by various contaminants. Conventional environmental monitoring typically requires extensive sampling grids covering several media including air, water, soil and vegetation. A far more efficient, innovative and inexpensive tactic has been found using honeybees as sampling mechanisms. Members from a single bee colony forage over large areas ({approx}2 x 10{sup 6} m{sup 2}), making tens of thousands of trips per day, and return to a fixed location where sampling can be conveniently conducted. The bees are in direct contact with the air, water, soil and vegetation where they encounter and collect any contaminants that are present in gaseous, liquid and particulate form. The monitoring of honeybees when they return to the hive provides a rapid method to assess chemical distributions and impacts (1). The primary goal of this technology is to evaluate the efficiency of the transport mechanism (honeybees) to the hive using preconcentrators to collect samples. Once the extent and nature of the contaminant exposure has been characterized, resources can be distributed and environmental monitoring designs efficiently directed to the most appropriate locations. Methyl salicylate, a chemical agent surrogate was used as the target compound in this study.
Using intense magnetic pressure, a method was developed to launch flyer plates to velocities in excess of 20 km/s. This technique was used to perform plate-impact, shock wave experiments on cryogenic liquid deuterium (LD{sub 2}) to examine its high-pressure equation of state (EOS). Using an impedance matching method, Hugoniot measurements were obtained in the pressure range of 30-70 GPa. The results of these experiments disagree with previously reported Hugoniot measurements of LD{sub 2} in the pressure range above {approx}40 GPa, but are in good agreement with first principles, ab-initio models for hydrogen and its isotopes.
Sandstones that overlie or that are interbedded with evaporitic or other ductile strata commonly contain numerous localized domains of fractures, each covering an area of a few square miles. Fractures within the Entrada Sandstone at the Salt Valley Anticline are associated with salt mobility within the underlying Paradox Formation. The fracture relationships observed at Salt Valley (along with examples from Paleozoic strata at the southern edge of the Holbrook basin in northeastern Arizona, and sandstones of the Frontier Formation along the western edge of the Green River basin in southwestern Wyoming), show that although each fracture domain may contain consistently oriented fractures, the orientations and patterns of the fractures vary considerably from domain to domain. Most of the fracture patterns in the brittle sandstones are related to local stresses created by subtle, irregular flexures resulting from mobility of the associated, interbedded ductile strata (halite or shale). Sequential episodes of evaporite dissolution and/or mobility in different directions can result in multiple, superimposed fracture sets in the associated sandstones. Multiple sets of superimposed fractures create reservoir-quality fracture interconnectivity within restricted localities of a formation. However, it is difficult to predict the orientations and characteristics of this type of fracturing in the subsurface. This is primarily because the orientations and characteristics of these fractures typically have little relationship to the regional tectonic stresses that might be used to predict fracture characteristics prior to drilling. Nevertheless, the high probability of numerous, intersecting fractures in such settings attests to the importance of determining fracture orientations in these types of fractured reservoirs.
Carbon is an important support for heterogeneous catalysts, such as platinum supported on activated carbon (AC). An important property of these catalysts is that they decompose upon heating in air. Consequently, Pt/AC catalysts can be used in applications requiring rapid decomposition of a material, leaving little residue. This report describes the catalytic effects of platinum on carbon decomposition in an attempt to maximize decomposition rates. Catalysts were prepared by impregnating the AC with two different Pt precursors, Pt(NH{sub 3}){sub 4}(NO{sub 3}){sub 2} and H{sub 2}PtCl{sub 6}. Some catalysts were treated in flowing N{sub 2} or H{sub 2} at elevated temperatures to decompose the Pt precursor. The catalysts were analyzed for weight loss in air at temperatures ranging from 375 to 450 C, using thermogravimetric analysis (TGA). The following results were obtained: (1) Pt/AC decomposes much faster than pure carbon; (2) treatment of the as-prepared 1% Pt/AC samples in N{sub 2} or H{sub 2} enhances decomposition; (3) autocatalytic behavior is observed for 1% Pt/AC samples at temperatures {ge} 425 C; (4) oxygen is needed for decomposition to occur. Overall, the Pt/AC catalyst with the highest activity was impregnated with H{sub 2}PtCl{sub 6} dissolved in acetone, and then treated in H{sub 2}. However, further research and development should produce a more active Pt/AC material.
The Microsystems Subgrid Physics project is intended to address gaps between developing high-performance modeling and simulation capabilities and microdomain specific physics. The initial effort has focused on incorporating electrostatic excitations, adhesive surface interactions, and scale dependent material and thermal properties into existing modeling capabilities. Developments related to each of these efforts are summarized, and sample applications are presented. While detailed models of the relevant physics are still being developed, a general modeling framework is emerging that can be extended to incorporate evolving material and surface interaction modules.
Recently an innovative technique known as the Isentropic Compression Experiment (ICE) was developed that allows the dynamic compressibility curve of a material to be measured in a single experiment. Hence, ICE significantly reduces the cost and time required for generating and validating theoretical models of dynamic material response. ICE has been successfully demonstrated on several materials using the 20 MA Z accelerator, resulting in a large demand for its use. The present project has demonstrated its use on another accelerator, Saturn. In the course of this study, Saturn was tailored to produce a satisfactory drive time structure, and instrumented to produce velocity data. Pressure limits are observed to be approximately 10-15 GPa (''LP'' configuration) or 40-50 GPa (''HP'' configuration), depending on sample material. Drive reproducibility (panel to panel within a shot and between shots) is adequate for useful experimentation, but alignment fixturing problems make it difficult to achieve the same precision as is possible at Z. Other highlights included the useful comparison of slightly different PZT and ALOX compositions (neutron generator materials), temperature measurement using optical pyrometry, and the development of a new technique for preheating samples. 28 ICE tests have been conducted at Saturn to date, including the experiments described herein.
Sandia is investigating the shock response of single-crystal diamond up to several Mbar pressure in a collaborative effort with the Institute for Shock Physics (ISP) at Washington State University (WSU). This is project intended to determine (i) the usefulness of diamond as a window material for high pressure velocity interferometry measurements, (ii) the maximum stress level at which diamond remains transparent in the visible region, (iii) if a two-wave structure can be detected and analyzed, and if so, (iv) the Hugoniot elastic limit (HEL) for the [110] orientation of diamond. To this end experiments have been designed and performed, scoping the shock response in diamond in the 2-3 Mbar pressure range using conventional velocity interferometry techniques (conventional VISAR diagnostic). In order to perform more detailed and highly resolved measurements, an improved line-imaging VISAR has been developed and experiments using this technique have been designed. Prior to performing these more detailed experiments, additional scoping experiments are being performed using conventional techniques at WSU to refine the experimental design.
Explosive charges placed on the fuze end of a drained chemical munition are expected to be used as a means to destroy the fuze and burster charges of the munition. Analyses are presented to evaluate the effect of these additional initiation charges on the fragmentation characteristics for the M121A1 155mm chemical munition, modeled with a T244 fuze attached, and to assess the consequences of these fragment impacts on the walls of a containment chamber--the Burster Detonation Vessel. A numerical shock physics code (CTH) is used to characterize the mass and velocity of munition fragments. Both two- and three-dimensional simulations of the munition have been completed in this study. Based on threshold fragment velocity/mass results drawn from both previous and current analyses, it is determined that under all fragment impact conditions from the munition configurations considered in this study, no perforation of the inner chamber wall will occur, and the integrity of the Burster Detonation Vessel is retained. However, the munition case fragments have sufficient mass and velocity to locally damage the surface of the inner wall of the containment vessel.
Schells, Regina L.; Bogdan, Carolyn W.; Wix, Steven D.
This document describes the High Performance Electrical Modeling and Simulation (HPEMS) Global Verification Test Suite (VERTS). The VERTS is a regression test suite used for verification of the electrical circuit simulation codes currently being developed by the HPEMS code development team. This document contains descriptions of the Tier I test cases.
The Geometric Search Engine is a software system for storing and searching a database of geometric models. The database maybe searched for modeled objects similar in shape to a target model supplied by the user. The database models are generally from CAD models while the target model may be either a CAD model or a model generated from range data collected from a physical object. This document describes key generation, database layout, and search of the database.
This report provides a summary of the work completed in the Source Code Assurance Tool project. This work was done as part of the Laboratory Directed Research and Development program.
This report provides a preliminary functional description of a novel software application, the Source Code Assurance Tool, which would assist a system analyst in the software assessment process. An overview is given of the tool's functionality and design; and how the analyst would use it to assess a body of source code. This work was done as part of a Laboratory Directed Research and Development project.
In this paper we describe a new language, Visual Structure Language (VSL), designed to describe the structure of a program and explain its pieces. This new language is built on top of a general-purpose language, such as C. The language consists of three extensions: explanations, nesting, and arcs. Explanations are comments explicitly associated with code segments. These explanations can be nested. And arcs can be inserted between explanations to show data- or control-flow. The value of VSL is that it enables a developer to better control a code. The developer can represent the structure via nested explanations, using arcs to indicate the flow of data and control. The explanations provide a ''second opinion'' about the code so that at any level, the developer can confirm that the code operates as it is intended to do. We believe that VSL enables a programmer to use in a computer language the same model--a hierarchy of components--that they use in their heads when they conceptualize systems.
We present the tool we built as part of a Laboratory Directed Research and Development (LDRD) project. This tool consists of a commercially-available, graphical editor front-end, combined with a back end ''slicer.'' The significance of the tool is that it shows how to slice across system components. This is an advance from slicing across program components.
This report details experimental data useful in validating radiative transfer codes involving participating media, particularly for cases involving combustion. Special emphasis is on data for pool fires. Features sought in the references are: Flame geometry and fuel that approximate conditions for a pool fire or a well-defined flame geometry and characteristics that can be completely modeled; detailed information that could be used as code input data, including species concentration and temperature profiles and associated absorption coefficients, soot morphology and concentration profiles, associated scattering coefficients and phase functions, specification of system geometry, and system boundary conditions; detailed information that could be compared against code output predictions, including measured boundary radiative energy flux distributions (preferably spectral) and/or boundary temperature distributions; and a careful experimental error analysis so that code predictions could be rationally compared with experimental measurements. Reference data were gathered from more than 35 persons known to be active in the field of radiative transfer and combustion, particularly in experimental work. A literature search was carried out using key words. Additionally, the reference lists in papers/reports were pursued for additional leads. The report presents extended abstracts of the cited references, with comments on available and missing data for code validation, and comments on reported error. A graphic for quick reference is added to each abstract that indicates the completeness of data and how well the data mimics a large-scale pool fire. The references are organized into Lab-Scale Pool Fires, Large-Scale Pool Fires, Momentum-Driven Diffusion Flames, and Enclosure Fires. As an additional aid to report users, the Tables in Appendix A show the types of data included in each reference. The organization of the tables follows that used for the abstracts.
Sandia National Laboratories is developing innovative alternative technology to replace open burn/open detonation (OB/OD) operations for the destruction and disposal of obsolete, excess, and off-spec energetic materials. Alternatives to OB/OD are necessary to comply with increasingly stringent regulations. This program is developing an alternative technology to destruct energetic materials using organic amines with minimal discharge of toxic chemicals to the environment and defining the application of the by-products for the manufacture of structural materials.
Wire explosion experiments have been carried out at the University of Nevada, Reno. These experiments investigated the explosion phase of wires with properties and current-driving conditions comparable to that used in the initial stage of wire array z-pinch implosions on the Z machine at Sandia National Laboratories. Specifically, current pulses similar to and faster than the pre-pulse current on Z (current prior to fast rise in current pulse) were applied to single wire loads to study wire heating and the early development of plasmas in the wire initiation process. Understanding such issues are important to larger pulsed power machines that implode cylindrical wire array loads comprised of many wires. It is thought that the topology of an array prior to its acceleration influences the implosion and final stagnation properties, and therefore may depend on the initiation phase of the wires. Single wires ranging from 4 to 40 pm in diameter and comprised of material ranging from AI to W were investigated. Several diagnostics were employed to determine wire current, voltage, total emitted-light energy and power, along with the wire expansion velocity throughout the explosion. In a number of cases, the explosion process was also observed with x-ray backlighting using x-pinches. The experimental data indicates that the characteristics of a wire explosion depend dramatically on the rate of rise of the current, on the diameter of the wire, and on the heat of vaporization of the wire material. In this report, these characteristics will be described in detail. Of particular interest is the result that a faster current rise produces a higher energy deposition into the wire prior to explosion. This result introduces a different means of increasing the efficiency of wire heating. In this case, the energy deposition along the wire and its subsequent expansion, is uniform compared to a ''slow'' current rise (170 A/ns compared to 22 A /s current rise into a short circuit) and the expansion velocity is larger. The energy deposition and wire expansion is further modified by the wire diameter and material. Investigations of wire diameter indicate that the diameter primarily effects the expansion velocity and energy deposition; thicker wires explode with greater velocities but absorb less energy per atom. The heat of vaporization also categorizes the wire explosion; wires with a low heat of vaporization expand faster and emit less radiation than their high heat of vaporization counterparts.
An important capability in conducting underground nuclear tests is to be able to determine the nuclear test yield accurately within hours after a test. Due to a nuclear test moratorium, the seismic method that has been used in the past has not been exercised since a non-proliferation high explosive test in 1993. Since that time, the seismic recording system and the computing environment have been replaced with modern equipment. This report describes the actions that have been taken to preserve the capability for determining seismic yield, in the event that nuclear testing should resume. Specifically, this report describes actions taken to preserve seismic data, actions taken to modernize software, and actions taken to document procedures. It concludes with a summary of the current state of the data system and makes recommendations for maintaining this system in the future.
This report describes testing of prototype InfiniBand{trademark} host channel adapters from Intel Corporation, using the Linux(reg sign) operating system. Three generations of prototype hardware were obtained, and Linux device drivers were written which exercised the data movement capabilities of the cards. Latency and throughput results obtained were similar to other SAN technologies, but not significantly better.
This project set out to scientifically-tailor ''smart'' interfacial films and 3-D composite nanostructures to exhibit photochromic responses to specific, highly-localized chemical and/or mechanical stimuli, and to integrate them into optical microsystems. The project involved the design of functionalized chromophoric self-assembled materials that possessed intense and environmentally-sensitive optical properties (absorbance, fluorescence) enabling their use as detectors of specific stimuli and transducers when interfaced with optical probes. The conjugated polymer polydiacetylene (PDA) proved to be the most promising material in many respects, although it had some drawbacks concerning reversibility. Throughout his work we used multi-task scanning probes (AFM, NSOM), offering simultaneous optical and interfacial force capabilities, to actuate and characterize the PDA with localized and specific interactions for detailed characterization of physical mechanisms and parameters. In addition to forming high quality mono-, bi-, and tri-layers of PDA via Langmuir-Blodgett deposition, we were successful in using the diacetylene monomer precursor as a surfactant that directed the self-assembly of an ordered, mesostructured inorganic host matrix. Remarkably, the diacetylene was polymerized in the matrix, thus providing a PDA-silica composite. The inorganic matrix serves as a perm-selective barrier to chemical and biological agents and provides structural support for improved material durability in microsystems. Our original goal was to use the composite films as a direct interface with microscale devices as optical elements (e.g., intracavity mirrors, diffraction gratings), taking advantage of the very high sensitivity of device performance to real-time dielectric changes in the films. However, our optical physics colleagues (M. Crawford and S. Kemme) were unsuccessful in these efforts, mainly due to the poor optical quality of the composite films.
The intention of this project was to collaborate with Harvard University in the general area of nanoscale structures, biomolecular materials and their application in support of Sandia's MEMS technology. The expertise at Harvard was crucial in fostering these fundamentally interdisciplinary developments. Areas that were of interest included: (1) nanofabrication that exploits traditional methods (from Si technology) and developing new methods; (2) self-assembly of organic and inorganic systems; (3) assembly and dynamics of membranes and microfluidics; (4) study of the hierarchy of scales in assembly; (5) innovative imaging methods; and (6) hard (engineering)/soft (biological) interfaces. Specifically, we decided to work with Harvard to design and construct an experimental test station to measure molecular transport through single nanopores. The pore may be of natural origin, such as a self-assembled bacterial protein in a lipid bilayer, or an artificial structure in silicon or silicon nitride.
This report documents work supporting the Sandia National Laboratories initiative in Distributed Energy Resources (DERs) and Supervisory Control and Data Acquisition (SCADA) systems. One approach for real-time control of power generation assets using feedback control, Quantitative feedback theory (QFT), has recently been applied to voltage, frequency, and phase-control of power systems at Sandia. QFT provided a simple yet powerful philosophy for designing the control systems--allowing the designer to optimize the system by making design tradeoffs without getting lost in complex mathematics. The feedback systems were effective in reducing sensitivity to large and sudden changes in the power grid system. Voltage, frequency, and phase were accurately controlled, even with large disturbances to the power grid system.
This report is divided into two parts: a study of the glass transition in confined geometries, and formation mechanisms of block copolymer mesophases by solvent evaporation-induced self-assembly. The effect of geometrical confinement on the glass transition of polymers is a very important consideration for applications of polymers in nanotechnology applications. We hypothesize that the shift of the glass transition temperature of polymers in confined geometries can be attributed to the inhomogeneous density profile of the liquid. Accordingly, we assume that the glass temperature in the inhomogeneous state can be approximated by the Tg of a corresponding homogeneous, bulk polymer, but at a density equal to the average density of the inhomogeneous system. Simple models based on this hypothesis give results that are in remarkable agreement with experimental measurements of the glass transition of confined liquids. Evaporation-induced self-assembly (EISA) of block copolymers is a versatile process for producing novel, nanostructured materials and is the focus of much of the experimental work at Sandia in the Brinker group. In the EISA process, as the solvent preferentially evaporates from a cast film, two possible scenarios can occur: microphase separation or micellization of the block copolymers in solution. In the present investigation, we established the conditions that dictate which scenario takes place. Our approach makes use of scaling arguments to determine whether the overlap concentration c* occurs before or after the critical micelle concentration (CMC). These theoretical arguments are used to interpret recent experimental results of Yu and collaborators on EISA experiments on Silica/PS-PEO systems.
In exploring the question of how humans reason in ambiguous situations or in the absence of complete information, we stumbled onto a body of knowledge that addresses issues beyond the original scope of our effort. We have begun to understand the importance that philosophy, in particular the work of C. S. Peirce, plays in developing models of human cognition and of information theory in general. We have a foundation that can serve as a basis for further studies in cognition and decision making. Peircean philosophy provides a foundation for understanding human reasoning and capturing behavioral characteristics of decision makers due to cultural, physiological, and psychological effects. The present paper describes this philosophical approach to understanding the underpinnings of human reasoning. We present the work of C. S. Peirce, and define sets of fundamental reasoning behavior that would be captured in the mathematical constructs of these newer technologies and would be able to interact in an agent type framework. Further, we propose the adoption of a hybrid reasoning model based on his work for future computational representations or emulations of human cognition.
This article summarizes information related to the automated course of action (COA) development effort. The information contained in this document puts the COA effort into an operational perspective that addresses command and control theory, as well as touching on the military planning concept known as effects-based operations. The sections relating to the COA effort detail the rationale behind the functional models developed and identify technologies that could support the process functions. The functional models include a section related to adversarial modeling, which adds a dynamic to the COA process that is missing in current combat simulations. The information contained in this article lays the foundation for building a unique analytic capability.
This report is an update to previous ''smart gun'' work and the corresponding report that were completed in 1996. It incorporates some new terminology and expanded definitions. This effort is the product of an open source look at what has happened to the ''smart gun'' technology landscape since the 1996 report was published.
The Comprehensive Test Ban Treaty of 1996 banned any future nuclear explosions or testing of nuclear weapons and created the CTBTO in Vienna to implement the treaty. The U.S. response to this was the cessation of all above and below ground nuclear testing. As such, all stockpile reliability assessments are now based on periodic testing of subsystems being stored in a wide variety of environments. This data provides a wealth of information and feeds a growing web of deterministic, physics-based computer models for assessment of stockpile reliability. Unfortunately until 1996 it was difficult to relate the deterministic materials aging test data to component reliability. Since that time we have made great strides in mathematical techniques and computer tools that permit explicit relationships between materials degradation, e.g. corrosion, thermo-mechanical fatigue, and reliability. The resulting suite of tools is known as CRAX and the mathematical library supporting these tools is Cassandra. However, these techniques ignore the historical data that is also available on similar systems in the nuclear stockpile, the DoD weapons complex and even in commercial applications. Traditional statistical techniques commonly used in classical re liability assessment do not permit data from these sources to be easily included in the overall assessment of system reliability. An older, alternative approach based on Bayesian probability theory permits the inclusion of data from all applicable sources. Data from a variety of sources is brought together in a logical fashion through the repeated application of inductive mathematics. This research brings together existing mathematical methods, modifies and expands those techniques as required, permitting data from a wide variety of sources to be combined in a logical fashion to increase the confidence in the reliability assessment of the nuclear weapons stockpile. The application of this research is limited to those systems composed of discrete components, e.g. those that can be characterized as operating or not operating. However, there is nothing unique about the underlying principles and the extension to continuous subsystem/systems is straightforward. The framework is also laid for the consideration of systems with multiple correlated failure modes. While an important consideration, time and resources limited the specific demonstration of these methods.