Energy Consumption of Personal Computing Including Portable
Namboodiri, Vinod
processing unit (CPU) processing power and capacity of mass storage devices doubles every 18 months. Such growth in both processing and storage capabilities fuels the production of ever more powerful portableEnergy Consumption of Personal Computing Including Portable Communication Devices Pavel Somavat1
Typologies of Computation and Computational Models
Mark Burgin; Gordana Dodig-Crnkovic
2013-12-09
We need much better understanding of information processing and computation as its primary form. Future progress of new computational devices capable of dealing with problems of big data, internet of things, semantic web, cognitive robotics and neuroinformatics depends on the adequate models of computation. In this article we first present the current state of the art through systematization of existing models and mechanisms, and outline basic structural framework of computation. We argue that defining computation as information processing, and given that there is no information without (physical) representation, the dynamics of information on the fundamental level is physical/ intrinsic/ natural computation. As a special case, intrinsic computation is used for designed computation in computing machinery. Intrinsic natural computation occurs on variety of levels of physical processes, containing the levels of computation of living organisms (including highly intelligent animals) as well as designed computational devices. The present article offers a typology of current models of computation and indicates future paths for the advancement of the field; both by the development of new computational models and by learning from nature how to better compute using different mechanisms of intrinsic computation.
Simple Model of Membrane Proteins Including Solvent
D. L. Pagan; A. Shiryayev; T. P. Connor; J. D. Gunton
2006-03-04
We report a numerical simulation for the phase diagram of a simple two dimensional model, similar to one proposed by Noro and Frenkel [J. Chem. Phys. \\textbf{114}, 2477 (2001)] for membrane proteins, but one that includes the role of the solvent. We first use Gibbs ensemble Monte Caro simulations to determine the phase behavior of particles interacting via a square-well potential in two dimensions for various values of the interaction range. A phenomenological model for the solute-solvent interactions is then studied to understand how the fluid-fluid coexistence curve is modified by solute-solvent interactions. It is shown that such a model can yield systems with liquid-liquid phase separation curves that have both upper and lower critical points, as well as closed loop phase diagrams, as is the case with the corresponding three dimensional model.
) Total 15 Total 14 Total 14 Total 16 3rd Year 4th Year FALL SPRING FALL SPRING CS 3311 3 CS 4411 4 CS your academic dept. Includes 3 units of co-curricular activities. Effective Fall 2010 1st Year 2nd Year
computational modeling of biological systems
$author.value
Faculty. Faculty listing for "computational modeling of biological systems" ... Research Interests: computational modeling of biological systems.
Parallel computing in enterprise modeling.
Goldsby, Michael E.; Armstrong, Robert C.; Shneider, Max S.; Vanderveen, Keith; Ray, Jaideep; Heath, Zach; Allan, Benjamin A.
2008-08-01
This report presents the results of our efforts to apply high-performance computing to entity-based simulations with a multi-use plugin for parallel computing. We use the term 'Entity-based simulation' to describe a class of simulation which includes both discrete event simulation and agent based simulation. What simulations of this class share, and what differs from more traditional models, is that the result sought is emergent from a large number of contributing entities. Logistic, economic and social simulations are members of this class where things or people are organized or self-organize to produce a solution. Entity-based problems never have an a priori ergodic principle that will greatly simplify calculations. Because the results of entity-based simulations can only be realized at scale, scalable computing is de rigueur for large problems. Having said that, the absence of a spatial organizing principal makes the decomposition of the problem onto processors problematic. In addition, practitioners in this domain commonly use the Java programming language which presents its own problems in a high-performance setting. The plugin we have developed, called the Parallel Particle Data Model, overcomes both of these obstacles and is now being used by two Sandia frameworks: the Decision Analysis Center, and the Seldon social simulation facility. While the ability to engage U.S.-sized problems is now available to the Decision Analysis Center, this plugin is central to the success of Seldon. Because Seldon relies on computationally intensive cognitive sub-models, this work is necessary to achieve the scale necessary for realistic results. With the recent upheavals in the financial markets, and the inscrutability of terrorist activity, this simulation domain will likely need a capability with ever greater fidelity. High-performance computing will play an important part in enabling that greater fidelity.
Computational Models for Understanding Weather
Muraki, David J.
Computational Models for Understanding Weather Mathematics for Atmospheric Science http://weather-S migration Dutton Conway zonal jetstream in unstable weather 6 #12;Baroclinic Instability Vortices
A coke oven model including thermal decomposition kinetics of tar
Munekane, Fuminori; Yamaguchi, Yukio [Mitsubishi Chemical Corp., Yokohama (Japan); Tanioka, Seiichi [Mitsubishi Chemical Corp., Sakaide (Japan)
1997-12-31
A new one-dimensional coke oven model has been developed for simulating the amount and the characteristics of by-products such as tar and gas as well as coke. This model consists of both heat transfer and chemical kinetics including thermal decomposition of coal and tar. The chemical kinetics constants are obtained by estimation based on the results of experiments conducted to investigate the thermal decomposition of both coal and tar. The calculation results using the new model are in good agreement with experimental ones.
Improved computer models support genetics research
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
Simple computer models unravel genetic stress reactions in cells Simple computer models unravel genetic stress reactions in cells Integrated biological and computational methods...
Accounting for the Energy Consumption of Personal Computing Including Portable Devices
Namboodiri, Vinod
that computing consumes more than 3% of the global electricity consumption. (ii) We characterize the powerAccounting for the Energy Consumption of Personal Computing Including Portable Devices Pavel.S.A vinod.namboodiri@wichita.edu ABSTRACT In light of the increased awareness of global energy consumption
LANL computer model boosts engine efficiency
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
LANL computer model boosts engine efficiency LANL computer model boosts engine efficiency The KIVA model has been instrumental in helping researchers and manufacturers understand...
Cummings, P. T.
2010-02-08
This document reports the outcomes of the Computational Nanoscience Project, "Integrated Multiscale Modeling of Molecular Computing Devices". It includes a list of participants and publications arising from the research supported.
Computational systems biology and in silico modeling of the
Borenstein, Elhanan
Computational systems biology and in silico modeling of the human microbiome Elhanan Borenstein Professor at the Santa Fe Institute. His research interests include computational and evolutionary systems is a complex biological system with numerous interacting components across multiple organizational levels
Improved computer models support genetics research
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
February Simple computer models unravel genetic stress reactions in cells Simple computer models unravel genetic stress reactions in cells Integrated biological and...
Robust Resource Allocations in Parallel Computing Systems: Model and Heuristics
Maciejewski, Anthony A. "Tony"
Robust Resource Allocations in Parallel Computing Systems: Model and Heuristics Vladimir Shestak1 in parallel computer systems (including heterogeneous clusters) should be allocated to the computational was supported by the Colorado State University Center for Robustness in Computer Systems (funded by the Colorado
Computer Graphics Hierarchical Modeling
for sta9c models, and vital for anima9on! · Key is to structure the transforma9 Transforma9ons Translate base by (5,0,0) Translate lower arm by (5,0,0) Translate upper arm by (5,0,0) Translate hammer by (5,0,0) ... q Each part
Elastic–Plastic Spherical Contact Modeling Including Roughness Effects
Li, L.; Etsion, I.; Talke, F. E.
2010-01-01
A multilevel model for elastic–plastic contact between ajunction growth of an elastic–plastic spherical contact. J.nite element based elastic–plastic model for the contact of
Optimization of neuron models using grid computing
Vella, Mike
2011-09-09
stream_source_info Mike_Vella.ppt.txt stream_content_type text/plain stream_size 2634 Content-Encoding UTF-8 stream_name Mike_Vella.ppt.txt Content-Type text/plain; charset=UTF-8 Optimization of neuron models using grid computing... channels Set of information to be included in a model is large Single neuron multi-compartment models Why optimize? Single neuron models provide a basis for understanding cell and local circuit function Maximal conductances, compartment capacitances...
Comparison of Joint Modeling Approaches Including Eulerian Sliding...
Office of Scientific and Technical Information (OSTI)
the geologic model are well suited for such an analysis. Unfortunately, current meshing tools are unable to automatically generate adequate hexahedral meshes for large numbers of...
Cosmic Logic: a Computational Model
Vitaly Vanchurin
2015-07-05
We describe a simple computational model of cosmic logic suitable for analysis of, for example, discretized cosmological systems. The construction is based on a particular model of computation, developed by Alan Turing, with cosmic observers (CO), cosmic measures (CM) and cosmic symmetries (CS) described by Turing machines. CO machines always start with a blank tape and CM machines take CO's Turing number (also known as description number or G{\\" o}del number) as input and output the corresponding probability. Similarly, CS machines take CO's Turing number as input, but output either one if the CO machines are in the same equivalence class or zero otherwise. We argue that CS machines are more fundamental than CM machines and, thus, should be used as building blocks in constructing CM machines. We prove the non-computability of a CS machine which discriminates between two classes of CO machines: mortal that halts in finite time and immortal that runs forever. In context of eternal inflation this result implies that it is impossible to construct CM machines to compute probabilities using cut-off prescriptions or that all of the cut-off measures are non-computable.
Including model uncertainty in risk-informed decision-making
Reinert, Joshua M
2005-01-01
Model uncertainties can have a significant impact on decisions regarding licensing basis changes. We present a methodology to identify basic events in the risk assessment that have the potential to change the decision and ...
A framework for modelling trojans and computer virus infection
Cairns, Paul
A framework for modelling trojans and computer virus infection Harold Thimbleby1 , Stuart Anderson2 world, including the possibility of Trojan Horse programs and computer viruses, as simply a finite realisation of a Turing Machine. We consider the actions of Trojan Horses and viruses in real computer systems
Quantum computation beyond the circuit model
Jordan, Stephen Paul
2008-01-01
The quantum circuit model is the most widely used model of quantum computation. It provides both a framework for formulating quantum algorithms and an architecture for the physical construction of quantum computers. However, ...
Cosmic Logic: a Computational Model
Vanchurin, Vitaly
2015-01-01
We describe a simple computational model of cosmic logic suitable for analysis of, for example, discretized cosmological systems. The construction is based on a particular model of computation, developed by Alan Turing, with cosmic observers (CO), cosmic measures (CM) and cosmic symmetries (CS) described by Turing machines. CO machines always start with a blank tape and CM machines take CO's Turing number (also known as description number or G{\\" o}del number) as input and output the corresponding probability. Similarly, CS machines take CO's Turing number as input, but output either one if the CO machines are in the same equivalence class or zero otherwise. We argue that CS machines are more fundamental than CM machines and, thus, should be used as building blocks in constructing CM machines. We prove the non-computability of a CS machine which discriminates between two classes of CO machines: mortal that halts in finite time and immortal that runs forever. In context of eternal inflation this result implies...
Grace, T. M.; Wag, K. J.; Horton, R. R.; Frederick, W. J.
1994-01-01
This paper describes an improved model of char burning during black liquor combustion that is capable of predicting net rates of sulfate reduction to sulfide as well as carbon burnup rates. Enhancements include a proper ...
RELAP5-3D Code Includes Athena Features and Models
Richard A. Riemke; Cliff B. Davis; Richard R. Schultz
2006-07-01
Version 2.3 of the RELAP5-3D computer program includes all features and models previously available only in the ATHENA version of the code. These include the addition of new working fluids (i.e., ammonia, blood, carbon dioxide, glycerol, helium, hydrogen, lead-bismuth, lithium, lithium-lead, nitrogen, potassium, sodium, and sodium-potassium) and a magnetohydrodynamic model that expands the capability of the code to model many more thermal-hydraulic systems. In addition to the new working fluids along with the standard working fluid water, one or more noncondensable gases (e.g., air, argon, carbon dioxide, carbon monoxide, helium, hydrogen, krypton, nitrogen, oxygen, sf6, xenon) can be specified as part of the vapor/gas phase of the working fluid. These noncondensable gases were in previous versions of RELAP5- 3D. Recently four molten salts have been added as working fluids to RELAP5-3D Version 2.4, which has had limited release. These molten salts will be in RELAP5-3D Version 2.5, which will have a general release like RELAP5-3D Version 2.3. Applications that use these new features and models are discussed in this paper.
Cupola Furnace Computer Process Model
Seymour Katz
2004-12-31
The cupola furnace generates more than 50% of the liquid iron used to produce the 9+ million tons of castings annually. The cupola converts iron and steel into cast iron. The main advantages of the cupola furnace are lower energy costs than those of competing furnaces (electric) and the ability to melt less expensive metallic scrap than the competing furnaces. However the chemical and physical processes that take place in the cupola furnace are highly complex making it difficult to operate the furnace in optimal fashion. The results are low energy efficiency and poor recovery of important and expensive alloy elements due to oxidation. Between 1990 and 2004 under the auspices of the Department of Energy, the American Foundry Society and General Motors Corp. a computer simulation of the cupola furnace was developed that accurately describes the complex behavior of the furnace. When provided with the furnace input conditions the model provides accurate values of the output conditions in a matter of seconds. It also provides key diagnostics. Using clues from the diagnostics a trained specialist can infer changes in the operation that will move the system toward higher efficiency. Repeating the process in an iterative fashion leads to near optimum operating conditions with just a few iterations. More advanced uses of the program have been examined. The program is currently being combined with an ''Expert System'' to permit optimization in real time. The program has been combined with ''neural network'' programs to affect very easy scanning of a wide range of furnace operation. Rudimentary efforts were successfully made to operate the furnace using a computer. References to these more advanced systems will be found in the ''Cupola Handbook''. Chapter 27, American Foundry Society, Des Plaines, IL (1999).
Modeling Computer Viruses MSc Thesis (Afstudeerscriptie)
Amsterdam, University of
Modeling Computer Viruses MSc Thesis (Afstudeerscriptie) written by Luite Menno Pieter van Zelst About half a year ago, Alban Ponse, my thesis supervisor, suggested that the topic of `computer viruses indus- try and the creators of computer viruses. After all, the anti-virus industry stands to lose a lot
Integrating Numerical Computation into the Modeling Instruction Curriculum
Caballero, Marcos D; Aiken, John M; Douglas, Scott S; Scanlon, Erin M; Thoms, Brian; Schatz, Michael F
2012-01-01
We describe a way to introduce physics high school students with no background in programming to computational problem-solving experiences. Our approach builds on the great strides made by the Modeling Instruction reform curriculum. This approach emphasizes the practices of "Developing and using models" and "Computational thinking" highlighted by the NRC K-12 science standards framework. We taught 9th-grade students in a Modeling-Instruction-based physics course to construct computational models using the VPython programming environment. Numerical computation within the Modeling Instruction curriculum provides coherence among the curriculum's different force and motion models, links the various representations which the curriculum employs, and extends the curriculum to include real-world problems that are inaccessible to a purely analytic approach.
Climate Modeling using High-Performance Computing
Mirin, A A
2007-02-05
The Center for Applied Scientific Computing (CASC) and the LLNL Climate and Carbon Science Group of Energy and Environment (E and E) are working together to improve predictions of future climate by applying the best available computational methods and computer resources to this problem. Over the last decade, researchers at the Lawrence Livermore National Laboratory (LLNL) have developed a number of climate models that provide state-of-the-art simulations on a wide variety of massively parallel computers. We are now developing and applying a second generation of high-performance climate models. Through the addition of relevant physical processes, we are developing an earth systems modeling capability as well.
Computational modeling of composite material fires.
Brown, Alexander L.; Erickson, Kenneth L.; Hubbard, Joshua Allen; Dodd, Amanda B.
2010-10-01
Composite materials behave differently from conventional fuel sources and have the potential to smolder and burn for extended time periods. As the amount of composite materials on modern aircraft continues to increase, understanding the response of composites in fire environments becomes increasingly important. An effort is ongoing to enhance the capability to simulate composite material response in fires including the decomposition of the composite and the interaction with a fire. To adequately model composite material in a fire, two physical model development tasks are necessary; first, the decomposition model for the composite material and second, the interaction with a fire. A porous media approach for the decomposition model including a time dependent formulation with the effects of heat, mass, species, and momentum transfer of the porous solid and gas phase is being implemented in an engineering code, ARIA. ARIA is a Sandia National Laboratories multiphysics code including a range of capabilities such as incompressible Navier-Stokes equations, energy transport equations, species transport equations, non-Newtonian fluid rheology, linear elastic solid mechanics, and electro-statics. To simulate the fire, FUEGO, also a Sandia National Laboratories code, is coupled to ARIA. FUEGO represents the turbulent, buoyantly driven incompressible flow, heat transfer, mass transfer, and combustion. FUEGO and ARIA are uniquely able to solve this problem because they were designed using a common architecture (SIERRA) that enhances multiphysics coupling and both codes are capable of massively parallel calculations, enhancing performance. The decomposition reaction model is developed from small scale experimental data including thermogravimetric analysis (TGA) and Differential Scanning Calorimetry (DSC) in both nitrogen and air for a range of heating rates and from available data in the literature. The response of the composite material subject to a radiant heat flux boundary condition is examined to study the propagation of decomposition fronts of the epoxy and carbon fiber and their dependence on the ambient conditions such as oxygen concentration, surface flow velocity, and radiant heat flux. In addition to the computational effort, small scaled experimental efforts to attain adequate data used to validate model predictions is ongoing. The goal of this paper is to demonstrate the progress of the capability for a typical composite material and emphasize the path forward.
Destination Choice Model including panel data using WiFi localization in a pedestrian facility
Bierlaire, Michel
Destination Choice Model including panel data using WiFi localization in a pedestrian facility Loïc data using WiFi localization in a pedestrian facility April 2015 EPFL Destination Choice Model including panel data using WiFi localization in a pedestrian facility Loïc Tinguely, Antonin Danalet
Modelling energy efficiency for computation
Reams, Charles
2012-11-13
. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 130 10 List of Tables 3.1 Energy usage breakdown for computing equipment in the United States. Energy figures are in billions of kWh. . . . . . . . . . . . . . . . . . . . . . 40 5.1 Average solution time... understanding of the underlying behavioural properties will inevitably lead to improvements in the practicality of NTC, and practical NTC-purposed cores have now been constructed; for example, the Phoenix processor, which operates in the near-threshold region...
Computer aided nuclear reactor modeling
Warraich, Khalid Sarwar
1995-01-01
Nuclear reactor modeling is an important activity that lets us analyze existing as well as proposed systems for safety, correct operation, etc. The quality of a analysis is directly proportional to the quality of the model used. In this work we look...
Regional weather modeling on parallel computers.
Baillie, C.; Michalakes, J.; Skalin, R.; Mathematics and Computer Science; NOAA Forecast Systems Lab.; Norwegian Meteorological Inst.
1997-01-01
This special issue on 'regional weather models' complements the October 1995 special issue on 'climate and weather modeling', which focused on global models. In this introduction we review the similarities and differences between regional and global atmospheric models. Next, the structure of regional models is described and we consider how the basic algorithms applied in these models influence the parallelization strategy. Finally, we give a brief overview of the eight articles in this issue and discuss some remaining challenges in the area of adapting regional weather models to parallel computers.
Computationally Efficient Modeling of High-Efficiency Clean Combustion...
Office of Energy Efficiency and Renewable Energy (EERE) Indexed Site
& Publications Computationally Efficient Modeling of High-Efficiency Clean Combustion Engines Computationally Efficient Modeling of High-Efficiency Clean Combustion Engines...
Modeling-Computer Simulations At Fish Lake Valley Area (Deymonaz...
Modeling-Computer Simulations At Fish Lake Valley Area (Deymonaz, Et Al., 2008) Jump to: navigation, search GEOTHERMAL ENERGYGeothermal Home Exploration Activity: Modeling-Computer...
HIV virus spread and evolution studied through computer modeling
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
HIV and evolution studied through computer modeling HIV virus spread and evolution studied through computer modeling This approach distinguishes between susceptible and infected...
Modeling-Computer Simulations At Nw Basin & Range Region (Pritchett...
Modeling-Computer Simulations At Nw Basin & Range Region (Pritchett, 2004) Jump to: navigation, search GEOTHERMAL ENERGYGeothermal Home Exploration Activity: Modeling-Computer...
Modeling-Computer Simulations At Nw Basin & Range Region (Biasi...
Modeling-Computer Simulations At Nw Basin & Range Region (Biasi, Et Al., 2009) Jump to: navigation, search GEOTHERMAL ENERGYGeothermal Home Exploration Activity: Modeling-Computer...
Computer modeling reveals how surprisingly potent hepatitis C...
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
Hepatitis C computer modeling Computer modeling reveals how surprisingly potent hepatitis C drug works A study reveals how daclatasvir targets one of its proteins and causes the...
Modeling and Computational Strategies for Optimal Development Planning of Offshore
Grossmann, Ignacio E.
1 Modeling and Computational Strategies for Optimal Development Planning of Offshore Oilfields for offshore oil and gas fields as a basis to include the generic fiscal rules with ringfencing provisions-integer programming. 1 Introduction Offshore oil and gas field development planning has received significant attention
A Computational Model for Adaptive Emotion Regulation
Treur, Jan
A Computational Model for Adaptive Emotion Regulation Tibor Bosse, Matthijs Pontier, and Jan Treur} Abstract. Emotion regulation describes how a subject can use certain strategies to affect emotion response levels. Usually, models for emotion regulation as- sume mechanisms based on feedback loops that indicate
Computationally Efficient Model for Dopant Precipitation Kinetics
Dunham, Scott
Computationally Efficient Model for Dopant Precipitation Kinetics Iuval Clejan and Scott T. Dunham and precipitates. Dopant deactivation is typically modeled using a steadystate solid solubility or clustering such as dopant activation/deactivation, it is essential to consider the fact that precipitation involves a range
Mechanistic Models in Computational Social Science
Holme, Petter
2015-01-01
Quantitative social science is not only about regression analysis or, in general, data inference. Computer simulations of social mechanisms have an over 60 years long history. They have been used for many different purposes -- to test scenarios, to test the consistency of descriptive theories (proof-of-concept models), to explore emerging phenomena, for forecasting, etc. In this essay, we sketch these historical developments, the role of mechanistic models in the social sciences and the influences from natural and formal sciences. We argue that mechanistic computational models form a natural common ground for social and natural sciences, and look forward to possible future information flow across the social-natural divide.
Decision Model for Cloud Computing
Kondo, Derrick
with different pricing models for cost-cutting, resource-hungry users. Second, prices can differ dynamically (as Grenoble, France 1 #12;Trade-offs Supercomputers Performance Reliability Cost ($) low high high high Instances Â· "Spot" instance price varies dynamically Â· Spot instance provided when user's bid is greater
Modeling dynamic crack propagation in fiber reinforced composites including frictional effects q
Espinosa, Horacio D.
Modeling dynamic crack propagation in fiber reinforced composites including frictional effects q S Abstract Dynamic crack propagation in a unidirectional carbon/epoxy composite is studied through finite deformation anisotropic visco-plastic model is used to describe the constitutive response of the composite
A non-isothermal PEM fuel cell model including two water transport mechanisms in the
Münster, Westfälische Wilhelms-Universität
A non-isothermal PEM fuel cell model including two water transport mechanisms in the membrane K Freiburg Germany A dynamic two-phase flow model for proton exchange mem- brane (PEM) fuel cells and the species concentrations. In order to describe the charge transport in the fuel cell the Poisson equations
Including Source-Specific Phosphorus Mobility in a Nonpoint Source Pollution Model
Walter, M.Todd
pollution; Nonpoint pollution. Introduction In order to protect water quality, watershed managers tasked with implementing strategies for controlling nonpoint source NPS pollution need water quality models that canIncluding Source-Specific Phosphorus Mobility in a Nonpoint Source Pollution Model for Agricultural
CDF computing and event data models
Snider, F.D.; /Fermilab
2005-12-01
The authors discuss the computing systems, usage patterns and event data models used to analyze Run II data from the CDF-II experiment at the Tevatron collider. A critical analysis of the current implementation and design reveals some of the stronger and weaker elements of the system, which serve as lessons for future experiments. They highlight a need to maintain simplicity for users in the face of an increasingly complex computing environment.
District-heating strategy model: computer programmer's manual
Kuzanek, J.F.
1982-05-01
The US Department of Housing and Urban Development (HUD) and the US Department of Energy (DOE) cosponsor a program aimed at increasing the number of district heating and cooling (DHC) systems. Such systems can reduce the amount and costs of fuels used to heat and cool buildings in a district. Twenty-eight communities have agreed to aid HUD in a national feasibility assessment of DHC systems. The HUD/DOE program entails technical assistance by Argonne National Laboratory and Oak Ridge National Laboratory. The assistance includes a computer program, called the district heating strategy model (DHSM), that performs preliminary calculations to analyze potential DHC systems. This report describes the general capabilities of the DHSM, provides historical background on its development, and explains the computer installation and operation of the model - including the data file structures and the options. Sample problems illustrate the structure of the various input data files, the interactive computer-output listings. The report is written primarily for computer programmers responsible for installing the model on their computer systems, entering data, running the model, and implementing local modifications to the code.
Modeling Computations in a Semantic Network
Marko A. Rodriguez; Johan Bollen
2007-05-31
Semantic network research has seen a resurgence from its early history in the cognitive sciences with the inception of the Semantic Web initiative. The Semantic Web effort has brought forth an array of technologies that support the encoding, storage, and querying of the semantic network data structure at the world stage. Currently, the popular conception of the Semantic Web is that of a data modeling medium where real and conceptual entities are related in semantically meaningful ways. However, new models have emerged that explicitly encode procedural information within the semantic network substrate. With these new technologies, the Semantic Web has evolved from a data modeling medium to a computational medium. This article provides a classification of existing computational modeling efforts and the requirements of supporting technologies that will aid in the further growth of this burgeoning domain.
High performance computing and numerical modelling
,
2014-01-01
Numerical methods play an ever more important role in astrophysics. This is especially true in theoretical works, but of course, even in purely observational projects, data analysis without massive use of computational methods has become unthinkable. The key utility of computer simulations comes from their ability to solve complex systems of equations that are either intractable with analytic techniques or only amenable to highly approximative treatments. Simulations are best viewed as a powerful complement to analytic reasoning, and as the method of choice to model systems that feature enormous physical complexity such as star formation in evolving galaxies, the topic of this 43rd Saas Fee Advanced Course. The organizers asked me to lecture about high performance computing and numerical modelling in this winter school, and to specifically cover the basics of numerically treating gravity and hydrodynamics in the context of galaxy evolution. This is still a vast field, and I necessarily had to select a subset ...
Cummings, Peter [Vanderbilt University] [Vanderbilt University
2009-11-15
The document is the final report of the DOE Computational Nanoscience Project DE-FG02-03ER46096: Integrated Multiscale Modeling of Molecular Computing Devices. It included references to 62 publications that were supported by the grant.
Scot Martin
2013-01-31
The chemical evolution of secondary-organic-aerosol (SOA) particles and how this evolution alters their cloud-nucleating properties were studied. Simplified forms of full Koehler theory were targeted, specifically forms that contain only those aspects essential to describing the laboratory observations, because of the requirement to minimize computational burden for use in integrated climate and chemistry models. The associated data analysis and interpretation have therefore focused on model development in the framework of modified kappa-Koehler theory. Kappa is a single parameter describing effective hygroscopicity, grouping together several separate physicochemical parameters (e.g., molar volume, surface tension, and van't Hoff factor) that otherwise must be tracked and evaluated in an iterative full-Koehler equation in a large-scale model. A major finding of the project was that secondary organic materials produced by the oxidation of a range of biogenic volatile organic compounds for diverse conditions have kappa values bracketed in the range of 0.10 +/- 0.05. In these same experiments, somewhat incongruently there was significant chemical variation in the secondary organic material, especially oxidation state, as was indicated by changes in the particle mass spectra. Taken together, these findings then support the use of kappa as a simplified yet accurate general parameter to represent the CCN activation of secondary organic material in large-scale atmospheric and climate models, thereby greatly reducing the computational burden while simultaneously including the most recent mechanistic findings of laboratory studies.
Kim, G.; Pesaran, A.; Smith, K.; Graf, P.; Jun, M.; Yang, C.; Li, G.; Li, S.; Hochman, A.; Tselepidakis, D.; White, J.
2014-06-01
This presentation discusses the significant enhancement of computational efficiency in nonlinear multiscale battery model for computer aided engineering in current research at NREL.
A wave equation including leptons and quarks for the standard model of quantum physics in
Boyer, Edmond
A wave equation including leptons and quarks for the standard model of quantum physics in Clifford-m@orange.fr August 27, 2014 Abstract A wave equation with mass term is studied for all particles and an- tiparticles of color and antiquarks u and d. This wave equation is form invariant under the Cl 3 group generalizing
Discrimination of the Native from Misfolded Protein Models with an Energy Function Including
Lazaridis, Themis
Discrimination of the Native from Misfolded Protein Models with an Energy Function Including for theoretical protein structure prediction is an energy function that can discriminate the native from non by the widespread belief that they are superior for such discrimination to physics-based energy functions
Grossmann, Ignacio E.
Method Two objectives are included in the model: A cost minimization and an environmental impact minimization. The environmental impact is evaluated using the Eco- indicator 99 method. A set of Pareto optimal solutions is generated using the -constraint method. Conclusions · Environmental impact can be reduced by up
Sensitivity Analysis Methodology for a Complex System Computational Model
1 Sensitivity Analysis Methodology for a Complex System Computational Model James J. Filliben of computational models to serve as predictive surrogates for the system. The use of such models increasingly) of a computational model for a complex system is always an essential component in accepting/rejecting such a model
Computational Modeling of Self-organization of Dislocations and...
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
Computational Modeling of Self-organization of Dislocations and Mesoscale Deformation of Metals Event Sponsor: Mathematics and Computing Science - LANS Seminar Start Date: Jun 19...
Computer Modelling of 3D Geological Surface
Kodge, B G
2011-01-01
The geological surveying presently uses methods and tools for the computer modeling of 3D-structures of the geographical subsurface and geotechnical characterization as well as the application of geoinformation systems for management and analysis of spatial data, and their cartographic presentation. The objectives of this paper are to present a 3D geological surface model of Latur district in Maharashtra state of India. This study is undertaken through the several processes which are discussed in this paper to generate and visualize the automated 3D geological surface model of a projected area.
Baer, Ferdinand
Optimizing Computations in Weather and Climate Prediction Models* F. BAER, BANGLIN ZHANG, AND BING scenarios for many time scales, more computer power than is currently available will be needed. One and sometimes with a biosphere included, are very complex and require so much computing power on available
Wild Fire Computer Model Helps Firefighters
Canfield, Jesse
2012-09-04
A high-tech computer model called HIGRAD/FIRETEC, the cornerstone of a collaborative effort between U.S. Forest Service Rocky Mountain Research Station and Los Alamos National Laboratory, provides insights that are essential for front-line fire fighters. The science team is looking into levels of bark beetle-induced conditions that lead to drastic changes in fire behavior and how variable or erratic the behavior is likely to be.
Wild Fire Computer Model Helps Firefighters
Canfield, Jesse
2014-06-02
A high-tech computer model called HIGRAD/FIRETEC, the cornerstone of a collaborative effort between U.S. Forest Service Rocky Mountain Research Station and Los Alamos National Laboratory, provides insights that are essential for front-line fire fighters. The science team is looking into levels of bark beetle-induced conditions that lead to drastic changes in fire behavior and how variable or erratic the behavior is likely to be.
COMPUTATIONAL MODELING OF CIRCULATING FLUIDIZED BED REACTORS
Ibrahim, Essam A
2013-01-09
Details of numerical simulations of two-phase gas-solid turbulent flow in the riser section of Circulating Fluidized Bed Reactor (CFBR) using Computational Fluid Dynamics (CFD) technique are reported. Two CFBR riser configurations are considered and modeled. Each of these two riser models consist of inlet, exit, connecting elbows and a main pipe. Both riser configurations are cylindrical and have the same diameter but differ in their inlet lengths and main pipe height to enable investigation of riser geometrical scaling effects. In addition, two types of solid particles are exploited in the solid phase of the two-phase gas-solid riser flow simulations to study the influence of solid loading ratio on flow patterns. The gaseous phase in the two-phase flow is represented by standard atmospheric air. The CFD-based FLUENT software is employed to obtain steady state and transient solutions for flow modulations in the riser. The physical dimensions, types and numbers of computation meshes, and solution methodology utilized in the present work are stated. Flow parameters, such as static and dynamic pressure, species velocity, and volume fractions are monitored and analyzed. The differences in the computational results between the two models, under steady and transient conditions, are compared, contrasted, and discussed.
Dang, Zhe
Bond Computing Systems: a Biologically Inspired and High-level Dynamics Model for Pervasive. Targeting at modeling the high-level dynamics of pervasive comput- ing systems, we introduce Bond Computing are regular, and study their computation power and verification problems. Among other results, we show
Christian W. Straka; Pierre Demarque; D. B. Guenther; Linghuai Li; Frank J. Robinson
2005-09-14
The space telescope MOST is now providing us with extremely accurate low frequency p-mode oscillation data for the star Eta Boo. We demonstrate in this paper that these data, when combined with ground based measurements of the high frequency p-mode spectrum, can be reproduced with stellar models that include the effects of turbulence in their outer layers. Without turbulence, the l=0 modes of our models deviate from either the ground based or the space data by about 1.5-4.0 micro Hz. This discrepancy can be completely removed by including turbulence in the models and we can exactly match 12 out of 13 MOST frequencies that we identified as l=0 modes in addition to 13 out of 21 ground based frequencies within their observational 2 sigma tolerances. The better agreement between model frequencies and observed ones depends for the most part on the turbulent kinetic energy which was taken from a 3D convection simulation for the Sun.
The Food Crises: A quantitative model of food prices including speculators and ethanol conversion
Lagi, Marco; Bertrand, Karla Z; Bar-Yam, Yaneer
2011-01-01
Recent increases in basic food prices are severely impacting vulnerable populations worldwide. Proposed causes such as shortages of grain due to adverse weather, increasing meat consumption in China and India, conversion of corn to ethanol in the US, and investor speculation on commodity markets lead to widely differing implications for policy. A lack of clarity about which factors are responsible reinforces policy inaction. Here, for the first time, we construct a dynamic model that quantitatively agrees with food prices. The results show that the dominant causes of price increases are investor speculation and ethanol conversion. Models that just treat supply and demand are not consistent with the actual price dynamics. The two sharp peaks in 2007/2008 and 2010/2011 are specifically due to investor speculation, while an underlying upward trend is due to increasing demand from ethanol conversion. The model includes investor trend following as well as shifting between commodities, equities and bonds to take ad...
7. Business Models LearningsfromfoundingaComputerVisionStartup
Solem, Jan Erik
7. Business Models #12;LearningsfromfoundingaComputerVisionStartup Flickr:dystopos How are you models ! ! (not only technology) #12;LearningsfromfoundingaComputerVisionStartup Auction business model! Bricks and clicks business model! Collective business models! Component business model! Cutting out
Model for the catalytic oxidation of CO, including gas-phase impurities and CO desorption
Buendia, G M
2013-01-01
We present results of kinetic Monte Carlo simulations of a modified Ziff-Gulari-Barshad model for the reaction CO+O --> CO_2 on a catalytic surface. Our model includes impurities in the gas phase, CO desorption, and a modification known to eliminate the unphysical O poisoned phase. The impurities can adsorb and desorb on the surface, but otherwise remain inert. In a previous work that did not include CO desorption [G. M. Buendia and P. A. Rikvold, Phys. Rev. E, 85 031143 (2012)], we found that the impurities have very distinctive effects on the phase diagram and greatly diminish the reactivity of the system. If the impurities do not desorb, once the system reaches a stationary state, the CO_2 production disappears. When the impurities are allowed to desorb, there are regions where the CO_2 reaction window reappears, although greatly reduced. Following experimental evidence that indicates that temperature effects are crucial in many catalytic processes, here we further analyze these effects by including a CO d...
Wind energy conversion system analysis model (WECSAM) computer program documentation
Downey, W T; Hendrick, P L
1982-07-01
Described is a computer-based wind energy conversion system analysis model (WECSAM) developed to predict the technical and economic performance of wind energy conversion systems (WECS). The model is written in CDC FORTRAN V. The version described accesses a data base containing wind resource data, application loads, WECS performance characteristics, utility rates, state taxes, and state subsidies for a six state region (Minnesota, Michigan, Wisconsin, Illinois, Ohio, and Indiana). The model is designed for analysis at the county level. The computer model includes a technical performance module and an economic evaluation module. The modules can be run separately or together. The model can be run for any single user-selected county within the region or looped automatically through all counties within the region. In addition, the model has a restart capability that allows the user to modify any data-base value written to a scratch file prior to the technical or economic evaluation. Thus, any user-supplied data for WECS performance, application load, utility rates, or wind resource may be entered into the scratch file to override the default data-base value. After the model and the inputs required from the user and derived from the data base are described, the model output and the various output options that can be exercised by the user are detailed. The general operation is set forth and suggestions are made for efficient modes of operation. Sample listings of various input, output, and data-base files are appended. (LEW)
Modeling-Computer Simulations At San Juan Volcanic Field Area...
Area Exploration Technique Modeling-Computer Simulations Activity Date Usefulness useful DOE-funding Unknown Notes In this study we combine thermal maturation models, based on the...
Computational models of intergroup competition and warfare.
Letendre, Kenneth (University of New Mexico); Abbott, Robert G.
2011-11-01
This document reports on the research of Kenneth Letendre, the recipient of a Sandia Graduate Research Fellowship at the University of New Mexico. Warfare is an extreme form of intergroup competition in which individuals make extreme sacrifices for the benefit of their nation or other group to which they belong. Among animals, limited, non-lethal competition is the norm. It is not fully understood what factors lead to warfare. We studied the global variation in the frequency of civil conflict among countries of the world, and its positive association with variation in the intensity of infectious disease. We demonstrated that the burden of human infectious disease importantly predicts the frequency of civil conflict and tested a causal model for this association based on the parasite-stress theory of sociality. We also investigated the organization of social foraging by colonies of harvester ants in the genus Pogonomyrmex, using both field studies and computer models.
Dang, Zhe
Bond Computing Systems: a Biologically Inspired and High-level Dynamics Model for Pervasive their com- putation power and verification problems. Among other results, we show that the computing power) techniques for pervasive computing systems. At a high-level, there are at least two views in modeling
Preliminary Phase Field Computational Model Development
Li, Yulan; Hu, Shenyang Y.; Xu, Ke; Suter, Jonathan D.; McCloy, John S.; Johnson, Bradley R.; Ramuhalli, Pradeep
2014-12-15
This interim report presents progress towards the development of meso-scale models of magnetic behavior that incorporate microstructural information. Modeling magnetic signatures in irradiated materials with complex microstructures (such as structural steels) is a significant challenge. The complexity is addressed incrementally, using the monocrystalline Fe (i.e., ferrite) film as model systems to develop and validate initial models, followed by polycrystalline Fe films, and by more complicated and representative alloys. In addition, the modeling incrementally addresses inclusion of other major phases (e.g., martensite, austenite), minor magnetic phases (e.g., carbides, FeCr precipitates), and minor nonmagnetic phases (e.g., Cu precipitates, voids). The focus of the magnetic modeling is on phase-field models. The models are based on the numerical solution to the Landau-Lifshitz-Gilbert equation. From the computational standpoint, phase-field modeling allows the simulation of large enough systems that relevant defect structures and their effects on functional properties like magnetism can be simulated. To date, two phase-field models have been generated in support of this work. First, a bulk iron model with periodic boundary conditions was generated as a proof-of-concept to investigate major loop effects of single versus polycrystalline bulk iron and effects of single non-magnetic defects. More recently, to support the experimental program herein using iron thin films, a new model was generated that uses finite boundary conditions representing surfaces and edges. This model has provided key insights into the domain structures observed in magnetic force microscopy (MFM) measurements. Simulation results for single crystal thin-film iron indicate the feasibility of the model for determining magnetic domain wall thickness and mobility in an externally applied field. Because the phase-field model dimensions are limited relative to the size of most specimens used in experiments, special experimental methods were devised to create similar boundary conditions in the iron films. Preliminary MFM studies conducted on single and polycrystalline iron films with small sub-areas created with focused ion beam have correlated quite well qualitatively with phase-field simulations. However, phase-field model dimensions are still small relative to experiments thus far. We are in the process of increasing the size of the models and decreasing specimen size so both have identical dimensions. Ongoing research is focused on validation of the phase-field model. Validation is being accomplished through comparison with experimentally obtained MFM images (in progress), and planned measurements of major hysteresis loops and first order reversal curves. Extrapolation of simulation sizes to represent a more stochastic bulk-like system will require sampling of various simulations (i.e., with single non-magnetic defect, single magnetic defect, single grain boundary, single dislocation, etc.) with distributions of input parameters. These outputs can then be compared to laboratory magnetic measurements and ultimately to simulate magnetic Barkhausen noise signals.
Theoretical Neuroscience: Computational and Mathematical Modeling of Neural Systems
Olshausen, Bruno
Theoretical Neuroscience: Computational and Mathematical Modeling of Neural Systems Peter Dayan: Computational and Mathematical Modeling of Neural Systems. The emergence of this book represents more than- scientist was brought up on ``Kandel and Schwartz.'' Now, at last, the field of computational neuroscience
What can computational models tell us about face processing?
Cottrell, Garrison W.
What can computational models tell us about face processing? Garrison W. Cottrell Gary about face processing? Garrison W. Cottrell Gary's Unbelievable Research Unit (GURU) Computer Science, Lingyun Zhang What can computational models tell us about face processing? Garrison W. Cottrell Gary
Center for Programming Models for Scalable Parallel Computing
John Mellor-Crummey
2008-02-29
Rice University's achievements as part of the Center for Programming Models for Scalable Parallel Computing include: (1) design and implemention of cafc, the first multi-platform CAF compiler for distributed and shared-memory machines, (2) performance studies of the efficiency of programs written using the CAF and UPC programming models, (3) a novel technique to analyze explicitly-parallel SPMD programs that facilitates optimization, (4) design, implementation, and evaluation of new language features for CAF, including communication topologies, multi-version variables, and distributed multithreading to simplify development of high-performance codes in CAF, and (5) a synchronization strength reduction transformation for automatically replacing barrier-based synchronization with more efficient point-to-point synchronization. The prototype Co-array Fortran compiler cafc developed in this project is available as open source software from http://www.hipersoft.rice.edu/caf.
Conrad, Patrick Raymond
2014-01-01
Computational models of complex phenomena are an important resource for scientists and engineers. However, many state-of-the-art simulations of physical systems are computationally expensive to evaluate and are black ...
Integration of engineering models in computer-aided preliminary design
Lajoie, Ronnie M.
The problems of the integration of engineering models in computer-aided preliminary design are reviewed. This paper details the research, development, and testing of modifications to Paper Airplane, a LISP-based computer ...
Computational model of miniature pulsating heat pipes.
Martinez, Mario J.; Givler, Richard C.
2013-01-01
The modeling work described herein represents Sandia National Laboratories (SNL) portion of a collaborative three-year project with Northrop Grumman Electronic Systems (NGES) and the University of Missouri to develop an advanced, thermal ground-plane (TGP), which is a device, of planar configuration, that delivers heat from a source to an ambient environment with high efficiency. Work at all three institutions was funded by DARPA/MTO; Sandia was funded under DARPA/MTO project number 015070924. This is the final report on this project for SNL. This report presents a numerical model of a pulsating heat pipe, a device employing a two phase (liquid and its vapor) working fluid confined in a closed loop channel etched/milled into a serpentine configuration in a solid metal plate. The device delivers heat from an evaporator (hot zone) to a condenser (cold zone). This new model includes key physical processes important to the operation of flat plate pulsating heat pipes (e.g. dynamic bubble nucleation, evaporation and condensation), together with conjugate heat transfer with the solid portion of the device. The model qualitatively and quantitatively predicts performance characteristics and metrics, which was demonstrated by favorable comparisons with experimental results on similar configurations. Application of the model also corroborated many previous performance observations with respect to key parameters such as heat load, fill ratio and orientation.
Predictive Capability Maturity Model for computational modeling and simulation.
Oberkampf, William Louis; Trucano, Timothy Guy; Pilch, Martin M.
2007-10-01
The Predictive Capability Maturity Model (PCMM) is a new model that can be used to assess the level of maturity of computational modeling and simulation (M&S) efforts. The development of the model is based on both the authors experience and their analysis of similar investigations in the past. The perspective taken in this report is one of judging the usefulness of a predictive capability that relies on the numerical solution to partial differential equations to better inform and improve decision making. The review of past investigations, such as the Software Engineering Institute's Capability Maturity Model Integration and the National Aeronautics and Space Administration and Department of Defense Technology Readiness Levels, indicates that a more restricted, more interpretable method is needed to assess the maturity of an M&S effort. The PCMM addresses six contributing elements to M&S: (1) representation and geometric fidelity, (2) physics and material model fidelity, (3) code verification, (4) solution verification, (5) model validation, and (6) uncertainty quantification and sensitivity analysis. For each of these elements, attributes are identified that characterize four increasing levels of maturity. Importantly, the PCMM is a structured method for assessing the maturity of an M&S effort that is directed toward an engineering application of interest. The PCMM does not assess whether the M&S effort, the accuracy of the predictions, or the performance of the engineering system satisfies or does not satisfy specified application requirements.
A Vast Machine Computer Models, Climate Data, and the Politics of Global Warming
Edwards, Paul N.
A Vast Machine Computer Models, Climate Data, and the Politics of Global Warming Paul N. Edwards models, climate data, and the politics of global warming / Paul N. Edwards. p. cm. Includes this: Global warming is a myth. It's all model predictions, nothing but simulations. Before you believe
Computer Virus Propagation Models Giuseppe Serazzi and Stefano Zanero
Zanero, Stefano
Computer Virus Propagation Models Giuseppe Serazzi and Stefano Zanero Dipartimento di Elettronica e.zanero@polimi.it Abstract. The availability of reliable models of computer virus propa- gation would prove useful. In this pa- per, we review the most popular models of virus propagation, analyzing the underlying assumptions
Department of Computing CSP||B modelling for railway verification
Doran, Simon J.
University of Surrey Department of Computing Computing Sciences Report CS-12-03 CSP||B modelling Schneider Helen Treharne March 30th 2012 #12;CSP||B modelling for railway verification: the double junction work in verifying railway systems through CSP k B modelling and analysis. In particular we consider
Disruptive technology business models in cloud computing
Krikos, Alexis Christopher
2010-01-01
Cloud computing, a term whose origins have been in existence for more than a decade, has come into fruition due to technological capabilities and marketplace demands. Cloud computing can be defined as a scalable and flexible ...
A Unified Computational Model for Solar and Stellar Flares
Allred, Joel C; Carlsson, Mats
2015-01-01
We present a unified computational framework which can be used to describe impulsive flares on the Sun and on dMe stars. The models assume that the flare impulsive phase is caused by a beam of charged particles that is accelerated in the corona and propagates downward depositing energy and momentum along the way. This rapidly heats the lower stellar atmosphere causing it to explosively expand and dramatically brighten. Our models consist of flux tubes that extend from the sub-photosphere into the corona. We simulate how flare-accelerated charged particles propagate down one-dimensional flux tubes and heat the stellar atmosphere using the Fokker-Planck kinetic theory. Detailed radiative transfer is included so that model predictions can be directly compared with observations. The flux of flare-accelerated particles drives return currents which additionally heat the stellar atmosphere. These effects are also included in our models. We examine the impact of the flare-accelerated particle beams on model solar and...
Modeling-Computer Simulations At Nevada Test And Training Range...
Modeling-Computer Simulations At Nevada Test And Training Range Area (Sabin, Et Al., 2004) Jump to: navigation, search GEOTHERMAL ENERGYGeothermal Home Exploration Activity:...
Modeling-Computer Simulations At Fenton Hill HDR Geothermal Area...
Modeling-Computer Simulations At Fenton Hill HDR Geothermal Area (Brown & DuTeaux, 1997) Jump to: navigation, search GEOTHERMAL ENERGYGeothermal Home Exploration Activity:...
Modeling-Computer Simulations At Walker-Lane Transitional Zone...
Pritchett, 2004) Exploration Activity Details Location Walker-Lane Transition Zone Geothermal Region Exploration Technique Modeling-Computer Simulations Activity Date Usefulness...
New partnership uses advanced computer science modeling to address...
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
partnership uses advanced computer science modeling to address climate change | National Nuclear Security Administration Facebook Twitter Youtube Flickr RSS People Mission Managing...
COMPUTER MODELING OF NUCLIDE ADSORPTION ON GEOLOGIC MATERIALS
Silva, R.J.
2010-01-01
11899 G COMPUTER MODELING OF NUCLIDE ADSORPTION ON GEOLOGICdefined as the amount of nuclide adsorbed per gram of claydivided by the amount of nuclide per milliliter of solution,
Modeling-Computer Simulations At Central Nevada Seismic Zone...
Modeling-Computer Simulations At Central Nevada Seismic Zone Region (Pritchett, 2004) Jump to: navigation, search GEOTHERMAL ENERGYGeothermal Home Exploration Activity:...
Modeling-Computer Simulations At Central Nevada Seismic Zone...
Modeling-Computer Simulations At Central Nevada Seismic Zone Region (Biasi, Et Al., 2009) Jump to: navigation, search GEOTHERMAL ENERGYGeothermal Home Exploration Activity:...
MaRIE theory, modeling and computation roadmap executive summary...
Office of Scientific and Technical Information (OSTI)
discovery, with theory and high performance computing, itself co-designed by constrained optimization of hardware and software, and experiments. MaRIE's theory, modeling, and...
Modeling-Computer Simulations At Stillwater Area (Wisian & Blackwell...
Exploration Activity Details Location Stillwater Area Exploration Technique Modeling-Computer Simulations Activity Date Usefulness not indicated DOE-funding Unknown References...
Modeling-Computer Simulations At Desert Peak Area (Wisian & Blackwell...
Exploration Activity Details Location Desert Peak Area Exploration Technique Modeling-Computer Simulations Activity Date Usefulness not indicated DOE-funding Unknown References...
Parameter Discovery for Stochastic Computational Models in Systems Biology Using Bayesian Model
Parameter Discovery for Stochastic Computational Models in Systems Biology Using Bayesian Model--Parameterized probabilistic complex computational (P2 C2 ) models are increasingly used in computational systems biology to study biochemical and physiological systems. A key challenge is to build mechanistic P2 C2 models
"Creating computational models of biological systems to better combat
Zhigilei, Leonid V.
"Creating computational models of biological systems to better combat dangerous pathogens and human of Biomedical Engineering University of Virginia Charlottesville, VA 434.924.8195 Computational Systems Biology system in biofuel and nutraceutical production. With the aid of computational techniques, we can predict
Bytecode unification of geospatial computable models Bytecode unification of geospatial
KÃ¶bben, Barend
heterogeneous to fix and reuse. Field-based and objects-based geospatial models of- ten share common GIS data and object-based data models, and other challenges re- garding synergy of geospatial systems that need to use both types of data models. Keywords: field-based model, object-based model, computability, managed
Towards Real Earth Models --Computational Geophysics on Unstructured Tetrahedral Meshes?
Farquharson, Colin G.
Towards Real Earth Models -- Computational Geophysics on Unstructured Tetrahedral Meshes? Colin tetrahedral meshes. EM geophysics on unstructured tetrahedral meshes. Disadvantages, difficulties, challenges. Conclusions. #12;Outline: Geological models! Advantages of unstructured tetrahedral meshes. EM geophysics
Applying High Performance Computing to Analyzing by Probabilistic Model Checking
Schneider, Carsten
Applying High Performance Computing to Analyzing by Probabilistic Model Checking Mobile Cellular on the use of high performance computing in order to analyze with the proba- bilistic model checker PRISM. The Figure Generation Script 22 2 #12;1. Introduction We report in this paper on the use of high performance
Overview of ASC Capability Computing System Governance Model
Doebling, Scott W. [Los Alamos National Laboratory
2012-07-11
This document contains a description of the Advanced Simulation and Computing Program's Capability Computing System Governance Model. Objectives of the Governance Model are to ensure that the capability system resources are allocated on a priority-driven basis according to the Program requirements; and to utilize ASC Capability Systems for the large capability jobs for which they were designed and procured.
An Interactive Computer Model of Two-Country Trade
Hamlen, Kevin W.
country when free trade is made available. One of the most important learning lessons for the students91 An Interactive Computer Model of Two-Country Trade Bill Hamlen and Kevin Hamlen Abstract We introduce an interactive computer model of two-country trade that allows students to investigate
ONE-DIMENSIONAL PSEUDO-HOMOGENEOUS PACKED BED REACTOR MODELING INCLUDING NO-CO KINETICS
Srinivasan, Anand
2011-08-31
-homogeneous packed bed reactor model for this type of reactor setup built in collaboration with the Chemical and Petroleum Engineering Department at the University of Kansas. A brief summary of the pseudo-homogeneous model is presented in order to properly develop...
Los Alamos CCS (Center for Computer Security) formal computer security model
Dreicer, J.S.; Hunteman, W.J. (Los Alamos National Lab., NM (USA))
1989-01-01
This paper provides a brief presentation of the formal computer security model currently being developed at the Los Alamos Department of Energy (DOE) Center for Computer Security (CCS). The initial motivation for this effort was the need to provide a method by which DOE computer security policy implementation could be tested and verified. The actual analytical model was a result of the integration of current research in computer security and previous modeling and research experiences. The model is being developed to define a generic view of the computer and network security domains, to provide a theoretical basis for the design of a security model, and to address the limitations of present models. Formal mathematical models for computer security have been designed and developed in conjunction with attempts to build secure computer systems since the early 70's. The foundation of the Los Alamos DOE CCS model is a series of functionally dependent probability equations, relations, and expressions. The mathematical basis appears to be justified and is undergoing continued discrimination and evolution. We expect to apply the model to the discipline of the Bell-Lapadula abstract sets of objects and subjects. 5 refs.
Pandoe, Wahyu Widodo
2004-09-30
provides a basis for determining how the water circulation three-dimensionally controls the hydrodynamics of the system and ultimately transports the suspended and soluble materials due to combined currents and waves. A three-dimensional circulation model...
A model for including Arduino microcontroller programming in the introductory physics lab
Haugen, Andrew J
2014-01-01
The paper describes a curricular framework for introducing microcontroller programming in the University Physics lab. The approach makes use of Modeling Instruction, an effective approach for teaching science at the secondary level in which student learn the standard material by developing and deploying models of the physical world. In our approach, students engage with a context-rich problem that can be solved with one or more sensors and a microcontroller. The solution path we describe then consists of developing a mathematical model for how the sensors' data can be mapped to a meaningful measurement, and further, developing an algorithmic model that will be implemented in the microcontroller. Once the system is developed and implemented, students are given an array of similar problems in which they can deploy their data collection system. Results from the implementation of this idea, in two University Physics sections, using Arduino microcontrollers, are also described.
Error models in quantum computation: an application of model selection
Lucia Schwarz; Steven van Enk
2013-09-04
Threshold theorems for fault-tolerant quantum computing assume that errors are of certain types. But how would one detect whether errors of the "wrong" type occur in one's experiment, especially if one does not even know what type of error to look for? The problem is that for many qubits a full state description is impossible to analyze, and a full process description is even more impossible to analyze. As a result, one simply cannot detect all types of errors. Here we show through a quantum state estimation example (on up to 25 qubits) how to attack this problem using model selection. We use, in particular, the Akaike Information Criterion. The example indicates that the number of measurements that one has to perform before noticing errors of the wrong type scales polynomially both with the number of qubits and with the error size.
A Vast Machine Computer Models, Climate Data, and the Politics of Global Warming
Edwards, Paul N.
A Vast Machine Computer Models, Climate Data, and the Politics of Global Warming Paul N. Edwards models, climate data, and the politics of global warming / Paul N. Edwards. p. cm. Includes. Climatology--History. 3. Meteorology--History. 4. Climatology--Technological innovation. 5. Global temperature
Grujicic, Mica
Computational Modeling of theComputational Modeling of the Vacuum Assisted Resin Transfer MoldingVacuum Assisted Resin Transfer Molding (VARTM) Process(VARTM) Process April 2004April 2004 DepartmentMS Thesis Advisor: Dr. Grujicic #12;What is VARTM?What is VARTM? Vacuum Assisted Resin Transfer Molding
Computing the Electricity Market Equilibrium: Uses of market equilibrium models
Baldick, Ross
1 Computing the Electricity Market Equilibrium: Uses of market equilibrium models Ross Baldick Abstract--In this paper we consider the formulation and uses of electric- ity market equilibrium models. Keywords--Electricity market, Equilibrium models I. INTRODUCTION Electricity market equilibrium modelling
ComputerScience,TexasA&MUniversity Modeling Heterogeneous User
Loguinov, Dmitri
1 ComputerScience,TexasA&MUniversity Modeling Heterogeneous User Churn and Local Resilience of Unstructured P2P Networks Modeling Heterogeneous UserModeling Heterogeneous User Churn and Local Resilience ofChurn Churn and Local Resilience of Unstructured P2P Networks Modeling Heterogeneous User
Groleau, Julie; Marecaux, Christophe; Payrard, Natacha; Segaud, Brice; Rochette, Michel; Perrier, Pascal; Payan, Yohan
2008-01-01
A 3D biomechanical finite element model of the face is presented. Muscles are represented by piece-wise uniaxial tension cable elements linking the insertion points. Such insertion points are specific entities differing from nodes of the finite element mesh, which makes possible to change either the mesh or the muscle implementation totally independently of each other. Lip/teeth and upper lip/lower lip contacts are also modeled. Simulations of smiling and of an Orbicularis Oris activation are presented and interpreted. The importance of a proper account of contacts and of an accurate anatomical description is shown
A stepped leader model for lightning including charge distribution in branched channels
Shi, Wei; Zhang, Li [School of Electrical Engineering, Shandong University, Jinan 250061 (China); Li, Qingmin, E-mail: lqmeee@ncepu.edu.cn [Beijing Key Lab of HV and EMC, North China Electric Power University, Beijing 102206 (China); State Key Lab of Alternate Electrical Power System with Renewable Energy Sources, Beijing 102206 (China)
2014-09-14
The stepped leader process in negative cloud-to-ground lightning plays a vital role in lightning protection analysis. As lightning discharge usually presents significant branched or tortuous channels, the charge distribution along the branched channels and the stochastic feature of stepped leader propagation were investigated in this paper. The charge density along the leader channel and the charge in the leader tip for each lightning branch were approximated by introducing branch correlation coefficients. In combination with geometric characteristics of natural lightning discharge, a stochastic stepped leader propagation model was presented based on the fractal theory. By comparing simulation results with the statistics of natural lightning discharges, it was found that the fractal dimension of lightning trajectory in simulation was in the range of that observed in nature and the calculation results of electric field at ground level were in good agreement with the measurements of a negative flash, which shows the validity of this proposed model. Furthermore, a new equation to estimate the lightning striking distance to flat ground was suggested based on the present model. The striking distance obtained by this new equation is smaller than the value estimated by previous equations, which indicates that the traditional equations may somewhat overestimate the attractive effect of the ground.
Hiroaki Sugiyama; Koji Tsumura; Hiroshi Yokoya
2012-10-05
The doubly charged scalar boson (H^{\\pm\\pm}) is introduced in several models of the new physics beyond the standard model. The H^{\\pm\\pm} has Yukawa interactions with two left-handed charged leptons or two right-handed charged leptons depending on the models. We study kinematical properties of H^{\\pm\\pm} decay products through tau leptons in order to discriminate the chiral structures of the new Yukawa interaction. The chirality of tau leptons can be measured by the energy distributions of the tau decay products, and thus the chiral structure of the new Yukawa interaction can be traced in the invariant-mass distributions of the H^{\\pm\\pm} decay products. We perform simulation studies for the typical decay patterns of the H^{\\pm\\pm} with simple event selections and tau-tagging procedures, and show that the chiral structure of the Yukawa interactions of H^{\\pm\\pm} can be distinguished by measuring the invariant-mass distributions.
Computational Approaches for Modelling Elastohydrodynamic Lubrication Using Multiphysics Software
Jimack, Peter
the Integral Approach (IA). The computational cost of performing this calculation is , for N points over developing highly specialised, bespoke software are highlighted. In order to calculate the complex nonlinear partial differential equation system representing the full problem. This includes
H. J. Haubold; D. Kumar
2007-08-16
The Maxwell-Boltzmannian approach to nuclear reaction rate theory is extended to cover Tsallis statistics (Tsallis, 1988) and more general cases of distribution functions. An analytical study of respective thermonuclear functions is being conducted with the help of statistical techniques. The pathway model, recently introduced by Mathai (2005), is utilized for thermonuclear functions and closed-form representations are obtained in terms of H-functions and G-functions. Maxwell-Boltzmannian thermonuclear functions become particular cases of the extended thermonuclear functions. A brief review on the development of the theory of analytic representations of nuclear reaction rates is given.
Ambient temperature modelling with soft computing techniques
Bertini, Ilaria; Ceravolo, Francesco; Citterio, Marco; Di Pietra, Biagio; Margiotta, Francesca; Pizzuti, Stefano; Puglisi, Giovanni; De Felice, Matteo
2010-07-15
This paper proposes a hybrid approach based on soft computing techniques in order to estimate monthly and daily ambient temperature. Indeed, we combine the back-propagation (BP) algorithm and the simple Genetic Algorithm (GA) in order to effectively train artificial neural networks (ANN) in such a way that the BP algorithm initialises a few individuals of the GA's population. Experiments concerned monthly temperature estimation of unknown places and daily temperature estimation for thermal load computation. Results have shown remarkable improvements in accuracy compared to traditional methods. (author)
Computational Modeling and Optimization of Proton Exchange Membrane Fuel Cells
Victoria, University of
Computational Modeling and Optimization of Proton Exchange Membrane Fuel Cells by Marc Secanell and Optimization of Proton Exchange Membrane Fuel Cells by Marc Secanell Gallart Bachelor in Engineering cells. In this thesis, a computational framework for fuel cell analysis and optimization is presented
Clique-detection Models in Computational Biochemistry and Genomics
Butenko, Sergiy
Clique-detection Models in Computational Biochemistry and Genomics S. Butenko and W. E. Wilhelm,wilhelm}@tamu.edu Abstract Many important problems arising in computational biochemistry and genomics have been formulated and genomic aspects of the problems as well as to the graph-theoretic aspects of the solution approaches. Each
Inverse Modelling in Geology by Interactive Evolutionary Computation
Boschetti, Fabio
Inverse Modelling in Geology by Interactive Evolutionary Computation Chris Wijns a,b,, Fabio of geological processes, in the absence of established numerical criteria to act as inversion targets, requires evolutionary computation provides for the inclusion of qualitative geological expertise within a rigorous
Computational load in model physics of the parallel NCAR community climate model
Michalakes, J.G.; Nanjundiah, R.S.
1994-11-01
Maintaining a balance of computational load over processors is a crucial issue in parallel computing. For efficient parallel implementation, complex codes such as climate models need to be analyzed for load imbalances. In the present study we focus on the load imbalances in the physics portion of the community climate model`s (CCM2) distributed-memory parallel implementation on the Intel Touchstone DELTA computer. We note that the major source of load imbalance is the diurnal variation in the computation of solar radiation. Convective weather patterns also cause some load imbalance. Land-ocean contrast is seen to have little effect on computational load in the present version of the model.
COMPUTATIONAL FLUID DYNAMICS MODELING OF SOLID OXIDE FUEL CELLS
COMPUTATIONAL FLUID DYNAMICS MODELING OF SOLID OXIDE FUEL CELLS Ugur Pasaogullari and Chao-dimensional model has been developed to simulate solid oxide fuel cells (SOFC). The model fully couples current density operation. INTRODUCTION Solid oxide fuel cells (SOFC) are among possible candidates
Quantum Computers: Noise Propagation and Adversarial Noise Models
Kalai, Gil
Quantum Computers: Noise Propagation and Adversarial Noise Models Gil Kalai Hebrew University of Jerusalem and Yale University April 21, 2009 Abstract In this paper we consider adversarial noise models." Detrimental noise is modeled after familiar properties of noise propagation. However, it can have various
Computational Nanophotonics: modeling optical interactions and...
Office of Scientific and Technical Information (OSTI)
provides a powerful approach for performing these studies, so much of this research project involved the development of methods for solving Maxwell's equations, including...
Modeling-Computer Simulations At Valles Caldera - Sulphur Springs...
DOE-funding Unknown Notes A computer program capable of two-dimensional modeling of gravity data was used in interpreting gravity observations along profiles A--A' and B--B'...
Modeling-Computer Simulations At Valles Caldera - Redondo Geothermal...
DOE-funding Unknown Notes A computer program capable of two-dimensional modeling of gravity data was used in interpreting gravity observations along profiles A--A' and B--B'...
Computational Model of Forward and Opposed Smoldering Combustion in Microgravity
Rein, Guillermo; Fernandez-Pello, Carlos; Urban, David
2006-08-06
A novel computational model of smoldering combustion capable of predicting both forward and opposed propagation is developed. This is accomplished by considering the one-dimensional, transient, governing equations for ...
Language acquisition and implication for language change: A computational model.
Clark, Robert A J
1997-01-01
Computer modeling techniques, when applied to language acquisition problems, give an often unrealized insight into the diachronic change that occurs in language over successive generations. This paper shows that using ...
Computer simulation and topological modeling of radiation effects in zircon
Zhang, Yi, 1979-
2006-01-01
The purpose of this study is to understand on atomic level the structural response of zircon (ZrSiO4) to irradiation using molecular dynamics (MD) computer simulations, and to develop topological models that can describe ...
15.094 Systems Optimization: Models and Computation, Spring 2002
Freund, Robert Michael
A computational and application-oriented introduction to the modeling of large-scale systems in a wide variety of decision-making domains and the optimization of such systems using state-of-the-art optimization software. ...
Computational Modeling for the American Chemical Society | GE...
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
Computational Modeling for the American Chemical Society Click to email this to a friend (Opens in new window) Share on Facebook (Opens in new window) Click to share (Opens in new...
A Variable Refrigerant Flow Heat Pump Computer Model in EnergyPlus
Raustad, Richard A. [Florida Solar Energy Center
2013-01-01
This paper provides an overview of the variable refrigerant flow heat pump computer model included with the Department of Energy's EnergyPlusTM whole-building energy simulation software. The mathematical model for a variable refrigerant flow heat pump operating in cooling or heating mode, and a detailed model for the variable refrigerant flow direct-expansion (DX) cooling coil are described in detail.
Computational models of early visual processing layers
Shan, Honghao
2010-01-01
prevailing view of retinal processing. However, as discussedsimplified) model of retinal processing. A Retinal Codingretinal coding, the pre-cortical stage of visual processing,
Gering, Kevin L.
2013-01-01
A system includes an electrochemical cell, monitoring hardware, and a computing system. The monitoring hardware samples performance characteristics of the electrochemical cell. The computing system determines cell information from the performance characteristics. The computing system also analyzes the cell information of the electrochemical cell with a Butler-Volmer (BV) expression modified to determine exchange current density of the electrochemical cell by including kinetic performance information related to pulse-time dependence, electrode surface availability, or a combination thereof. A set of sigmoid-based expressions may be included with the modified-BV expression to determine kinetic performance as a function of pulse time. The determined exchange current density may be used with the modified-BV expression, with or without the sigmoid expressions, to analyze other characteristics of the electrochemical cell. Model parameters can be defined in terms of cell aging, making the overall kinetics model amenable to predictive estimates of cell kinetic performance along the aging timeline.
Sandia Energy - Computational Modeling & Simulation
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
AFDC Printable Version Share this resource Send a link to EERE: Alternative Fuels Data Center Home Page to someone by E-mail Share EERE: Alternative Fuels Data Center Home Page on Facebook Tweet about EERE: Alternative Fuels Data Center Home Page on Twitter Bookmark EERE: Alternative Fuels Data Center Homesum_a_epg0_fpd_mmcf_m.xls" ,"Available from WebQuantityBonneville Power AdministrationRobust, High-Throughput Analysis ofSample SULIColin Humphreys Home Colin HumphreysComputational
Presently, the PV industry is switching to the selective emitter design, where the phosphorus densityIMPROVING THE PREDICTIVE POWER OF MODELING THE EMITTER DIFFUSION BY FULLY INCLUDING or a spin-on source. In the selective emitter design, the phosphorus density is significantly reduced
Discrete transfinite computation models School of Mathematics,
Welch, Philip
such as Davies [Davies (2001)], and the models of Beggs and Tucker [Beggs and Tucker (2007)] that attempt' functions or oracles, such as is done in [Beggs et al. (2008)]. We also shall not particularly consider
CARD No. 23 Models and Computer Codes
that are used to demonstrate that the WIPP will comply with the radioactive waste disposal regulations at 40 CFR, and radionuclide transport in the repository and overlying rock formations. Numerical models are then created
Lagi, Marco; Bertrand, Karla Z; Bar-Yam, Yaneer
2012-01-01
Increases in global food prices have led to widespread hunger and social unrest---and an imperative to understand their causes. In a previous paper published in September 2011, we constructed for the first time a dynamic model that quantitatively agreed with food prices. Specifically, the model fit the FAO Food Price Index time series from January 2004 to March 2011, inclusive. The results showed that the dominant causes of price increases during this period were investor speculation and ethanol conversion. The model included investor trend following as well as shifting between commodities, equities and bonds to take advantage of increased expected returns. Here, we extend the food prices model to January 2012, without modifying the model but simply continuing its dynamics. The agreement is still precise, validating both the descriptive and predictive abilities of the analysis. Policy actions are needed to avoid a third speculative bubble that would cause prices to rise above recent peaks by the end of 2012.
3D Bone Microarchitecture Modeling and Fracture Risk Department of Computer
Buffalo, State University of New York
3D Bone Microarchitecture Modeling and Fracture Risk Prediction Hui Li Department of Computer will also rise. It calls for innovative research on understanding of osteoporo- sis and fracture mechanisms-of-the-art probabilistic approach to analyze bone fracture risk factors including demographic attributes and life styles
Computational Fluid Dynamics (CFD) Modelling on Soot Yield for Fire
Computational Fluid Dynamics (CFD) Modelling on Soot Yield for Fire Engineering Assessment Yong S (CFD) Modelling is now widely used by fire safety engineers throughout the world as a tool of the smoke control design as part of the performance based fire safety design in the current industry
Computational Modeling of Chord Fingering for String Instruments
Radicioni, Daniele
Computational Modeling of Chord Fingering for String Instruments Daniele P. Radicioni (radicion for the fingering process with string instruments, based on a constraint satisfaction approach. The model is imple-mechanical aspects of the performer's hand in its interaction with the musical instrument. Introduction Music
Personal Computer-Based Model for Cool Storage Performance Simulation
Kasprowicz, L. M.; Jones, J. W.; Hitzfelder, J.
1990-01-01
A personal computer based hourly simulation model was developed based on the CBS/ICE routines in the DOE-2.1 mainframe building simulation software. The menu driven new model employs more efficient data and information handling than the previous...
Methodology for characterizing modeling and discretization uncertainties in computational simulation
ALVIN,KENNETH F.; OBERKAMPF,WILLIAM L.; RUTHERFORD,BRIAN M.; DIEGERT,KATHLEEN V.
2000-03-01
This research effort focuses on methodology for quantifying the effects of model uncertainty and discretization error on computational modeling and simulation. The work is directed towards developing methodologies which treat model form assumptions within an overall framework for uncertainty quantification, for the purpose of developing estimates of total prediction uncertainty. The present effort consists of work in three areas: framework development for sources of uncertainty and error in the modeling and simulation process which impact model structure; model uncertainty assessment and propagation through Bayesian inference methods; and discretization error estimation within the context of non-deterministic analysis.
Gering, Kevin L
2013-08-27
A system includes an electrochemical cell, monitoring hardware, and a computing system. The monitoring hardware periodically samples performance characteristics of the electrochemical cell. The computing system determines cell information from the performance characteristics of the electrochemical cell. The computing system also develops a mechanistic level model of the electrochemical cell to determine performance fade characteristics of the electrochemical cell and analyzing the mechanistic level model to estimate performance fade characteristics over aging of a similar electrochemical cell. The mechanistic level model uses first constant-current pulses applied to the electrochemical cell at a first aging period and at three or more current values bracketing a first exchange current density. The mechanistic level model also is based on second constant-current pulses applied to the electrochemical cell at a second aging period and at three or more current values bracketing the second exchange current density.
Computing Biological Model Parameters by Parallel Statistical Model Checking
Tronci, Enrico
of Treatments for Infertility Related Endocrinological Diseases, 600773). #12;patient-specific model parameters
Models of quantum computation and quantum programming languages
J. A. Miszczak
2011-12-03
The goal of the presented paper is to provide an introduction to the basic computational models used in quantum information theory. We review various models of quantum Turing machine, quantum circuits and quantum random access machine (QRAM) along with their classical counterparts. We also provide an introduction to quantum programming languages, which are developed using the QRAM model. We review the syntax of several existing quantum programming languages and discuss their features and limitations.
A computer model to compare greenhouse structures
Chiou, Jiun-Shian
1989-01-01
systems include pad alone, water distribution system, sumps, pumps, and plumbing kit. 7. Heating system: The two typical heating systems are either central or unit. The fuel can be either gas or oil. (1). Part one lists the prices of central heating... 36 48 24 $88 $114 $228 $32 36 48 $103 (3). Fan Jets Tube size (inch) HP Price Motorized Price shutter (4). Plastic tubes 18 24 30 1/4 420 18lbs 1/3 534 23lbs 1/2 622 33lbs Tube size (inch) 12 18 24 30 166 176 194 Price (ft...
Quantum Computers: Noise Propagation and Adversarial Noise Models
Gil Kalai
2009-04-21
In this paper we consider adversarial noise models that will fail quantum error correction and fault-tolerant quantum computation. We describe known results regarding high-rate noise, sequential computation, and reversible noisy computation. We continue by discussing highly correlated noise and the "boundary," in terms of correlation of errors, of the "threshold theorem." Next, we draw a picture of adversarial forms of noise called (collectively) "detrimental noise." Detrimental noise is modeled after familiar properties of noise propagation. However, it can have various causes. We start by pointing out the difference between detrimental noise and standard noise models for two qubits and proceed to a discussion of highly entangled states, the rate of noise, and general noisy quantum systems.
A gas kick model for the personal computer
Miller, Clayton Lowell
1987-01-01
A GAS KICK MODEL FOR THE PERSONAL COMPUTER A Thesis by CLAYTON LOWELL MILLER Submitted to the Graduate College of Texas A6M University in partial fulfillment of the requirements for the degree of MASTER OF SCIENCE May 1987 Major Subject...: Petroleum Engineering A GAS KICK MODEL FOR THE PERSONAL COMPUTER A Thesis by CLAYTON LOWELL MILLER Approved as to style and content by: Wana C. vkam-Wold (Chair f Committee) Robert W. Heine (Member) Tibor G. Rozg yi (Member) W. D. Von Gonten Head...
The U-Machine: A Model of Generalized Computation
MacLennan, Bruce
classification: 68Q05. This report may be used for any non-profit purpose provided that the source is credited in the broad- est sense, including analog and digital, and provides a framework for computation using novel known, the end of Moore's Law is in sight, and we are approaching the limits of digital electronics
Concurrent multiscale computational modeling for dense dry granular materials interfacing
Regueiro, Richard A.
of interfacial mechanics between granular soil and tire, tool, or penetrometer, while properly representing far computational modeling of interfacial mechanics between granular materials and deformable solid bodies, agricultural grains (in silo flows), dry soils (sand, silt, gravel), and lunar and martian regolith (soil found
ADVANCED WUFI COMPUTER MODELING WORKSHOP FOR WALL DESIGN AND PERFORMANCE
Oak Ridge National Laboratory
TRANSFER IN BUILDING ENVELOPES) Napa, CA, January 30 - February 1, 2012 WUFI/ORNL1 Program made availableADVANCED WUFI COMPUTER MODELING WORKSHOP FOR WALL DESIGN AND PERFORMANCE (HEAT AND MOISTURE-Institut für Bauphysik (IBP) and co-sponsored by the National Institute of Building Sciences (NIBS)/ Building
Computational model of aortic valve surgical repair using grafted pericardium
1 Computational model of aortic valve surgical repair using grafted pericardium Peter E. Hammer1, aortic valve repair, membrane, surgical planning, leaflet graft, pericardium ABSTRACT Aortic valve leaflets. Difficulty is largely due to the complex geometry and function of the valve and the lower
A Computational Model for Sound Field Absorption by Acoustic Arrays
. We then formulate the acoustic wave equation with the absorption boundary coeÆcient in the frequency the sound absorption property of arrays of micro-acoustic actuators at a control surface. We use the waveA Computational Model for Sound Field Absorption by Acoustic Arrays H. T. Banks #3; D. G. Cole z K
Faceted Models of Blog Feeds Department of Computer
Meng, Weiyi
Faceted Models of Blog Feeds Lifeng Jia Department of Computer Science University of Illinois@cs.binghamton.edu ABSTRACT Faceted blog distillation aims at retrieving the blogs that are not only relevant to a query blogs depict various topics related to the personal experiences of bloggers while official blogs deliver
Open Learner Models at Birmingham Electronic, Electrical and Computer Engineering
Bull, Susan
and Engineering Principles and Methods EE2D2 Introduction to Communications EE2F1 Speech and Audio Technology EE2G EE1B2 Circuits, Devices and Fields EE1E1&2 C Programming EE1F1 Introduction to InformationOpen Learner Models at Birmingham Electronic, Electrical and Computer Engineering University
innovati nNREL Computer Models Integrate Wind Turbines with
innovati nNREL Computer Models Integrate Wind Turbines with Floating Platforms Far off the shores for today's seabed-mounted offshore wind turbines. For the United States to tap into these vast offshore wind energy resources, wind turbines must be mounted on floating platforms to be cost effective
Computational Model for Forced Expiration from Asymmetric Normal Lungs
Lutchen, Kenneth
Computational Model for Forced Expiration from Asymmetric Normal Lungs ADAM G. POLAK 1 losses along the airway branches. Calculations done for succeeding lung volumes result in the semidynamic to the choke points, characteristic differences of lung regional pressures and volumes, and a shape
Computer Modeling of Crystalline Electrolytes Lithium Thiophosphates and Phosphates
Holzwarth, Natalie
migration. I. Introduction During the last 5 years, lithium thiophosphate solid electrolyte materials haveComputer Modeling of Crystalline Electrolytes Lithium Thiophosphates and Phosphates N. D. Lepley properties of (thio)phosphate electrolyte materials, focusing on the "superionic" electrolyte Li7P3S11. We
Challenges for the CMS computing model in the first year
Fisk, I.; /Fermilab
2009-05-01
CMS is in the process of commissioning a complex detector and a globally distributed computing infrastructure simultaneously. This represents a unique challenge. Even at the beginning there is not sufficient analysis or organized processing resources at CERN alone. In this presentation we discuss the unique computing challenges CMS expects to face during the first year of running and how they influence the baseline computing model decisions. During the early accelerator commissioning periods, CMS will attempt to collect as many events as possible when the beam is on in order to provide adequate early commissioning data. Some of these plans involve overdriving the Tier-0 infrastructure during data collection with recovery when the beam is off. In addition to the larger number of triggered events, there will be pressure in the first year to collect and analyze more complete data formats as the summarized formats mature. The large event formats impact the required storage, bandwidth, and processing capacity across all the computing centers. While the understanding of the detector and the event selections is being improved, there will likely be a larger number of reconstruction passes and skims performed by both central operations and individual users. We discuss how these additional stresses impact the allocation of resources and the changes from the baseline computing model.
A New Perspective for the Calibration of Computational Predictor Models.
Crespo, Luis Guillermo
2014-11-01
This paper presents a framework for calibrating computational models using data from sev- eral and possibly dissimilar validation experiments. The offset between model predictions and observations, which might be caused by measurement noise, model-form uncertainty, and numerical error, drives the process by which uncertainty in the models parameters is characterized. The resulting description of uncertainty along with the computational model constitute a predictor model. Two types of predictor models are studied: Interval Predictor Models (IPMs) and Random Predictor Models (RPMs). IPMs use sets to characterize uncer- tainty, whereas RPMs use random vectors. The propagation of a set through a model makes the response an interval valued function of the state, whereas the propagation of a random vector yields a random process. Optimization-based strategies for calculating both types of predictor models are proposed. Whereas the formulations used to calculate IPMs target solutions leading to the interval value function of minimal spread containing all observations, those for RPMs seek to maximize the models' ability to reproduce the distribution of obser- vations. Regarding RPMs, we choose a structure for the random vector (i.e., the assignment of probability to points in the parameter space) solely dependent on the prediction error. As such, the probabilistic description of uncertainty is not a subjective assignment of belief, nor is it expected to asymptotically converge to a fixed value, but instead it is a description of the model's ability to reproduce the experimental data. This framework enables evaluating the spread and distribution of the predicted response of target applications depending on the same parameters beyond the validation domain (i.e., roll-up and extrapolation).
ONSET OF CHAOS IN A MODEL OF QUANTUM COMPUTATION G. BERMAN; ET...
Office of Scientific and Technical Information (OSTI)
OF CHAOS IN A MODEL OF QUANTUM COMPUTATION G. BERMAN; ET AL 71 CLASSICAL AND QUANTUM MECHANICS, GENERAL PHYSICS; 99 GENERAL AND MISCELLANEOUSMATHEMATICS, COMPUTING, AND...
Liu, Huan
Second International Workshop on Social Computing, Behavioral Modeling, and Prediction Phoenix, Arizona March 31 - April 1, 2009 Phoenix, Arizona Proceedings published by Springer Social computing
Modeling of BWR core meltdown accidents - for application in the MELRPI. MOD2 computer code
Koh, B R; Kim, S H; Taleyarkhan, R P; Podowski, M Z; Lahey, Jr, R T
1985-04-01
This report summarizes improvements and modifications made in the MELRPI computer code. A major difference between this new, updated version of the code, called MELRPI.MOD2, and the one reported previously, concerns the inclusion of a model for the BWR emergency core cooling systems (ECCS). This model and its computer implementation, the ECCRPI subroutine, account for various emergency injection modes, for both intact and rubblized geometries. Other changes to MELRPI deal with an improved model for canister wall oxidation, rubble bed modeling, and numerical integration of system equations. A complete documentation of the entire MELRPI.MOD2 code is also given, including an input guide, list of subroutines, sample input/output and program listing.
Vandersall, Jennifer A.; Gardner, Shea N.; Clague, David S.
2010-05-04
A computational method and computer-based system of modeling DNA synthesis for the design and interpretation of PCR amplification, parallel DNA synthesis, and microarray chip analysis. The method and system include modules that address the bioinformatics, kinetics, and thermodynamics of DNA amplification and synthesis. Specifically, the steps of DNA selection, as well as the kinetics and thermodynamics of DNA hybridization and extensions, are addressed, which enable the optimization of the processing and the prediction of the products as a function of DNA sequence, mixing protocol, time, temperature and concentration of species.
The origins of computer weather prediction and climate modeling
Lynch, Peter [Meteorology and Climate Centre, School of Mathematical Sciences, University College Dublin, Belfield (Ireland)], E-mail: Peter.Lynch@ucd.ie
2008-03-20
Numerical simulation of an ever-increasing range of geophysical phenomena is adding enormously to our understanding of complex processes in the Earth system. The consequences for mankind of ongoing climate change will be far-reaching. Earth System Models are capable of replicating climate regimes of past millennia and are the best means we have of predicting the future of our climate. The basic ideas of numerical forecasting and climate modeling were developed about a century ago, long before the first electronic computer was constructed. There were several major practical obstacles to be overcome before numerical prediction could be put into practice. A fuller understanding of atmospheric dynamics allowed the development of simplified systems of equations; regular radiosonde observations of the free atmosphere and, later, satellite data, provided the initial conditions; stable finite difference schemes were developed; and powerful electronic computers provided a practical means of carrying out the prodigious calculations required to predict the changes in the weather. Progress in weather forecasting and in climate modeling over the past 50 years has been dramatic. In this presentation, we will trace the history of computer forecasting through the ENIAC integrations to the present day. The useful range of deterministic prediction is increasing by about one day each decade, and our understanding of climate change is growing rapidly as Earth System Models of ever-increasing sophistication are developed.
A General Hippocampal Computational Model Combining Episodic and Spatial Memory in a Spiking Model
Aguiar, Paulo de Castro
The hippocampus, in humans and rats, plays crucial roles in spatial tasks and nonspatial tasks involving episodic-type memory. This thesis presents a novel computational model of the hippocampus (CA1, CA3 and dentate ...
Final Report for Integrated Multiscale Modeling of Molecular Computing Devices
Glotzer, Sharon C.
2013-08-28
In collaboration with researchers at Vanderbilt University, North Carolina State University, Princeton and Oakridge National Laboratory we developed multiscale modeling and simulation methods capable of modeling the synthesis, assembly, and operation of molecular electronics devices. Our role in this project included the development of coarse-grained molecular and mesoscale models and simulation methods capable of simulating the assembly of millions of organic conducting molecules and other molecular components into nanowires, crossbars, and other organized patterns.
Computer Modeling of Violent Intent: A Content Analysis Approach
Sanfilippo, Antonio P.; Mcgrath, Liam R.; Bell, Eric B.
2014-01-03
We present a computational approach to modeling the intent of a communication source representing a group or an individual to engage in violent behavior. Our aim is to identify and rank aspects of radical rhetoric that are endogenously related to violent intent to predict the potential for violence as encoded in written or spoken language. We use correlations between contentious rhetoric and the propensity for violent behavior found in documents from radical terrorist and non-terrorist groups and individuals to train and evaluate models of violent intent. We then apply these models to unseen instances of linguistic behavior to detect signs of contention that have a positive correlation with violent intent factors. Of particular interest is the application of violent intent models to social media, such as Twitter, that have proved to serve as effective channels in furthering sociopolitical change.
Comprehensive computer model for magnetron sputtering. II. Charged particle transport
Jimenez, Francisco J., E-mail: fjimenez@ualberta.ca; Dew, Steven K. [Department of Electrical and Computer Engineering, University of Alberta, Edmonton T6G 2V4 (Canada); Field, David J. [Smith and Nephew (Alberta) Inc., Fort Saskatchewan T8L 4K4 (Canada)
2014-11-01
Discharges for magnetron sputter thin film deposition systems involve complex plasmas that are sensitively dependent on magnetic field configuration and strength, working gas species and pressure, chamber geometry, and discharge power. The authors present a numerical formulation for the general solution of these plasmas as a component of a comprehensive simulation capability for planar magnetron sputtering. This is an extensible, fully three-dimensional model supporting realistic magnetic fields and is self-consistently solvable on a desktop computer. The plasma model features a hybrid approach involving a Monte Carlo treatment of energetic electrons and ions, along with a coupled fluid model for thermalized particles. Validation against a well-known one-dimensional system is presented. Various strategies for improving numerical stability are investigated as is the sensitivity of the solution to various model and process parameters. In particular, the effect of magnetic field, argon gas pressure, and discharge power are studied.
Computer Modeling and Simulation of an Active Vision Camera System MingChin Lu
Subbarao, Murali "Rao"
Computer Modeling and Simulation of an Active Vision Camera System MingChin Lu Communications of computer simulation systems. Computer simulation avoids the necessity of building actual camera systems. Based on the proposed model and algorithms, a computer simulation system called Active Vision Simulator
EconoGrid: A detailed Simulation Model of a Standards-based Grid Compute Economy
EconoGrid: A detailed Simulation Model of a Standards-based Grid Compute Economy EconoGrid is a detailed simulation model, implemented in SLX1 , of a grid compute economy that implements selected of users. In a grid compute economy, computing resources are sold to users in a market where price
Computational Modeling of Blood Flow in the TrapEase Inferior Vena Cava Filter
Singer, M A; Henshaw, W D; Wang, S L
2008-02-04
To evaluate the flow hemodynamics of the TrapEase vena cava filter using three dimensional computational fluid dynamics, including simulated thrombi of multiple shapes, sizes, and trapping positions. The study was performed to identify potential areas of recirculation and stagnation and areas in which trapped thrombi may influence intrafilter thrombosis. Computer models of the TrapEase filter, thrombi (volumes ranging from 0.25mL to 2mL, 3 different shapes), and a 23mm diameter cava were constructed. The hemodynamics of steady-state flow at Reynolds number 600 was examined for the unoccluded and partially occluded filter. Axial velocity contours and wall shear stresses were computed. Flow in the unoccluded TrapEase filter experienced minimal disruption, except near the superior and inferior tips where low velocity flow was observed. For spherical thrombi in the superior trapping position, stagnant and recirculating flow was observed downstream of the thrombus; the volume of stagnant flow and the peak wall shear stress increased monotonically with thrombus volume. For inferiorly trapped spherical thrombi, marked disruption to the flow was observed along the cava wall ipsilateral to the thrombus and in the interior of the filter. Spherically shaped thrombus produced a lower peak wall shear stress than conically shaped thrombus and a larger peak stress than ellipsoidal thrombus. We have designed and constructed a computer model of the flow hemodynamics of the TrapEase IVC filter with varying shapes, sizes, and positions of thrombi. The computer model offers several advantages over in vitro techniques including: improved resolution, ease of evaluating different thrombus sizes and shapes, and easy adaptation for new filter designs and flow parameters. Results from the model also support a previously reported finding from photochromic experiments that suggest the inferior trapping position of the TrapEase IVC filter leads to an intra-filter region of recirculating/stagnant flow with very low shear stress that may be thrombogenic.
Integrated Core-SOL-Divertor Modelling for ITER Including Impurity: Effect of Tungsten on Fusion Performance in H-mode and Hybrid Scenario
A Vast Machine: Computer Models, Climate Data, and the Politics of Global Warming
Maret, Susan
2011-01-01
Review: A Vast Machine: Computer Models, Climate Data, andEdwards, Paul N. A Vast Machine: Computer Models, ClimateEdwards, Paul N. 2004. "A vast machine: standards as social
Friedman, Carey
We use the global 3-D chemical transport model GEOS-Chem to simulate long-range atmospheric transport of polycyclic aromatic hydrocarbons (PAHs). To evaluate the model’s ability to simulate PAHs with different volatilities, ...
Models and People: an alternative view of the emergent properties of computational models
Boschetti, Fabio
be used to address scientific questions or support real-world decision making. At these higher levels ecological model may be used i) to address a scientific question and ii) how the scientific insight so gained management problem (Figure 1, top) 11-16 . 2 Computational models and scientific questions Within
Not Available
2012-08-01
Cation degradation insights obtained by computational modeling could result in better performance and longer lifetime for alkaline membrane fuel cells.
Compensator models for fluence field modulated computed tomography
Bartolac, Steven; Jaffray, David; Radiation Medicine Program, Princess Margaret Hospital Department of Radiation Oncology, University of Toronto, Toronto, Ontario M5G 2M9
2013-12-15
Purpose: Fluence field modulated computed tomography (FFMCT) presents a novel approach for acquiring CT images, whereby a patient model guides dynamically changing fluence patterns in an attempt to achieve task-based, user-prescribed, regional variations in image quality, while also controlling dose to the patient. This work aims to compare the relative effectiveness of FFMCT applied to different thoracic imaging tasks (routine diagnostic CT, lung cancer screening, and cardiac CT) when the modulator is subject to limiting constraints, such as might be present in realistic implementations.Methods: An image quality plan was defined for a simulated anthropomorphic chest slice, including regions of high and low image quality, for each of the thoracic imaging tasks. Modulated fluence patterns were generated using a simulated annealing optimization script, which attempts to achieve the image quality plan under a global dosimetric constraint. Optimization was repeated under different types of modulation constraints (e.g., fixed or gantry angle dependent patterns, continuous or comprised of discrete apertures) with the most limiting case being a fixed conventional bowtie filter. For each thoracic imaging task, an image quality map (IQM{sub sd}) representing the regionally varying standard deviation is predicted for each modulation method and compared to the prescribed image quality plan as well as against results from uniform fluence fields. Relative integral dose measures were also compared.Results: Each IQM{sub sd} resulting from FFMCT showed improved agreement with planned objectives compared to those from uniform fluence fields for all cases. Dynamically changing modulation patterns yielded better uniformity, improved image quality, and lower dose compared to fixed filter patterns with optimized tube current. For the latter fixed filter cases, the optimal choice of tube current modulation was found to depend heavily on the task. Average integral dose reduction compared to a uniform fluence field ranged from 10% using a bowtie filter to 40% or greater using an idealized modulator.Conclusions: The results support that FFMCT may achieve regionally varying image quality distributions in good agreement with user-prescribed values, while limiting dose. The imposition of constraints inhibits dose reduction capacity and agreement with image quality plans but still yields significant improvement over what is afforded by conventional dose minimization techniques. These results suggest that FFMCT can be implemented effectively even when the modulator has limited modulation capabilities.
DIMITRI GIDASPOW
1997-08-15
The objective of this study is to develop a predictive experimentally verified computational fluid dynamic (CFD) three phase model. It predicts the gas, liquid and solid hold-ups (volume fractions) and flow patterns in the industrially important bubble-coalesced (churn-turbulent) regime. The input into the model can be either particulate viscosities as measured with a Brookfield viscometer or effective restitution coefficient for particles. A combination of x-ray and {gamma}-ray densitometers was used to measure solid and liquid volume fractions. There is a fair agreement between the theory and the experiment. A CCD camera was used to measure instantaneous particle velocities. There is a good agreement between the computed time average velocities and the measurements. There is an excellent agreement between the viscosity of 800 {micro}m glass beads obtained from measurement of granular temperature (random kinetic energy of particles) and the measurement using a Brookfield viscometer. A relation between particle Reynolds stresses and granular temperature was found for developed flow. Such measurement and computations gave a restitution coefficient for a methanol catalyst to be about 0.9. A transient, two-dimensional hydrodynamic model for production of methanol from syn-gas in an Air Products/DOE LaPorte slurry bubble column reactor was developed. The model predicts downflow of catalyst at the walls and oscillatory particle and gas flow at the center, with a frequency of about 0.7 Hertz. The computed temperature variation in the rector with heat exchangers was only about 5 K, indicating good thermal management. The computed slurry height, the gas holdup and the rate of methanol production agree with LaPorte's reported data. Unlike the previous models in the literature, this model computes the gas and the particle holdups and the particle rheology. The only adjustable parameter in the model is the effective particle restitution coefficient.
FINITE VOLUME METHODS APPLIED TO THE COMPUTATIONAL MODELLING OF WELDING PHENOMENA
Taylor, Gary
1 FINITE VOLUME METHODS APPLIED TO THE COMPUTATIONAL MODELLING OF WELDING PHENOMENA Gareth A.Taylor@brunel.ac.uk ABSTRACT This paper presents the computational modelling of welding phenomena within a versatile numerical) and Computational Solid Mechanics (CSM). With regard to the CFD modelling of the weld pool fluid dynamics, heat
How Computational Models Predict the Behavior of Complex Systems John Symons 1
Boschetti, Fabio
How Computational Models Predict the Behavior of Complex Systems John Symons 1 Fabio Boschetti2,3 1 of prediction in the use of computational models in science. We focus on the consequences of the irreversibility of computational models and on the conditional or ceteris paribus, nature of the kinds of their predictions
Boutchko, R.
2014-01-01
emission tomography systems and computational fluid dynamicsa computational ?uid dynamics (CFD) model of the systemthe computational domain. A Cartesian coordinate system was
Even-Parity S_(N) Adjoint Method Including SP_(N) Model Error and Iterative Efficiency
Zhang, Yunhuang
2014-08-10
In this Dissertation, we analyze an adjoint-based approach for assessing the model error of SP_(N) equations (low fidelity model) by comparing it against S_(N) equations (high fidelity model). Three model error estimation methods, namely, direct...
HIGH RESOLUTION FORWARD AND INVERSE EARTHQUAKE MODELING ON TERASCALE COMPUTERS
Shewchuk, Jonathan
highly populated seismic region in the U.S., it has well- characterized geological structures (including in characterizing earthquake source and basin material properties, a critical remaining challenge is to invert basin geology and earthquake sources, and to use this capability to model and forecast strong ground
Computing Limb Darkening Coefficients from Stellar Atmosphere Models
David Heyrovsky
2006-10-24
We explore the sensitivity of limb darkening coefficients computed from stellar atmosphere models to different least-squares fitting methods. We demonstrate that conventional methods are strongly biased to fitting the stellar limb. Our suggested method of fitting by minimizing the radially integrated squared residual yields improved fits with better flux conservation. The differences of the obtained coefficients from commonly used values are observationally significant. We show that the new values are in better agreement with solar limb darkening measurements as well as with coefficients reported from analyses of eclipsing binary light curves.
Computing model independent perturbations in dark energy and modified gravity
Battye, Richard A. [Jodrell Bank Centre for Astrophysics, School of Physics and Astronomy, The University of Manchester, Manchester M13 9PL (United Kingdom); Pearson, Jonathan A., E-mail: richard.battye@manchester.ac.uk, E-mail: jonathan.pearson@durham.ac.uk [Department of Mathematical Sciences, Durham University, South Road, Durham, DH1 3LE (United Kingdom)
2014-03-01
We present a methodology for computing model independent perturbations in dark energy and modified gravity. This is done from the Lagrangian for perturbations, by showing how field content, symmetries, and physical principles are often sufficient ingredients for closing the set of perturbed fluid equations. The fluid equations close once ''equations of state for perturbations'' are identified: these are linear combinations of fluid and metric perturbations which construct gauge invariant entropy and anisotropic stress perturbations for broad classes of theories. Our main results are the proof of the equation of state for perturbations presented in a previous paper, and the development of the required calculational tools.
Modeling the Fracture of Ice Sheets on Parallel Computers
Waisman, Haim; Tuminaro, Ray
2013-10-10
The objective of this project was to investigate the complex fracture of ice and understand its role within larger ice sheet simulations and global climate change. This objective was achieved by developing novel physics based models for ice, novel numerical tools to enable the modeling of the physics and by collaboration with the ice community experts. At the present time, ice fracture is not explicitly considered within ice sheet models due in part to large computational costs associated with the accurate modeling of this complex phenomena. However, fracture not only plays an extremely important role in regional behavior but also influences ice dynamics over much larger zones in ways that are currently not well understood. To this end, our research findings through this project offers significant advancement to the field and closes a large gap of knowledge in understanding and modeling the fracture of ice sheets in the polar regions. Thus, we believe that our objective has been achieved and our research accomplishments are significant. This is corroborated through a set of published papers, posters and presentations at technical conferences in the field. In particular significant progress has been made in the mechanics of ice, fracture of ice sheets and ice shelves in polar regions and sophisticated numerical methods that enable the solution of the physics in an efficient way.
Computational fluid dynamic modeling of fluidized-bed polymerization reactors
Rokkam, Ram
2012-11-02
Polyethylene is one of the most widely used plastics, and over 60 million tons are produced worldwide every year. Polyethylene is obtained by the catalytic polymerization of ethylene in gas and liquid phase reactors. The gas phase processes are more advantageous, and use fluidized-bed reactors for production of polyethylene. Since they operate so close to the melting point of the polymer, agglomeration is an operational concern in all slurry and gas polymerization processes. Electrostatics and hot spot formation are the main factors that contribute to agglomeration in gas-phase processes. Electrostatic charges in gas phase polymerization fluidized bed reactors are known to influence the bed hydrodynamics, particle elutriation, bubble size, bubble shape etc. Accumulation of electrostatic charges in the fluidized-bed can lead to operational issues. In this work a first-principles electrostatic model is developed and coupled with a multi-fluid computational fluid dynamic (CFD) model to understand the effect of electrostatics on the dynamics of a fluidized-bed. The multi-fluid CFD model for gas-particle flow is based on the kinetic theory of granular flows closures. The electrostatic model is developed based on a fixed, size-dependent charge for each type of particle (catalyst, polymer, polymer fines) phase. The combined CFD model is first verified using simple test cases, validated with experiments and applied to a pilot-scale polymerization fluidized-bed reactor. The CFD model reproduced qualitative trends in particle segregation and entrainment due to electrostatic charges observed in experiments. For the scale up of fluidized bed reactor, filtered models are developed and implemented on pilot scale reactor.
Modeling and Synthesizing Task Placement Constraints in Google Compute Clusters
Cortes, Corinna
characterization for high performance computing and grids focus on task resource require- ments for CPU, memory of compute clusters. Existing workload characteri- zations for high performance computing and grids focus
Shekhar, Ravi
2009-05-15
and amplitude variation with offset (AVO) results for our example model predicts that CO2 is easier to detect than brine in the fractured reservoirs. The effects of geochemical processes on seismics are simulated by time-lapse modeling for t = 1000 years. My...
Computational model for simulation small testing launcher, technical solution
Chelaru, Teodor-Viorel; Cristian, Barbu; Chelaru, Adrian
2014-12-10
The purpose of this paper is to present some aspects regarding the computational model and technical solutions for multistage suborbital launcher for testing (SLT) used to test spatial equipment and scientific measurements. The computational model consists in numerical simulation of SLT evolution for different start conditions. The launcher model presented will be with six degrees of freedom (6DOF) and variable mass. The results analysed will be the flight parameters and ballistic performances. The discussions area will focus around the technical possibility to realize a small multi-stage launcher, by recycling military rocket motors. From technical point of view, the paper is focused on national project 'Suborbital Launcher for Testing' (SLT), which is based on hybrid propulsion and control systems, obtained through an original design. Therefore, while classical suborbital sounding rockets are unguided and they use as propulsion solid fuel motor having an uncontrolled ballistic flight, SLT project is introducing a different approach, by proposing the creation of a guided suborbital launcher, which is basically a satellite launcher at a smaller scale, containing its main subsystems. This is why the project itself can be considered an intermediary step in the development of a wider range of launching systems based on hybrid propulsion technology, which may have a major impact in the future European launchers programs. SLT project, as it is shown in the title, has two major objectives: first, a short term objective, which consists in obtaining a suborbital launching system which will be able to go into service in a predictable period of time, and a long term objective that consists in the development and testing of some unconventional sub-systems which will be integrated later in the satellite launcher as a part of the European space program. This is why the technical content of the project must be carried out beyond the range of the existing suborbital vehicle programs towards the current technological necessities in the space field, especially the European one.
Ortman, Daniel William
1982-01-01
with reasonable heat transfer character1stics were available, no air from the axial air com- pressor would need to be utilized. Steam is a widely suggested alter- native cool1ng med1um to air because 1ts h1gher specific heat suggests that it would prov1de...
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
AFDC Printable Version Share this resource Send a link to EERE: Alternative Fuels Data Center Home Page to someone by E-mail Share EERE: Alternative Fuels Data Center Home Page on Facebook Tweet about EERE: Alternative Fuels Data Center Home Page on Twitter Bookmark EERE: Alternative Fuels Data Center Homesum_a_epg0_fpd_mmcf_m.xls" ,"Available from WebQuantityBonneville Power Administration would like submit the following commentsMethodsCompositional VariationCompressed
Compare Energy Use in Variable Refrigerant Flow Heat Pumps Field Demonstration and Computer Model
Sharma, Chandan; Raustad, Richard
2013-06-01
Variable Refrigerant Flow (VRF) heat pumps are often regarded as energy efficient air-conditioning systems which offer electricity savings as well as reduction in peak electric demand while providing improved individual zone setpoint control. One of the key advantages of VRF systems is minimal duct losses which provide significant reduction in energy use and duct space. However, there is limited data available to show their actual performance in the field. Since VRF systems are increasingly gaining market share in the US, it is highly desirable to have more actual field performance data of these systems. An effort was made in this direction to monitor VRF system performance over an extended period of time in a US national lab test facility. Due to increasing demand by the energy modeling community, an empirical model to simulate VRF systems was implemented in the building simulation program EnergyPlus. This paper presents the comparison of energy consumption as measured in the national lab and as predicted by the program. For increased accuracy in the comparison, a customized weather file was created by using measured outdoor temperature and relative humidity at the test facility. Other inputs to the model included building construction, VRF system model based on lab measured performance, occupancy of the building, lighting/plug loads, and thermostat set-points etc. Infiltration model inputs were adjusted in the beginning to tune the computer model and then subsequent field measurements were compared to the simulation results. Differences between the computer model results and actual field measurements are discussed. The computer generated VRF performance closely resembled the field measurements.
Computational Biology and Bioinformatics 10.10 Models of substitution I: Basic Models A
Goldschmidt, Christina
& stochastic grammars 7.11 RNA structures 9.11 Finding signals in sequences 14.11 Challenges in genome of structure & movements & shapes & grammars 28.11 Integrative genomics: the omics DNA mRNA Protein Metabolite Phenotype 30.11 Integrative genomics: mapping #12;Computational Biology and Bioinformatics 10.10 Models
Loya, Sudarshan Kedarnath
2011-12-31
these postulations have been effective in the past, they might not work with new versions of catalytic converters and the architectures being proposed. In particular, classical models neglect viscosity, conductivity and diffusion in the bulk gas phase. However...
Fernández-Nieto, E D; Narbona-Reina, G; Zabsonré, J D
2015-01-01
In this work we present a deduction of the Saint-Venant-Exner model through an asymptotic analysis of the Navier-Stokes equations. A multi-scale analysis is performed in order to take into account that the velocity of the sediment layer is smaller than the one of the fluid layer. This leads us to consider a shallow water type system for the fluid layer and a lubrication Reynolds equation for the sediment one. This deduction provides some improvements with respect to the classical Saint-Venant-Exner model: (i) the deduced model has an associated energy. Moreover, it allows us to explain why classical models do not have an associated energy and how to modify them in order to recover a model with this property. (ii) The model incorporates naturally a necessary modification that must be taken into account in order to be applied to arbitrarily sloping beds. Furthermore, we show that this modification is different of the ones considered classically, and that it coincides with a classical one only if the solution ha...
A computational model of the motivation-learning interface Manish Saggar (mishu@cs.utexas.edu)
Maddox, W. Todd
A computational model of the motivation-learning interface Manish Saggar (mishu the influence of motivation on learning observed by Markman, Baldwin and Maddox (2005). They showed was confirmed. These results constitute a first computational step towards understanding how motivation
Efficient Computation of Info-Gap Robustness for Finite Element Models
Stull, Christopher J.; Hemez, Francois M.; Williams, Brian J.
2012-07-05
A recent research effort at LANL proposed info-gap decision theory as a framework by which to measure the predictive maturity of numerical models. Info-gap theory explores the trade-offs between accuracy, that is, the extent to which predictions reproduce the physical measurements, and robustness, that is, the extent to which predictions are insensitive to modeling assumptions. Both accuracy and robustness are necessary to demonstrate predictive maturity. However, conducting an info-gap analysis can present a formidable challenge, from the standpoint of the required computational resources. This is because a robustness function requires the resolution of multiple optimization problems. This report offers an alternative, adjoint methodology to assess the info-gap robustness of Ax = b-like numerical models solved for a solution x. Two situations that can arise in structural analysis and design are briefly described and contextualized within the info-gap decision theory framework. The treatments of the info-gap problems, using the adjoint methodology are outlined in detail, and the latter problem is solved for four separate finite element models. As compared to statistical sampling, the proposed methodology offers highly accurate approximations of info-gap robustness functions for the finite element models considered in the report, at a small fraction of the computational cost. It is noted that this report considers only linear systems; a natural follow-on study would extend the methodologies described herein to include nonlinear systems.
A Polarizable QM/MM Explicit Solvent Model for Computational Electrochemistry in Water
Wang, Lee-Ping
We present a quantum mechanical/molecular mechanical (QM/MM) explicit solvent model for the computation of standard reduction potentials E[subscript 0]. The QM/MM model uses density functional theory (DFT) to model the ...
Computation Modeling and Assessment of Nanocoatings for Ultra Supercritical Boilers
J. Shingledecker; D. Gandy; N. Cheruvu; R. Wei; K. Chan
2011-06-21
Forced outages and boiler unavailability of coal-fired fossil plants is most often caused by fire-side corrosion of boiler waterwalls and tubing. Reliable coatings are required for Ultrasupercritical (USC) application to mitigate corrosion since these boilers will operate at a much higher temperatures and pressures than in supercritical (565 C {at} 24 MPa) boilers. Computational modeling efforts have been undertaken to design and assess potential Fe-Cr-Ni-Al systems to produce stable nanocrystalline coatings that form a protective, continuous scale of either Al{sub 2}O{sub 3} or Cr{sub 2}O{sub 3}. The computational modeling results identified a new series of Fe-25Cr-40Ni with or without 10 wt.% Al nanocrystalline coatings that maintain long-term stability by forming a diffusion barrier layer at the coating/substrate interface. The computational modeling predictions of microstructure, formation of continuous Al{sub 2}O{sub 3} scale, inward Al diffusion, grain growth, and sintering behavior were validated with experimental results. Advanced coatings, such as MCrAl (where M is Fe, Ni, or Co) nanocrystalline coatings, have been processed using different magnetron sputtering deposition techniques. Several coating trials were performed and among the processing methods evaluated, the DC pulsed magnetron sputtering technique produced the best quality coating with a minimum number of shallow defects and the results of multiple deposition trials showed that the process is repeatable. scale, inward Al diffusion, grain growth, and sintering behavior were validated with experimental results. The cyclic oxidation test results revealed that the nanocrystalline coatings offer better oxidation resistance, in terms of weight loss, localized oxidation, and formation of mixed oxides in the Al{sub 2}O{sub 3} scale, than widely used MCrAlY coatings. However, the ultra-fine grain structure in these coatings, consistent with the computational model predictions, resulted in accelerated Al diffusion from the coating into the substrate. An effective diffusion barrier interlayer coating was developed to prevent inward Al diffusion. The fire-side corrosion test results showed that the nanocrystalline coatings with a minimum number of defects have a great potential in providing corrosion protection. The coating tested in the most aggressive environment showed no evidence of coating spallation and/or corrosion attack after 1050 hours exposure. In contrast, evidence of coating spallation in isolated areas and corrosion attack of the base metal in the spalled areas were observed after 500 hours. These contrasting results after 500 and 1050 hours exposure suggest that the premature coating spallation in isolated areas may be related to the variation of defects in the coating between the samples. It is suspected that the cauliflower-type defects in the coating were presumably responsible for coating spallation in isolated areas. Thus, a defect free good quality coating is the key for the long-term durability of nanocrystalline coatings in corrosive environments. Thus, additional process optimization work is required to produce defect-free coatings prior to development of a coating application method for production parts.
MOMDIS: a Glauber model computer code for knockout reactions
C. A. Bertulani; A. Gade
2006-04-12
A computer program is described to calculate momentum distributions in stripping and diffraction dissociation reactions. A Glauber model is used with the scattering wavefunctions calculated in the eikonal approximation. The program is appropriate for knockout reactions at intermediate energy collisions (30 MeV $\\leq$ E$_{lab}/$nucleon $\\leq 2000$ MeV). It is particularly useful for reactions involving unstable nuclear beams, or exotic nuclei (e.g. neutron-rich nuclei), and studies of single-particle occupancy probabilities (spectroscopic factors) and other related physical observables. Such studies are an essential part of the scientific program of radioactive beam facilities, as in for instance the proposed RIA (Rare Isotope Accelerator) facility in the US.
National Nuclear Security Administration (NNSA)
Each successive generation of computing system has provided greater computing power and energy efficiency.
CTS-1 clusters will support NNSA's Life Extension Program and...
High-Performance Computer Modeling of the Cosmos-Iridium Collision
Olivier, S; Cook, K; Fasenfest, B; Jefferson, D; Jiang, M; Leek, J; Levatin, J; Nikolaev, S; Pertica, A; Phillion, D; Springer, K; De Vries, W
2009-08-28
This paper describes the application of a new, integrated modeling and simulation framework, encompassing the space situational awareness (SSA) enterprise, to the recent Cosmos-Iridium collision. This framework is based on a flexible, scalable architecture to enable efficient simulation of the current SSA enterprise, and to accommodate future advancements in SSA systems. In particular, the code is designed to take advantage of massively parallel, high-performance computer systems available, for example, at Lawrence Livermore National Laboratory. We will describe the application of this framework to the recent collision of the Cosmos and Iridium satellites, including (1) detailed hydrodynamic modeling of the satellite collision and resulting debris generation, (2) orbital propagation of the simulated debris and analysis of the increased risk to other satellites (3) calculation of the radar and optical signatures of the simulated debris and modeling of debris detection with space surveillance radar and optical systems (4) determination of simulated debris orbits from modeled space surveillance observations and analysis of the resulting orbital accuracy, (5) comparison of these modeling and simulation results with Space Surveillance Network observations. We will also discuss the use of this integrated modeling and simulation framework to analyze the risks and consequences of future satellite collisions and to assess strategies for mitigating or avoiding future incidents, including the addition of new sensor systems, used in conjunction with the Space Surveillance Network, for improving space situational awareness.
Go-Smart: Web-based Computational Modeling of Minimally Invasive Cancer Treatments
Weir, Phil; Ellerweg, Roland; Alhonnoro, Tuomas; Pollari, Mika; Voglreiter, Philip; Mariappan, Panchatcharam; Flanagan, Ronan; Park, Chang Sub; Payne, Stephen; Staerk, Elmar; Voigt, Peter; Moche, Michael; Kolesnik, Marina
2015-01-01
The web-based Go-Smart environment is a scalable system that allows the prediction of minimally invasive cancer treatment. Interventional radiologists create a patient-specific 3D model by semi-automatic segmentation and registration of pre-interventional CT (Computed Tomography) and/or MRI (Magnetic Resonance Imaging) images in a 2D/3D browser environment. This model is used to compare patient-specific treatment plans and device performance via built-in simulation tools. Go-Smart includes evaluation techniques for comparing simulated treatment with real ablation lesions segmented from follow-up scans. The framework is highly extensible, allowing manufacturers and researchers to incorporate new ablation devices, mathematical models and physical parameters.
Madiraju, Praveen
discovery process lightweight and secure. In this paper we present details of the trust and risk models. We of devices running in a pervasive computing environment [10]. The resource discovery process demands modelsSSRD+: A Privacy-aware Trust and Security Model for Resource Discovery in Pervasive Computing
A Risk-aware Trust Based Secure Resource Discovery (RTSRD) Model for Pervasive Computing
Madiraju, Praveen
security threat to them. Thus, the resource discovery process demands models that ensure the privacyA Risk-aware Trust Based Secure Resource Discovery (RTSRD) Model for Pervasive Computing Sheikh I-hoc network of pervasive computing, a resource discovery model is needed that can resolve security and privacy
Using Parallel MCMC Sampling to Calibrate a Computer Model of a Geothermal Reservoir
Fox, Colin
Using Parallel MCMC Sampling to Calibrate a Computer Model of a Geothermal Reservoir by T. Cui, C. 686 ISSN 1178-360 #12;Using Parallel MCMC Sampling to Calibrate a Computer Model of a Geothermal of a geothermal field to achieve model `calibration' from measured well-test data. We explore three scenarios
Ray tracing computations in the smoothed SEG/EAGE Salt Model
Cerveny, Vlastislav
Ray tracing computations in the smoothed SEG/EAGE Salt Model V#19;aclav Bucha Department to compute rays and synthetic seismograms of refracted and re ected P-waves in the smoothed SEG/EAGE Salt The original 3-D SEG/EAGE Salt Model (Aminzadeh et al. 1997) is very complex model and cannot be used for ray
A Three-Dimensional Computational Model of PEM Fuel Cell with Serpentine Gas Channels
Victoria, University of
A Three-Dimensional Computational Model of PEM Fuel Cell with Serpentine Gas Channels by Phong ABSTRACT A three-dimensional computational fluid dynamics model of a Polymer Electrolyte Membrane (PEM) fuel cell with serpentine gas flow channels is presented in this thesis. This comprehensive model
A network based model for heterogeneous parallel computation
Sathye, Adwait B.
1993-01-01
The computational requirements of science and engineering demand computational resources orders of magnitude of the current day sequential machines. Most of the research effort has been concentrated upon the creation of parallel algorithms...
A New Model for Image-Based Humanities Computing
Brown, Jacob Hohmann
2009-05-15
Image-based humanities computing, the computer-assisted study of digitallyrepresented “objects or artifacts of cultural heritage,” is an increasingly popular yet “established practice” located at the most recent intersections ...
AIR INGRESS ANALYSIS: PART 2 – COMPUTATIONAL FLUID DYNAMIC MODELS
Chang H. Oh; Eung S. Kim; Richard Schultz; Hans Gougar; David Petti; Hyung S. Kang
2011-01-01
The Idaho National Laboratory (INL), under the auspices of the U.S. Department of Energy, is performing research and development that focuses on key phenomena important during potential scenarios that may occur in very high temperature reactors (VHTRs). Phenomena Identification and Ranking Studies to date have ranked an air ingress event, following on the heels of a VHTR depressurization, as important with regard to core safety. Consequently, the development of advanced air ingress-related models and verification and validation data are a very high priority. Following a loss of coolant and system depressurization incident, air will enter the core of the High Temperature Gas Cooled Reactor through the break, possibly causing oxidation of the in-the core and reflector graphite structure. Simple core and plant models indicate that, under certain circumstances, the oxidation may proceed at an elevated rate with additional heat generated from the oxidation reaction itself. Under postulated conditions of fluid flow and temperature, excessive degradation of the lower plenum graphite can lead to a loss of structural support. Excessive oxidation of core graphite can also lead to the release of fission products into the confinement, which could be detrimental to a reactor safety. Computational fluid dynamic model developed in this study will improve our understanding of this phenomenon. This paper presents two-dimensional and three-dimensional CFD results for the quantitative assessment of the air ingress phenomena. A portion of results of the density-driven stratified flow in the inlet pipe will be compared with results of the experimental results.
A model for computer frustration: the role of instrumental and dispositional factors
Shneiderman, Ben
A model for computer frustration: the role of instrumental and dispositional factors on incident 20742, USA Available online 16 April 2004 Abstract Frustration is almost universally accepted
Designing Computing System Architecture and Models for the HL-LHC era
Lothar Bauerdick; Brian Bockelman; Peter Elmer; Stephen Gowdy; Matevz Tadel; Frank Wuerthwein
2015-07-20
This paper describes a programme to study the computing model in CMS after the next long shutdown near the end of the decade.
Designing Computing System Architecture and Models for the HL-LHC era
Bauerdick, Lothar; Elmer, Peter; Gowdy, Stephen; Tadel, Matevz; Wuerthwein, Frank
2015-01-01
This paper describes a programme to study the computing model in CMS after the next long shutdown near the end of the decade.
and NCAR in the development of a comprehensive, earth systems model. This model incorporates the most-performance climate models. Through the addition of relevant physical processes, we are developing an earth systems modeling capability as well. Our collaborators in climate research include the National Center
Computational Modeling and Assessment Of Nanocoatings for Ultra Supercritical Boilers
David W. Gandy; John P. Shingledecker
2011-04-11
Forced outages and boiler unavailability in conventional coal-fired fossil power plants is most often caused by fireside corrosion of boiler waterwalls. Industry-wide, the rate of wall thickness corrosion wastage of fireside waterwalls in fossil-fired boilers has been of concern for many years. It is significant that the introduction of nitrogen oxide (NOx) emission controls with staged burners systems has increased reported waterwall wastage rates to as much as 120 mils (3 mm) per year. Moreover, the reducing environment produced by the low-NOx combustion process is the primary cause of accelerated corrosion rates of waterwall tubes made of carbon and low alloy steels. Improved coatings, such as the MCrAl nanocoatings evaluated here (where M is Fe, Ni, and Co), are needed to reduce/eliminate waterwall damage in subcritical, supercritical, and ultra-supercritical (USC) boilers. The first two tasks of this six-task project-jointly sponsored by EPRI and the U.S. Department of Energy (DE-FC26-07NT43096)-have focused on computational modeling of an advanced MCrAl nanocoating system and evaluation of two nanocrystalline (iron and nickel base) coatings, which will significantly improve the corrosion and erosion performance of tubing used in USC boilers. The computational model results showed that about 40 wt.% is required in Fe based nanocrystalline coatings for long-term durability, leading to a coating composition of Fe-25Cr-40Ni-10 wt.% Al. In addition, the long term thermal exposure test results further showed accelerated inward diffusion of Al from the nanocrystalline coatings into the substrate. In order to enhance the durability of these coatings, it is necessary to develop a diffusion barrier interlayer coating such TiN and/or AlN. The third task 'Process Advanced MCrAl Nanocoating Systems' of the six-task project jointly sponsored by the Electric Power Research Institute, EPRI and the U.S. Department of Energy (DE-FC26-07NT43096)- has focused on processing of advanced nanocrystalline coating systems and development of diffusion barrier interlayer coatings. Among the diffusion interlayer coatings evaluated, the TiN interlayer coating was found to be the optimum one. This report describes the research conducted under the Task 3 workscope.
A Mathematical Model for Virus Infection in a System of Interacting Computers
Cipolatti, Rolci
A Mathematical Model for Virus Infection in a System of Interacting Computers J. L´opez Gondar & R are explored and enlightened in this paper. 1. Introduction The infection of computers by virtual viruses of virtual viruses in a system of interacting computers could be compared with a disease transmitted
Modeling Computational Security in Long-Lived Systems, Version 2 Ran Canetti1,2
International Association for Cryptologic Research (IACR)
Modeling Computational Security in Long-Lived Systems, Version 2 Ran Canetti1,2 , Ling Cheung2 Introduction Computational security in long-lived systems: Security properties of cryptographic protocols computational power. This type of security degrades progressively over the lifetime of a protocol. However, some
Modeling Computational Security in Long-Lived Systems Ran Canetti1,2
International Association for Cryptologic Research (IACR)
Modeling Computational Security in Long-Lived Systems Ran Canetti1,2 , Ling Cheung2 , Dilsun Kaynar Introduction Computational security in long-lived systems: Security properties of cryptographic protocols protocols, security relies on the assumption that adversarial entities have lim- ited computational power
Novel properties generated by interacting computational systems: A minimal model Fabio Boschetti1,2
Boschetti, Fabio
Novel properties generated by interacting computational systems: A minimal model Fabio questions: First, what is the smallest number of components a computational system needs in order such as selforganisation and emergence have been discussed in computational terms within Complex System Science
Pedram, Massoud
Trace-Based Analysis and Prediction of Cloud Computing User Behavior Using the Fractal Modeling and technology. In this paper, we investigate the characteristics of the cloud computing requests received the alpha- stable distribution. Keywords- cloud computing; alpha-stable distribution; fractional order
CPT: An Energy-Efficiency Model for Multi-core Computer Systems
Shi, Weisong
CPT: An Energy-Efficiency Model for Multi-core Computer Systems Weisong Shi, Shinan Wang and Bing efficiency of computer systems. These techniques affect the energy efficiency across different layers metric that represents the energy efficiency of a computer system, for a specific configuration, given
Michael V. Glazoff; Jeong-Whan Yoon
2013-08-01
In this report (prepared in collaboration with Prof. Jeong Whan Yoon, Deakin University, Melbourne, Australia) a research effort was made to develop a non associated flow rule for zirconium. Since Zr is a hexagonally close packed (hcp) material, it is impossible to describe its plastic response under arbitrary loading conditions with any associated flow rule (e.g. von Mises). As a result of strong tension compression asymmetry of the yield stress and anisotropy, zirconium displays plastic behavior that requires a more sophisticated approach. Consequently, a new general asymmetric yield function has been developed which accommodates mathematically the four directional anisotropies along 0 degrees, 45 degrees, 90 degrees, and biaxial, under tension and compression. Stress anisotropy has been completely decoupled from the r value by using non associated flow plasticity, where yield function and plastic potential have been treated separately to take care of stress and r value directionalities, respectively. This theoretical development has been verified using Zr alloys at room temperature as an example as these materials have very strong SD (Strength Differential) effect. The proposed yield function reasonably well models the evolution of yield surfaces for a zirconium clock rolled plate during in plane and through thickness compression. It has been found that this function can predict both tension and compression asymmetry mathematically without any numerical tolerance and shows the significant improvement compared to any reported functions. Finally, in the end of the report, a program of further research is outlined aimed at constructing tensorial relationships for the temperature and fluence dependent creep surfaces for Zr, Zircaloy 2, and Zircaloy 4.
CASTING DEFECT MODELING IN AN INTEGRATED COMPUTATIONAL MATERIALS ENGINEERING APPROACH
Sabau, Adrian S [ORNL
2015-01-01
To accelerate the introduction of new cast alloys, the simultaneous modeling and simulation of multiphysical phenomena needs to be considered in the design and optimization of mechanical properties of cast components. The required models related to casting defects, such as microporosity and hot tears, are reviewed. Three aluminum alloys are considered A356, 356 and 319. The data on calculated solidification shrinkage is presented and its effects on microporosity levels discussed. Examples are given for predicting microporosity defects and microstructure distribution for a plate casting. Models to predict fatigue life and yield stress are briefly highlighted here for the sake of completion and to illustrate how the length scales of the microstructure features as well as porosity defects are taken into account for modeling the mechanical properties. Thus, the data on casting defects, including microstructure features, is crucial for evaluating the final performance-related properties of the component. ACKNOWLEDGEMENTS This work was performed under a Cooperative Research and Development Agreement (CRADA) with the Nemak Inc., and Chrysler Co. for the project "High Performance Cast Aluminum Alloys for Next Generation Passenger Vehicle Engines. The author would also like to thank Amit Shyam for reviewing the paper and Andres Rodriguez of Nemak Inc. Research sponsored by the U. S. Department of Energy, Office of Energy Efficiency and Renewable Energy, Vehicle Technologies Office, as part of the Propulsion Materials Program under contract DE-AC05-00OR22725 with UT-Battelle, LLC. Part of this research was conducted through the Oak Ridge National Laboratory's High Temperature Materials Laboratory User Program, which is sponsored by the U. S. Department of Energy, Office of Energy Efficiency and Renewable Energy, Vehicle Technologies Program.
A Calibrated Computer Model for the Thermal Simulation of Courtyard Microclimates
Bagneid, A.; Haberl, J.
2006-01-01
This paper describes a calibrated stand-alone courtyard microclimate model. This model is considered to be the fIrst calibrated computer program for the simulation of courtyard microclimates. In order to accomplish this a calibrated simplif...
Cielo Computational Environment Usage Model With Mappings to...
Office of Scientific and Technical Information (OSTI)
Environment (ACE) Version 8.0 requirements. The ACE requirements reflect the high performance computing requirements for the Production Readiness Milestone user environment...
Mathematical and Computer Modelling 53 (2011) 716730 Contents lists available at ScienceDirect
Berges, John A.
2011-01-01
Direct Mathematical and Computer Modelling journal homepage: www.elsevier.com/locate/mcm Dynamics of a virusMathematical and Computer Modelling 53 (2011) 716730 Contents lists available at Science: Received 19 January 2010 Accepted 17 October 2010 Keywords: Virushost dynamics Quota Bacteriophage
8/30/2001 Parallel Programming -Fall 2001 1 Models of Parallel Computation
Browne, James C.
8/30/2001 Parallel Programming - Fall 2001 1 Models of Parallel Computation Philosophy Parallel of parallel programming. #12;8/30/2001 Parallel Programming - Fall 2001 2 Models of Parallel Computation will discuss parallelism from the viewpoint of programming but with connections to other domains. #12;8/30/2001
Computing Estimates in the Proportional Odds Model David R. Hunter1
Hunter, David
Computing Estimates in the Proportional Odds Model David R. Hunter1 Kenneth Lange2 Department running head: Computing MLE for proportional odds Submitted to Annals of the Institute of Statistical The semiparametric proportional odds model for survival data is useful when mortality rates of different groups
Dessouky, Maged
A Hierarchical Task Model for Dispatching in Computer- Assisted Demand-Responsive Paratransit Model for Dispatching in Computer-Assisted Demand-Responsive Paratransit Operation ABSTRACT, Dispatch Training #12;1 INTRODUCTION Demand-responsive paratransit service is on the rise. For example
Computer modeling of the spatial resolution properties of a dedicated breast CT system
Yang Kai; Kwan, Alexander L. C.; Boone, John M. [Department of Radiology, University of California, Davis Medical Center, 4860 Y Street, Suite 3100 Ellison Building, Sacramento, California 95817 (United States) and Department of Biomedical Engineering, University of California, Davis, California, 95616 (United States); Department of Radiology, University of California, Davis Medical Center, 4860 Y Street, Suite 3100 Ellison Building, Sacramento, California 95817 (United States); Department of Radiology, University of California, Davis Medical Center, 4860 Y Street, Suite 3100 Ellison Building, Sacramento, California 95817 (United States) and Department of Biomedical Engineering, University of California, Davis, California 95616 (United States)
2007-06-15
Computer simulation methods were used to evaluate the spatial resolution properties of a dedicated cone-beam breast CT system. x-ray projection data of a 70 {mu}m nickel-chromium wire were simulated. The modulation transfer function (MTF) was calculated from the reconstructed axial images at different radial positions from the isocenter to study the spatial dependency of the spatial resolution of the breast CT scanner. The MTF was also calculated in both the radial and azimuthal directions. Subcomponents of the cone beam CT system that affect the MTF were modeled in the computer simulation in a serial manner, including the x-ray focal spot distribution, gantry rotation under the condition of continuous fluoroscopy, detector lag, and detector spatial resolution. Comparison between the computer simulated and physically measured MTF values demonstrates reasonable accuracy in the simulation process, with a small systematic difference ({approx}9.5{+-}6.4% difference, due to unavoidable uncertainties from physical measurement and system calibration). The intrinsic resolution in the radial direction determined by simulation was about 2.0 mm{sup -1} uniformly through the field of view. The intrinsic resolution in the azimuthal direction degrades from 2.0 mm{sup -1} at the isocenter to 1.0 mm{sup -1} at the periphery with 76.9 mm from the isocenter. The results elucidate the intrinsic spatial resolution properties of the prototype breast CT system, and suggest ways in which spatial resolution can be improved with system modification.
Tarragon : a programming model for latency-hiding scientific computations
Cicotti, Pietro
2011-01-01
Chapter 2 Programming Model . . . . . . . . . 2.1vi Chapter 6 Dynamic programming . . . . . . . . . . 6.1 Theof related programming model implementations. . . . . .
Nishino, Takafumi
2012-01-01
Modelling of turbine blade-induced turbulence (BIT) is discussed within the framework of three-dimensional Reynolds-averaged Navier-Stokes (RANS) actuator disk computations. We first propose a generic (baseline) BIT model, which is applied only to the actuator disk surface, does not include any model coefficients (other than those used in the original RANS turbulence model) and is expected to be valid in the limiting case where BIT is fully isotropic and in energy equilibrium. The baseline model is then combined with correction functions applied to the region behind the disk to account for the effect of rotor tip vortices causing a mismatch of Reynolds shear stress between short- and long-time averaged flow fields. Results are compared with wake measurements of a two-bladed wind turbine model of Medici and Alfredsson [Wind Energy, Vol. 9, 2006, pp. 219-236] to demonstrate the capability of the new model.
Improvements in fast-response flood modeling: desktop parallel computing and domain tracking
Judi, David R; Mcpherson, Timothy N; Burian, Steven J
2009-01-01
It is becoming increasingly important to have the ability to accurately forecast flooding, as flooding accounts for the most losses due to natural disasters in the world and the United States. Flood inundation modeling has been dominated by one-dimensional approaches. These models are computationally efficient and are considered by many engineers to produce reasonably accurate water surface profiles. However, because the profiles estimated in these models must be superimposed on digital elevation data to create a two-dimensional map, the result may be sensitive to the ability of the elevation data to capture relevant features (e.g. dikes/levees, roads, walls, etc...). Moreover, one-dimensional models do not explicitly represent the complex flow processes present in floodplains and urban environments and because two-dimensional models based on the shallow water equations have significantly greater ability to determine flow velocity and direction, the National Research Council (NRC) has recommended that two-dimensional models be used over one-dimensional models for flood inundation studies. This paper has shown that two-dimensional flood modeling computational time can be greatly reduced through the use of Java multithreading on multi-core computers which effectively provides a means for parallel computing on a desktop computer. In addition, this paper has shown that when desktop parallel computing is coupled with a domain tracking algorithm, significant computation time can be eliminated when computations are completed only on inundated cells. The drastic reduction in computational time shown here enhances the ability of two-dimensional flood inundation models to be used as a near-real time flood forecasting tool, engineering, design tool, or planning tool. Perhaps even of greater significance, the reduction in computation time makes the incorporation of risk and uncertainty/ensemble forecasting more feasible for flood inundation modeling (NRC 2000; Sayers et al. 2000).
Validation of the thermospheric vector spherical harmonic (VSH) computer model. Master's thesis
Davis, J.L.
1991-01-01
A semi-empirical computer model of the lower thermosphere has been developed that provides a description of the composition and dynamics of the thermosphere (Killeen et al., 1992). Input variables needed to run the VSH model include time, space and geophysical conditions. One of the output variables the model provides, neutral density, is of particular interest to the U.S. Air Force. Neutral densities vary both as a result of change in solar flux (eg. the solar cycle) and as a result of changes in the magnetosphere (eg. large changes occur in neutral density during geomagnetic storms). Satellites in earth orbit experience aerodynamic drag due to the atmospheric density of the thermosphere. Variability in the neutral density described above affects the drag a satellite experiences and as a result can change the orbital characteristics of the satellite. These changes make it difficult to track the satellite's position. Therefore, it is particularly important to insure that the accuracy of the model's neutral density is optimized for all input parameters. To accomplish this, a validation program was developed to evaluate the strengths and weaknesses of the model's density output by comparing it to SETA-2 (satellite electrostatic accelerometer) total mass density measurements.
Towards a Model for Computing in European Astroparticle Physics
T. Berghöfer; I. Agrafioti; B. Allen; V. Beckmann; T. Chiarusi; M. Delfino; S. Hesping; J. Chudoba; L. Dell'Agnello; S. Katsanevas; G. Lamanna; R. Lemrani; A. Margiotta; G. Maron; C. Palomba; G. Russo; P. Wegner
2015-12-03
Current and future astroparticle physics experiments are operated or are being built to observe highly energetic particles, high energy electromagnetic radiation and gravitational waves originating from all kinds of cosmic sources. The data volumes taken by the experiments are large and expected to grow significantly during the coming years. This is a result of advanced research possibilities and improved detector technology. To cope with the substantially increasing data volumes of astroparticle physics projects it is important to understand the future needs for computing resources in this field. Providing these resources constitutes a larger fraction of the overall running costs of future infrastructures. This document presents the results of a survey made by APPEC with the help of computing experts of major projects and future initiatives in astroparticle physics, representatives of current Tier-1 and Tier-2 LHC computing centers, as well as specifically astroparticle physics computing centers, e.g. the Albert Einstein Institute for gravitational waves analysis in Hanover. In summary, the overall CPU usage and short-term disk and long-term (tape) storage space currently available for astroparticle physics projects' computing services is of the order of one third of the central computing available for LHC data at the Tier-0 center at CERN. Till the end of the decade the requirements for computing resources are estimated to increase by a factor of 10. Furthermore, this document shall describe the diversity of astroparticle physics data handling and serve as a basis to estimate a distribution of computing and storage tasks among the major computing centers. (Abridged)
Math 484: Mathematical & Computational Modeling Course Information and Syllabus Spring 2013
Math 484: Mathematical & Computational Modeling Course Information and Syllabus Spring 2013 students how to build and analyze mathematical models. Most concepts presented in the context of physical students how to build mathematical models to physical problems. 2. Be able to solve modeling problems
Math 484: Mathematical & Computational Modeling Course Information and Syllabus Spring 2012
Math 484: Mathematical & Computational Modeling Course Information and Syllabus Spring 2012 students how to build and analyze mathematical models. Most concepts presented in the context of physical students how to build mathematical models to physical problems. 2. Be able to solve modeling problems
JACKSON VL
2011-08-31
The primary purpose of the tank mixing and sampling demonstration program is to mitigate the technical risks associated with the ability of the Hanford tank farm delivery and celtification systems to measure and deliver a uniformly mixed high-level waste (HLW) feed to the Waste Treatment and Immobilization Plant (WTP) Uniform feed to the WTP is a requirement of 24590-WTP-ICD-MG-01-019, ICD-19 - Interface Control Document for Waste Feed, although the exact definition of uniform is evolving in this context. Computational Fluid Dynamics (CFD) modeling has been used to assist in evaluating scaleup issues, study operational parameters, and predict mixing performance at full-scale.
A Fire Model for 2-D Computer Animation
Yu, J-H.; Patterson, J.W.
Yu,J-H. Patterson,J.W. Proceedings of the EUROGRAPHICS Workshop on Computer Animation '96, Poitiers, France. Published in Eurographics Series, (Boulic R. and Hegron, G., Eds.). pp 49-60 Springer
Edinburgh Research Explorer Computational models in systems biology
Millar, Andrew J.
called process algebras and Petri nets offer alternative ways of constructing computational systems networks in an abstract way that is independent of particular mathe- matical techniques of analysis. Access
The Need for Biological Computation System Models | GE Global...
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
2012.10.09 Hello everyone, I'm Maria Zavodszky and I work in the Computational Biology and Biostatistics Lab at GE Global Research in Niskayuna, New York. This being our...
Multiscale Computational Modeling of Multiphase Composites with Damage
Cheng, Feifei
2013-11-01
A multiscale computational framework for multiphase composites considering damage is developed in this research. In micro-scale, micromechanics based homogenization methods are used to estimate effective elastic moduli of graded Ti_(2)Al...
Automatic Symmetry Detection for Model Checking Using Computational Group Theory
Donaldson, A.F.; Miller, A.
Donaldson,A.F. Miller,A. Proceedings of the 13th International Symposium on Formal Methods Europe (FME 2005). Lecture Notes in Computing Science volume 3582. pp 481--496 Springer
Applications of Computer Modelling to Fire Safety Design
Torero, Jose L; Steinhaus, Thomas
Tools in support of fire safety engineering design have proliferated in the last few years due to the increased performance of computers. These tools are currently being used in a generalized manner in areas such as egress, ...
Cloud computing adoption model for governments and large enterprises
Trivedi, Hrishikesh
2013-01-01
Cloud Computing has held organizations across the globe spell bound with its promise. As it moves from being a buzz word and hype into adoption, organizations are faced with question of how to best adopt cloud. Existing ...
Call for Papers ACM Transactions on Modeling and Computer Simulation
L'Ecuyer, Pierre
, and smart energy grids. Important issues of interest include: building appropriate and reasonably realistic
Effective Design And Use Of Computer Decision Models.
Fuerst, William L.
1984-03-01
of decision models such as simulation have not been demonstrated. This paper looks at recent literature regarding decision model deficiencies, evaluates selected financial simulation model packages, and suggests design needs for expanding the use of decision...
Modeling-Computer Simulations At Dixie Valley Geothermal Area...
W. Wisian, David D. Blackwell (2004) Numerical Modeling Of Basin And Range Geothermal Systems Additional References Retrieved from "http:en.openei.orgwindex.php?titleModel...
Center for Programming Models for Scalable Parallel Computing: Future Programming Models
Gao, Guang, R.
2008-07-24
The mission of the pmodel center project is to develop software technology to support scalable parallel programming models for terascale systems. The goal of the specific UD subproject is in the context developing an efficient and robust methodology and tools for HPC programming. More specifically, the focus is on developing new programming models which facilitate programmers in porting their application onto parallel high performance computing systems. During the course of the research in the past 5 years, the landscape of microprocessor chip architecture has witnessed a fundamental change – the emergence of multi-core/many-core chip architecture appear to become the mainstream technology and will have a major impact to for future generation parallel machines. The programming model for shared-address space machines is becoming critical to such multi-core architectures. Our research highlight is the in-depth study of proposed fine-grain parallelism/multithreading support on such future generation multi-core architectures. Our research has demonstrated the significant impact such fine-grain multithreading model can have on the productivity of parallel programming models and their efficient implementation.
Tesfatsion, Leigh
Bounded Computing Capacity ·· Explicit SpaceExplicit Space ·· Local InteractionsLocal Interactions ·· Non explanatory notion.notion. #12;4 SugarscapeSugarscape ·· Events unfold on a landscape of renewableEvents
Kwicklis, Edward Michael [Los Alamos National Laboratory; Keating, Elizabeth H [Los Alamos National Laboratory
2010-12-02
Much progress has been made in the last several years in modeling radionuclide transport from tests conducted both in the unsaturated zone and saturated volcanic rocks of Yucca Flat, Nevada. The presentations to the DOE NNSA pre-emptive review panel contained herein document the progress to date, and discuss preliminary conclusions regarding the present and future extents of contamination resulting from past nuclear tests. The presentations also discuss possible strategies for addressing uncertainty in the model results.
Computer Modeling VRF Heat Pumps in Commercial Buildings using EnergyPlus
Raustad, Richard
2013-06-01
Variable Refrigerant Flow (VRF) heat pumps are increasingly used in commercial buildings in the United States. Monitored energy use of field installations have shown, in some cases, savings exceeding 30% compared to conventional heating, ventilating, and air-conditioning (HVAC) systems. A simulation study was conducted to identify the installation or operational characteristics that lead to energy savings for VRF systems. The study used the Department of Energy EnergyPlus? building simulation software and four reference building models. Computer simulations were performed in eight U.S. climate zones. The baseline reference HVAC system incorporated packaged single-zone direct-expansion cooling with gas heating (PSZ-AC) or variable-air-volume systems (VAV with reheat). An alternate baseline HVAC system using a heat pump (PSZ-HP) was included for some buildings to directly compare gas and electric heating results. These baseline systems were compared to a VRF heat pump model to identify differences in energy use. VRF systems combine multiple indoor units with one or more outdoor unit(s). These systems move refrigerant between the outdoor and indoor units which eliminates the need for duct work in most cases. Since many applications install duct work in unconditioned spaces, this leads to installation differences between VRF systems and conventional HVAC systems. To characterize installation differences, a duct heat gain model was included to identify the energy impacts of installing ducts in unconditioned spaces. The configuration of variable refrigerant flow heat pumps will ultimately eliminate or significantly reduce energy use due to duct heat transfer. Fan energy is also studied to identify savings associated with non-ducted VRF terminal units. VRF systems incorporate a variable-speed compressor which may lead to operational differences compared to single-speed compression systems. To characterize operational differences, the computer model performance curves used to simulate cooling operation are also evaluated. The information in this paper is intended to provide a relative difference in system energy use and compare various installation practices that can impact performance. Comparative results of VRF versus conventional HVAC systems include energy use differences due to duct location, differences in fan energy when ducts are eliminated, and differences associated with electric versus fossil fuel type heating systems.
A.24-1 A.24 ENHANCING THE CAPABILITY OF COMPUTATIONAL EARTH SYSTEM MODELS AND NASA DATA) computational support of Earth system modeling. #12;A.24-2 2.1 Acceleration of Operational Use of Research Data
Computer modeling reveals how surprisingly potent hepatitis C drug works
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
AFDC Printable Version Share this resource Send a link to EERE: Alternative Fuels Data Center Home Page to someone by E-mail Share EERE: Alternative Fuels Data Center Home Page on Facebook Tweet about EERE: Alternative Fuels Data Center Home Page on Twitter Bookmark EERE: Alternative Fuels Data Center Homesum_a_epg0_fpd_mmcf_m.xls" ,"Available from WebQuantityBonneville Power Administration would like submit the following commentsMethodsCompositional6 ComputationalComputerHepatitis C
Computer-Aided Construction of Combustion Chemistry Models
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
AFDC Printable Version Share this resource Send a link to EERE: Alternative Fuels Data Center Home Page to someone by E-mail Share EERE: Alternative Fuels Data Center Home Page on Facebook Tweet about EERE: Alternative Fuels Data Center Home Page on Twitter Bookmark EERE: Alternative Fuels Data Center Homesum_a_epg0_fpd_mmcf_m.xls" ,"Available from WebQuantityBonneville Power Administration would like submit the following commentsMethodsCompositional6 ComputationalComputerHepatitis
WUFI COMPUTER MODELING WORKSHOP FOR WALL DESIGN AND PERFORMANCE
Oak Ridge National Laboratory
from rain, solar radiation and other crucial weather events on an hourly basis. Both vapor and liquid which can easily be selected from a map. The temporal behavior of the computed quantities (temperatures tell me performance predicting · Mold growth predictions new post-processing modules · Development
Modeling Living Systems for Computer Vision Demetri Terzopoulos
Terzopoulos, Demetri
, for computer vision. First, I present a new breed of artificial animals in a physicsbased virtual marine world vision. 1 These advances center around the idea of artificial animals, or ``animats'' a term coined and 3 present our work on artificial fishes and animat vision. The basic idea in a nutshell
Modelling Photochemical Pollution using Parallel and Distributed Computing Platforms
Abramson, David
of photochemical air pollution (smog) in industrialised cities. However, computational hardware demands can that have been used as part of an air pollution study being conducted in Melbourne, Australia. We also necessary to perform real air pollution studies. The system is used as part of the Melbourne Airshed study
Modeling-Computer Simulations At Kilauea East Rift Geothermal...
East Rift Zone Notes Three models were made from data collected in the exploratory well HGP-A. The models simulated constant heat sources from a vertical dike, a magma chamber,...
An Axiomatisation of Computationally Adequate Domain Theoretic Models of FPC
Fiore, Marcelo P; Plotkin, Gordon
1994-01-01
Categorical models of the metalanguage FPC (a type theory with sums, products, exponentials and recursive types) are defined. Then, domain-theoretic models of FPC are axiomatised and a wide subclass of them —the ...
Bradley, D.R.; Gardner, D.R.; Brockmann, J.E.; Griffith, R.O. [Sandia National Labs., Albuquerque, NM (United States)
1993-10-01
The CORCON-Mod3 computer code was developed to mechanistically model the important core-concrete interaction phenomena, including those phenomena relevant to the assessment of containment failure and radionuclide release. The code can be applied to a wide range of severe accident scenarios and reactor plants. The code represents the current state of the art for simulating core debris interactions with concrete. This document comprises the user`s manual and gives a brief description of the models and the assumptions and limitations in the code. Also discussed are the input parameters and the code output. Two sample problems are also given.
Experimental Evaluations of Expert and Non-expert Computer Users' Mental Models of Security Risks
Camp, L. Jean
Experimental Evaluations of Expert and Non-expert Computer Users' Mental Models of Security Risks risks and thereby enable informed decisions by naive users. Yet computer security has not been en- gaged with the scholarship of risk communication. While the existence of malicious actors may appear at first to distinguish
Creative Commons Copyright 2013 Some Rights Reserved CMC: A Model Computer Science Curriculum
Iyer, Sridhar
for K-12 Schools 3rd Edition, Released June 2013 Technical Report: TR-CSE-2013-52 Department of Computer-12 Schools Creative Commons Copyright © 2013 Some Rights Reserved 2 CMC: A Model Computer Science Curriculum for K-12 Schools 3rd Edition June, 2013 Authors Sridhar Iyer*, Farida Khan, Sahana Murthy
An Exact Modeling of Signal Statistics in Energy-integrating X-ray Computed Tomography
used by modern computed tomography (CT) scanners and has been an interesting research topicAn Exact Modeling of Signal Statistics in Energy-integrating X-ray Computed Tomography Yi Fan1.i.d.), such as Gamma, Gaussian, etc, would be valid. A comparison study was performed to estimate the introduced errors
Internship Parallel Computer Evaluation Parallelization of a Lagrangian Particle Diffusion Model
use case is a nuclear accident like a core meltdown at a atomic power plant, where atomic radiation emits in the air. The Lagrangian model can predict how the nuclear cloud spreads under different that will be computed. Particle: One single molecule floating in the wind field. Compute unit: One unit that runs
A computational model for nanoscale adhesion between deformable solids and its application to gecko adhesion Roger A. Sauer 1 Aachen Institute for Advanced Study in Computational Engineering Science (AICES), RWTH Aachen University, Templergaben 55, 52056 Aachen, Germany Published2 in the Journal of Adhesion
Author's personal copy Calibration procedures for a computational model of ductile fracture
Hutchinson, John W.
Author's personal copy Calibration procedures for a computational model of ductile fracture Z. Xue fracture Computational fracture Shear fracture Damage parameters a b s t r a c t A recent extension of the cup-cone fracture mode in the neck of a round tensile bar. Ductility of a notched round bar provides
A Computational Model of Knowledge-Intensive Learning and Problem Solving1
Aamodt, Agnar
1 A Computational Model of Knowledge-Intensive Learning and Problem Solving1 Agnar Aamodt Knowledge. If knowledge-based systems are to become more competent and robust in solving real world problems, they need model - a framework -for knowledge-intensive problem solving and learning from experience. The model has
COMPUTERS AND BIOMEDICAL RESEARCH 17, 580-589 ( 1984) Multicompartment Model of Lung Dynamics*
Longtin, André
COMPUTERS AND BIOMEDICAL RESEARCH 17, 580-589 ( 1984) Multicompartment Model of Lung Dynamics* B of the lungs. The lungs are represented by 24 compartments each corresponding to a generation of the Weibel model A. In the model it is assumed that gases are transported in the lungs by convection and diffusion
Lecture Notes in Computer Science 1 Data Reduction Using Multiple Models Integration
Obradovic, Zoran
Lecture Notes in Computer Science 1 Data Reduction Using Multiple Models Integration Aleksandar the models constructed on previously considered data samples. In addition to random sampling, controllable sampling based on the boosting algorithm is proposed, where the models are combined using a weighted voting
Computational Fluid Dynamics Modeling of a Lithium/Thionyl Chloride Battery with Electrolyte Flow
Wang, Chao-Yang
Computational Fluid Dynamics Modeling of a Lithium/Thionyl Chloride Battery with Electrolyte Flow W-dimensional model is developed to simulate discharge of a primary lithium/thionyl chloride battery. The model to the first task with important examples of lead-acid,1-3 nickel-metal hydride,4-8 and lithium-based batteries
DEVELOPMENT OF A COMPUTER SIMULATION MODEL FOR BLOWING GLASS CONTAINERS C. G. Giannopapa
Eindhoven, Technische Universiteit
1 DEVELOPMENT OF A COMPUTER SIMULATION MODEL FOR BLOWING GLASS CONTAINERS C. G. Giannopapa Dept to be used for industrial purposes that accurately captures the blowing step of glass containers. The model,2]. This paper concentrates on modeling the blowing stage of the forming process of glass containers
Ortiz, Michael
mechanical re- sponse of a ceramic matrix composite is simulated by a numerical model for a ®ber-matrix unitComputational modeling of damage evolution in unidirectional fiber reinforced ceramic matrix evolution in brittle matrix composites was developed. This modeling is based on an axisymmetric unit cell
Gedeon, Tomas
, from those appearing in physiology and ecology to Earth systems modeling, often experience critical
Computational Modeling of Conventionally Reinforced Concrete Coupling Beams
Shastri, Ajay Seshadri
2012-02-14
. The model is developed in the finite element analysis software ABAQUS. The concrete damaged plasticity model was used to simulate the behavior of concrete. A calibration model using a cantilever beam was produced to generate key parameters in the model... Stress (ABAQUS 2008)??????.????...61 Fig. 3.9. CPS8 Element Used for Modeling Concrete (ABAQUS 2008)?????64 Fig. 4.1. Elevation and Cross-Section of the Cantilever Beam?????????66 Fig. 4.2. Compressive Stress-Strain Behavior of Concrete...
Michael V. Glazoff; Piyush Sabharwall; Akira Tokuhiro
2014-09-01
An evaluation of thermodynamic aspects of hot corrosion of the superalloys Haynes 242 and HastelloyTM N in the eutectic mixtures of KF and ZrF4 is carried out for development of Advanced High Temperature Reactor (AHTR). This work models the behavior of several superalloys, potential candidates for the AHTR, using computational thermodynamics tool (ThermoCalc), leading to the development of thermodynamic description of the molten salt eutectic mixtures, and on that basis, mechanistic prediction of hot corrosion. The results from these studies indicated that the principal mechanism of hot corrosion was associated with chromium leaching for all of the superalloys described above. However, HastelloyTM N displayed the best hot corrosion performance. This was not surprising given it was developed originally to withstand the harsh conditions of molten salt environment. However, the results obtained in this study provided confidence in the employed methods of computational thermodynamics and could be further used for future alloy design efforts. Finally, several potential solutions to mitigate hot corrosion were proposed for further exploration, including coating development and controlled scaling of intermediate compounds in the KF-ZrF4 system.
Emergency Response Equipment and Related Training: Airborne Radiological Computer System (Model II)
David P. Colton
2007-02-28
The materials included in the Airborne Radiological Computer System, Model-II (ARCS-II) were assembled with several considerations in mind. First, the system was designed to measure and record the airborne gamma radiation levels and the corresponding latitude and longitude coordinates, and to provide a first overview look of the extent and severity of an accident's impact. Second, the portable system had to be light enough and durable enough that it could be mounted in an aircraft, ground vehicle, or watercraft. Third, the system must control the collection and storage of the data, as well as provide a real-time display of the data collection results to the operator. The notebook computer and color graphics printer components of the system would only be used for analyzing and plotting the data. In essence, the provided equipment is composed of an acquisition system and an analysis system. The data can be transferred from the acquisition system to the analysis system at the end of the data collection or at some other agreeable time.
Seagraves, Andrew Nathan
2010-01-01
In this thesis a new parallel computational method is proposed for modeling threedimensional dynamic fracture of brittle solids. The method is based on a combination of the discontinuous Galerkin (DG) formulation of the ...
Curry, Benjamin David
A goal of Artificial Intelligence is to develop computational models of what would be considered intelligent behaviour in a human. One such task is that of musical performance. This research specifically focuses on aspects ...
A Computational Market Model for Distributed Configuration Design Michael P. Wellman
Wellman, Michael P.
economies" constitutes the market solution to the original problem. After defining the configuration design. Consider a hyper-simplified scenario in aircraft design. (We choose this not as a serious exemplar ¡ ¢ £ £ ¤ ¢ ¥ ¦ § ¨ © ¨ A Computational Market Model for Distributed Configuration
THEORETICAL MODELING AND COMPUTATIONAL SIMULATION OF ROBUST CONTROL FOR MARS AIRCRAFT
Oh, Seyool
2014-05-31
The focus of this dissertation is the development of control system design algorithms for autonomous operation of an aircraft in the Martian atmosphere. This research will show theoretical modeling and computational ...
Ma, Yongting
2011-01-11
This thesis presents development of mathematical models for multi-media interaction process using Eulerian description and associated computational infrastructure to obtain numerical solution of the initial value problems ...
Protein translocation without specific quality control in a computational model of the Tat system
Chitra R. Nayak; Aidan I. Brown; Andrew D. Rutenberg
2014-08-20
The twin-arginine translocation (Tat) system transports folded proteins of various sizes across both bacterial and plant thylakoid membranes. The membrane-associated TatA protein is an essential component of the Tat translocon, and a broad distribution of different sized TatA-clusters is observed in bacterial membranes. We assume that the size dynamics of TatA clusters are affected by substrate binding, unbinding, and translocation to associated TatBC clusters, where clusters with bound translocation substrates favour growth and those without associated substrates favour shrinkage. With a stochastic model of substrate binding and cluster dynamics, we numerically determine the TatA cluster size distribution. We include a proportion of targeted but non-translocatable (NT) substrates, with the simplifying hypothesis that the substrate translocatability does not directly affect cluster dynamical rate constants or substrate binding or unbinding rates. This amounts to a translocation model without specific quality control. Nevertheless, NT substrates will remain associated with TatA clusters until unbound and so will affect cluster sizes and translocation rates. We find that the number of larger TatA clusters depends on the NT fraction $f$. The translocation rate can be optimized by tuning the rate of spontaneous substrate unbinding, $\\Gamma_U$. We present an analytically solvable three-state model of substrate translocation without cluster size dynamics that follows our computed translocation rates, and that is consistent with {\\em in vitro} Tat-translocation data in the presence of NT substrates.
Westbrook, C K; Mizobuchi, Y; Poinsot, T J; Smith, P J; Warnatz, J
2004-08-26
Progress in the field of computational combustion over the past 50 years is reviewed. Particular attention is given to those classes of models that are common to most system modeling efforts, including fluid dynamics, chemical kinetics, liquid sprays, and turbulent flame models. The developments in combustion modeling are placed into the time-dependent context of the accompanying exponential growth in computer capabilities and Moore's Law. Superimposed on this steady growth, the occasional sudden advances in modeling capabilities are identified and their impacts are discussed. Integration of submodels into system models for spark ignition, diesel and homogeneous charge, compression ignition engines, surface and catalytic combustion, pulse combustion, and detonations are described. Finally, the current state of combustion modeling is illustrated by descriptions of a very large jet lifted 3D turbulent hydrogen flame with direct numerical simulation and 3D large eddy simulations of practical gas burner combustion devices.
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
AFDC Printable Version Share this resource Send a link to EERE: Alternative Fuels Data Center Home Page to someone by E-mail Share EERE: Alternative Fuels Data Center Home Page on Facebook Tweet about EERE: Alternative Fuels Data Center Home Page on Twitter Bookmark EERE: Alternative Fuels Data Center Homesum_a_epg0_fpd_mmcf_m.xls" ,"Available from WebQuantityBonneville Power Administration would like submit the following commentsMethodsCompositional6Energy Computers,Computing and
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
AFDC Printable Version Share this resource Send a link to EERE: Alternative Fuels Data Center Home Page to someone by E-mail Share EERE: Alternative Fuels Data Center Home Page on Facebook Tweet about EERE: Alternative Fuels Data Center Home Page on Twitter Bookmark EERE: Alternative Fuels Data Center Homesum_a_epg0_fpd_mmcf_m.xls" ,"Available from WebQuantityBonneville Power Administration would like submit the following commentsMethodsCompositional6Energy Computers,Computing
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
AFDC Printable Version Share this resource Send a link to EERE: Alternative Fuels Data Center Home Page to someone by E-mail Share EERE: Alternative Fuels Data Center Home Page on Facebook Tweet about EERE: Alternative Fuels Data Center Home Page on Twitter Bookmark EERE: Alternative Fuels Data Center Homesum_a_epg0_fpd_mmcf_m.xls" ,"Available from WebQuantity ofkandz-cm11 Outreach Home Room News PublicationsAuditsCluster Compatibilitydefault Changes TheComputeComputing
Formal computational models and non-standard finiteness
Ayala-RincÃ³n, Mauricio
(f) #12;Finiteness in Computation ^ without NNO T : Pos - f : P(Q Ã? Pos Ã? Tfin) - P(Q Ã? Posfin Ã? Q), such that Tfin = {h/fin({p Pos : h(p) = b})} X = P(Q Ã? Pos Ã? Tfin), St(X) = {S X : f(S) S} i : St(X) P(X) has = { x , y /Pos , yT > ^ Pos , xT > and final(q)} #12;Finitenes
Modeling-Computer Simulations At Long Valley Caldera Geothermal...
surrounding a vertically dipping prolate spheroid source during an active period of time-dependent deformation between 1995 and 2000 at Long Valley caldera. We model a rapid...
Modeling-Computer Simulations At Long Valley Caldera Geothermal...
and seismic data was conducted in 2003 to investigate the cause of recent uplift of the resurgent dome. Notes Modeling of deformation and microgravity data suggests...
Advanced Computing Tools and Models for Accelerator Physics
Ryne, Robert D.
2008-01-01
MODELS FOR ACCELERATOR PHYSICS * Robert D. Ryne, Lawrencetools for accelerator physics. Following an introduction Icomputing in accelerator physics. INTRODUCTION To begin I
Mathematical modeling and computer simulation of processes in energy systems
Hanjalic, K.C. )
1990-01-01
This book is divided into the following chapters. Modeling techniques and tools (fundamental concepts of modeling); 2. Fluid flow, heat and mass transfer, chemical reactions, and combustion; 3. Processes in energy equipment and plant components (boilers, steam and gas turbines, IC engines, heat exchangers, pumps and compressors, nuclear reactors, steam generators and separators, energy transport equipment, energy convertors, etc.); 4. New thermal energy conversion technologies (MHD, coal gasification and liquefaction fluidized-bed combustion, pulse-combustors, multistage combustion, etc.); 5. Combined cycles and plants, cogeneration; 6. Dynamics of energy systems and their components; 7. Integrated approach to energy systems modeling, and 8. Application of modeling in energy expert systems.
Modeling-Computer Simulations At Long Valley Caldera Geothermal...
LVEW are best matched using modeled solutions for a flow system consisting of a rock matrix with finite hydraulic conductivity cut by a steeply dipping fracture with infinite...
Vassilis Geroyannis; Georgios Kleftogiannis
2014-06-14
We revisit the problem of radial pulsations of neutron stars by computing four general-relativistic polytropic models, in which "density" and "adiabatic index" are involved with their discrete meanings: (i) "rest-mass density" or (ii) "mass-energy density" regarding the density, and (i) "constant" or (ii) "variable" regarding the adiabatic index. Considering the resulting four discrete combinations, we construct corresponding models and compute for each model the frequencies of the lowest three radial modes. Comparisons with previous results are made. The deviations of respective frequencies of the resolved models seem to exhibit a systematic behavior, an issue discussed here in detail.
Gaussian Process Modeling and Computation in Engineering Applications
Pourhabib, Arash
2014-07-08
; and predictive modeling for large datasets. First, we develop a spatial-temporal model for local wind fields in a wind farm with more than 200 wind turbines. Our framework utilizes the correlation among the derivatives of wind speeds to find a neighborhood...
Modeling civil violence: An agent-based computational approach
Tesfatsion, Leigh
," I do so advisedly, recognizing that no political or social order is represented in the model of revolutions properly speaking. The dynamics of decentralized upheaval, rather than its political substance Against Central Authority This model involves two categories of actors. ``Agents'' are members
Huang, Su-Yun
factors · Simulation codes with calibration parameters 8 #12;Example: Designing Cellular Heat Exchangers in Qian et al. (2006, ASME) Related to the autoregressive model in Kennedy and O'Hagan (2000) · x = (x1
Didactyl: Toward a Useful Computational Model of Piano Fingering
, ergonomics, cognition, and habit. Theoretically, the search space for an ideal) to define evaluation corpora to allow comparisons between the models, and (3 and evaluation corpora to reduce opportunity costs for future researchers working
Computational tools for modeling and measuring chromosome structure
Ross, Brian Christopher
2012-01-01
DNA conformation within cells has many important biological implications, but there are challenges both in modeling DNA due to the need for specialized techniques, and experimentally since tracing out in vivo conformations ...
When Model Checking Met Deduction Computer Science Laboratory
Clarke, Edmund M.
Park, CA Sep 19, 2014 #12;Alan Turing It is of course important that some efforts be made to verify hold in each case. Alan Turing (quoted by D. MacKenzie in Risk and Reason) N. Shankar Model checking
Scalable computational architecture for integrating biological pathway models
Shiva, V. A
2007-01-01
A grand challenge of systems biology is to model the cell. The cell is an integrated network of cellular functions. Each cellular function, such as immune response, cell division, metabolism or apoptosis, is defined by an ...
A Computational Model of How the Basal Ganglia Produce Sequences
Berns, Gregory S.
closely on known anatomy and physiology. First, we assume that the thalamic targets, which relay ascend the external globus pallidus (GPe) and the subthalamic nucleus (STN). As a test of the model, the system
Computational Modeling of Combined Steam Pyrolysis and Hydrogasification of Ethanol
Singh, S; Park, C S; Norbeck, J N
2005-01-01
Approximate modelling of coal pyrolysis. Fuel, 78(7), 825-9. Gonzalez, JF. (2003). Pyrolysis of cherry stones: energyof analytical and applied pyrolysis, 67(1), 165-190. 10.
Computer support to run models of the atmosphere. Final report
Fung, I.
1996-08-30
This research is focused on a better quantification of the variations in CO{sub 2} exchanges between the atmosphere and biosphere and the factors responsible for these exchangers. The principal approach is to infer the variations in the exchanges from variations in the atmospheric CO{sub 2} distribution. The principal tool involves using a global three-dimensional tracer transport model to advect and convect CO{sub 2} in the atmosphere. The tracer model the authors used was developed at the Goddard institute for Space Studies (GISS) and is derived from the GISS atmospheric general circulation model. A special run of the GCM is made to save high-frequency winds and mixing statistics for the tracer model.
Modeling-Computer Simulations At White Mountains Area (Goff ...
useful DOE-funding Unknown Notes Review and identification of 24 potential sites for EGS development across the U.S., as well as modeling of the representative geologic systems...
Modeling-Computer Simulations At Fenton Hill HDR Geothermal Area...
useful DOE-funding Unknown Notes Review and identification of 24 potential sites for EGS development across the U.S., as well as modeling of the representative geologic systems...
Continuum-based Multiscale Computational Damage Modeling of Cementitous Composites
Kim, Sun-Myung
2011-08-08
, aggregates, and interfacial transition zone (ITZ) and interaction among components at meso-scale, and the interaction between reinforcements, such as fiber and carbon nanotubes (CNTs) and mortar matrix or the ITZ at nano scale in order to predict more... of Advisory Committee: Dr. Rashid K. Abu Al-Rub Based on continuum damage mechanics (CDM), an isotropic and anisotropic damage model coupled with a novel plasticity model for plain concrete is proposed in this research. Two different damage evolution laws...
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
AFDC Printable Version Share this resource Send a link to EERE: Alternative Fuels Data Center Home Page to someone by E-mail Share EERE: Alternative Fuels Data Center Home Page on Facebook Tweet about EERE: Alternative Fuels Data Center Home Page on Twitter Bookmark EERE: Alternative Fuels Data Center Homesum_a_epg0_fpd_mmcf_m.xls" ,"Available from WebQuantityBonneville Power Administration would like submit the following commentsMethodsCompositional6 Computational CSEEnergy
Slepoy, Alexander; Mitchell, Scott A.; Backus, George A.; McNamara, Laura A.; Trucano, Timothy Guy
2008-09-01
Sandia National Laboratories is investing in projects that aim to develop computational modeling and simulation applications that explore human cognitive and social phenomena. While some of these modeling and simulation projects are explicitly research oriented, others are intended to support or provide insight for people involved in high consequence decision-making. This raises the issue of how to evaluate computational modeling and simulation applications in both research and applied settings where human behavior is the focus of the model: when is a simulation 'good enough' for the goals its designers want to achieve? In this report, we discuss two years' worth of review and assessment of the ASC program's approach to computational model verification and validation, uncertainty quantification, and decision making. We present a framework that extends the principles of the ASC approach into the area of computational social and cognitive modeling and simulation. In doing so, we argue that the potential for evaluation is a function of how the modeling and simulation software will be used in a particular setting. In making this argument, we move from strict, engineering and physics oriented approaches to V&V to a broader project of model evaluation, which asserts that the systematic, rigorous, and transparent accumulation of evidence about a model's performance under conditions of uncertainty is a reasonable and necessary goal for model evaluation, regardless of discipline. How to achieve the accumulation of evidence in areas outside physics and engineering is a significant research challenge, but one that requires addressing as modeling and simulation tools move out of research laboratories and into the hands of decision makers. This report provides an assessment of our thinking on ASC Verification and Validation, and argues for further extending V&V research in the physical and engineering sciences toward a broader program of model evaluation in situations of high consequence decision-making.
Theoretical and computer models of detonation in solid explosives
Tarver, C.M.; Urtiew, P.A.
1997-10-01
Recent experimental and theoretical advances in understanding energy transfer and chemical kinetics have led to improved models of detonation waves in solid explosives. The Nonequilibrium Zeldovich - von Neumann - Doring (NEZND) model is supported by picosecond laser experiments and molecular dynamics simulations of the multiphonon up-pumping and internal vibrational energy redistribution (IVR) processes by which the unreacted explosive molecules are excited to the transition state(s) preceding reaction behind the leading shock front(s). High temperature, high density transition state theory calculates the induction times measured by laser interferometric techniques. Exothermic chain reactions form product gases in highly excited vibrational states, which have been demonstrated to rapidly equilibrate via supercollisions. Embedded gauge and Fabry-Perot techniques measure the rates of reaction product expansion as thermal and chemical equilibrium is approached. Detonation reaction zone lengths in carbon-rich condensed phase explosives depend on the relatively slow formation of solid graphite or diamond. The Ignition and Growth reactive flow model based on pressure dependent reaction rates and Jones-Wilkins-Lee (JWL) equations of state has reproduced this nanosecond time resolved experimental data and thus has yielded accurate average reaction zone descriptions in one-, two- and three- dimensional hydrodynamic code calculations. The next generation reactive flow model requires improved equations of state and temperature dependent chemical kinetics. Such a model is being developed for the ALE3D hydrodynamic code, in which heat transfer and Arrhenius kinetics are intimately linked to the hydrodynamics.
Computational modeling of thermal conductivity of single walled carbon nanotube polymer composites
Maruyama, Shigeo
was developed to study the thermal conductivity of single walled carbon nanotube (SWNT)-polymer composites1 Computational modeling of thermal conductivity of single walled carbon nanotube polymer resistance on effective conductivity of composites were quantified. The present model is a useful tool
The Application of L-systems and Developmental Models to Computer Art,
McCormack, Jon
The Application of L-systems and Developmental Models to Computer Art, Animation, and Music Synthesis Jon McCormack, B.Sc.(Hons), Grad Dip. Art (Film & Television) A Thesis Submitted for the Degree.3.1 Generative Modelling Systems ................................................ 8 1.3.2 Critical and Art
Toward Cost-Sensitive Modeling for Intrusion Detection Computer Science Department
Toward Cost-Sensitive Modeling for Intrusion Detection Wenke Lee Computer Science Department North,ezk,weaselg@cs.columbia.edu Abstract Intrusion detection systems need to maximize security while minimizing costs. In this paper, we study the problem of building cost-sensitive intrusion detection models. We examine the major cost
STATISTICAL MODELING OF THE LUNG NODULES IN LOW DOSE COMPUTED TOMOGRAPHY SCANS OF THE CHEST
Louisville, University of
STATISTICAL MODELING OF THE LUNG NODULES IN LOW DOSE COMPUTED TOMOGRAPHY SCANS OF THE CHEST Amal in automatic detection of the lung nodules and is compared with respect to parametric nodule models in terms appearing in low dose CT (LDCT) scans of the human chest. Four types of common lung nodules are analyzed
A two-layer granular landslide model for tsunami wave generation: Theory and computation
Kirby, James T.
A two-layer granular landslide model for tsunami wave generation: Theory and computation Gangfeng for granular landslide motion and tsunami wave generation. The landslide, either submarine or subaerial experiments on impulsive wave generation by subaerial granular landslides. Model results illustrate a complex
Grossmann, Ignacio E.
1 Computational Strategies for Large-Scale MILP Transshipment Models for Heat Exchanger Network Determining the minimum number of units is an important step in heat exchanger network synthesis (HENS Words heat exchanger network synthesis (HENS), transshipment model, mixed-integer linear programming
Stallard, Alisha M.
2011-08-08
. This prompted a reconstruction of the dose to the worker’s hands. The computer code MCNP was chosen to model the tasks that the worker performed to evaluate the potential nonuniform hand dose distribution. A model was constructed similar to the worker’s hands...
A computational model for predicting damage evolution in laminated composite plates
Phillips, Mark Lane
1999-01-01
computationally tenable is shown herein. Due to the complicated nature of the many cracks and their interactions, a multi-scale micro-meso-local-global methodology is employed in order to model damage modes. Interface degradation is first modeled analytically...
Modeling and Design of RF MEMS Structures Using Computationally Efficient Numerical Techniques
Tentzeris, Manos
Modeling and Design of RF MEMS Structures Using Computationally Efficient Numerical Techniques N. A Abstract The modeling of MEMS structures using MRTD is presented. Many complex RF structures have been communication systems efficiently and accurately. Specifically, micromachined structures such as MEMS
Tentzeris, Manos
Modeling and Optimization of RF-MEMS Reconfigurable Tuners with Computationally Efficient Time of Technology, Atlanta, GA 30332 2 Raytheon Company, Tucson AZ, 85734 Abstract -- Modern RF-MEMS device design methods in which the FDTD technique can be used to model a reconfigurable RF-MEMS tuner. A new method
An efficient computational model for macroscale simulations of moving contact lines
Boyer, Edmond
with CO2, for example). A major challenge in numerical simulations of moving contact linesAn efficient computational model for macroscale simulations of moving contact lines Y. Sui1 simulation of moving contact lines. The main purpose is to formulate and test a model wherein the macroscale
Coupling remote sensing with computational fluid dynamics modelling to estimate lake chlorophyll form 17 October 2000; accepted 1 June 2001 Abstract A remotely sensed image of Loch Leven, a shallow in the remotely sensed image. It is proposed that CFD modelling benefits the interpretation of remotely sensed
GWU Department of Mathematics Topics in Model Theory: Classical and Computable
Harizanov, Valentina S.
framework for the notions of language, meaning, and truth. A model, a concept used in all of sciences course will be, in some sense, self-contained. We will start by reviewing the fundamental conceptsÂ194 (survey chapter without proofs). (3) V. Harizanov, "Pure computable model theory," in the volume: Handbook
A Novel Computational Model for Tilting Pad Journal Bearings with Soft Pivot Stiffnesses
Tao, Yujiao 1988-
2012-12-10
A novel tilting pad journal bearing model including pivot flexibility as well as temporal fluid inertia effects on the thin film fluid flow aims to accurately predict the bearing forced performance. The predictive model also accounts for the thermal...
CASE FOR SUPPORT: Computational Modeling of Salience Sensitive Control
Heinke, Dietmar
in neural network modeling, machine learning, adaptive systems in general and self-organising systems] and verification of real-time systems [6]. A large amount of this research has been performed using the CADP verification environment, which is one of the most powerful tool suites available, boasting a spectrum
A quantitative model of computation Dan R. Ghica
Ghica, Dan
languages. We define a Hyland-Ong-style games framework called slot games, which consists of HO games be nevertheless difficult, if not impossible, to handle using known operational techniques. Categories and Subject]: Modeling techniques. General Terms: Languages, Performance, Theory, Verifi- cation Keywords: Game semantics
Computational models for the berry phase in semiconductor quantum dots
Prabhakar, S. Melnik, R. V. N.; Sebetci, A.
2014-10-06
By developing a new model and its finite element implementation, we analyze the Berry phase low-dimensional semiconductor nanostructures, focusing on quantum dots (QDs). In particular, we solve the Schrödinger equation and investigate the evolution of the spin dynamics during the adiabatic transport of the QDs in the 2D plane along circular trajectory. Based on this study, we reveal that the Berry phase is highly sensitive to the Rashba and Dresselhaus spin-orbit lengths.
Pruess, K.
2011-05-15
Storage of CO{sub 2} in saline aquifers is intended to be at supercritical pressure and temperature conditions, but CO{sub 2} leaking from a geologic storage reservoir and migrating toward the land surface (through faults, fractures, or improperly abandoned wells) would reach subcritical conditions at depths shallower than 500-750 m. At these and shallower depths, subcritical CO{sub 2} can form two-phase mixtures of liquid and gaseous CO{sub 2}, with significant latent heat effects during boiling and condensation. Additional strongly non-isothermal effects can arise from decompression of gas-like subcritical CO{sub 2}, the so-called Joule-Thomson effect. Integrated modeling of CO{sub 2} storage and leakage requires the ability to model non-isothermal flows of brine and CO{sub 2} at conditions that range from supercritical to subcritical, including three-phase flow of aqueous phase, and both liquid and gaseous CO{sub 2}. In this paper, we describe and demonstrate comprehensive simulation capabilities that can cope with all possible phase conditions in brine-CO{sub 2} systems. Our model formulation includes: (1) an accurate description of thermophysical properties of aqueous and CO{sub 2}-rich phases as functions of temperature, pressure, salinity and CO{sub 2} content, including the mutual dissolution of CO{sub 2} and H{sub 2}O; (2) transitions between super- and subcritical conditions, including phase change between liquid and gaseous CO{sub 2}; (3) one-, two-, and three-phase flow of brine-CO{sub 2} mixtures, including heat flow; (4) non-isothermal effects associated with phase change, mutual dissolution of CO{sub 2} and water, and (de-) compression effects; and (5) the effects of dissolved NaCl, and the possibility of precipitating solid halite, with associated porosity and permeability change. Applications to specific leakage scenarios demonstrate that the peculiar thermophysical properties of CO{sub 2} provide a potential for positive as well as negative feedbacks on leakage rates, with a combination of self-enhancing and self-limiting effects. Lower viscosity and density of CO{sub 2} as compared to aqueous fluids provides a potential for self-enhancing effects during leakage, while strong cooling effects from liquid CO{sub 2} boiling into gas, and from expansion of gas rising towards the land surface, act to self-limit discharges. Strong interference between fluid phases under three-phase conditions (aqueous - liquid CO{sub 2} - gaseous CO{sub 2}) also tends to reduce CO{sub 2} fluxes. Feedback on different space and time scales can induce non-monotonic behavior of CO{sub 2} flow rates.
Koniges, A; Eder, E; Liu, W; Barnard, J; Friedman, A; Logan, G; Fisher, A; Masers, N; Bertozzi, A
2011-11-04
The Neutralized Drift Compression Experiment II (NDCX II) is an induction accelerator planned for initial commissioning in 2012. The final design calls for a 3 MeV, Li+ ion beam, delivered in a bunch with characteristic pulse duration of 1 ns, and transverse dimension of order 1 mm. The NDCX II will be used in studies of material in the warm dense matter (WDM) regime, and ion beam/hydrodynamic coupling experiments relevant to heavy ion based inertial fusion energy. We discuss recent efforts to adapt the 3D ALE-AMR code to model WDM experiments on NDCX II. The code, which combines Arbitrary Lagrangian Eulerian (ALE) hydrodynamics with Adaptive Mesh Refinement (AMR), has physics models that include ion deposition, radiation hydrodynamics, thermal diffusion, anisotropic material strength with material time history, and advanced models for fragmentation. Experiments at NDCX-II will explore the process of bubble and droplet formation (two-phase expansion) of superheated metal solids using ion beams. Experiments at higher temperatures will explore equation of state and heavy ion fusion beam-to-target energy coupling efficiency. Ion beams allow precise control of local beam energy deposition providing uniform volumetric heating on a timescale shorter than that of hydrodynamic expansion. The ALE-AMR code does not have any export control restrictions and is currently running at the National Energy Research Scientific Computing Center (NERSC) at LBNL and has been shown to scale well to thousands of CPUs. New surface tension models that are being implemented and applied to WDM experiments. Some of the approaches use a diffuse interface surface tension model that is based on the advective Cahn-Hilliard equations, which allows for droplet breakup in divergent velocity fields without the need for imposed perturbations. Other methods require seeding or other methods for droplet breakup. We also briefly discuss the effects of the move to exascale computing and related computational changes on general modeling codes in fusion energy.
Verification of a VRF Heat Pump Computer Model in EnergyPlus
Nigusse, Bereket; Raustad, Richard
2013-06-01
This paper provides verification results of the EnergyPlus variable refrigerant flow (VRF) heat pump computer model using manufacturer's performance data. The paper provides an overview of the VRF model, presents the verification methodology, and discusses the results. The verification provides quantitative comparison of full and part-load performance to manufacturer's data in cooling-only and heating-only modes of operation. The VRF heat pump computer model uses dual range bi-quadratic performance curves to represent capacity and Energy Input Ratio (EIR) as a function of indoor and outdoor air temperatures, and dual range quadratic performance curves as a function of part-load-ratio for modeling part-load performance. These performance curves are generated directly from manufacturer's published performance data. The verification compared the simulation output directly to manufacturer's performance data, and found that the dual range equation fit VRF heat pump computer model predicts the manufacturer's performance data very well over a wide range of indoor and outdoor temperatures and part-load conditions. The predicted capacity and electric power deviations are comparbale to equation-fit HVAC computer models commonly used for packaged and split unitary HVAC equipment.
Computational Human Performance Modeling For Alarm System Design
Jacques Hugo
2012-07-01
The introduction of new technologies like adaptive automation systems and advanced alarms processing and presentation techniques in nuclear power plants is already having an impact on the safety and effectiveness of plant operations and also the role of the control room operator. This impact is expected to escalate dramatically as more and more nuclear power utilities embark on upgrade projects in order to extend the lifetime of their plants. One of the most visible impacts in control rooms will be the need to replace aging alarm systems. Because most of these alarm systems use obsolete technologies, the methods, techniques and tools that were used to design the previous generation of alarm system designs are no longer effective and need to be updated. The same applies to the need to analyze and redefine operators’ alarm handling tasks. In the past, methods for analyzing human tasks and workload have relied on crude, paper-based methods that often lacked traceability. New approaches are needed to allow analysts to model and represent the new concepts of alarm operation and human-system interaction. State-of-the-art task simulation tools are now available that offer a cost-effective and efficient method for examining the effect of operator performance in different conditions and operational scenarios. A discrete event simulation system was used by human factors researchers at the Idaho National Laboratory to develop a generic alarm handling model to examine the effect of operator performance with simulated modern alarm system. It allowed analysts to evaluate alarm generation patterns as well as critical task times and human workload predicted by the system.
College of Science Bachelor of Science in Computational Modeling and Data Analytics
Zallen, Richard
(Pre: MATH 1225) (4)( ) MATH 2114 Introduction to Linear Algebra (3)( ) MATH 2204 Intro Multivariable)( ) MATH 4144* Linear Algebra (Pre: MATH 3144) (3)( ) MATH 4425* Fourier Series PDE (Pre: MATH 2214, 2204 for computing "in-major GPA." Also included: additional STAT, MATH, CS courses at 3000 or 4000-level
The Nuclear Energy Advanced Modeling and Simulation Enabling Computational Technologies FY09 Report
Diachin, L F; Garaizar, F X; Henson, V E; Pope, G
2009-10-12
In this document we report on the status of the Nuclear Energy Advanced Modeling and Simulation (NEAMS) Enabling Computational Technologies (ECT) effort. In particular, we provide the context for ECT In the broader NEAMS program and describe the three pillars of the ECT effort, namely, (1) tools and libraries, (2) software quality assurance, and (3) computational facility (computers, storage, etc) needs. We report on our FY09 deliverables to determine the needs of the integrated performance and safety codes (IPSCs) in these three areas and lay out the general plan for software quality assurance to meet the requirements of DOE and the DOE Advanced Fuel Cycle Initiative (AFCI). We conclude with a brief description of our interactions with the Idaho National Laboratory computer center to determine what is needed to expand their role as a NEAMS user facility.
Medical Nuclear Supply Chain Design: A Tractable Network Model and Computational
Nagurney, Anna
importance of considering waste management. Anna Nagurney and Ladimer S. Nagurney Medical Nuclear SupplyMedical Nuclear Supply Chain Design: A Tractable Network Model and Computational Approach Anna Nagurney1 and Ladimer S. Nagurney2 1John F. Smith Memorial Professor - Isenberg School of Management
Medical Nuclear Supply Chain Design: A Tractable Network Model and Computational Approach
Nagurney, Anna
operational cost minimization, and the minimization of cost associated with nuclear waste discarding, coupled, the underlying physics of radioactive decay, and the inclusion of waste management. We focus on Molybdenum-99 dueMedical Nuclear Supply Chain Design: A Tractable Network Model and Computational Approach Anna
Many Task Computing for Modeling the Fate of Oil Discharged from the Deep Water Horizon Well
Many Task Computing for Modeling the Fate of Oil Discharged from the Deep Water Horizon Well@jhu.edu Abstract--The Deep Water Horizon well blowout on April 20th 2010 discharged between 40,000 - 1.2 million@med.miami.edu O. M. Knio Dept of Mechanical Engineering Johns Hopkins University Baltimore, MD knio
Computational Modeling of Neural Plasticity for Self-Organization of Neural Networks
Jin, Yaochu
Computational Modeling of Neural Plasticity for Self-Organization of Neural Networks Joseph Chrol on the learning per- formance of neural networks for accomplishing machine learning tasks such as classication, dynamics and learning per- formance of neural networks remains elusive. The purpose of this article
In-Vehicle Testing and Computer Modeling of Electric Vehicle Batteries
Wang, Chao-Yang
In-Vehicle Testing and Computer Modeling of Electric Vehicle Batteries B. Thomas, W.B. Gu, J.edu Abstract A combined simulation and testing approach has been developed to evaluate battery packs in real accelerates battery development cycle, and enables innovative battery design and optimization. Several
Breitling, Rainer
, Continuous Petri nets Computational modelling of intracellular biochemical networks has become a growth topic and analysis of such networks, as well as an increase in the quality and amount of experimentally determined such as bifurcations, robustness to interference, or oscillations are not obvious from the network topology
Computational Modeling of Electrolyte/Cathode Interfaces in Proton Exchange Membrane Fuel Cells
Bjørnstad, Ottar Nordal
Computational Modeling of Electrolyte/Cathode Interfaces in Proton Exchange Membrane Fuel Cells Dr Proton exchange membrane fuel cells (PEMFCs) are alternative energy conversion devices that efficiently. The fundamental relationship between operating conditions and device performance will help to optimize the device
A computational strategy for multiscale systems with applications to Lorenz 96 model
Van Den Eijnden, Eric
A computational strategy for multiscale systems with applications to Lorenz 96 model Ibrahim 2004 Available online Abstract Numerical schemes for systems with multiple spatio-temporal scales are investigated. The multiscale schemes use asymptotic results for this type of systems which guarantee
P&P: a Combined Push-Pull Model for Resource Monitoring in Cloud Computing Environment
Wang, Liqiang
P&P: a Combined Push-Pull Model for Resource Monitoring in Cloud Computing Environment He Huang, various platforms and software. Resource monitoring involves collecting information of system resources to facilitate decision making by other components in Cloud environ- ment. It is the foundation of many major
COMPUTATIONAL CHALLENGES IN THE NUMERICAL TREATMENT OF LARGE AIR POLLUTION MODELS
Ostromsky, Tzvetan
COMPUTATIONAL CHALLENGES IN THE NUMERICAL TREATMENT OF LARGE AIR POLLUTION MODELS I. DIMOV , K. GEORGIEVy, TZ. OSTROMSKY , R. J. VAN DER PASz, AND Z. ZLATEVx Abstract. The air pollution, and especially the reduction of the air pollution to some acceptable levels, is an important environmental problem, which
MATHEMATICAL PERGAMON Mathematical and Computer Modelling 35 (2002) 1371-1375
Gorban, Alexander N.
2002-01-01
Application to the Efficiency of Free Flow Turbines A. GORBAN' Institute of Computational Modeling, Russian obstacle is considered. Its application to estimating the efficiency of free flow turbines is discussed hydraulic turbines, i.e., the turbines that work without dams [l]. For this kind of turbine, the term
A Computational Model of Aging and Calcification in the Aortic Heart Valve
Mofrad, Mohammad R. K.
A Computational Model of Aging and Calcification in the Aortic Heart Valve Eli J. Weinberg1 of America Abstract The aortic heart valve undergoes geometric and mechanical changes over time. The cusps of a normal, healthy valve thicken and become less extensible over time. In the disease calcific aortic
Computational Modeling and the Experimental Plasma Research Program A White Paper Submitted of the fusion energy program. The experimental plasma research (EPR) program is well positioned to make major in fusion development and promote scientific discovery. Experimental plasma research projects explore
PhD Position Available: integrative biomechanics, computational modeling, nonlinear dynamics
Clewley, Robert
PhD Position Available: integrative biomechanics, computational modeling, nonlinear dynamics and mathematical analysis of biomechanical and neural control systems. We are looking for an excellent and highly.edu/~biodhe/#Research). These are being used to study the Crayfish swim escape mechanism as a case study in integrative biomechanical
Computational framework for modeling the dynamic evolution of large-scale multi-agent organizations
Lazar, Alina
disciplines as well. This is because the emergence of such a social structure can have a profound impact 48202 ABSTRACT The process by which complex social entities such as the state emerged from lower level on the societies' physical and social environment. However, the task of developing realistic computational models
Broader source: Energy.gov [DOE]
The objective of this Funding Opportunity Announcement (FOA) is to leverage scientific advancements in mathematics and computation for application to power system models and software tools, with the long-term goal of enabling real-time protection and control based on wide-area sensor measurements.
Buyya, Rajkumar
infrastructure (hardware, software, services) for different application and service models under varying load problem to tackle. To simplify this process, in this paper we propose CloudSim: a new generalized Cloud computing infrastructures and management services. The simulation framework has the following
Margaliot, Michael
1 Entrainment to Periodic Initiation and Transition Rates in a Computational Model for Gene, the biological system must entrain or phase-lock to the periodic excitation. Entrainment is also important in synthetic biology. For example, connecting several artificial biological systems that entrain to a common
Sontag, Eduardo
Entrainment to Periodic Initiation and Transition Rates in a Computational Model for Gene to the solar day. In the terminology of systems theory, the biological system must entrain or phase-lock to the periodic excitation. Entrainment is also important in synthetic biology. For example, connecting several
Baker, Jack W.
Conditional Spectrum Computation Incorporating Multiple Causal Earthquakes and Ground-Motion Prediction Models by Ting Lin, Stephen C. Harmsen, Jack W. Baker, and Nicolas Luco Abstract The conditional uncertainties in all earthquake scenarios and resulting ground motions, as well as the epistemic uncertainties
International Association for Cryptologic Research (IACR)
Modeling Computational Security in LongLived Systems # ## Ran Canetti 1,2 , Ling Cheung 2 , Dilsun Introduction Computational security in longlived systems: Security properties of cryptographic protocols computational power. This type of security degrades progressively over the lifetime of a protocol. However, some
Buyya, Rajkumar
CloudAnalyst: A CloudSim-based Visual Modeller for Analysing Cloud Computing Environments and Applications Bhathiya Wickremasinghe1 , Rodrigo N. Calheiros2 , and Rajkumar Buyya1 1 The Cloud Computing and Distributed Systems (CLOUDS) Laboratory Department of Computer Science and Software Engineering The University
Zhou, Shujia; Duffy, Daniel; Clune, Thomas; Suarez, Max; Williams, Samuel; Halem, Milton
2009-01-10
The call for ever-increasing model resolutions and physical processes in climate and weather models demands a continual increase in computing power. The IBM Cell processor's order-of-magnitude peak performance increase over conventional processors makes it very attractive to fulfill this requirement. However, the Cell's characteristics, 256KB local memory per SPE and the new low-level communication mechanism, make it very challenging to port an application. As a trial, we selected the solar radiation component of the NASA GEOS-5 climate model, which: (1) is representative of column physics components (half the total computational time), (2) has an extremely high computational intensity: the ratio of computational load to main memory transfers, and (3) exhibits embarrassingly parallel column computations. In this paper, we converted the baseline code (single-precision Fortran) to C and ported it to an IBM BladeCenter QS20. For performance, we manually SIMDize four independent columns and include several unrolling optimizations. Our results show that when compared with the baseline implementation running on one core of Intel's Xeon Woodcrest, Dempsey, and Itanium2, the Cell is approximately 8.8x, 11.6x, and 12.8x faster, respectively. Our preliminary analysis shows that the Cell can also accelerate the dynamics component (~;;25percent total computational time). We believe these dramatic performance improvements make the Cell processor very competitive as an accelerator.
Technical Review of the CENWP Computational Fluid Dynamics Model of the John Day Dam Forebay
Rakowski, Cynthia L.; Serkowski, John A.; Richmond, Marshall C.
2010-12-01
The US Army Corps of Engineers Portland District (CENWP) has developed a computational fluid dynamics (CFD) model of the John Day forebay on the Columbia River to aid in the development and design of alternatives to improve juvenile salmon passage at the John Day Project. At the request of CENWP, Pacific Northwest National Laboratory (PNNL) Hydrology Group has conducted a technical review of CENWP's CFD model run in CFD solver software, STAR-CD. PNNL has extensive experience developing and applying 3D CFD models run in STAR-CD for Columbia River hydroelectric projects. The John Day forebay model developed by CENWP is adequately configured and validated. The model is ready for use simulating forebay hydraulics for structural and operational alternatives. The approach and method are sound, however CENWP has identified some improvements that need to be made for future models and for modifications to this existing model.
bioenergetics models were expanded to the population- level and dynamically coupled to the lower trophic levels (LTL) of the NEMURO model. The individual fish bioenergetics model and the one-way coupling to NEMURO (i.e. NEMURO is run first and output is used to force the fish bioenergetics model) are described
Marinos, Alexandros
2009-01-01
Cloud Computing is rising fast, with its data centres growing at an unprecedented rate. However, this has come with concerns over privacy, efficiency at the expense of resilience, and environmental sustainability, because of the dependence on Cloud vendors such as Google, Amazon and Microsoft. Our response is an alternative model for the Cloud conceptualisation, providing a paradigm for Clouds in the community, utilising networked personal computers for liberation from the centralised vendor model. Community Cloud Computing (C3) offers an alternative architecture, created by combing the Cloud with paradigms from Grid Computing, principles from Digital Ecosystems, and sustainability from Green Computing, while remaining true to the original vision of the Internet. It is more technically challenging than Cloud Computing, having to deal with distributed computing issues, including heterogeneous nodes, varying quality of service, and additional security constraints. However, these are not insurmountable challenge...
Spycher, N.; Oldenburg, C.M.
2014-01-01
This study uses modeling and simulation approaches to investigate the impacts on injectivity of trace amounts of mercury (Hg) in a carbon dioxide (CO{sub 2}) stream injected for geologic carbon sequestration in a sandstone reservoir at ~2.5 km depth. At the range of Hg concentrations expected (7-190 ppbV, or ~ 0.06-1.6 mg/std.m{sup 3}CO{sub 2}), the total volumetric plugging that could occur due to complete condensation of Hg, or due to complete precipitation of Hg as cinnabar, results in a very small porosity change. In addition, Hg concentration much higher than the concentrations considered here would be required for Hg condensation to even occur. Concentration of aqueous Hg by water evaporation into CO{sub 2} is also unlikely because the higher volatility of Hg relative to H{sub 2}O at reservoir conditions prevents the Hg concentration from increasing in groundwater as dry CO{sub 2} sweeps through, volatilizing both H{sub 2}O and Hg. Using a model-derived aqueous solution to represent the formation water, batch reactive geochemical modeling show that the reaction of the formation water with the CO{sub 2}-Hg mixture causes the pH to drop to about 4.7 and then become buffered near 5.2 upon reaction with the sediments, with a negligible net volume change from mineral dissolution and precipitation. Cinnabar (HgS(s)) is found to be thermodynamically stable as soon as the Hg-bearing CO{sub 2} reacts with the formation water which contains small amounts of dissolved sulfide. Liquid mercury (Hg(l)) is not found to be thermodynamically stable at any point during the simulation. Two-dimensional radial reactive transport simulations of CO{sub 2} injection at a rate of 14.8 kg/s into a 400 m-thick formation at isothermal conditions of 106°C and average pressure near 215 bar, with varying amounts of Hg and H{sub 2}S trace gases, show generally that porosity changes only by about ±0.05% (absolute, i.e., new porosity = initial porosity ±0.0005) with Hg predicted to readily precipitate from the CO{sub 2} as cinnabar in a zone mostly matching the single-phase CO{sub 2} plume. The precipitation of minerals other than cinnabar, however, dominates the evolution of porosity. Main reactions include the replacement of primarily Fe-chlorite by siderite, of calcite by dolomite, and of K-feldspar by muscovite. Chalcedony is also predicted to precipitate from the dissolution of feldspars and quartz. Although the range of predicted porosity change is quite small, the amount of dissolution and precipitation predicted for these individual minerals is not negligible. These reactive transport simulations assume that Hg gas behaves ideally. To examine effects of non-ideality on these simulations, approximate calculations of the fugacity coefficient of Hg in CO{sub 2} were made. Results suggest that Hg condensation could be significantly overestimated when assuming ideal gas behavior, making our simulation results conservative with respect to impacts on injectivity. The effect of pressure on Henry’s constant for Hg is estimated to yield Hg solubilities about 10% lower than when this effect is not considered, a change that is considered too small to affect the conclusions of this report. Although all results in this study are based on relatively mature data and modeling approaches, in the absence of experimental data and more detailed site-specific information, it is not possible to fully validate the results and conclusions.
Computer modeling of electromagnetic edge containment in twin-roll casting
Chang, F.C.; Turner, L.R.; Hull, J.R.; Wang, Y.H.; Blazek, K.E.
1998-07-01
This paper presents modeling studies of magnetohydrodynamics (MHD) analysis in twin-roll casting. Argonne National Laboratory (ANL) and Inland Steel Company have worked together to develop a 3-D computer model that can predict eddy currents, fluid flows, and liquid metal containment for an electromagnetic (EM) edge containment device. This mathematical model can greatly shorten casting research on the use of EM fields for liquid metal containment and control. It can also optimize the existing casting processes and minimize expensive, time-consuming full-scale testing. The model was verified by comparing predictions with experimental results of liquid-metal containment and fluid flow in EM edge dams designed at Inland Steel for twin-roll casting. Numerical simulation was performed by coupling a three-dimensional (3-D) finite-element EM code (ELEKTRA) and a 3-D finite-difference fluids code (CaPS-EM) to solve Maxwell`s equations, Ohm`s law, Navier-Stokes equations, and transport equations of turbulence flow in a casting process that uses EM fields. ELEKTRA is able to predict the eddy-current distribution and electromagnetic forces in complex geometry. CaPS-EM is capable of modeling fluid flows with free-surfaces and dynamic rollers. The computed 3-D magnetic fields and induced eddy currents in ELEKTRA are used as input to flow-field computations in CaPS-EM. Results of the numerical simulation compared well with measurements obtained from both static and dynamic tests.
High-Performance Computing Modeling Advances Accelerator Science for High-Energy Physics
DOE Public Access Gateway for Energy & Science Beta (PAGES Beta)
Amundson, James; Macridin, Alexandru; Spentzouris, Panagiotis
2014-11-01
The development and optimization of particle accelerators are essential for advancing our understanding of the properties of matter, energy, space and time. Particle accelerators are complex devices whose behavior involves many physical effects on multiple scales. Therefore, advanced computational tools utilizing high-performance computing (HPC) are essential for accurately modeling them. In the past decade, the DOE SciDAC program has produced such accelerator-modeling tools, which have beem employed to tackle some of the most difficult accelerator science problems. In this article we discuss the Synergia beam-dynamics framework and its applications to high-intensity particle accelerator physics. Synergia is an accelerator simulation packagemore »capable of handling the entire spectrum of beam dynamics simulations. We present the design principles, key physical and numerical models in Synergia and its performance on HPC platforms. Finally, we present the results of Synergia applications for the Fermilab proton source upgrade, known as the Proton Improvement Plan (PIP).« less
High-Performance Computing Modeling Advances Accelerator Science for High-Energy Physics
Amundson, James; Macridin, Alexandru; Spentzouris, Panagiotis
2014-11-01
The development and optimization of particle accelerators are essential for advancing our understanding of the properties of matter, energy, space and time. Particle accelerators are complex devices whose behavior involves many physical effects on multiple scales. Therefore, advanced computational tools utilizing high-performance computing (HPC) are essential for accurately modeling them. In the past decade, the DOE SciDAC program has produced such accelerator-modeling tools, which have beem employed to tackle some of the most difficult accelerator science problems. In this article we discuss the Synergia beam-dynamics framework and its applications to high-intensity particle accelerator physics. Synergia is an accelerator simulation package capable of handling the entire spectrum of beam dynamics simulations. We present the design principles, key physical and numerical models in Synergia and its performance on HPC platforms. Finally, we present the results of Synergia applications for the Fermilab proton source upgrade, known as the Proton Improvement Plan (PIP).
Gabriel Álvarez; Luis Martínez Alonso; Elena Medina
2011-01-14
We present a method to compute the genus expansion of the free energy of Hermitian matrix models from the large N expansion of the recurrence coefficients of the associated family of orthogonal polynomials. The method is based on the Bleher-Its deformation of the model, on its associated integral representation of the free energy, and on a method for solving the string equation which uses the resolvent of the Lax operator of the underlying Toda hierarchy. As a byproduct we obtain an efficient algorithm to compute generating functions for the enumeration of labeled k-maps which does not require the explicit expressions of the coefficients of the topological expansion. Finally we discuss the regularization of singular one-cut models within this approach.
Bayesian model-based approaches with MCMC computation to some bioinformatics problems
Bae, Kyounghwa
2005-08-29
Bioinformatics applications can address the transfer of information at several stages of the central dogma of molecular biology, including transcription and translation. This dissertation focuses on using Bayesian models ...
Data-Driven Optimization for Modeling in Computer Graphics and Vision
Yu, Lap Fai
2013-01-01
optimization problem, including: Explicit Control: Users can have explicit control over the modeling process,process while having close, interactive control. For example, in the Clutterbrush project, an optimization-
Carbajo, Juan; Jeong, Hae-Yong; Wigeland, Roald; Corradini, Michael; Schmidt, Rodney Cannon; Thomas, Justin; Wei, Tom; Sofu, Tanju; Ludewig, Hans; Tobita, Yoshiharu; Ohshima, Hiroyuki; Serre, Frederic
2011-06-01
This report summarizes the results of an expert-opinion elicitation activity designed to qualitatively assess the status and capabilities of currently available computer codes and models for accident analysis and reactor safety calculations of advanced sodium fast reactors, and identify important gaps. The twelve-member panel consisted of representatives from five U.S. National Laboratories (SNL, ANL, INL, ORNL, and BNL), the University of Wisconsin, the KAERI, the JAEA, and the CEA. The major portion of this elicitation activity occurred during a two-day meeting held on Aug. 10-11, 2010 at Argonne National Laboratory. There were two primary objectives of this work: (1) Identify computer codes currently available for SFR accident analysis and reactor safety calculations; and (2) Assess the status and capability of current US computer codes to adequately model the required accident scenarios and associated phenomena, and identify important gaps. During the review, panel members identified over 60 computer codes that are currently available in the international community to perform different aspects of SFR safety analysis for various event scenarios and accident categories. A brief description of each of these codes together with references (when available) is provided. An adaptation of the Predictive Capability Maturity Model (PCMM) for computational modeling and simulation is described for use in this work. The panel's assessment of the available US codes is presented in the form of nine tables, organized into groups of three for each of three risk categories considered: anticipated operational occurrences (AOOs), design basis accidents (DBA), and beyond design basis accidents (BDBA). A set of summary conclusions are drawn from the results obtained. At the highest level, the panel judged that current US code capabilities are adequate for licensing given reasonable margins, but expressed concern that US code development activities had stagnated and that the experienced user-base and the experimental validation base was decaying away quickly.
Victoria, University of
On the Use of Computational Models for Wave Climate Assessment in Support of the Wave Energy On the Use of Computational Models for Wave Climate Assessment in Support of the Wave Energy Industry Effective, economic extraction of ocean wave energy requires an intimate under- standing of the ocean wave
Pump apparatus including deconsolidator
Sonwane, Chandrashekhar; Saunders, Timothy; Fitzsimmons, Mark Andrew
2014-10-07
A pump apparatus includes a particulate pump that defines a passage that extends from an inlet to an outlet. A duct is in flow communication with the outlet. The duct includes a deconsolidator configured to fragment particle agglomerates received from the passage.
DualTrust: A Trust Management Model for Swarm-Based Autonomic Computing Systems
Maiden, Wendy M.
2010-05-01
Trust management techniques must be adapted to the unique needs of the application architectures and problem domains to which they are applied. For autonomic computing systems that utilize mobile agents and ant colony algorithms for their sensor layer, certain characteristics of the mobile agent ant swarm -- their lightweight, ephemeral nature and indirect communication -- make this adaptation especially challenging. This thesis looks at the trust issues and opportunities in swarm-based autonomic computing systems and finds that by monitoring the trustworthiness of the autonomic managers rather than the swarming sensors, the trust management problem becomes much more scalable and still serves to protect the swarm. After analyzing the applicability of trust management research as it has been applied to architectures with similar characteristics, this thesis specifies the required characteristics for trust management mechanisms used to monitor the trustworthiness of entities in a swarm-based autonomic computing system and describes a trust model that meets these requirements.
Experimental validation of a kilovoltage x-ray source model for computing imaging dose
Poirier, Yannick, E-mail: yannick.poirier@cancercare.mb.ca [CancerCare Manitoba, 675 McDermot Ave, Winnipeg, Manitoba R3E 0V9 (Canada)] [CancerCare Manitoba, 675 McDermot Ave, Winnipeg, Manitoba R3E 0V9 (Canada); Kouznetsov, Alexei; Koger, Brandon [Department of Physics and Astronomy, University of Calgary, Calgary, Alberta T2N 1N4 (Canada)] [Department of Physics and Astronomy, University of Calgary, Calgary, Alberta T2N 1N4 (Canada); Tambasco, Mauro, E-mail: mtambasco@mail.sdsu.edu [Department of Physics, San Diego State University, San Diego, California 92182-1233 and Department of Physics and Astronomy and Department of Oncology, University of Calgary, Calgary, Alberta T2N 1N4 (Canada)] [Department of Physics, San Diego State University, San Diego, California 92182-1233 and Department of Physics and Astronomy and Department of Oncology, University of Calgary, Calgary, Alberta T2N 1N4 (Canada)
2014-04-15
Purpose: To introduce and validate a kilovoltage (kV) x-ray source model and characterization method to compute absorbed dose accrued from kV x-rays. Methods: The authors propose a simplified virtual point source model and characterization method for a kV x-ray source. The source is modeled by: (1) characterizing the spatial spectral and fluence distributions of the photons at a plane at the isocenter, and (2) creating a virtual point source from which photons are generated to yield the derived spatial spectral and fluence distribution at isocenter of an imaging system. The spatial photon distribution is determined by in-air relative dose measurements along the transverse (x) and radial (y) directions. The spectrum is characterized using transverse axis half-value layer measurements and the nominal peak potential (kVp). This source modeling approach is used to characterize a Varian{sup ®} on-board-imager (OBI{sup ®}) for four default cone-beam CT beam qualities: beams using a half bowtie filter (HBT) with 110 and 125 kVp, and a full bowtie filter (FBT) with 100 and 125 kVp. The source model and characterization method was validated by comparing dose computed by the authors’ inhouse software (kVDoseCalc) to relative dose measurements in a homogeneous and a heterogeneous block phantom comprised of tissue, bone, and lung-equivalent materials. Results: The characterized beam qualities and spatial photon distributions are comparable to reported values in the literature. Agreement between computed and measured percent depth-dose curves is ?2% in the homogeneous block phantom and ?2.5% in the heterogeneous block phantom. Transverse axis profiles taken at depths of 2 and 6 cm in the homogeneous block phantom show an agreement within 4%. All transverse axis dose profiles in water, in bone, and lung-equivalent materials for beams using a HBT, have an agreement within 5%. Measured profiles of FBT beams in bone and lung-equivalent materials were higher than their computed counterparts resulting in an agreement within 2.5%, 5%, and 8% within solid water, bone, and lung, respectively. Conclusions: The proposed virtual point source model and characterization method can be used to compute absorbed dose in both the homogeneous and heterogeneous block phantoms within of 2%–8% of measured values, depending on the phantom and the beam quality. The authors’ results also provide experimental validation for their kV dose computation software, kVDoseCalc.
Use of model calibration to achieve high accuracy in analysis of computer networks
Frogner, Bjorn; Guarro, Sergio; Scharf, Guy
2004-05-11
A system and method are provided for creating a network performance prediction model, and calibrating the prediction model, through application of network load statistical analyses. The method includes characterizing the measured load on the network, which may include background load data obtained over time, and may further include directed load data representative of a transaction-level event. Probabilistic representations of load data are derived to characterize the statistical persistence of the network performance variability and to determine delays throughout the network. The probabilistic representations are applied to the network performance prediction model to adapt the model for accurate prediction of network performance. Certain embodiments of the method and system may be used for analysis of the performance of a distributed application characterized as data packet streams.
Parametric Studies and Optimization of Eddy Current Techniques through Computer Modeling
Todorov, E. I. [EWI, Engineering and NDE, 1250 Arthur E. Adams Dr., Columbus, OH 43221-3585 (United States)
2007-03-21
The paper demonstrates the use of computer models for parametric studies and optimization of surface and subsurface eddy current techniques. The study with high-frequency probe investigates the effect of eddy current frequency and probe shape on the detectability of flaws in the steel substrate. The low-frequency sliding probe study addresses the effect of conductivity between the fastener and the hole, frequency and coil separation distance on detectability of flaws in subsurface layers.
NREL Computer Models Integrate Wind Turbines with Floating Platforms (Fact Sheet)
Not Available
2011-07-01
Far off the shores of energy-hungry coastal cities, powerful winds blow over the open ocean, where the water is too deep for today's seabed-mounted offshore wind turbines. For the United States to tap into these vast offshore wind energy resources, wind turbines must be mounted on floating platforms to be cost effective. Researchers at the National Renewable Energy Laboratory (NREL) are supporting that development with computer models that allow detailed analyses of such floating wind turbines.
An Additive Bivariate Hierarchical Model for Functional Data and Related Computations
Redd, Andrew Middleton
2011-10-21
the penalties 11 are given is computationally intensive. In addition the space for nding the penalty parameters is four dimensional, two for each predictor variable, one each for the mean and principal component functions. Optimizing over a four dimensional... ed by the structure of the formula used. I implement two new formula operators that only work with the pfda package; %&% bind together variables, on the left side of the formula indicates the paired model, on the right an additive variable...
Path2Models: large-scale generation of computational models from biochemical pathway maps
2013-01-01
C: Petri net modelling of biological networks. Brief Bioinfonetworks [14-16] to discrete algebra [17] and dif- ferential equations [18], Petri
Broader source: Energy.gov [DOE]
Presentation given by NREL at 2014 DOE Hydrogen and Fuel Cells Program and Vehicle Technologies Office Annual Merit Review and Peer Evaluation Meeting about significant enhancement of computational...
Chang, F.C.; Hull, J.R.; Wang, Y.H.; Blazek, K.E.
1996-02-01
A computer model was developed to predict eddy currents and fluid flows in molten steel. The model was verified by comparing predictions with experimental results of liquid-metal containment and fluid flow in electromagnetic (EM) edge dams (EMDs) designed at Inland Steel for twin-roll casting. The model can optimize the EMD design so it is suitable for application, and minimize expensive, time-consuming full-scale testing. Numerical simulation was performed by coupling a three-dimensional (3-D) finite-element EM code (ELEKTRA) and a 3-D finite-difference fluids code (CaPS-EM) to solve heat transfer, fluid flow, and turbulence transport in a casting process that involves EM fields. ELEKTRA is able to predict the eddy- current distribution and the electromagnetic forces in complex geometries. CaPS-EM is capable of modeling fluid flows with free surfaces. Results of the numerical simulation compared well with measurements obtained from a static test.
Computational fluid dynamics modeling of coal gasification in a pressurized spout-fluid bed
Zhongyi Deng; Rui Xiao; Baosheng Jin; He Huang; Laihong Shen; Qilei Song; Qianjun Li [Southeast University, Nanjing (China). Key Laboratory of Clean Coal Power Generation and Combustion Technology of Ministry of Education
2008-05-15
Computational fluid dynamics (CFD) modeling, which has recently proven to be an effective means of analysis and optimization of energy-conversion processes, has been extended to coal gasification in this paper. A 3D mathematical model has been developed to simulate the coal gasification process in a pressurized spout-fluid bed. This CFD model is composed of gas-solid hydrodynamics, coal pyrolysis, char gasification, and gas phase reaction submodels. The rates of heterogeneous reactions are determined by combining Arrhenius rate and diffusion rate. The homogeneous reactions of gas phase can be treated as secondary reactions. A comparison of the calculated and experimental data shows that most gasification performance parameters can be predicted accurately. This good agreement indicates that CFD modeling can be used for complex fluidized beds coal gasification processes. 37 refs., 7 figs., 5 tabs.
) Cocurricular Activity (1) Total 16 Total 16 Total 15-16 Total 16 3rd Year 4th Year FALL SPRING FALL SPRING CS Year 2nd Year FALL SPRING FALL SPRING CS 1000 1 CS 1141 2 CS 2311 3 CS 3141 3 CS 1131 5 CS 2321 3 MA
) Cocurricular Activity (1) Total 14 Total 17 Total 14-15 Total 16 3rd Year 4th Year FALL SPRING FALL SPRING CS Year 2nd Year FALL SPRING FALL SPRING CS 1000 1 CS 1122 3 CS 1141 2 CS 2311 3 CS 1121 3 MA 2160 4 CS
Activity (1) Total 17 Total 15 Total 14-15 Total 16-17 3rd Year 4th Year FALL SPRING FALL SPRING CS 3311 3 your academic dept. Includes 3 units of co-curricular activities. Effective Fall 2010 1st Year 2nd Year
Rapidly re-computable EEG (electroencephalography) forward models for realistic head shapes
Ermer, J. J.; Mosher, J. C.; Baillet, S.; Leahy, R. M.
2001-01-01
Solution of the EEG source localization (inverse) problem utilizing model-based methods typically requires a significant number of forward model evaluations. For subspace based inverse methods like MUSIC [6], the total number of forward model evaluations can often approach an order of 10{sup 3} or 10{sup 4}. Techniques based on least-squares minimization may require significantly more evaluations. The observed set of measurements over an M-sensor array is often expressed as a linear forward spatio-temporal model of the form: F = GQ + N (1) where the observed forward field F (M-sensors x N-time samples) can be expressed in terms of the forward model G, a set of dipole moment(s) Q (3xP-dipoles x N-time samples) and additive noise N. Because of their simplicity, ease of computation, and relatively good accuracy, multi-layer spherical models [7] (or fast approximations described in [1], [7]) have traditionally been the 'forward model of choice' for approximating the human head. However, approximation of the human head via a spherical model does have several key drawbacks. By its very shape, the use of a spherical model distorts the true distribution of passive currents in the skull cavity. Spherical models also require that the sensor positions be projected onto the fitted sphere (Fig. 1), resulting in a distortion of the true sensor-dipole spatial geometry (and ultimately the computed surface potential). The use of a single 'best-fitted' sphere has the added drawback of incomplete coverage of the inner skull region, often ignoring areas such as the frontal cortex. In practice, this problem is typically countered by fitting additional sphere(s) to those region(s) not covered by the primary sphere. The use of these additional spheres results in added complication to the forward model. Using high-resolution spatial information obtained via X-ray CT or MR imaging, a realistic head model can be formed by tessellating the head into a set of contiguous regions (typically the scalp, outer skull, and inner skull surfaces). Since accurate in vivo determination of internal conductivities is currently not currently possible, the head is typically assumed to consist of a set of contiguous isotropic regions, each with constant conductivity.
Ellis, Matthew
2015-01-01
efficiency. . . . . . . . . . . . . . . . . . . . . . . .Performance and Computational Efficiency . . 4.4.1 Class ofestimation and computational efficiency. Journal of Process
Stack and cell modelling with SOFC3D: a computer program for the 3D simulations of
Herbin, Raphaèle
Stack and cell modelling with SOFC3D: a computer program for the 3D simulations of solid oxide fuel, France 1 Introduction SOFC3D is a computer program, which simulates the behaviour of a solid oxide fuel or the channels, the electrical potential \\Phi at any point of the solid part of the SOFC, and the molar fractions
Dr. Chenn Zhou
2008-10-15
Pulverized coal injection (PCI) into the blast furnace (BF) has been recognized as an effective way to decrease the coke and total energy consumption along with minimization of environmental impacts. However, increasing the amount of coal injected into the BF is currently limited by the lack of knowledge of some issues related to the process. It is therefore important to understand the complex physical and chemical phenomena in the PCI process. Due to the difficulty in attaining trus BF measurements, Computational fluid dynamics (CFD) modeling has been identified as a useful technology to provide such knowledge. CFD simulation is powerful for providing detailed information on flow properties and performing parametric studies for process design and optimization. In this project, comprehensive 3-D CFD models have been developed to simulate the PCI process under actual furnace conditions. These models provide raceway size and flow property distributions. The results have provided guidance for optimizing the PCI process.
Briceno, Luis Diego [Colorado State University, Fort Collins; Khemka, Bhavesh [Colorado State University, Fort Collins; Siegel, Howard Jay [Colorado State University, Fort Collins; Maciejewski, Anthony A [ORNL; Groer, Christopher S [ORNL; Koenig, Gregory A [ORNL; Okonski, Gene D [ORNL; Poole, Stephen W [ORNL
2011-01-01
This study considers a heterogeneous computing system and corresponding workload being investigated by the Extreme Scale Systems Center (ESSC) at Oak Ridge National Laboratory (ORNL). The ESSC is part of a collaborative effort between the Department of Energy (DOE) and the Department of Defense (DoD) to deliver research, tools, software, and technologies that can be integrated, deployed, and used in both DOE and DoD environments. The heterogeneous system and workload described here are representative of a prototypical computing environment being studied as part of this collaboration. Each task can exhibit a time-varying importance or utility to the overall enterprise. In this system, an arriving task has an associated priority and precedence. The priority is used to describe the importance of a task, and precedence is used to describe how soon the task must be executed. These two metrics are combined to create a utility function curve that indicates how valuable it is for the system to complete a task at any given moment. This research focuses on using time-utility functions to generate a metric that can be used to compare the performance of different resource schedulers in a heterogeneous computing system. The contributions of this paper are: (a) a mathematical model of a heterogeneous computing system where tasks arrive dynamically and need to be assigned based on their priority, precedence, utility characteristic class, and task execution type, (b) the use of priority and precedence to generate time-utility functions that describe the value a task has at any given time, (c) the derivation of a metric based on the total utility gained from completing tasks to measure the performance of the computing environment, and (d) a comparison of the performance of resource allocation heuristics in this environment.
IEEE TRANSACTIONS ON NANOBIOSCIENCE, VOL. 5, NO. 1, MARCH 2006 41 Computational Modeling of a New Truong, Member, IEEE Abstract--The class of fluorescence resonance energy transfer (FRET) protein
Zhou, Shujia
2009-01-01
Acceleration of Numerical Weather Prediction,” ProceedingsComputer Systems for Climate and Weather Models Shujia Zhouprocesses in climate and weather models demands a continual
Enabling a Highly-Scalable Global Address Space Model for Petascale Computing
Apra, Edoardo; Vetter, Jeffrey S; Yu, Weikuan
2010-01-01
Over the past decade, the trajectory to the petascale has been built on increased complexity and scale of the underlying parallel architectures. Meanwhile, software de- velopers have struggled to provide tools that maintain the productivity of computational science teams using these new systems. In this regard, Global Address Space (GAS) programming models provide a straightforward and easy to use addressing model, which can lead to improved produc- tivity. However, the scalability of GAS depends directly on the design and implementation of the runtime system on the target petascale distributed-memory architecture. In this paper, we describe the design, implementation, and optimization of the Aggregate Remote Memory Copy Interface (ARMCI) runtime library on the Cray XT5 2.3 PetaFLOPs computer at Oak Ridge National Laboratory. We optimized our implementation with the flow intimation technique that we have introduced in this paper. Our optimized ARMCI implementation improves scalability of both the Global Arrays (GA) programming model and a real-world chemistry application NWChem from small jobs up through 180,000 cores.
Computational Model of Population Dynamics Based on the Cell Cycle and Local Interactions
Oprisan, Sorinel Adrian; Oprisan, Ana
2005-03-31
Our study bridges cellular (mesoscopic) level interactions and global population (macroscopic) dynamics of carcinoma. The morphological differences and transitions between well and smooth defined benign tumors and tentacular malignat tumors suggest a theoretical analysis of tumor invasion based on the development of mathematical models exhibiting bifurcations of spatial patterns in the density of tumor cells. Our computational model views the most representative and clinically relevant features of oncogenesis as a fight between two distinct sub-systems: the immune system of the host and the neoplastic system. We implemented the neoplastic sub-system using a three-stage cell cycle: active, dormant, and necrosis. The second considered sub-system consists of cytotoxic active (effector) cells -- EC, with a very broad phenotype ranging from NK cells to CTL cells, macrophages, etc. Based on extensive numerical simulations, we correlated the fractal dimensions for carcinoma, which could be obtained from tumor imaging, with the malignat stage. Our computational model was able to also simulate the effects of surgical, chemotherapeutical, and radiotherapeutical treatments.
Swelling in light water reactor internal components: Insights from computational modeling
Stoller, Roger E.; Barashev, Alexander V.; Golubov, Stanislav I.
2015-08-01
A modern cluster dynamics model has been used to investigate the materials and irradiation parameters that control microstructural evolution under the relatively low-temperature exposure conditions that are representative of the operating environment for in-core light water reactor components. The focus is on components fabricated from austenitic stainless steel. The model accounts for the synergistic interaction between radiation-produced vacancies and the helium that is produced by nuclear transmutation reactions. Cavity nucleation rates are shown to be relatively high in this temperature regime (275 to 325°C), but are sensitive to assumptions about the fine scale microstructure produced under low-temperature irradiation. The cavity nucleation rates observed run counter to the expectation that void swelling would not occur under these conditions. This expectation was based on previous research on void swelling in austenitic steels in fast reactors. This misleading impression arose primarily from an absence of relevant data. The results of the computational modeling are generally consistent with recent data obtained by examining ex-service components. However, it has been shown that the sensitivity of the model s predictions of low-temperature swelling behavior to assumptions about the primary damage source term and specification of the mean-field sink strengths is somewhat greater that that observed at higher temperatures. Further assessment of the mathematical model is underway to meet the long-term objective of this research, which is to provide a predictive model of void swelling at relevant lifetime exposures to support extended reactor operations.
Leishear, R.; Poirier, M.; Fowley, M.
2011-05-26
The Salt Disposition Integration (SDI) portfolio of projects provides the infrastructure within existing Liquid Waste facilities to support the startup and long term operation of the Salt Waste Processing Facility (SWPF). Within SDI, the Blend and Feed Project will equip existing waste tanks in the Tank Farms to serve as Blend Tanks where 300,000-800,000 gallons of salt solution will be blended in 1.3 million gallon tanks and qualified for use as feedstock for SWPF. Blending requires the miscible salt solutions from potentially multiple source tanks per batch to be well mixed without disturbing settled sludge solids that may be present in a Blend Tank. Disturbing solids may be problematic both from a feed quality perspective as well as from a process safety perspective where hydrogen release from the sludge is a potential flammability concern. To develop the necessary technical basis for the design and operation of blending equipment, Savannah River National Laboratory (SRNL) completed scaled blending and transfer pump tests and computational fluid dynamics (CFD) modeling. A 94 inch diameter pilot-scale blending tank, including tank internals such as the blending pump, transfer pump, removable cooling coils, and center column, were used in this research. The test tank represents a 1/10.85 scaled version of an 85 foot diameter, Type IIIA, nuclear waste tank that may be typical of Blend Tanks used in SDI. Specifically, Tank 50 was selected as the tank to be modeled per the SRR, Project Engineering Manager. SRNL blending tests investigated various fixed position, non-rotating, dual nozzle pump designs, including a blending pump model provided by the blend pump vendor, Curtiss Wright (CW). Primary research goals were to assess blending times and to evaluate incipient sludge disturbance for waste tanks. Incipient sludge disturbance was defined by SRR and SRNL as minor blending of settled sludge from the tank bottom into suspension due to blending pump operation, where the sludge level was shown to remain constant. To experimentally model the sludge layer, a very thin, pourable, sludge simulant was conservatively used for all testing. To experimentally model the liquid, supernate layer above the sludge in waste tanks, two salt solution simulants were used, which provided a bounding range of supernate properties. One solution was water (H{sub 2}O + NaOH), and the other was an inhibited, more viscous salt solution. The research performed and data obtained significantly advances the understanding of fluid mechanics, mixing theory and CFD modeling for nuclear waste tanks by benchmarking CFD results to actual experimental data. This research significantly bridges the gap between previous CFD models and actual field experiences in real waste tanks. A finding of the 2009, DOE, Slurry Retrieval, Pipeline Transport and Plugging, and Mixing Workshop was that CFD models were inadequate to assess blending processes in nuclear waste tanks. One recommendation from that Workshop was that a validation, or bench marking program be performed for CFD modeling versus experiment. This research provided experimental data to validate and correct CFD models as they apply to mixing and blending in nuclear waste tanks. Extensive SDI research was a significant step toward bench marking and applying CFD modeling. This research showed that CFD models not only agreed with experiment, but demonstrated that the large variance in actual experimental data accounts for misunderstood discrepancies between CFD models and experiments. Having documented this finding, SRNL was able to provide correction factors to be used with CFD models to statistically bound full scale CFD results. Through the use of pilot scale tests performed for both types of pumps and available engineering literature, SRNL demonstrated how to effectively apply CFD results to salt batch mixing in full scale waste tanks. In other words, CFD models were in error prior to development of experimental correction factors determined during this research, which provided a technique to use CFD models fo
Lunds Universitet
Alignments for the IBM-3 Translation Model Thomas Schoenemann Centre for Mathematical Sciences Lund University, Sweden Abstract Prior work on training the IBM-3 transla- tion model is based on suboptimal meth from a statistical viewpoint and introduced five probability models, known as IBM 1-5. Their models
G. R. Odette; G. E. Lucas
2005-11-15
This final report on "In-Service Design & Performance Prediction of Advanced Fusion Material Systems by Computational Modeling and Simulation" (DE-FG03-01ER54632) consists of a series of summaries of work that has been published, or presented at meetings, or both. It briefly describes results on the following topics: 1) A Transport and Fate Model for Helium and Helium Management; 2) Atomistic Studies of Point Defect Energetics, Dynamics and Interactions; 3) Multiscale Modeling of Fracture consisting of: 3a) A Micromechanical Model of the Master Curve (MC) Universal Fracture Toughness-Temperature Curve Relation, KJc(T - To), 3b) An Embrittlement DTo Prediction Model for the Irradiation Hardening Dominated Regime, 3c) Non-hardening Irradiation Assisted Thermal and Helium Embrittlement of 8Cr Tempered Martensitic Steels: Compilation and Analysis of Existing Data, 3d) A Model for the KJc(T) of a High Strength NFA MA957, 3e) Cracked Body Size and Geometry Effects of Measured and Effective Fracture Toughness-Model Based MC and To Evaluations of F82H and Eurofer 97, 3-f) Size and Geometry Effects on the Effective Toughness of Cracked Fusion Structures; 4) Modeling the Multiscale Mechanics of Flow Localization-Ductility Loss in Irradiation Damaged BCC Alloys; and 5) A Universal Relation Between Indentation Hardness and True Stress-Strain Constitutive Behavior. Further details can be found in the cited references or presentations that generally can be accessed on the internet, or provided upon request to the authors. Finally, it is noted that this effort was integrated with our base program in fusion materials, also funded by the DOE OFES.
Joseph, Earl C.; Conway, Steve; Dekate, Chirag
2013-09-30
This study investigated how high-performance computing (HPC) investments can improve economic success and increase scientific innovation. This research focused on the common good and provided uses for DOE, other government agencies, industry, and academia. The study created two unique economic models and an innovation index: 1 A macroeconomic model that depicts the way HPC investments result in economic advancements in the form of ROI in revenue (GDP), profits (and cost savings), and jobs. 2 A macroeconomic model that depicts the way HPC investments result in basic and applied innovations, looking at variations by sector, industry, country, and organization size. ? A new innovation index that provides a means of measuring and comparing innovation levels. Key findings of the pilot study include: IDC collected the required data across a broad set of organizations, with enough detail to create these models and the innovation index. The research also developed an expansive list of HPC success stories.
Ellis, Matthew
2015-01-01
integration. Computers & Chemical Engineering, 20:315–323,control. Computers & Chemical Engineering, 58:334–343,reactors: a review. Chemical Engineering Communications, 1:
2004-01-01
Mathematics and Computers in Simulation 65 (2004) 557577 Parallel runs of a large air pollution 20 January 2004; accepted 21 January 2004 Abstract Large-scale air pollution models can successfully. The mathematical description of a large-scale air pollution model will be discussed in this paper. The principles
Shepelyansky, Dima
Quantum computing of quantum chaos in the kicked rotator model B. LeÂ´vi, B. Georgeot, and D. L model, a system that displays rich physical properties and enables to study problems of quantum chaos are robust in presence of imperfections. This implies that the algorithm can simulate the dynamics of quantum
Hawick, Ken
0 Computational Science Technical Note CSTN-163 Transients in a Forest-Fire Simulation Model with Varying Combustion Neighbourhoods and Watercourse Firebreaks K. A. Hawick 2012 Forest and bush fires-Schwabl Forest-fire model is investigated with various localised combustion neighbourhoods. The transient
Cerveny, Vlastislav
Ray tracing computations in the smoothed SEG/EAGE Salt Model V. Bucha*, Department of Geophysics in comparison with precise methods, e.g., finite differences or finite elements, are the speed of computation, because ray tracing computations need smooth velocity macro models (Bucha, Bulant & Klimes 2003
Donna Post Guillen; Tami Grimmett; Anastasia M. Gribik; Steven P. Antal
2010-09-01
The Hybrid Energy Systems Testing (HYTEST) Laboratory is being established at the Idaho National Laboratory to develop and test hybrid energy systems with the principal objective to safeguard U.S. Energy Security by reducing dependence on foreign petroleum. A central component of the HYTEST is the slurry bubble column reactor (SBCR) in which the gas-to-liquid reactions will be performed to synthesize transportation fuels using the Fischer Tropsch (FT) process. SBCRs are cylindrical vessels in which gaseous reactants (for example, synthesis gas or syngas) is sparged into a slurry of liquid reaction products and finely dispersed catalyst particles. The catalyst particles are suspended in the slurry by the rising gas bubbles and serve to promote the chemical reaction that converts syngas to a spectrum of longer chain hydrocarbon products, which can be upgraded to gasoline, diesel or jet fuel. These SBCRs operate in the churn-turbulent flow regime which is characterized by complex hydrodynamics, coupled with reacting flow chemistry and heat transfer, that effect reactor performance. The purpose of this work is to develop a computational multiphase fluid dynamic (CMFD) model to aid in understanding the physico-chemical processes occurring in the SBCR. Our team is developing a robust methodology to couple reaction kinetics and mass transfer into a four-field model (consisting of the bulk liquid, small bubbles, large bubbles and solid catalyst particles) that includes twelve species: (1) CO reactant, (2) H2 reactant, (3) hydrocarbon product, and (4) H2O product in small bubbles, large bubbles, and the bulk fluid. Properties of the hydrocarbon product were specified by vapor liquid equilibrium calculations. The absorption and kinetic models, specifically changes in species concentrations, have been incorporated into the mass continuity equation. The reaction rate is determined based on the macrokinetic model for a cobalt catalyst developed by Yates and Satterfield [1]. The model includes heat generation due to the exothermic chemical reaction, as well as heat removal from a constant temperature heat exchanger. Results of the CMFD simulations (similar to those shown in Figure 1) will be presented.
The Computational Sciences. Research
Christensen, Dan
The Computational Sciences. Research activities range from the theoretical foundations. The teaching mission of the computational sciences includes almost every student in the University for computational hardware and software. The computational sciences are undergoing explosive growth worldwide
Economou, Tassos
understanding how it functions; correspondingly, one of the holy grails of computational struc- tural biology
Ghosh, Somnath
. New research initiatives like the Materials Genome Initiative (MGI) and the Integrated Computational Materials Science & Engineering (ICMSE) are creating unprecedented opportunities for unraveling new1 PREFACE The recent times have seen a surge in computational modeling of materials and processes
Kaper, Tasso J. Kramer, Mark A.; Rotstein, Horacio G.
2013-12-15
Rhythmic neuronal oscillations across a broad range of frequencies, as well as spatiotemporal phenomena, such as waves and bumps, have been observed in various areas of the brain and proposed as critical to brain function. While there is a long and distinguished history of studying rhythms in nerve cells and neuronal networks in healthy organisms, the association and analysis of rhythms to diseases are more recent developments. Indeed, it is now thought that certain aspects of diseases of the nervous system, such as epilepsy, schizophrenia, Parkinson's, and sleep disorders, are associated with transitions or disruptions of neurological rhythms. This focus issue brings together articles presenting modeling, computational, analytical, and experimental perspectives about rhythms and dynamic transitions between them that are associated to various diseases.
Sasaki, Y.; Kitatsuji, Y.; Hirata, M.; Kimura, T.; Yoshizuka, K.
2008-07-01
Multi-dentate diamides have been synthesized and examined for actinide (An) extractions. Bi- and tridentate extractants are the focus in this work. The extraction of actinides was performed from 0.1-6 M HNO{sub 3} to organic solvents. It was obvious that N,N,N',N'-tetra-alkyl-diglycolamide (DGA) derivatives, 2,2'-(methylimino)bis(N,N-dioctyl-acetamide) (MIDOA), and N,N'-dimethyl-N,N'-dioctyl-2-(3-oxa-pentadecane)-malonamide (DMDOOPDMA) have relatively high D values (D(Pu) > 70). The following notable results using DGA extractants were obtained: (1) DGAs with short alkyl chains give higher D values than those with long alkyl chain, (2) DGAs with long alkyl chain have high solubility in n-dodecane. Computational molecular modeling was also used to elucidate the effects of structural and electronic properties of the reagents on their different extractabilities. (authors)
Anderson, C.A.
1981-01-01
Six years ago the Reactor Safety Research Division of the Nuclear Regulatory Commission (NRC) approached the Los Alamos National Laboratory to develop a comprehensive concrete structural analysis code to predict the static and dynamic behavior of Prestressed Concrete Reactor Vessels (PCRVs) that serve as the containment structure of a High-Temperature Gas-Cooled Reactor. The PCRV is a complex concrete structure that must be modeled in three dimensions and posseses other complicating features such as a steel liner for the reactor cavity and woven cables embedded vertically in the PCRV and wound circumferentially on the outside of the PCRV. The cables, or tendons, are used for prestressing the reactor vessel. In addition to developing the computational capability to predict inelastic three dimensional concrete structural behavior, the code response was verified against documented experiments on concrete structural behavior. This code development/verification effort is described.
GEOCITY: a computer model for systems analysis of geothermal district heating and cooling costs
Fassbender, L.L.; Bloomster, C.H.
1981-06-01
GEOCITY is a computer-simulation model developed to study the economics of district heating/cooling using geothermal energy. GEOCITY calculates the cost of district heating/cooling based on climate, population, resource characteristics, and financing conditions. The basis for our geothermal-energy cost analysis is the unit cost of energy which will recover all the costs of production. The calculation of the unit cost of energy is based on life-cycle costing and discounted-cash-flow analysis. A wide variation can be expected in the range of potential geothermal district heating and cooling costs. The range of costs is determined by the characteristics of the resource, the characteristics of the demand, and the distance separating the resource and the demand. GEOCITY is a useful tool for estimating costs for each of the main parts of the production process and for determining the sensitivity of these costs to several significant parameters under a consistent set of assumptions.
Pierre-Luc Dallaire-Demers; Frank K. Wilhelm
2015-08-18
Many phenomena of strongly correlated materials are encapsulated in the Fermi-Hubbard model whose thermodynamical properties can be computed from its grand canonical potential according to standard procedures. In general, there is no closed form solution for lattices of more than one spatial dimension, but solutions can be approximated with cluster perturbation theory. To model long-range effects such as order parameters, a powerful method to compute the cluster's Green's function consists in finding its self-energy through a variational principle of the grand canonical potential. This opens the possibility of studying various phase transitions at finite temperature in the Fermi-Hubbard model. However, a classical cluster solver quickly hits an exponential wall in the memory (or computation time) required to store the computation variables. Here it is shown theoretically that that the cluster solver can be mapped to a subroutine on a quantum computer whose quantum memory scales as the number of orbitals in the simulated cluster. A quantum computer with a few tens of qubits could therefore simulate the thermodynamical properties of complex fermionic lattices inaccessible to classical supercomputers.
Pierre-Luc Dallaire-Demers; Frank K. Wilhelm
2015-11-27
Many phenomena of strongly correlated materials are encapsulated in the Fermi-Hubbard model whose thermodynamical properties can be computed from its grand canonical potential according to standard procedures. In general, there is no closed form solution for lattices of more than one spatial dimension, but solutions can be approximated with cluster perturbation theory. To model long-range effects such as order parameters, a powerful method to compute the cluster's Green's function consists in finding its self-energy through a variational principle of the grand canonical potential. This opens the possibility of studying various phase transitions at finite temperature in the Fermi-Hubbard model. However, a classical cluster solver quickly hits an exponential wall in the memory (or computation time) required to store the computation variables. Here it is shown theoretically that that the cluster solver can be mapped to a subroutine on a quantum computer whose quantum memory scales as the number of orbitals in the simulated cluster. A quantum computer with a few tens of qubits could therefore simulate the thermodynamical properties of complex fermionic lattices inaccessible to classical supercomputers.
DAVENPORT, J.
2006-11-01
Computational Science is an integral component of Brookhaven's multi science mission, and is a reflection of the increased role of computation across all of science. Brookhaven currently has major efforts in data storage and analysis for the Relativistic Heavy Ion Collider (RHIC) and the ATLAS detector at CERN, and in quantum chromodynamics. The Laboratory is host for the QCDOC machines (quantum chromodynamics on a chip), 10 teraflop/s computers which boast 12,288 processors each. There are two here, one for the Riken/BNL Research Center and the other supported by DOE for the US Lattice Gauge Community and other scientific users. A 100 teraflop/s supercomputer will be installed at Brookhaven in the coming year, managed jointly by Brookhaven and Stony Brook, and funded by a grant from New York State. This machine will be used for computational science across Brookhaven's entire research program, and also by researchers at Stony Brook and across New York State. With Stony Brook, Brookhaven has formed the New York Center for Computational Science (NYCCS) as a focal point for interdisciplinary computational science, which is closely linked to Brookhaven's Computational Science Center (CSC). The CSC has established a strong program in computational science, with an emphasis on nanoscale electronic structure and molecular dynamics, accelerator design, computational fluid dynamics, medical imaging, parallel computing and numerical algorithms. We have been an active participant in DOES SciDAC program (Scientific Discovery through Advanced Computing). We are also planning a major expansion in computational biology in keeping with Laboratory initiatives. Additional laboratory initiatives with a dependence on a high level of computation include the development of hydrodynamics models for the interpretation of RHIC data, computational models for the atmospheric transport of aerosols, and models for combustion and for energy utilization. The CSC was formed to bring together researchers in these areas and to provide a focal point for the development of computational expertise at the Laboratory. These efforts will connect to and support the Department of Energy's long range plans to provide Leadership class computing to researchers throughout the Nation. Recruitment for six new positions at Stony Brook to strengthen its computational science programs is underway. We expect some of these to be held jointly with BNL.
Johnson, J. D.; Oberkampf, William Louis; Helton, Jon Craig (Arizona State University, Tempe, AZ); Storlie, Curtis B. (North Carolina State University, Raleigh, NC)
2006-10-01
Evidence theory provides an alternative to probability theory for the representation of epistemic uncertainty in model predictions that derives from epistemic uncertainty in model inputs, where the descriptor epistemic is used to indicate uncertainty that derives from a lack of knowledge with respect to the appropriate values to use for various inputs to the model. The potential benefit, and hence appeal, of evidence theory is that it allows a less restrictive specification of uncertainty than is possible within the axiomatic structure on which probability theory is based. Unfortunately, the propagation of an evidence theory representation for uncertainty through a model is more computationally demanding than the propagation of a probabilistic representation for uncertainty, with this difficulty constituting a serious obstacle to the use of evidence theory in the representation of uncertainty in predictions obtained from computationally intensive models. This presentation describes and illustrates a sampling-based computational strategy for the representation of epistemic uncertainty in model predictions with evidence theory. Preliminary trials indicate that the presented strategy can be used to propagate uncertainty representations based on evidence theory in analysis situations where naive sampling-based (i.e., unsophisticated Monte Carlo) procedures are impracticable due to computational cost.
A computer code for calculations in the algebraic collective model of the atomic nucleus
T. A. Welsh; D. J. Rowe
2015-02-20
A Maple code is presented for algebraic collective model (ACM) calculations. The ACM is an algebraic version of the Bohr model of the atomic nucleus, in which all required matrix elements are derived by exploiting the model's SU(1,1) x SO(5) dynamical group. This, in particular, obviates the use of coefficients of fractional parentage. This paper reviews the mathematical formulation of the ACM, and serves as a manual for the code. The code enables a wide range of model Hamiltonians to be analysed. This range includes essentially all Hamiltonians that are rational functions of the model's quadrupole moments $q_M$ and are at most quadratic in the corresponding conjugate momenta $\\pi_N$ ($-2\\le M,N\\le 2$). The code makes use of expressions for matrix elements derived elsewhere and newly derived matrix elements of the operators $[\\pi\\otimes q \\otimes\\pi]_0$ and $[\\pi\\otimes\\pi]_{LM}$. The code also provides ready access to SO(3)-reduced SO(5) Clebsch-Gordan coefficients through data files provided with the code.
Harting, Jens
Steering in computational science: mesoscale modelling and simulation J. CHIN{, J. HARTING{, S. JHA steering for high performance computing applications. Lattice-Boltzmann mesoscale fluid simulations, there is currently considerable interest in mesoscale models. These models coarse grain most of the atomic
Wind Turbine Modeling for Computational Fluid Dynamics: December 2010 - December 2012
Tossas, L. A. M.; Leonardi, S.
2013-07-01
With the shortage of fossil fuel and the increasing environmental awareness, wind energy is becoming more and more important. As the market for wind energy grows, wind turbines and wind farms are becoming larger. Current utility-scale turbines extend a significant distance into the atmospheric boundary layer. Therefore, the interaction between the atmospheric boundary layer and the turbines and their wakes needs to be better understood. The turbulent wakes of upstream turbines affect the flow field of the turbines behind them, decreasing power production and increasing mechanical loading. With a better understanding of this type of flow, wind farm developers could plan better-performing, less maintenance-intensive wind farms. Simulating this flow using computational fluid dynamics is one important way to gain a better understanding of wind farm flows. In this study, we compare the performance of actuator disc and actuator line models in producing wind turbine wakes and the wake-turbine interaction between multiple turbines. We also examine parameters that affect the performance of these models, such as grid resolution, the use of a tip-loss correction, and the way in which the turbine force is projected onto the flow field.
Computer modelling of the reduction of rare earth dopants in barium aluminate
Rezende, Marcos V. dos S; Valerio, Mario E.G. [Department of Physics, Federal University of Sergipe, 49100-000 Sao Cristovao, SE (Brazil); Jackson, Robert A., E-mail: r.a.jackson@chem.keele.ac.uk [School of Physical and Geographical Sciences, Keele University, Keele, Staffordshire ST5 5BG (United Kingdom)
2011-08-15
Long lasting phosphorescence in barium aluminates can be achieved by doping with rare earth ions in divalent charge states. The rare earth ions are initially in a trivalent charge state, but are reduced to a divalent charge state before being doped into the material. In this paper, the reduction of trivalent rare earth ions in the BaAl{sub 2}O{sub 4} lattice is studied by computer simulation, with the energetics of the whole reduction and doping process being modelled by two methods, one based on single ion doping and one which allows dopant concentrations to be taken into account. A range of different reduction schemes are considered and the most energetically favourable schemes identified. - Graphical abstract: The doping and subsequent reduction of a rare earth ion into the barium aluminate lattice. Highlights: > The doping of barium aluminate with rare earth ions reduced in a range of atmospheres has been modelled. > The overall solution energy for the doping process for each ion in each reducing atmosphere is calculated using two methods. > The lowest energy reduction process is predicted and compared with experimental results.
Development of Computational Tools for Metabolic Model Curation, Flux Elucidation and Strain Design
Maranas, Costas D
2012-05-21
An overarching goal of the Department of EnergyÂ? mission is the efficient deployment and engineering of microbial and plant systems to enable biomass conversion in pursuit of high energy density liquid biofuels. This has spurred the pace at which new organisms are sequenced and annotated. This torrent of genomic information has opened the door to understanding metabolism in not just skeletal pathways and a handful of microorganisms but for truly genome-scale reconstructions derived for hundreds of microbes and plants. Understanding and redirecting metabolism is crucial because metabolic fluxes are unique descriptors of cellular physiology that directly assess the current cellular state and quantify the effect of genetic engineering interventions. At the same time, however, trying to keep pace with the rate of genomic data generation has ushered in a number of modeling and computational challenges related to (i) the automated assembly, testing and correction of genome-scale metabolic models, (ii) metabolic flux elucidation using labeled isotopes, and (iii) comprehensive identification of engineering interventions leading to the desired metabolism redirection.
Anastasia Gribik; Doona Guillen, PhD; Daniel Ginosar, PhD
2008-09-01
Currently multi-tubular fixed bed reactors, fluidized bed reactors, and slurry bubble column reactors (SBCRs) are used in commercial Fischer Tropsch (FT) synthesis. There are a number of advantages of the SBCR compared to fixed and fluidized bed reactors. The main advantage of the SBCR is that temperature control and heat recovery are more easily achieved. The SBCR is a multiphase chemical reactor where a synthesis gas, comprised mainly of H2 and CO, is bubbled through a liquid hydrocarbon wax containing solid catalyst particles to produce specialty chemicals, lubricants, or fuels. The FT synthesis reaction is the polymerization of methylene groups [-(CH2)-] forming mainly linear alkanes and alkenes, ranging from methane to high molecular weight waxes. The Idaho National Laboratory is developing a computational multiphase fluid dynamics (CMFD) model of the FT process in a SBCR. This paper discusses the incorporation of absorption and reaction kinetics into the current hydrodynamic model. A phased approach for incorporation of the reaction kinetics into a CMFD model is presented here. Initially, a simple kinetic model is coupled to the hydrodynamic model, with increasing levels of complexity added in stages. The first phase of the model includes incorporation of the absorption of gas species from both large and small bubbles into the bulk liquid phase. The driving force for the gas across the gas liquid interface into the bulk liquid is dependent upon the interfacial gas concentration in both small and large bubbles. However, because it is difficult to measure the concentration at the gas-liquid interface, coefficients for convective mass transfer have been developed for the overall driving force between the bulk concentrations in the gas and liquid phases. It is assumed that there are no temperature effects from mass transfer of the gas phases to the bulk liquid phase, since there are only small amounts of dissolved gas in the liquid phase. The product from the incorporation of absorption is the steady state concentration profile of the absorbed gas species in the bulk liquid phase. The second phase of the model incorporates a simplified macrokinetic model to the mass balance equation in the CMFD code. Initially, the model assumes that the catalyst particles are sufficiently small such that external and internal mass and heat transfer are not rate limiting. The model is developed utilizing the macrokinetic rate expression developed by Yates and Satterfield (1991). Initially, the model assumes that the only species formed other than water in the FT reaction is C27H56. Change in moles of the reacting species and the resulting temperature of the catalyst and fluid phases is solved simultaneously. The macrokinetic model is solved in conjunction with the species transport equations in a separate module which is incorporated into the CMFD code.
Modeling a Reconfigurable System for Computing the FFT in Place via RewritingLogic #
Ayala-Rincón, Mauricio
hardware implementation of the Fast Fourier Transform -- FFT using rewritinglogic. It is shown on general purpose processors, reconfigurable computing delivers more processing power due electronic market. There are several taxonomies applied to reconfigurable computing. Concerning the specific
Huang, Yongxin
2010-01-16
ferromagnets, there exist two main challenges: the complicated microelasticity due to the magnetostrictive strain, and very expensive computation mainly caused by the calculation of long-range magnetostatic and elastic interactions. A parallel computing...
A Computational Model which Learns to Selectively Attend in Category Learning
Cottrell, Garrison W.
Garrison W. Cottrell lingyun,gary@cs.ucsd.edu UCSD Computer Science and Engineering 9500 Gilman Dr., La
Stockman, Mark [Georgia State University Research Foundation] [Georgia State University Research Foundation; Gray, Steven [Argon National Laboratory] [Argon National Laboratory
2014-02-21
The program is directed toward development of new computational approaches to photoprocesses in nanostructures whose geometry and composition are tailored to obtain desirable optical responses. The emphasis of this specific program is on the development of computational methods and prediction and computational theory of new phenomena of optical energy transfer and transformation on the extreme nanoscale (down to a few nanometers).
Bergman, Keren
at the scale of high performance computer clusters and warehouse scale data centers, system level simulations and results for rack scale photonic interconnection networks for high performance computing. Keywords: optical to the newsworthy power consumption [3], latency [4] and bandwidth challenges [5] of high performance computing (HPC
BatteryConscious Task Sequencing for Portable Devices Including Voltage/Clock Scaling
Kambhampati, Subbarao
model and validated it with measurements taken on a real lithiumion bat tery used in a pocket computerBatteryConscious Task Sequencing for Portable Devices Including Voltage/Clock Scaling #3; Daler Department, Arizona State University Tempe, Arizona 85287 chaitali@asu.edu ABSTRACT Operation of battery
Gas-Phase Lubrication of ta-C by Glycerol and Hydrogen Peroxide. Experimental and Computer Modeling
Goddard III, William A.
Gas-Phase Lubrication of ta-C by Glycerol and Hydrogen Peroxide. Experimental and Computer Modeling lubrication conditions at 80 Â°C in presence of OH-containing molecules. To understand the mechanism of ultralow friction, we performed gas-phase lubrication experiments followed by time-of-flight secondary ion
Geller, Michael R.
2013-01-01
code. Section III maps the calculation of the ground-state energy for the TIM onto a quantum phase in the transverse Ising model (TIM) [12], there is a large number of physical qubits and lengthy computational time]. Here we investigate the quantum simulation of the TIM ground-state energy on a surface code quantum
Computational Biology | More Science | ORNL
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
Computational Biology SHARE Computational Biology Computational Biology research encompasses many important aspects including molecular biophysics for bio-energy, genetic level...
Program Analyses for Cloud Computations
Tetali, Sai Deep
2015-01-01
search. ” In CCSW 09: Cloud Computing Security Workshop, pp.ACM workshop on Cloud computing security workshop, CCSW ’11,aspects of cloud computing, including security, performance
Musial, W.; Lawson, M.; Rooney, S.
2013-02-01
The Marine and Hydrokinetic Technology (MHK) Instrumentation, Measurement, and Computer Modeling Workshop was hosted by the National Renewable Energy Laboratory (NREL) in Broomfield, Colorado, July 9-10, 2012. The workshop brought together over 60 experts in marine energy technologies to disseminate technical information to the marine energy community and collect information to help identify ways in which the development of a commercially viable marine energy industry can be accelerated. The workshop was comprised of plenary sessions that reviewed the state of the marine energy industry and technical sessions that covered specific topics of relevance. Each session consisted of presentations, followed by facilitated discussions. During the facilitated discussions, the session chairs posed several prepared questions to the presenters and audience to encourage communication and the exchange of ideas between technical experts. Following the workshop, attendees were asked to provide written feedback on their takeaways and their best ideas on how to accelerate the pace of marine energy technology development. The first four sections of this document give a general overview of the workshop format, provide presentation abstracts and discussion session notes, and list responses to the post-workshop questions. The final section presents key findings and conclusions from the workshop that suggest how the U.S. Department of Energy and national laboratory resources can be utilized to most effectively assist the marine energy industry.
Musial, W.; Lawson, M.; Rooney, S.
2013-02-01
The Marine and Hydrokinetic Technology (MHK) Instrumentation, Measurement, and Computer Modeling Workshop was hosted by the National Renewable Energy Laboratory (NREL) in Broomfield, Colorado, July 9–10, 2012. The workshop brought together over 60 experts in marine energy technologies to disseminate technical information to the marine energy community, and to collect information to help identify ways in which the development of a commercially viable marine energy industry can be accelerated. The workshop was comprised of plenary sessions that reviewed the state of the marine energy industry and technical sessions that covered specific topics of relevance. Each session consisted of presentations, followed by facilitated discussions. During the facilitated discussions, the session chairs posed several prepared questions to the presenters and audience to encourage communication and the exchange of ideas between technical experts. Following the workshop, attendees were asked to provide written feedback on their takeaways from the workshop and their best ideas on how to accelerate the pace of marine energy technology development. The first four sections of this document give a general overview of the workshop format, provide presentation abstracts, supply discussion session notes, and list responses to the post-workshop questions. The final section presents key findings and conclusions from the workshop that suggest what the most pressing MHK technology needs are and how the U.S. Department of Energy (DOE) and national laboratory resources can be utilized to assist the marine energy industry in the most effective manner.
James P. Crutchfield; Christopher J. Ellison; Ryan G. James; John R. Mahoney
2010-07-30
We adapt tools from information theory to analyze how an observer comes to synchronize with the hidden states of a finitary, stationary stochastic process. We show that synchronization is determined by both the process's internal organization and by an observer's model of it. We analyze these components using the convergence of state-block and block-state entropies, comparing them to the previously known convergence properties of the Shannon block entropy. Along the way, we introduce a hierarchy of information quantifiers as derivatives and integrals of these entropies, which parallels a similar hierarchy introduced for block entropy. We also draw out the duality between synchronization properties and a process's controllability. The tools lead to a new classification of a process's alternative representations in terms of minimality, synchronizability, and unifilarity.
Pesaran, A.; Kim, G. H.; Smith, K.; Santhanagopalan, S.; Lee, K. J.
2012-05-01
This 2012 Annual Merit Review presentation gives an overview of the Computer-Aided Engineering of Batteries (CAEBAT) project and introduces the Multi-Scale, Multi-Dimensional model for modeling lithium-ion batteries for electric vehicles.
Data-Driven Optimization for Modeling in Computer Graphics and Vision
Yu, Lap Fai
2013-01-01
the recent trend of 3D printing in computer-aided design.3D display, and 3D printing have made these technologies
Data-Driven Optimization for Modeling in Computer Graphics and Vision
Yu, Lap Fai
2013-01-01
in 3D scanning, 3D display, and 3D printing have made thesethe recent trend of 3D printing in computer-aided design.
M.F. Simpson; K.-R. Kim
2010-12-01
In support of closing the nuclear fuel cycle using non-aqueous separations technology, this project aims to develop computational models of electrorefiners based on fundamental chemical and physical processes. Spent driver fuel from Experimental Breeder Reactor-II (EBR-II) is currently being electrorefined in the Fuel Conditioning Facility (FCF) at Idaho National Laboratory (INL). And Korea Atomic Energy Research Institute (KAERI) is developing electrorefining technology for future application to spent fuel treatment and management in the Republic of Korea (ROK). Electrorefining is a critical component of pyroprocessing, a non-aqueous chemical process which separates spent fuel into four streams: (1) uranium metal, (2) U/TRU metal, (3) metallic high-level waste containing cladding hulls and noble metal fission products, and (4) ceramic high-level waste containing sodium and active metal fission products. Having rigorous yet flexible electrorefiner models will facilitate process optimization and assist in trouble-shooting as necessary. To attain such models, INL/UI has focused on approaches to develop a computationally-light and portable two-dimensional (2D) model, while KAERI/SNU has investigated approaches to develop a computationally intensive three-dimensional (3D) model for detailed and fine-tuned simulation.
Illinois at Urbana-Champaign, University of
85721 Center for Reliable and High-Performance Computing Coordinated Science Laboratory UniversityFrom International Computer Performance and Dependability Symposium, Erlangen, Germany, April 1995, pp.285 294 MODELING RECYCLE: A CASE STUDY IN THE INDUSTRIAL USE OF MEASUREMENT AND MODELING Luai M
Eindhoven, Technische Universiteit
1 A COMPUTER SIMULATION MODEL FOR THE BLOW-BLOW FORMING PROCESS OF GLASS CONTAINERS C. G for industrial purposes. To achieve this both steps of the forming of glass containers, namely blow- blow needs conditions. In [1] the development of a computer model to be used for glass blowing was described
Larson, N. M.; Perey, F. G.
1980-11-01
A method is described for determining the parameters of a model from experimental data based upon the utilization of Bayes' theorem. This method has several advantages over the least-squares method as it is commonly used; one important advantage is that the assumptions under which the parameter values have been determined are more clearly evident than in many results based upon least squares. Bayes' method has been used to develop a computer code which can be utilized to analyze neutron cross-section data by means of the R-matrix theory. The required formulae from the R-matrix theory are presented, and the computer implementation of both Bayes' equations and R-matrix theory is described. Details about the computer code and compelte input/output information are given.
Coupling Multi-Component Models with MPH on Distributed Memory Computer Architectures
He, Yun; Ding, Chris
2005-01-01
Among these, NASA’s Earth System Models Framework (ESMF) [to facilitate coupling earth system model components and to
MODELING STRATEGIES TO COMPUTE NATURAL CIRCULATION USING CFD IN A VHTR AFTER A LOFA
Yu-Hsin Tung; Richard W. Johnson; Ching-Chang Chieng; Yuh-Ming Ferng
2012-11-01
A prismatic gas-cooled very high temperature reactor (VHTR) is being developed under the next generation nuclear plant program (NGNP) of the U.S. Department of Energy, Office of Nuclear Energy. In the design of the prismatic VHTR, hexagonal shaped graphite blocks are drilled to allow insertion of fuel pins, made of compacted TRISO fuel particles, and coolant channels for the helium coolant. One of the concerns for the reactor design is the effects of a loss of flow accident (LOFA) where the coolant circulators are lost for some reason, causing a loss of forced coolant flow through the core. In such an event, it is desired to know what happens to the (reduced) heat still being generated in the core and if it represents a problem for the fuel compacts, the graphite core or the reactor vessel (RV) walls. One of the mechanisms for the transport of heat out of the core is by the natural circulation of the coolant, which is still present. That is, how much heat may be transported by natural circulation through the core and upwards to the top of the upper plenum? It is beyond current capability for a computational fluid dynamic (CFD) analysis to perform a calculation on the whole RV with a sufficiently refined mesh to examine the full potential of natural circulation in the vessel. The present paper reports the investigation of several strategies to model the flow and heat transfer in the RV. It is found that it is necessary to employ representative geometries of the core to estimate the heat transfer. However, by taking advantage of global and local symmetries, a detailed estimate of the strength of the resulting natural circulation and the level of heat transfer to the top of the upper plenum is obtained.
Wang, S L; Singer, M A
2009-07-13
The purpose of this report is to evaluate the hemodynamic effects of renal vein inflow and filter position on unoccluded and partially occluded IVC filters using three-dimensional computational fluid dynamics. Three-dimensional models of the TrapEase and Gunther Celect IVC filters, spherical thrombi, and an IVC with renal veins were constructed. Hemodynamics of steady-state flow was examined for unoccluded and partially occluded TrapEase and Gunther Celect IVC filters in varying proximity to the renal veins. Flow past the unoccluded filters demonstrated minimal disruption. Natural regions of stagnant/recirculating flow in the IVC are observed superior to the bilateral renal vein inflows, and high flow velocities and elevated shear stresses are observed in the vicinity of renal inflow. Spherical thrombi induce stagnant and/or recirculating flow downstream of the thrombus. Placement of the TrapEase filter in the suprarenal vein position resulted in a large area of low shear stress/stagnant flow within the filter just downstream of thrombus trapped in the upstream trapping position. Filter position with respect to renal vein inflow influences the hemodynamics of filter trapping. Placement of the TrapEase filter in a suprarenal location may be thrombogenic with redundant areas of stagnant/recirculating flow and low shear stress along the caval wall due to the upstream trapping position and the naturally occurring region of stagnant flow from the renal veins. Infrarenal vein placement of IVC filters in a near juxtarenal position with the downstream cone near the renal vein inflow likely confers increased levels of mechanical lysis of trapped thrombi due to increased shear stress from renal vein inflow.
Protoplanetary disks including radiative feedback from accreting planets
Montesinos, Matias; Perez, Sebastian; Baruteau, Clement; Casassus, Simon
2015-01-01
While recent observational progress is converging on the detection of compact regions of thermal emission due to embedded protoplanets, further theoretical predictions are needed to understand the response of a protoplanetary disk to the planet formation radiative feedback. This is particularly important to make predictions for the observability of circumplanetary regions. In this work we use 2D hydrodynamical simulations to examine the evolution of a viscous protoplanetary disk in which a luminous Jupiter-mass planet is embedded. We use an energy equation which includes the radiative heating of the planet as an additional mechanism for planet formation feedback. Several models are computed for planet luminosities ranging from $10^{-5}$ to $10^{-3}$ Solar luminosities. We find that the planet radiative feedback enhances the disk's accretion rate at the planet's orbital radius, producing a hotter and more luminous environement around the planet, independently of the prescription used to model the disk's turbul...
Bauer, Roman; Stoll, Elizabeth
2015-01-01
Glioma is the most common form of primary brain tumor. Demographically, the risk of occurrence increases until old age. Here we present a novel computational model to reproduce the probability of glioma incidence across the lifespan. Previous mathematical models explaining glioma incidence are framed in a rather abstract way, and do not directly relate to empirical findings. To decrease this gap between theory and experimental observations, we incorporate recent data on cellular and molecular factors underlying gliomagenesis. Since evidence implicates the adult neural stem cell as the likely cell-of-origin of glioma, we have incorporated empirically-determined estimates of neural stem cell number, cell division rate, mutation rate and oncogenic potential into our model. We demonstrate that our model yields results which match actual demographic data in the human population. In particular, this model accounts for the observed peak incidence of glioma at approximately 80 years of age, without the need to assert...
DEVELOPING A NEW APPROACH OF COMPUTER USE `KISS MODELING' FOR DESIGN-IDEAS ALTERNATIVES OF FORM
, YOSHIHIRO KOBAYASHI South Valley University, Faculty of Fine Arts at Luxor, Egypt. wael, and form generation through computational power is more prominent in the two dimensions than the three
Computational Model of Forward and Opposed Smoldering Combustion with Improved Chemical Kinetics
Rein, Guillermo
A computational study has been carried out to investigate smoldering ignition and propagation in polyurethane foam. The onedimensional, transient, governing equations for smoldering combustion in a porous fuel are solved ...
Garba, M.T.; Gonzales-Velez, H.; Roach, D.L.
2010-11-26
This paper examines the initial parallel implementation of SCATTER, a computationally intensive inelastic neutron scattering routine with polycrystalline averaging capability, for the General Utility Lattice Program (GULP). Of particular importance...
100 IEEE TRANSACTIONS ON COMPUTERS, VOL. 45, NO. 10, OCTOBER 1996 An Analytical Model
Mudge, Trevor
to spend money is the cheapest (rather than the fastest) cache level, particularly with small system are with the Advanced Com puter Architecture Laboratory, Department of Electrical Engineer ing and Computer Science
100 IEEE TRANSACTIONS ON COMPUTERS, VOL. 45, NO. 10, OCTOBER 1996 An Analytical Model
Mudge, Trevor
is the cheapest rather than the fastest cache level, particularly with small system budgets. Another is that money, Department of Electrical Engineer- ing and Computer Science, University of Michigan, Ann Arbor, MI 48109
Using the FLUENT computational fluid dynamics code to model the NACOK corrosion test
Parks, Benjamin T
2004-01-01
As a part of advancing nuclear technology, computational fluid dynamics (CFD) analysis offers safer and lower-cost results relative to experimental work. Its use as a safety analysis tool is gaining much broader acceptance ...
Christopher Beetle; Benjamin Bromley; Richard H. Price
2006-02-08
The periodic standing wave approach to binary inspiral assumes rigid rotation of gravitational fields and hence helically symmetric solutions. To exploit the symmetry, numerical computations must solve for ``helical scalars,'' fields that are functions only of corotating coordinates, the labels on the helical Killing trajectories. Here we present the formalism for describing linearized general relativity in terms of helical scalars and we present solutions to the mixed partial differential equations of the linearized gravity problem (and to a toy nonlinear problem) using the adapted coordinates and numerical techniques previously developed for scalar periodic standing wave computations. We argue that the formalism developed may suffice for periodic standing wave computations for post-Minkowskian computations and for full general relativity.
Computer Energy Modeling Techniques for Simulation Large Scale Correctional Institutes in Texas
Heneghan, T.; Haberl, J. S.; Saman, N.; Bou-Saada, T. E.
1996-01-01
Building energy simulation programs have undergone an increase in use for evaluating energy consumption and energy conservation retrofits in buildings. Utilization of computer simulation programs for large facilities with multiple buildings, however...
A Hierarchical Model for Distributed Collaborative Computation in Wireless Sensor Networks
Hwang, Kai
with the physical world. Example applications include environment monitor- ing, target tracking and fire detection
A Hierarchical Model for Distributed Collaborative Computation in Wireless Sensor Networks
Hwang, Kai
include environment monitor- ing, target tracking and fire detection [4]. Low cost, ade- quate coverage
An experimental validation of the PRO model for parallel and distributed computation
Boyer, Edmond
model was introduced by Gebremedhin et al. [2002] as a frame- work for the design and analysis. The Parallel Resource-Optimal (PRO) model, introduced in Gebremedhin et al. [2002], is similar to the BSP model discussion of the PRO model, see Gebremedhin et al. [2002]. The PRO model The key features of PRO
Countries Gasoline Prices Including Taxes
Gasoline and Diesel Fuel Update (EIA)
Crude oil, gasoline, heating oil, diesel, propane, and other liquids including biofuels and natural gas liquids. Natural Gas Exploration and reserves, storage, imports and...
Refining and Extending the Business Model with Information Technology: Dell Computer Corporation
Kraemer, Kenneth L; Dedrick, Jason; Yamashiro, Sandra
1999-01-01
of Dell’s Direct Business Model Fuels Fifteenth ConsecutiveAND EXTENDING THE REFINING AND EXTENDING THE BUSINESS MODELBUSINESS MODEL CENTER FOR RESEARCH ON INFORMATION TECHNOLOGY
Abercrombie, Robert K; Sheldon, Federick T.; Schlicher, Bob G
2015-01-01
There are many influencing economic factors to weigh from the defender-practitioner stakeholder point-of-view that involve cost combined with development/deployment models. Some examples include the cost of countermeasures themselves, the cost of training and the cost of maintenance. Meanwhile, we must better anticipate the total cost from a compromise. The return on investment in countermeasures is essentially impact costs (i.e., the costs from violating availability, integrity and confidentiality / privacy requirements). The natural question arises about choosing the main risks that must be mitigated/controlled and monitored in deciding where to focus security investments. To answer this question, we have investigated the cost/benefits to the attacker/defender to better estimate risk exposure. In doing so, it s important to develop a sound basis for estimating the factors that derive risk exposure, such as likelihood that a threat will emerge and whether it will be thwarted. This impact assessment framework can provide key information for ranking cybersecurity threats and managing risk.
A non-CFD modeling system for computing 3D wind and concentration fields in urban environments
Nelson, Matthew A; Brown, Michael J; Williams, Michael D; Gowardhan, Akshay; Pardyjak, Eric R
2010-01-01
The Quick Urban & Industrial Complex (QUIC) Dispersion Modeling System has been developed to rapidly compute the transport and dispersion of toxic agent releases in the vicinity of buildings. It is composed of an empirical-diagnostic wind solver, an 'urbanized' Lagrangian random-walk model, and a graphical user interface. The code has been used for homeland security and environmental air pollution applications. In this paper, we discuss the wind solver methodology and improvements made to the original Roeckle schemes in order to better capture flow fields in dense built-up areas. The mode1-computed wind and concentration fields are then compared to measurements from several field experiments. Improvements to the QUIC Dispersion Modeling System have been made to account for the inhomogeneous and complex building layouts found in large cities. The logic that has been introduced into the code is described and comparisons of model output to full-scale outdoor urban measurements in Oklahoma City and New York City are given. Although far from perfect, the model agreed fairly well with measurements and in many cases performed equally to CFD codes.
Whitney, Daniel
, and Industry at Leeds University Daniel E. Whitney, Liaison Scientist, Manufacturing Summary The Computer Aided Engineering Unit in the Mechanical Engineering Department at Leeds University has built its expertise mechanics. Before coming to Leeds, de Pennington was at the Philips Production Automation Department
TECHNICAL REPORTS Today there are many well-established computer models that
Radcliffe, David
the monitoring data that is typically available. There is a need for more parsimonious models and monitoring data in the United States and Europe cover a spectrum from static risk-based index models such as the P-Index (Weld and Sharpley, 2007) and Source Apportionment (OSPAR, 2000) models to complex, process-based, dynamic models
Automated Analysis of Software Designs with Graphic Quality Models Department of Computer Science
Zhu, Hong
Automated Analysis of Software Designs with Graphic Quality Models QIAN ZHANG Department@brookes.ac.uk Abstract: Software quality models play a significant role in software quality assurance. Based on our previous work on graphic modelling of software quality, this paper extends the quality modelling language
Corley, Richard A.; Kabilan, Senthil; Kuprat, Andrew P.; Carson, James P.; Minard, Kevin R.; Jacob, Rick E.; Timchalk, Charles; Glenny, Robb W.; Pipavath, Sudhaker; Cox, Timothy C.; Wallis, Chris; Larson, Richard; Fanucchi, M.; Postlewait, Ed; Einstein, Daniel R.
2012-07-01
Coupling computational fluid dynamics (CFD) with physiologically based pharmacokinetic (PBPK) models is useful for predicting site-specific dosimetry of airborne materials in the respiratory tract and elucidating the importance of species differences in anatomy, physiology, and breathing patterns. Historically, these models were limited to discrete regions of the respiratory system. CFD/PBPK models have now been developed for the rat, monkey, and human that encompass airways from the nose or mouth to the lung. A PBPK model previously developed to describe acrolein uptake in nasal tissues was adapted to the extended airway models as an example application. Model parameters for each anatomic region were obtained from the literature, measured directly, or estimated from published data. Airflow and site-specific acrolein uptake patterns were determined under steadystate inhalation conditions to provide direct comparisons with prior data and nasalonly simulations. Results confirmed that regional uptake was dependent upon airflow rates and acrolein concentrations with nasal extraction efficiencies predicted to be greatest in the rat, followed by the monkey, then the human. For human oral-breathing simulations, acrolein uptake rates in oropharyngeal and laryngeal tissues were comparable to nasal tissues following nasal breathing under the same exposure conditions. For both breathing modes, higher uptake rates were predicted for lower tracheo-bronchial tissues of humans than either the rat or monkey. These extended airway models provide a unique foundation for comparing dosimetry across a significantly more extensive range of conducting airways in the rat, monkey, and human than prior CFD models.
Neutralino relic density including coannihilations
Paolo Gondolo; Joakim Edsjo
1997-11-25
We give an overview of our precise calculation of the relic density of the lightest neutralino, in which we included relativistic Boltzmann averaging, subthreshold and resonant annihilations, and coannihilation processes with charginos and neutralinos.
Rybkowski, Zofia K.; Wong, John-Michael; Ballard, Glenn; Tommelein, Iris D.
2008-01-01
1 Computer Push 20 46 55 0 3 0 58 Live Push 20 43 51 1 5 0 57* Computer Pull 12 46 1 0 1 0 2 Live Pull 12 39 1 1 0 0 2 *WS1 ran out of pieces at 5'20" DISCUSSION A purpose of many lean production simulations is to educate players... of Construction Engineering and Management, ASCE, 125(4), 265-276. Myers, J.L. and Well, A.D. (2003). Research Design and Statistical Analysis. 2nd Ed., Lawrence Erlbaum Assoc. Pubs., Mahwah, NJ. Pasqualini, F. and Zawislak, P.A. (2005). "Value Stream Mapping...
Stewart, James
, \\Shadows for bump-mapped surfaces", in Advanced Computer Graphics (Proceedings of Computer Graphics Tokyo. Hershberger, \\Finding the upper envelope of n line segments in O(nlogn) time", Inform. Process. Lett., vol. 33. The horizon computation can also be used to build horizon maps that are used in conjunction with bump maps
Stewart, James
. Max, ``Shadows for bumpmapped surfaces'', in Advanced Computer Graphics (Proceedings of Computer, pp. 92--99. [10] J. Hershberger, ``Finding the upper envelope of n line segments in O(n log n) time input to a progressive radiosity algorithm. The horizon computation can also be used to build horizon
Gribaudo, Marco
, Analysis, and Simulation of Computer and Telecommunications Systems (MASCOTS'04) 1526-7539/04 $20.00 © 2004, Analysis, and Simulation of Computer and Telecommunications Systems (MASCOTS'04) 1526-7539/04 $20.00 © 2004, Analysis, and Simulation of Computer and Telecommunications Systems (MASCOTS'04) 1526-7539/04 $20.00 © 2004
Gray, Alan; Harlen, Oliver G.; Harris, Sarah A.; Khalid, Syma; Leung, Yuk Ming; Lonsdale, Richard; Mulholland, Adrian J.; Pearson, Arwen R.; Read, Daniel J.; Richardson, Robin A.
2015-01-01
The current computational techniques available for biomolecular simulation are described, and the successes and limitations of each with reference to the experimental biophysical methods that they complement are presented. Despite huge advances in the computational techniques available for simulating biomolecules at the quantum-mechanical, atomistic and coarse-grained levels, there is still a widespread perception amongst the experimental community that these calculations are highly specialist and are not generally applicable by researchers outside the theoretical community. In this article, the successes and limitations of biomolecular simulation and the further developments that are likely in the near future are discussed. A brief overview is also provided of the experimental biophysical methods that are commonly used to probe biomolecular structure and dynamics, and the accuracy of the information that can be obtained from each is compared with that from modelling. It is concluded that progress towards an accurate spatial and temporal model of biomacromolecules requires a combination of all of these biophysical techniques, both experimental and computational.
Computer hardware fault administration
Archer, Charles J. (Rochester, MN); Megerian, Mark G. (Rochester, MN); Ratterman, Joseph D. (Rochester, MN); Smith, Brian E. (Rochester, MN)
2010-09-14
Computer hardware fault administration carried out in a parallel computer, where the parallel computer includes a plurality of compute nodes. The compute nodes are coupled for data communications by at least two independent data communications networks, where each data communications network includes data communications links connected to the compute nodes. Typical embodiments carry out hardware fault administration by identifying a location of a defective link in the first data communications network of the parallel computer and routing communications data around the defective link through the second data communications network of the parallel computer.
Computing dynamics of copper-based SMA via centre manifold reduction of 3D models q
Melnik, Roderick
of applications in such areas as aerospace engineering, medicine, manufacturing (in- cluding automotive industry is based on the authors' contribution to the European Conference on Computational Mechanics, Munich, 1999 mechanics, and applied mathematics communities [4±13]. This interest has two major
Norman, Gethin
probabilistically quantified properties pertaining to buffer sizes, delays, energy usage etc., for each derived and Rajesh Gupta3 1School of Computer Science, University of Birmingham, Birmingham, UK 2Bradley Department). In this paper, we focus on the latter. Dynamic power management (DPM) is a way to save energy in devices which
Creativity as Pastiche: A Computational Model of Dynamic Blending and Textual Collage,
Veale, Tony
Reference to the Use of Blending in Cinematic Narratives Tony Veale, School of Computer Applications, Dublin combination within a framework of interconnected mental spaces. Through an analysis of cinematic pastiche, as exemplified by its use in cinematic pastiche, a process in which two or more narratives are blended together
Ewing, Richard E.
and Information Volume 1, Number 1, Pages 116 HIGH PERFORMANCE COMPUTING IN PETROLEUM APPLICATIONS RICHARD E in petroleum applications is to try to optimize the recovery of hydrocarbon from permeable underground the complex geomechanical, physical, and multiphase fluid flow processes that accompany the various recov- ery
Liquid Computing in a Simplified Model of Cortical Layer IV: Learning to Balance a Ball
Maass, Wolfgang
Probst1,3 , Wolfgang Maass2 , Henry Markram1 , and Marc-Oliver Gewaltig1 1 Blue Brain Project, ´Ecole that the cortical network can robustly learn this task using a supervised learning rule that penalizes the error function is then constructed by readout neurons or action cells which compute a weighted sum of the high
MapReduce Programming Model for .NET-based Distributed Computing Chao Jin and Rajkumar Buyya
Buyya, Rajkumar
,000 to several million commodity machines dispersed across about 25 data centers [10]. Microsoft also presents of Virtual Machines through the web service interface. Although the popularity of data centers is increasing the basis of `data center computer' [5]. #12;2 .NET is the standard platform of Windows applications
Reyes, Dasia Ann
2009-05-15
-Stokes (RANS) method. Preliminary investigations of PANS show promising re-sults but there still exist computational and physical issues that must be addressed.This study investigates the performance of the PANS method for ow past acylinder at a Reynolds number...
Brown, Jed
are needed to cope Computer architecture is increasingly hierarchical algorithms should conform the task of systematic truncation error testing and accuracy estimation. -- 1986 #12;Diffusive cooling(nlogk) Project vector onto eigenbasis in O(nlogk) Optimal-order graph partitioning and clustering Continuation
Modeling a Reconfigurable System for Computing the FFT in Place via Rewriting-Logic
Ayala-Rincón, Mauricio
implementation of the Fast Fourier Transform FFT using rewriting-logic. It is shown that rewriting-logic can purpose processors, reconfigurable computing delivers more processing power due to the implementation design makes it a good solution to the rapid changing electronic market. There are several taxonomies
Stat 586 Interpretation of Data I (Modern Applied Statistical Modeling and Computing)
Xie, Minge
, B.D., Modern Applied Statistics with S-Plus (Springer) Cody, R.P. and Simth, J.K. Applied Statistics and the SAS Programming Language Elliott, R. J. Learning SAS in the Computer Lab (Duxbury Press). Based allows) Students should expect to use either R or SAS or Splus statistical software packages in data
Development and Validation of a Computer Model for Energy-Efficient Shaded Fenestration Design
Oh, Kie Whan
2000-01-01
transfer through an unshaded window at any orientation. A computerized simulation model, called the Shaded Fenestration Design (SFD) model, was developed that is capable of calculating the amount of solar radiation for a clear day as well as calculating...
Liquefied Natural Gas (LNG) Vapor Dispersion Modeling with Computational Fluid Dynamics Codes
Qi, Ruifeng
2012-10-19
Federal regulation 49 CFR 193 and standard NFPA 59A require the use of validated consequence models to determine the vapor cloud dispersion exclusion zones for accidental liquefied natural gas (LNG) releases. For modeling purposes, the physical...
A critical review on modelling formalisms and simulation tools in computational biosystems
Machado, Daniel
Integration of different kinds of biological processes is an ultimate goal for whole-cell modelling. We briefly review modelling formalisms that have been used in Systems Biology and identify the criteria that must be ...
Daigle, Matthew
, and availability. Prognos- tics deals with determining the health state of compo- nents, and projecting) predictions. Model-based prognos- tics approaches perform these tasks with the aid of a model that captures
Flow field computation of the NREL S809 airfoil using various turbulence models
Chang, Y.L.; Yang, S.L.; Arici, O. [Michigan Technological Univ., Houghton, MI (United States). Mechanical Engineering-Engineering Mechanics Dept.
1996-10-01
Performance comparison of three popular turbulence models, namely Baldwin-Lomas algebraic model, Chien`s Low-Reynolds-Number {kappa}-{epsilon} model, and Wilcox`s Low-Reynolds-Number {kappa}-{omega} model, is given. These models were applied to calculate the flow field around the National Renewable Energy Laboratory S809 airfoil using Total Variational Diminishing scheme. Numerical results of C{sub P}, C{sub L}, and C{sub D} are presented along with the Delft experimental data. It is shown that all three models perform well for attached flow, i.e., no flow separation at low angles of attack. However, at high angles of attack with flow separation, convergence characteristics show Wilcox`s model outperforms the other models. Results of this study will be used to guide the authors in their dynamic stall research.
Chen, Shu-Ching
Florida Public Hurricane Loss Model Purpose: To develop and maintain a public computer model to assess hurricane wind, surge and flood related risk and to project annual expected insured residential for user defined scenarios. This public model can also be used to quantify the cost benefits of hurricane
Mathematical and Computer Modelling 43 (2006) 718726 www.elsevier.com/locate/mcm
Bohner, Martin
2006-01-01
maximization problem is much more flexible and realistic. For example, a consumer receives income at one point unify the standard discrete and continuous models in economics, but also, for example, allows models are set up in either continuous or discrete time. For example, in a discrete model, a consumer
A Computer Model of Field-directed Morphogenesis Part I -Julia Sets
Levin, Michael
are viewed as a map from ¢¡ to £¡ , and are modeled by the complex number algebra. Functions in complex that gene interactions can be modeled as a multi-dimensional algebra, and that only two interacting genes with actual pictures of any particular real organism. A brief introduction to top-down models and positional
Computer resources Computer resources
Yang, Zong-Liang
Computer resources 1 Computer resources available to the LEAD group Cédric David 30 September 2009 #12;Ouline · UT computer resources and services · JSG computer resources and services · LEAD computers· LEAD computers 2 #12;UT Austin services UT EID and Password 3 https://utdirect.utexas.edu #12;UT Austin
Yock, Adam D. Kudchadker, Rajat J.; Rao, Arvind; Dong, Lei; Beadle, Beth M.; Garden, Adam S.; Court, Laurence E.
2014-05-15
Purpose: The purpose of this work was to develop and evaluate the accuracy of several predictive models of variation in tumor volume throughout the course of radiation therapy. Methods: Nineteen patients with oropharyngeal cancers were imaged daily with CT-on-rails for image-guided alignment per an institutional protocol. The daily volumes of 35 tumors in these 19 patients were determined and used to generate (1) a linear model in which tumor volume changed at a constant rate, (2) a general linear model that utilized the power fit relationship between the daily and initial tumor volumes, and (3) a functional general linear model that identified and exploited the primary modes of variation between time series describing the changing tumor volumes. Primary and nodal tumor volumes were examined separately. The accuracy of these models in predicting daily tumor volumes were compared with those of static and linear reference models using leave-one-out cross-validation. Results: In predicting the daily volume of primary tumors, the general linear model and the functional general linear model were more accurate than the static reference model by 9.9% (range: ?11.6%–23.8%) and 14.6% (range: ?7.3%–27.5%), respectively, and were more accurate than the linear reference model by 14.2% (range: ?6.8%–40.3%) and 13.1% (range: ?1.5%–52.5%), respectively. In predicting the daily volume of nodal tumors, only the 14.4% (range: ?11.1%–20.5%) improvement in accuracy of the functional general linear model compared to the static reference model was statistically significant. Conclusions: A general linear model and a functional general linear model trained on data from a small population of patients can predict the primary tumor volume throughout the course of radiation therapy with greater accuracy than standard reference models. These more accurate models may increase the prognostic value of information about the tumor garnered from pretreatment computed tomography images and facilitate improved treatment management.
Graduate Research Opportunities in SIGNALS, SYSTEMS & COMPUTATIONAL INTELLIGENCE
Rusu, Adrian
Communications · Computer Networks · Bioinformatics · Biomedical Signal Processing & Modeling · Image Processing in relevant areas (signal processing, probability and statistics, linear algebra, programming, bioinformatics background courses. Electives offered include: · Advanced Digital Signal Processing · Wavelets Transforms
Grigorieva, Ellina V.
2012-01-01
model of water treatment with includes a bounded control. In order to foretell possible outcomes that society is facing today is pollution of water, and in particularly the discharge of a sewage into the water. Biological treatment, which employs bacteria in order to eliminate pathogens and reduce