Notions of Computation Determine Monads
Plotkin, Gordon; Power, John
2002-01-01
We model notions of computation using algebraic operations and equations. We show that these generate several of the monads of pri- mary interest that have been used to model computational e ects, with the striking ...
computational modeling of biological systems
$author.value
Faculty. Faculty listing for "computational modeling of biological systems" ... Research Interests: computational modeling of biological systems.
Typologies of Computation and Computational Models
Mark Burgin; Gordana Dodig-Crnkovic
2013-12-09
We need much better understanding of information processing and computation as its primary form. Future progress of new computational devices capable of dealing with problems of big data, internet of things, semantic web, cognitive robotics and neuroinformatics depends on the adequate models of computation. In this article we first present the current state of the art through systematization of existing models and mechanisms, and outline basic structural framework of computation. We argue that defining computation as information processing, and given that there is no information without (physical) representation, the dynamics of information on the fundamental level is physical/ intrinsic/ natural computation. As a special case, intrinsic computation is used for designed computation in computing machinery. Intrinsic natural computation occurs on variety of levels of physical processes, containing the levels of computation of living organisms (including highly intelligent animals) as well as designed computational devices. The present article offers a typology of current models of computation and indicates future paths for the advancement of the field; both by the development of new computational models and by learning from nature how to better compute using different mechanisms of intrinsic computation.
Determine Ramsey numbers on a quantum computer
Hefeng Wang
2015-10-07
We present a quantum algorithm for computing the Ramsey numbers whose computational complexity grows super-exponentially with the number of vertices of a graph on a classical computer. The problem is mapped to a decision problem on a quantum computer, a probe qubit is coupled to a register that represents the problem and detects the energy levels of the problem Hamiltonian. The decision problem is solved by determining whether the probe qubit exhibits resonance dynamics. The algorithm shows a quadratic speedup over its classical counterparts, and the degenerate ground state problem in the adiabatic quantum evolution algorithm for this problem is avoided.
Computational Models for Understanding Weather
Muraki, David J.
Computational Models for Understanding Weather Mathematics for Atmospheric Science http://weather-S migration Dutton Conway zonal jetstream in unstable weather 6 #12;Baroclinic Instability Vortices
Improved computer models support genetics research
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
Simple computer models unravel genetic stress reactions in cells Simple computer models unravel genetic stress reactions in cells Integrated biological and computational methods...
LANL computer model boosts engine efficiency
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
LANL computer model boosts engine efficiency LANL computer model boosts engine efficiency The KIVA model has been instrumental in helping researchers and manufacturers understand...
2014-02-21 Issuance: Proposed Determination of Computer Servers...
Office of Environmental Management (EM)
Servers as a Covered Consumer Product; Withdrawal 2014-02-21 Issuance: Proposed Determination of Computer Servers as a Covered Consumer Product; Withdrawal This document is a...
Improved computer models support genetics research
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
February Simple computer models unravel genetic stress reactions in cells Simple computer models unravel genetic stress reactions in cells Integrated biological and...
Computer Graphics Hierarchical Modeling
for sta9c models, and vital for anima9on! · Key is to structure the transforma9 Transforma9ons Translate base by (5,0,0) Translate lower arm by (5,0,0) Translate upper arm by (5,0,0) Translate hammer by (5,0,0) ... q Each part
Cosmic Logic: a Computational Model
Vitaly Vanchurin
2015-07-05
We describe a simple computational model of cosmic logic suitable for analysis of, for example, discretized cosmological systems. The construction is based on a particular model of computation, developed by Alan Turing, with cosmic observers (CO), cosmic measures (CM) and cosmic symmetries (CS) described by Turing machines. CO machines always start with a blank tape and CM machines take CO's Turing number (also known as description number or G{\\" o}del number) as input and output the corresponding probability. Similarly, CS machines take CO's Turing number as input, but output either one if the CO machines are in the same equivalence class or zero otherwise. We argue that CS machines are more fundamental than CM machines and, thus, should be used as building blocks in constructing CM machines. We prove the non-computability of a CS machine which discriminates between two classes of CO machines: mortal that halts in finite time and immortal that runs forever. In context of eternal inflation this result implies that it is impossible to construct CM machines to compute probabilities using cut-off prescriptions or that all of the cut-off measures are non-computable.
Parallel computing in enterprise modeling.
Goldsby, Michael E.; Armstrong, Robert C.; Shneider, Max S.; Vanderveen, Keith; Ray, Jaideep; Heath, Zach; Allan, Benjamin A.
2008-08-01
This report presents the results of our efforts to apply high-performance computing to entity-based simulations with a multi-use plugin for parallel computing. We use the term 'Entity-based simulation' to describe a class of simulation which includes both discrete event simulation and agent based simulation. What simulations of this class share, and what differs from more traditional models, is that the result sought is emergent from a large number of contributing entities. Logistic, economic and social simulations are members of this class where things or people are organized or self-organize to produce a solution. Entity-based problems never have an a priori ergodic principle that will greatly simplify calculations. Because the results of entity-based simulations can only be realized at scale, scalable computing is de rigueur for large problems. Having said that, the absence of a spatial organizing principal makes the decomposition of the problem onto processors problematic. In addition, practitioners in this domain commonly use the Java programming language which presents its own problems in a high-performance setting. The plugin we have developed, called the Parallel Particle Data Model, overcomes both of these obstacles and is now being used by two Sandia frameworks: the Decision Analysis Center, and the Seldon social simulation facility. While the ability to engage U.S.-sized problems is now available to the Decision Analysis Center, this plugin is central to the success of Seldon. Because Seldon relies on computationally intensive cognitive sub-models, this work is necessary to achieve the scale necessary for realistic results. With the recent upheavals in the financial markets, and the inscrutability of terrorist activity, this simulation domain will likely need a capability with ever greater fidelity. High-performance computing will play an important part in enabling that greater fidelity.
Quantum computation beyond the circuit model
Jordan, Stephen Paul
2008-01-01
The quantum circuit model is the most widely used model of quantum computation. It provides both a framework for formulating quantum algorithms and an architecture for the physical construction of quantum computers. However, ...
Cosmic Logic: a Computational Model
Vanchurin, Vitaly
2015-01-01
We describe a simple computational model of cosmic logic suitable for analysis of, for example, discretized cosmological systems. The construction is based on a particular model of computation, developed by Alan Turing, with cosmic observers (CO), cosmic measures (CM) and cosmic symmetries (CS) described by Turing machines. CO machines always start with a blank tape and CM machines take CO's Turing number (also known as description number or G{\\" o}del number) as input and output the corresponding probability. Similarly, CS machines take CO's Turing number as input, but output either one if the CO machines are in the same equivalence class or zero otherwise. We argue that CS machines are more fundamental than CM machines and, thus, should be used as building blocks in constructing CM machines. We prove the non-computability of a CS machine which discriminates between two classes of CO machines: mortal that halts in finite time and immortal that runs forever. In context of eternal inflation this result implies...
Determining position inside building via laser rangefinder and handheld computer
Ramsey, Jr. James L. (Albuquerque, NM); Finley, Patrick (Albuquerque, NM); Melton, Brad (Albuquerque, NM)
2010-01-12
An apparatus, computer software, and a method of determining position inside a building comprising selecting on a PDA at least two walls of a room in a digitized map of a building or a portion of a building, pointing and firing a laser rangefinder at corresponding physical walls, transmitting collected range information to the PDA, and computing on the PDA a position of the laser rangefinder within the room.
Cupola Furnace Computer Process Model
Seymour Katz
2004-12-31
The cupola furnace generates more than 50% of the liquid iron used to produce the 9+ million tons of castings annually. The cupola converts iron and steel into cast iron. The main advantages of the cupola furnace are lower energy costs than those of competing furnaces (electric) and the ability to melt less expensive metallic scrap than the competing furnaces. However the chemical and physical processes that take place in the cupola furnace are highly complex making it difficult to operate the furnace in optimal fashion. The results are low energy efficiency and poor recovery of important and expensive alloy elements due to oxidation. Between 1990 and 2004 under the auspices of the Department of Energy, the American Foundry Society and General Motors Corp. a computer simulation of the cupola furnace was developed that accurately describes the complex behavior of the furnace. When provided with the furnace input conditions the model provides accurate values of the output conditions in a matter of seconds. It also provides key diagnostics. Using clues from the diagnostics a trained specialist can infer changes in the operation that will move the system toward higher efficiency. Repeating the process in an iterative fashion leads to near optimum operating conditions with just a few iterations. More advanced uses of the program have been examined. The program is currently being combined with an ''Expert System'' to permit optimization in real time. The program has been combined with ''neural network'' programs to affect very easy scanning of a wide range of furnace operation. Rudimentary efforts were successfully made to operate the furnace using a computer. References to these more advanced systems will be found in the ''Cupola Handbook''. Chapter 27, American Foundry Society, Des Plaines, IL (1999).
Theoretical, experimental, and computational aspects of optical property determination of
Mandelis, Andreas
-density-wave source is applied to data from model phantoms. The combined theoretical, experimental, and computational media uniquely, as compared with PPTR, which ex- hibits uniqueness problems. From data sets obtained applications, especially with tur- bid media such as tissue.36 In earlier studies,4,5 PPTR was used
Modeling Computer Viruses MSc Thesis (Afstudeerscriptie)
Amsterdam, University of
Modeling Computer Viruses MSc Thesis (Afstudeerscriptie) written by Luite Menno Pieter van Zelst About half a year ago, Alban Ponse, my thesis supervisor, suggested that the topic of `computer viruses indus- try and the creators of computer viruses. After all, the anti-virus industry stands to lose a lot
Determining collective barrier operation skew in a parallel computer
Faraj, Daniel A.
2015-12-24
Determining collective barrier operation skew in a parallel computer that includes a number of compute nodes organized into an operational group includes: for each of the nodes until each node has been selected as a delayed node: selecting one of the nodes as a delayed node; entering, by each node other than the delayed node, a collective barrier operation; entering, after a delay by the delayed node, the collective barrier operation; receiving an exit signal from a root of the collective barrier operation; and measuring, for the delayed node, a barrier completion time. The barrier operation skew is calculated by: identifying, from the compute nodes' barrier completion times, a maximum barrier completion time and a minimum barrier completion time and calculating the barrier operation skew as the difference of the maximum and the minimum barrier completion time.
Climate Modeling using High-Performance Computing
Mirin, A A
2007-02-05
The Center for Applied Scientific Computing (CASC) and the LLNL Climate and Carbon Science Group of Energy and Environment (E and E) are working together to improve predictions of future climate by applying the best available computational methods and computer resources to this problem. Over the last decade, researchers at the Lawrence Livermore National Laboratory (LLNL) have developed a number of climate models that provide state-of-the-art simulations on a wide variety of massively parallel computers. We are now developing and applying a second generation of high-performance climate models. Through the addition of relevant physical processes, we are developing an earth systems modeling capability as well.
Preliminary Phase Field Computational Model Development
Li, Yulan; Hu, Shenyang Y.; Xu, Ke; Suter, Jonathan D.; McCloy, John S.; Johnson, Bradley R.; Ramuhalli, Pradeep
2014-12-15
This interim report presents progress towards the development of meso-scale models of magnetic behavior that incorporate microstructural information. Modeling magnetic signatures in irradiated materials with complex microstructures (such as structural steels) is a significant challenge. The complexity is addressed incrementally, using the monocrystalline Fe (i.e., ferrite) film as model systems to develop and validate initial models, followed by polycrystalline Fe films, and by more complicated and representative alloys. In addition, the modeling incrementally addresses inclusion of other major phases (e.g., martensite, austenite), minor magnetic phases (e.g., carbides, FeCr precipitates), and minor nonmagnetic phases (e.g., Cu precipitates, voids). The focus of the magnetic modeling is on phase-field models. The models are based on the numerical solution to the Landau-Lifshitz-Gilbert equation. From the computational standpoint, phase-field modeling allows the simulation of large enough systems that relevant defect structures and their effects on functional properties like magnetism can be simulated. To date, two phase-field models have been generated in support of this work. First, a bulk iron model with periodic boundary conditions was generated as a proof-of-concept to investigate major loop effects of single versus polycrystalline bulk iron and effects of single non-magnetic defects. More recently, to support the experimental program herein using iron thin films, a new model was generated that uses finite boundary conditions representing surfaces and edges. This model has provided key insights into the domain structures observed in magnetic force microscopy (MFM) measurements. Simulation results for single crystal thin-film iron indicate the feasibility of the model for determining magnetic domain wall thickness and mobility in an externally applied field. Because the phase-field model dimensions are limited relative to the size of most specimens used in experiments, special experimental methods were devised to create similar boundary conditions in the iron films. Preliminary MFM studies conducted on single and polycrystalline iron films with small sub-areas created with focused ion beam have correlated quite well qualitatively with phase-field simulations. However, phase-field model dimensions are still small relative to experiments thus far. We are in the process of increasing the size of the models and decreasing specimen size so both have identical dimensions. Ongoing research is focused on validation of the phase-field model. Validation is being accomplished through comparison with experimentally obtained MFM images (in progress), and planned measurements of major hysteresis loops and first order reversal curves. Extrapolation of simulation sizes to represent a more stochastic bulk-like system will require sampling of various simulations (i.e., with single non-magnetic defect, single magnetic defect, single grain boundary, single dislocation, etc.) with distributions of input parameters. These outputs can then be compared to laboratory magnetic measurements and ultimately to simulate magnetic Barkhausen noise signals.
Modelling energy efficiency for computation
Reams, Charles
2012-11-13
. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 130 10 List of Tables 3.1 Energy usage breakdown for computing equipment in the United States. Energy figures are in billions of kWh. . . . . . . . . . . . . . . . . . . . . . 40 5.1 Average solution time... understanding of the underlying behavioural properties will inevitably lead to improvements in the practicality of NTC, and practical NTC-purposed cores have now been constructed; for example, the Phoenix processor, which operates in the near-threshold region...
Computer aided nuclear reactor modeling
Warraich, Khalid Sarwar
1995-01-01
Nuclear reactor modeling is an important activity that lets us analyze existing as well as proposed systems for safety, correct operation, etc. The quality of a analysis is directly proportional to the quality of the model used. In this work we look...
Regional weather modeling on parallel computers.
Baillie, C.; Michalakes, J.; Skalin, R.; Mathematics and Computer Science; NOAA Forecast Systems Lab.; Norwegian Meteorological Inst.
1997-01-01
This special issue on 'regional weather models' complements the October 1995 special issue on 'climate and weather modeling', which focused on global models. In this introduction we review the similarities and differences between regional and global atmospheric models. Next, the structure of regional models is described and we consider how the basic algorithms applied in these models influence the parallelization strategy. Finally, we give a brief overview of the eight articles in this issue and discuss some remaining challenges in the area of adapting regional weather models to parallel computers.
Computationally Efficient Modeling of High-Efficiency Clean Combustion...
Office of Energy Efficiency and Renewable Energy (EERE) Indexed Site
& Publications Computationally Efficient Modeling of High-Efficiency Clean Combustion Engines Computationally Efficient Modeling of High-Efficiency Clean Combustion Engines...
Modeling-Computer Simulations At Fish Lake Valley Area (Deymonaz...
Modeling-Computer Simulations At Fish Lake Valley Area (Deymonaz, Et Al., 2008) Jump to: navigation, search GEOTHERMAL ENERGYGeothermal Home Exploration Activity: Modeling-Computer...
HIV virus spread and evolution studied through computer modeling
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
HIV and evolution studied through computer modeling HIV virus spread and evolution studied through computer modeling This approach distinguishes between susceptible and infected...
Modeling-Computer Simulations At Nw Basin & Range Region (Pritchett...
Modeling-Computer Simulations At Nw Basin & Range Region (Pritchett, 2004) Jump to: navigation, search GEOTHERMAL ENERGYGeothermal Home Exploration Activity: Modeling-Computer...
Modeling-Computer Simulations At Nw Basin & Range Region (Biasi...
Modeling-Computer Simulations At Nw Basin & Range Region (Biasi, Et Al., 2009) Jump to: navigation, search GEOTHERMAL ENERGYGeothermal Home Exploration Activity: Modeling-Computer...
Computer modeling reveals how surprisingly potent hepatitis C...
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
Hepatitis C computer modeling Computer modeling reveals how surprisingly potent hepatitis C drug works A study reveals how daclatasvir targets one of its proteins and causes the...
A Computational Model for Adaptive Emotion Regulation
Treur, Jan
A Computational Model for Adaptive Emotion Regulation Tibor Bosse, Matthijs Pontier, and Jan Treur} Abstract. Emotion regulation describes how a subject can use certain strategies to affect emotion response levels. Usually, models for emotion regulation as- sume mechanisms based on feedback loops that indicate
Computationally Efficient Model for Dopant Precipitation Kinetics
Dunham, Scott
Computationally Efficient Model for Dopant Precipitation Kinetics Iuval Clejan and Scott T. Dunham and precipitates. Dopant deactivation is typically modeled using a steadystate solid solubility or clustering such as dopant activation/deactivation, it is essential to consider the fact that precipitation involves a range
Mason, Harris E. [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States); Walsh, Stuart D. C. [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States); DuFrane, Wyatt L. [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States); Carroll, Susan A. [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States)
2014-06-17
The development of accurate, predictive models for use in determining wellbore integrity requires detailed information about the chemical and mechanical changes occurring in hardened Portland cements. X-ray computed tomography (XRCT) provides a method that can nondestructively probe these changes in three dimensions. Here, we describe a method for extracting subvoxel mineralogical and chemical information from synchrotron XRCT images by combining advanced image segmentation with geochemical models of cement alteration. The method relies on determining “effective linear activity coefficients” (ELAC) for the white light source to generate calibration curves that relate the image grayscales to material composition. The resulting data set supports the modeling of cement alteration by CO_{2}-rich brine with discrete increases in calcium concentration at reaction boundaries. The results of these XRCT analyses can be used to further improve coupled geochemical and mechanical models of cement alteration in the wellbore environment.
Mechanistic Models in Computational Social Science
Holme, Petter
2015-01-01
Quantitative social science is not only about regression analysis or, in general, data inference. Computer simulations of social mechanisms have an over 60 years long history. They have been used for many different purposes -- to test scenarios, to test the consistency of descriptive theories (proof-of-concept models), to explore emerging phenomena, for forecasting, etc. In this essay, we sketch these historical developments, the role of mechanistic models in the social sciences and the influences from natural and formal sciences. We argue that mechanistic computational models form a natural common ground for social and natural sciences, and look forward to possible future information flow across the social-natural divide.
Decision Model for Cloud Computing
Kondo, Derrick
with different pricing models for cost-cutting, resource-hungry users. Second, prices can differ dynamically (as Grenoble, France 1 #12;Trade-offs Supercomputers Performance Reliability Cost ($) low high high high Instances Â· "Spot" instance price varies dynamically Â· Spot instance provided when user's bid is greater
CDF computing and event data models
Snider, F.D.; /Fermilab
2005-12-01
The authors discuss the computing systems, usage patterns and event data models used to analyze Run II data from the CDF-II experiment at the Tevatron collider. A critical analysis of the current implementation and design reveals some of the stronger and weaker elements of the system, which serve as lessons for future experiments. They highlight a need to maintain simplicity for users in the face of an increasingly complex computing environment.
Optimization of neuron models using grid computing
Vella, Mike
2011-09-09
stream_source_info Mike_Vella.ppt.txt stream_content_type text/plain stream_size 2634 Content-Encoding UTF-8 stream_name Mike_Vella.ppt.txt Content-Type text/plain; charset=UTF-8 Optimization of neuron models using grid computing... channels Set of information to be included in a model is large Single neuron multi-compartment models Why optimize? Single neuron models provide a basis for understanding cell and local circuit function Maximal conductances, compartment capacitances...
Modeling Computations in a Semantic Network
Marko A. Rodriguez; Johan Bollen
2007-05-31
Semantic network research has seen a resurgence from its early history in the cognitive sciences with the inception of the Semantic Web initiative. The Semantic Web effort has brought forth an array of technologies that support the encoding, storage, and querying of the semantic network data structure at the world stage. Currently, the popular conception of the Semantic Web is that of a data modeling medium where real and conceptual entities are related in semantically meaningful ways. However, new models have emerged that explicitly encode procedural information within the semantic network substrate. With these new technologies, the Semantic Web has evolved from a data modeling medium to a computational medium. This article provides a classification of existing computational modeling efforts and the requirements of supporting technologies that will aid in the further growth of this burgeoning domain.
High performance computing and numerical modelling
,
2014-01-01
Numerical methods play an ever more important role in astrophysics. This is especially true in theoretical works, but of course, even in purely observational projects, data analysis without massive use of computational methods has become unthinkable. The key utility of computer simulations comes from their ability to solve complex systems of equations that are either intractable with analytic techniques or only amenable to highly approximative treatments. Simulations are best viewed as a powerful complement to analytic reasoning, and as the method of choice to model systems that feature enormous physical complexity such as star formation in evolving galaxies, the topic of this 43rd Saas Fee Advanced Course. The organizers asked me to lecture about high performance computing and numerical modelling in this winter school, and to specifically cover the basics of numerically treating gravity and hydrodynamics in the context of galaxy evolution. This is still a vast field, and I necessarily had to select a subset ...
Who Wants to Know What When? Privacy Preference Determinants in Ubiquitous Computing
Madiraju, Praveen
Who Wants to Know What When? Privacy Preference Determinants in Ubiquitous Computing Scott Lederer disclosed through a ubiquitous computing system. We found that privacy preferences varied by inquirer more privacy in ubiquitous computing. Keywords Ubiquitous Computing, Privacy, Social and Legal Issues
Kim, G.; Pesaran, A.; Smith, K.; Graf, P.; Jun, M.; Yang, C.; Li, G.; Li, S.; Hochman, A.; Tselepidakis, D.; White, J.
2014-06-01
This presentation discusses the significant enhancement of computational efficiency in nonlinear multiscale battery model for computer aided engineering in current research at NREL.
Sensitivity Analysis Methodology for a Complex System Computational Model
1 Sensitivity Analysis Methodology for a Complex System Computational Model James J. Filliben of computational models to serve as predictive surrogates for the system. The use of such models increasingly) of a computational model for a complex system is always an essential component in accepting/rejecting such a model
Computational Modeling of Self-organization of Dislocations and...
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
Computational Modeling of Self-organization of Dislocations and Mesoscale Deformation of Metals Event Sponsor: Mathematics and Computing Science - LANS Seminar Start Date: Jun 19...
Computer Modelling of 3D Geological Surface
Kodge, B G
2011-01-01
The geological surveying presently uses methods and tools for the computer modeling of 3D-structures of the geographical subsurface and geotechnical characterization as well as the application of geoinformation systems for management and analysis of spatial data, and their cartographic presentation. The objectives of this paper are to present a 3D geological surface model of Latur district in Maharashtra state of India. This study is undertaken through the several processes which are discussed in this paper to generate and visualize the automated 3D geological surface model of a projected area.
Wild Fire Computer Model Helps Firefighters
Canfield, Jesse
2012-09-04
A high-tech computer model called HIGRAD/FIRETEC, the cornerstone of a collaborative effort between U.S. Forest Service Rocky Mountain Research Station and Los Alamos National Laboratory, provides insights that are essential for front-line fire fighters. The science team is looking into levels of bark beetle-induced conditions that lead to drastic changes in fire behavior and how variable or erratic the behavior is likely to be.
Wild Fire Computer Model Helps Firefighters
Canfield, Jesse
2014-06-02
A high-tech computer model called HIGRAD/FIRETEC, the cornerstone of a collaborative effort between U.S. Forest Service Rocky Mountain Research Station and Los Alamos National Laboratory, provides insights that are essential for front-line fire fighters. The science team is looking into levels of bark beetle-induced conditions that lead to drastic changes in fire behavior and how variable or erratic the behavior is likely to be.
COMPUTATIONAL MODELING OF CIRCULATING FLUIDIZED BED REACTORS
Ibrahim, Essam A
2013-01-09
Details of numerical simulations of two-phase gas-solid turbulent flow in the riser section of Circulating Fluidized Bed Reactor (CFBR) using Computational Fluid Dynamics (CFD) technique are reported. Two CFBR riser configurations are considered and modeled. Each of these two riser models consist of inlet, exit, connecting elbows and a main pipe. Both riser configurations are cylindrical and have the same diameter but differ in their inlet lengths and main pipe height to enable investigation of riser geometrical scaling effects. In addition, two types of solid particles are exploited in the solid phase of the two-phase gas-solid riser flow simulations to study the influence of solid loading ratio on flow patterns. The gaseous phase in the two-phase flow is represented by standard atmospheric air. The CFD-based FLUENT software is employed to obtain steady state and transient solutions for flow modulations in the riser. The physical dimensions, types and numbers of computation meshes, and solution methodology utilized in the present work are stated. Flow parameters, such as static and dynamic pressure, species velocity, and volume fractions are monitored and analyzed. The differences in the computational results between the two models, under steady and transient conditions, are compared, contrasted, and discussed.
Dang, Zhe
Bond Computing Systems: a Biologically Inspired and High-level Dynamics Model for Pervasive. Targeting at modeling the high-level dynamics of pervasive comput- ing systems, we introduce Bond Computing are regular, and study their computation power and verification problems. Among other results, we show
7. Business Models LearningsfromfoundingaComputerVisionStartup
Solem, Jan Erik
7. Business Models #12;LearningsfromfoundingaComputerVisionStartup Flickr:dystopos How are you models ! ! (not only technology) #12;LearningsfromfoundingaComputerVisionStartup Auction business model! Bricks and clicks business model! Collective business models! Component business model! Cutting out
Modeling-Computer Simulations At San Juan Volcanic Field Area...
Area Exploration Technique Modeling-Computer Simulations Activity Date Usefulness useful DOE-funding Unknown Notes In this study we combine thermal maturation models, based on the...
Computational models of intergroup competition and warfare.
Letendre, Kenneth (University of New Mexico); Abbott, Robert G.
2011-11-01
This document reports on the research of Kenneth Letendre, the recipient of a Sandia Graduate Research Fellowship at the University of New Mexico. Warfare is an extreme form of intergroup competition in which individuals make extreme sacrifices for the benefit of their nation or other group to which they belong. Among animals, limited, non-lethal competition is the norm. It is not fully understood what factors lead to warfare. We studied the global variation in the frequency of civil conflict among countries of the world, and its positive association with variation in the intensity of infectious disease. We demonstrated that the burden of human infectious disease importantly predicts the frequency of civil conflict and tested a causal model for this association based on the parasite-stress theory of sociality. We also investigated the organization of social foraging by colonies of harvester ants in the genus Pogonomyrmex, using both field studies and computer models.
Robust Resource Allocations in Parallel Computing Systems: Model and Heuristics
Maciejewski, Anthony A. "Tony"
Robust Resource Allocations in Parallel Computing Systems: Model and Heuristics Vladimir Shestak1 in parallel computer systems (including heterogeneous clusters) should be allocated to the computational was supported by the Colorado State University Center for Robustness in Computer Systems (funded by the Colorado
Dang, Zhe
Bond Computing Systems: a Biologically Inspired and High-level Dynamics Model for Pervasive their com- putation power and verification problems. Among other results, we show that the computing power) techniques for pervasive computing systems. At a high-level, there are at least two views in modeling
Theoretical Neuroscience: Computational and Mathematical Modeling of Neural Systems
Olshausen, Bruno
Theoretical Neuroscience: Computational and Mathematical Modeling of Neural Systems Peter Dayan: Computational and Mathematical Modeling of Neural Systems. The emergence of this book represents more than- scientist was brought up on ``Kandel and Schwartz.'' Now, at last, the field of computational neuroscience
What can computational models tell us about face processing?
Cottrell, Garrison W.
What can computational models tell us about face processing? Garrison W. Cottrell Gary about face processing? Garrison W. Cottrell Gary's Unbelievable Research Unit (GURU) Computer Science, Lingyun Zhang What can computational models tell us about face processing? Garrison W. Cottrell Gary
Conrad, Patrick Raymond
2014-01-01
Computational models of complex phenomena are an important resource for scientists and engineers. However, many state-of-the-art simulations of physical systems are computationally expensive to evaluate and are black ...
Cummings, P. T.
2010-02-08
This document reports the outcomes of the Computational Nanoscience Project, "Integrated Multiscale Modeling of Molecular Computing Devices". It includes a list of participants and publications arising from the research supported.
Computational systems biology and in silico modeling of the
Borenstein, Elhanan
Computational systems biology and in silico modeling of the human microbiome Elhanan Borenstein Professor at the Santa Fe Institute. His research interests include computational and evolutionary systems is a complex biological system with numerous interacting components across multiple organizational levels
Integration of engineering models in computer-aided preliminary design
Lajoie, Ronnie M.
The problems of the integration of engineering models in computer-aided preliminary design are reviewed. This paper details the research, development, and testing of modifications to Paper Airplane, a LISP-based computer ...
Determinant Formulas for Matrix Model Free Energy
D. Vasiliev
2005-07-11
The paper contains a new non-perturbative representation for subleading contribution to the free energy of multicut solution for hermitian matrix model. This representation is a generalisation of the formula, proposed by Klemm, Marino and Theisen for two cut solution, which was obtained by comparing the cubic matrix model with the topological B-model on the local Calabi-Yau geometry $\\hat {II}$ and was checked perturbatively. In this paper we give a direct proof of their formula and generalise it to the general multicut solution.
Predictive Capability Maturity Model for computational modeling and simulation.
Oberkampf, William Louis; Trucano, Timothy Guy; Pilch, Martin M.
2007-10-01
The Predictive Capability Maturity Model (PCMM) is a new model that can be used to assess the level of maturity of computational modeling and simulation (M&S) efforts. The development of the model is based on both the authors experience and their analysis of similar investigations in the past. The perspective taken in this report is one of judging the usefulness of a predictive capability that relies on the numerical solution to partial differential equations to better inform and improve decision making. The review of past investigations, such as the Software Engineering Institute's Capability Maturity Model Integration and the National Aeronautics and Space Administration and Department of Defense Technology Readiness Levels, indicates that a more restricted, more interpretable method is needed to assess the maturity of an M&S effort. The PCMM addresses six contributing elements to M&S: (1) representation and geometric fidelity, (2) physics and material model fidelity, (3) code verification, (4) solution verification, (5) model validation, and (6) uncertainty quantification and sensitivity analysis. For each of these elements, attributes are identified that characterize four increasing levels of maturity. Importantly, the PCMM is a structured method for assessing the maturity of an M&S effort that is directed toward an engineering application of interest. The PCMM does not assess whether the M&S effort, the accuracy of the predictions, or the performance of the engineering system satisfies or does not satisfy specified application requirements.
Computer Virus Propagation Models Giuseppe Serazzi and Stefano Zanero
Zanero, Stefano
Computer Virus Propagation Models Giuseppe Serazzi and Stefano Zanero Dipartimento di Elettronica e.zanero@polimi.it Abstract. The availability of reliable models of computer virus propa- gation would prove useful. In this pa- per, we review the most popular models of virus propagation, analyzing the underlying assumptions
Department of Computing CSP||B modelling for railway verification
Doran, Simon J.
University of Surrey Department of Computing Computing Sciences Report CS-12-03 CSP||B modelling Schneider Helen Treharne March 30th 2012 #12;CSP||B modelling for railway verification: the double junction work in verifying railway systems through CSP k B modelling and analysis. In particular we consider
Computational modeling of composite material fires.
Brown, Alexander L.; Erickson, Kenneth L.; Hubbard, Joshua Allen; Dodd, Amanda B.
2010-10-01
Composite materials behave differently from conventional fuel sources and have the potential to smolder and burn for extended time periods. As the amount of composite materials on modern aircraft continues to increase, understanding the response of composites in fire environments becomes increasingly important. An effort is ongoing to enhance the capability to simulate composite material response in fires including the decomposition of the composite and the interaction with a fire. To adequately model composite material in a fire, two physical model development tasks are necessary; first, the decomposition model for the composite material and second, the interaction with a fire. A porous media approach for the decomposition model including a time dependent formulation with the effects of heat, mass, species, and momentum transfer of the porous solid and gas phase is being implemented in an engineering code, ARIA. ARIA is a Sandia National Laboratories multiphysics code including a range of capabilities such as incompressible Navier-Stokes equations, energy transport equations, species transport equations, non-Newtonian fluid rheology, linear elastic solid mechanics, and electro-statics. To simulate the fire, FUEGO, also a Sandia National Laboratories code, is coupled to ARIA. FUEGO represents the turbulent, buoyantly driven incompressible flow, heat transfer, mass transfer, and combustion. FUEGO and ARIA are uniquely able to solve this problem because they were designed using a common architecture (SIERRA) that enhances multiphysics coupling and both codes are capable of massively parallel calculations, enhancing performance. The decomposition reaction model is developed from small scale experimental data including thermogravimetric analysis (TGA) and Differential Scanning Calorimetry (DSC) in both nitrogen and air for a range of heating rates and from available data in the literature. The response of the composite material subject to a radiant heat flux boundary condition is examined to study the propagation of decomposition fronts of the epoxy and carbon fiber and their dependence on the ambient conditions such as oxygen concentration, surface flow velocity, and radiant heat flux. In addition to the computational effort, small scaled experimental efforts to attain adequate data used to validate model predictions is ongoing. The goal of this paper is to demonstrate the progress of the capability for a typical composite material and emphasize the path forward.
Disruptive technology business models in cloud computing
Krikos, Alexis Christopher
2010-01-01
Cloud computing, a term whose origins have been in existence for more than a decade, has come into fruition due to technological capabilities and marketplace demands. Cloud computing can be defined as a scalable and flexible ...
On the computation of steady hopper flows I: stress determination for Coulomb materials.1
importance for many manufacturing industries, where vast quantities of raw materials are stored and handledOn the computation of steady hopper flows I: stress determination for Coulomb materials.1 Pierre A@math.ncsu.edu,jvmatthe@eos.ncsu.edu The problem of determining the steady state flow of granular materials in silos under the action of gravity
On the computation of steady hopper flows I: stress determination for Coulomb materials. 1
importance for many manufacturing industries, where vast quantities of raw materials are stored and handledOn the computation of steady hopper flows I: stress determination for Coulomb materials. 1 Pierre A@eos.ncsu.edu The problem of determining the steady state flow of granular materials in silos under the action of gravity
Modeling-Computer Simulations At Nevada Test And Training Range...
Modeling-Computer Simulations At Nevada Test And Training Range Area (Sabin, Et Al., 2004) Jump to: navigation, search GEOTHERMAL ENERGYGeothermal Home Exploration Activity:...
Modeling-Computer Simulations At Fenton Hill HDR Geothermal Area...
Modeling-Computer Simulations At Fenton Hill HDR Geothermal Area (Brown & DuTeaux, 1997) Jump to: navigation, search GEOTHERMAL ENERGYGeothermal Home Exploration Activity:...
Modeling-Computer Simulations At Walker-Lane Transitional Zone...
Pritchett, 2004) Exploration Activity Details Location Walker-Lane Transition Zone Geothermal Region Exploration Technique Modeling-Computer Simulations Activity Date Usefulness...
New partnership uses advanced computer science modeling to address...
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
partnership uses advanced computer science modeling to address climate change | National Nuclear Security Administration Facebook Twitter Youtube Flickr RSS People Mission Managing...
COMPUTER MODELING OF NUCLIDE ADSORPTION ON GEOLOGIC MATERIALS
Silva, R.J.
2010-01-01
11899 G COMPUTER MODELING OF NUCLIDE ADSORPTION ON GEOLOGICdefined as the amount of nuclide adsorbed per gram of claydivided by the amount of nuclide per milliliter of solution,
Modeling-Computer Simulations At Central Nevada Seismic Zone...
Modeling-Computer Simulations At Central Nevada Seismic Zone Region (Pritchett, 2004) Jump to: navigation, search GEOTHERMAL ENERGYGeothermal Home Exploration Activity:...
Modeling-Computer Simulations At Central Nevada Seismic Zone...
Modeling-Computer Simulations At Central Nevada Seismic Zone Region (Biasi, Et Al., 2009) Jump to: navigation, search GEOTHERMAL ENERGYGeothermal Home Exploration Activity:...
MaRIE theory, modeling and computation roadmap executive summary...
Office of Scientific and Technical Information (OSTI)
discovery, with theory and high performance computing, itself co-designed by constrained optimization of hardware and software, and experiments. MaRIE's theory, modeling, and...
Modeling-Computer Simulations At Stillwater Area (Wisian & Blackwell...
Exploration Activity Details Location Stillwater Area Exploration Technique Modeling-Computer Simulations Activity Date Usefulness not indicated DOE-funding Unknown References...
Modeling-Computer Simulations At Desert Peak Area (Wisian & Blackwell...
Exploration Activity Details Location Desert Peak Area Exploration Technique Modeling-Computer Simulations Activity Date Usefulness not indicated DOE-funding Unknown References...
Grossmann, Ignacio E.
1 Computational Strategies for Large-Scale MILP Transshipment Models for Heat Exchanger Network Determining the minimum number of units is an important step in heat exchanger network synthesis (HENS Words heat exchanger network synthesis (HENS), transshipment model, mixed-integer linear programming
Parameter Discovery for Stochastic Computational Models in Systems Biology Using Bayesian Model
Parameter Discovery for Stochastic Computational Models in Systems Biology Using Bayesian Model--Parameterized probabilistic complex computational (P2 C2 ) models are increasingly used in computational systems biology to study biochemical and physiological systems. A key challenge is to build mechanistic P2 C2 models
"Creating computational models of biological systems to better combat
Zhigilei, Leonid V.
"Creating computational models of biological systems to better combat dangerous pathogens and human of Biomedical Engineering University of Virginia Charlottesville, VA 434.924.8195 Computational Systems Biology system in biofuel and nutraceutical production. With the aid of computational techniques, we can predict
A framework for modelling trojans and computer virus infection
Cairns, Paul
A framework for modelling trojans and computer virus infection Harold Thimbleby1 , Stuart Anderson2 world, including the possibility of Trojan Horse programs and computer viruses, as simply a finite realisation of a Turing Machine. We consider the actions of Trojan Horses and viruses in real computer systems
Gering, Kevin L.
2013-01-01
A system includes an electrochemical cell, monitoring hardware, and a computing system. The monitoring hardware samples performance characteristics of the electrochemical cell. The computing system determines cell information from the performance characteristics. The computing system also analyzes the cell information of the electrochemical cell with a Butler-Volmer (BV) expression modified to determine exchange current density of the electrochemical cell by including kinetic performance information related to pulse-time dependence, electrode surface availability, or a combination thereof. A set of sigmoid-based expressions may be included with the modified-BV expression to determine kinetic performance as a function of pulse time. The determined exchange current density may be used with the modified-BV expression, with or without the sigmoid expressions, to analyze other characteristics of the electrochemical cell. Model parameters can be defined in terms of cell aging, making the overall kinetics model amenable to predictive estimates of cell kinetic performance along the aging timeline.
Model independent determination of the {sigma} pole
Leutwyler, H.
2008-08-31
The first part of this report reviews recent developments at the interface between lattice work on QCD with light dynamical quarks, effective field theory and low energy precision experiments. Then I discuss how dispersion theory can be used to analyze the low energy structure of the {pi}{pi} scattering amplitude in a model independent manner. This leads to an exact formula for the mass and width of the lowest few resonances, in terms of observable quantities. As an application, I consider the pole position of the {sigma}, paying particular to error propagation in the numerical analysis. The report is based on work done in collaboration with Irinel Caprini and Gilberto Colangelo.
Bytecode unification of geospatial computable models Bytecode unification of geospatial
KÃ¶bben, Barend
heterogeneous to fix and reuse. Field-based and objects-based geospatial models of- ten share common GIS data and object-based data models, and other challenges re- garding synergy of geospatial systems that need to use both types of data models. Keywords: field-based model, object-based model, computability, managed
Improvements in fast-response flood modeling: desktop parallel computing and domain tracking
Judi, David R; Mcpherson, Timothy N; Burian, Steven J
2009-01-01
It is becoming increasingly important to have the ability to accurately forecast flooding, as flooding accounts for the most losses due to natural disasters in the world and the United States. Flood inundation modeling has been dominated by one-dimensional approaches. These models are computationally efficient and are considered by many engineers to produce reasonably accurate water surface profiles. However, because the profiles estimated in these models must be superimposed on digital elevation data to create a two-dimensional map, the result may be sensitive to the ability of the elevation data to capture relevant features (e.g. dikes/levees, roads, walls, etc...). Moreover, one-dimensional models do not explicitly represent the complex flow processes present in floodplains and urban environments and because two-dimensional models based on the shallow water equations have significantly greater ability to determine flow velocity and direction, the National Research Council (NRC) has recommended that two-dimensional models be used over one-dimensional models for flood inundation studies. This paper has shown that two-dimensional flood modeling computational time can be greatly reduced through the use of Java multithreading on multi-core computers which effectively provides a means for parallel computing on a desktop computer. In addition, this paper has shown that when desktop parallel computing is coupled with a domain tracking algorithm, significant computation time can be eliminated when computations are completed only on inundated cells. The drastic reduction in computational time shown here enhances the ability of two-dimensional flood inundation models to be used as a near-real time flood forecasting tool, engineering, design tool, or planning tool. Perhaps even of greater significance, the reduction in computation time makes the incorporation of risk and uncertainty/ensemble forecasting more feasible for flood inundation modeling (NRC 2000; Sayers et al. 2000).
Towards Real Earth Models --Computational Geophysics on Unstructured Tetrahedral Meshes?
Farquharson, Colin G.
Towards Real Earth Models -- Computational Geophysics on Unstructured Tetrahedral Meshes? Colin tetrahedral meshes. EM geophysics on unstructured tetrahedral meshes. Disadvantages, difficulties, challenges. Conclusions. #12;Outline: Geological models! Advantages of unstructured tetrahedral meshes. EM geophysics
Applying High Performance Computing to Analyzing by Probabilistic Model Checking
Schneider, Carsten
Applying High Performance Computing to Analyzing by Probabilistic Model Checking Mobile Cellular on the use of high performance computing in order to analyze with the proba- bilistic model checker PRISM. The Figure Generation Script 22 2 #12;1. Introduction We report in this paper on the use of high performance
Overview of ASC Capability Computing System Governance Model
Doebling, Scott W. [Los Alamos National Laboratory
2012-07-11
This document contains a description of the Advanced Simulation and Computing Program's Capability Computing System Governance Model. Objectives of the Governance Model are to ensure that the capability system resources are allocated on a priority-driven basis according to the Program requirements; and to utilize ASC Capability Systems for the large capability jobs for which they were designed and procured.
An Interactive Computer Model of Two-Country Trade
Hamlen, Kevin W.
country when free trade is made available. One of the most important learning lessons for the students91 An Interactive Computer Model of Two-Country Trade Bill Hamlen and Kevin Hamlen Abstract We introduce an interactive computer model of two-country trade that allows students to investigate
Los Alamos CCS (Center for Computer Security) formal computer security model
Dreicer, J.S.; Hunteman, W.J. (Los Alamos National Lab., NM (USA))
1989-01-01
This paper provides a brief presentation of the formal computer security model currently being developed at the Los Alamos Department of Energy (DOE) Center for Computer Security (CCS). The initial motivation for this effort was the need to provide a method by which DOE computer security policy implementation could be tested and verified. The actual analytical model was a result of the integration of current research in computer security and previous modeling and research experiences. The model is being developed to define a generic view of the computer and network security domains, to provide a theoretical basis for the design of a security model, and to address the limitations of present models. Formal mathematical models for computer security have been designed and developed in conjunction with attempts to build secure computer systems since the early 70's. The foundation of the Los Alamos DOE CCS model is a series of functionally dependent probability equations, relations, and expressions. The mathematical basis appears to be justified and is undergoing continued discrimination and evolution. We expect to apply the model to the discipline of the Bell-Lapadula abstract sets of objects and subjects. 5 refs.
Gering, Kevin L
2013-08-27
A system includes an electrochemical cell, monitoring hardware, and a computing system. The monitoring hardware periodically samples performance characteristics of the electrochemical cell. The computing system determines cell information from the performance characteristics of the electrochemical cell. The computing system also develops a mechanistic level model of the electrochemical cell to determine performance fade characteristics of the electrochemical cell and analyzing the mechanistic level model to estimate performance fade characteristics over aging of a similar electrochemical cell. The mechanistic level model uses first constant-current pulses applied to the electrochemical cell at a first aging period and at three or more current values bracketing a first exchange current density. The mechanistic level model also is based on second constant-current pulses applied to the electrochemical cell at a second aging period and at three or more current values bracketing the second exchange current density.
Error models in quantum computation: an application of model selection
Lucia Schwarz; Steven van Enk
2013-09-04
Threshold theorems for fault-tolerant quantum computing assume that errors are of certain types. But how would one detect whether errors of the "wrong" type occur in one's experiment, especially if one does not even know what type of error to look for? The problem is that for many qubits a full state description is impossible to analyze, and a full process description is even more impossible to analyze. As a result, one simply cannot detect all types of errors. Here we show through a quantum state estimation example (on up to 25 qubits) how to attack this problem using model selection. We use, in particular, the Akaike Information Criterion. The example indicates that the number of measurements that one has to perform before noticing errors of the wrong type scales polynomially both with the number of qubits and with the error size.
Grujicic, Mica
Computational Modeling of theComputational Modeling of the Vacuum Assisted Resin Transfer MoldingVacuum Assisted Resin Transfer Molding (VARTM) Process(VARTM) Process April 2004April 2004 DepartmentMS Thesis Advisor: Dr. Grujicic #12;What is VARTM?What is VARTM? Vacuum Assisted Resin Transfer Molding
Computing the Electricity Market Equilibrium: Uses of market equilibrium models
Baldick, Ross
1 Computing the Electricity Market Equilibrium: Uses of market equilibrium models Ross Baldick Abstract--In this paper we consider the formulation and uses of electric- ity market equilibrium models. Keywords--Electricity market, Equilibrium models I. INTRODUCTION Electricity market equilibrium modelling
ComputerScience,TexasA&MUniversity Modeling Heterogeneous User
Loguinov, Dmitri
1 ComputerScience,TexasA&MUniversity Modeling Heterogeneous User Churn and Local Resilience of Unstructured P2P Networks Modeling Heterogeneous UserModeling Heterogeneous User Churn and Local Resilience ofChurn Churn and Local Resilience of Unstructured P2P Networks Modeling Heterogeneous User
Integrating Numerical Computation into the Modeling Instruction Curriculum
Caballero, Marcos D; Aiken, John M; Douglas, Scott S; Scanlon, Erin M; Thoms, Brian; Schatz, Michael F
2012-01-01
We describe a way to introduce physics high school students with no background in programming to computational problem-solving experiences. Our approach builds on the great strides made by the Modeling Instruction reform curriculum. This approach emphasizes the practices of "Developing and using models" and "Computational thinking" highlighted by the NRC K-12 science standards framework. We taught 9th-grade students in a Modeling-Instruction-based physics course to construct computational models using the VPython programming environment. Numerical computation within the Modeling Instruction curriculum provides coherence among the curriculum's different force and motion models, links the various representations which the curriculum employs, and extends the curriculum to include real-world problems that are inaccessible to a purely analytic approach.
Ambient temperature modelling with soft computing techniques
Bertini, Ilaria; Ceravolo, Francesco; Citterio, Marco; Di Pietra, Biagio; Margiotta, Francesca; Pizzuti, Stefano; Puglisi, Giovanni; De Felice, Matteo
2010-07-15
This paper proposes a hybrid approach based on soft computing techniques in order to estimate monthly and daily ambient temperature. Indeed, we combine the back-propagation (BP) algorithm and the simple Genetic Algorithm (GA) in order to effectively train artificial neural networks (ANN) in such a way that the BP algorithm initialises a few individuals of the GA's population. Experiments concerned monthly temperature estimation of unknown places and daily temperature estimation for thermal load computation. Results have shown remarkable improvements in accuracy compared to traditional methods. (author)
Breitling, Rainer
, Continuous Petri nets Computational modelling of intracellular biochemical networks has become a growth topic and analysis of such networks, as well as an increase in the quality and amount of experimentally determined such as bifurcations, robustness to interference, or oscillations are not obvious from the network topology
Computational Modeling and Optimization of Proton Exchange Membrane Fuel Cells
Victoria, University of
Computational Modeling and Optimization of Proton Exchange Membrane Fuel Cells by Marc Secanell and Optimization of Proton Exchange Membrane Fuel Cells by Marc Secanell Gallart Bachelor in Engineering cells. In this thesis, a computational framework for fuel cell analysis and optimization is presented
Clique-detection Models in Computational Biochemistry and Genomics
Butenko, Sergiy
Clique-detection Models in Computational Biochemistry and Genomics S. Butenko and W. E. Wilhelm,wilhelm}@tamu.edu Abstract Many important problems arising in computational biochemistry and genomics have been formulated and genomic aspects of the problems as well as to the graph-theoretic aspects of the solution approaches. Each
Inverse Modelling in Geology by Interactive Evolutionary Computation
Boschetti, Fabio
Inverse Modelling in Geology by Interactive Evolutionary Computation Chris Wijns a,b,, Fabio of geological processes, in the absence of established numerical criteria to act as inversion targets, requires evolutionary computation provides for the inclusion of qualitative geological expertise within a rigorous
Computational load in model physics of the parallel NCAR community climate model
Michalakes, J.G.; Nanjundiah, R.S.
1994-11-01
Maintaining a balance of computational load over processors is a crucial issue in parallel computing. For efficient parallel implementation, complex codes such as climate models need to be analyzed for load imbalances. In the present study we focus on the load imbalances in the physics portion of the community climate model`s (CCM2) distributed-memory parallel implementation on the Intel Touchstone DELTA computer. We note that the major source of load imbalance is the diurnal variation in the computation of solar radiation. Convective weather patterns also cause some load imbalance. Land-ocean contrast is seen to have little effect on computational load in the present version of the model.
COMPUTATIONAL FLUID DYNAMICS MODELING OF SOLID OXIDE FUEL CELLS
COMPUTATIONAL FLUID DYNAMICS MODELING OF SOLID OXIDE FUEL CELLS Ugur Pasaogullari and Chao-dimensional model has been developed to simulate solid oxide fuel cells (SOFC). The model fully couples current density operation. INTRODUCTION Solid oxide fuel cells (SOFC) are among possible candidates
Quantum Computers: Noise Propagation and Adversarial Noise Models
Kalai, Gil
Quantum Computers: Noise Propagation and Adversarial Noise Models Gil Kalai Hebrew University of Jerusalem and Yale University April 21, 2009 Abstract In this paper we consider adversarial noise models." Detrimental noise is modeled after familiar properties of noise propagation. However, it can have various
Modeling-Computer Simulations At Valles Caldera - Sulphur Springs...
DOE-funding Unknown Notes A computer program capable of two-dimensional modeling of gravity data was used in interpreting gravity observations along profiles A--A' and B--B'...
Modeling-Computer Simulations At Valles Caldera - Redondo Geothermal...
DOE-funding Unknown Notes A computer program capable of two-dimensional modeling of gravity data was used in interpreting gravity observations along profiles A--A' and B--B'...
Computational Model of Forward and Opposed Smoldering Combustion in Microgravity
Rein, Guillermo; Fernandez-Pello, Carlos; Urban, David
2006-08-06
A novel computational model of smoldering combustion capable of predicting both forward and opposed propagation is developed. This is accomplished by considering the one-dimensional, transient, governing equations for ...
Language acquisition and implication for language change: A computational model.
Clark, Robert A J
1997-01-01
Computer modeling techniques, when applied to language acquisition problems, give an often unrealized insight into the diachronic change that occurs in language over successive generations. This paper shows that using ...
Computer simulation and topological modeling of radiation effects in zircon
Zhang, Yi, 1979-
2006-01-01
The purpose of this study is to understand on atomic level the structural response of zircon (ZrSiO4) to irradiation using molecular dynamics (MD) computer simulations, and to develop topological models that can describe ...
15.094 Systems Optimization: Models and Computation, Spring 2002
Freund, Robert Michael
A computational and application-oriented introduction to the modeling of large-scale systems in a wide variety of decision-making domains and the optimization of such systems using state-of-the-art optimization software. ...
Computational Modeling for the American Chemical Society | GE...
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
Computational Modeling for the American Chemical Society Click to email this to a friend (Opens in new window) Share on Facebook (Opens in new window) Click to share (Opens in new...
The Nuclear Energy Advanced Modeling and Simulation Enabling Computational Technologies FY09 Report
Diachin, L F; Garaizar, F X; Henson, V E; Pope, G
2009-10-12
In this document we report on the status of the Nuclear Energy Advanced Modeling and Simulation (NEAMS) Enabling Computational Technologies (ECT) effort. In particular, we provide the context for ECT In the broader NEAMS program and describe the three pillars of the ECT effort, namely, (1) tools and libraries, (2) software quality assurance, and (3) computational facility (computers, storage, etc) needs. We report on our FY09 deliverables to determine the needs of the integrated performance and safety codes (IPSCs) in these three areas and lay out the general plan for software quality assurance to meet the requirements of DOE and the DOE Advanced Fuel Cycle Initiative (AFCI). We conclude with a brief description of our interactions with the Idaho National Laboratory computer center to determine what is needed to expand their role as a NEAMS user facility.
Computational models of early visual processing layers
Shan, Honghao
2010-01-01
prevailing view of retinal processing. However, as discussedsimplified) model of retinal processing. A Retinal Codingretinal coding, the pre-cortical stage of visual processing,
Sandia Energy - Computational Modeling & Simulation
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
AFDC Printable Version Share this resource Send a link to EERE: Alternative Fuels Data Center Home Page to someone by E-mail Share EERE: Alternative Fuels Data Center Home Page on Facebook Tweet about EERE: Alternative Fuels Data Center Home Page on Twitter Bookmark EERE: Alternative Fuels Data Center Homesum_a_epg0_fpd_mmcf_m.xls" ,"Available from WebQuantityBonneville Power AdministrationRobust, High-Throughput Analysis ofSample SULIColin Humphreys Home Colin HumphreysComputational
A computer program to determine the specific power of prismatic-core reactors
Dobranich, D.
1987-05-01
A computer program has been developed to determine the maximum specific power for prismatic-core reactors as a function of maximum allowable fuel temperature, core pressure drop, and coolant velocity. The prismatic-core reactors consist of hexagonally shaped fuel elements grouped together to form a cylindrically shaped core. A gas coolant flows axially through circular channels within the elements, and the fuel is dispersed within the solid element material either as a composite or in the form of coated pellets. Different coolant, fuel, coating, and element materials can be selected to represent different prismatic-core concepts. The computer program allows the user to divide the core into any arbitrary number of axial levels to account for different axial power shapes. An option in the program allows the automatic determination of the core height that results in the maximum specific power. The results of parametric specific power calculations using this program are presented for various reactor concepts.
Discrete transfinite computation models School of Mathematics,
Welch, Philip
such as Davies [Davies (2001)], and the models of Beggs and Tucker [Beggs and Tucker (2007)] that attempt' functions or oracles, such as is done in [Beggs et al. (2008)]. We also shall not particularly consider
CARD No. 23 Models and Computer Codes
that are used to demonstrate that the WIPP will comply with the radioactive waste disposal regulations at 40 CFR, and radionuclide transport in the repository and overlying rock formations. Numerical models are then created
Computational model of miniature pulsating heat pipes.
Martinez, Mario J.; Givler, Richard C.
2013-01-01
The modeling work described herein represents Sandia National Laboratories (SNL) portion of a collaborative three-year project with Northrop Grumman Electronic Systems (NGES) and the University of Missouri to develop an advanced, thermal ground-plane (TGP), which is a device, of planar configuration, that delivers heat from a source to an ambient environment with high efficiency. Work at all three institutions was funded by DARPA/MTO; Sandia was funded under DARPA/MTO project number 015070924. This is the final report on this project for SNL. This report presents a numerical model of a pulsating heat pipe, a device employing a two phase (liquid and its vapor) working fluid confined in a closed loop channel etched/milled into a serpentine configuration in a solid metal plate. The device delivers heat from an evaporator (hot zone) to a condenser (cold zone). This new model includes key physical processes important to the operation of flat plate pulsating heat pipes (e.g. dynamic bubble nucleation, evaporation and condensation), together with conjugate heat transfer with the solid portion of the device. The model qualitatively and quantitatively predicts performance characteristics and metrics, which was demonstrated by favorable comparisons with experimental results on similar configurations. Application of the model also corroborated many previous performance observations with respect to key parameters such as heat load, fill ratio and orientation.
Computational Fluid Dynamics (CFD) Modelling on Soot Yield for Fire
Computational Fluid Dynamics (CFD) Modelling on Soot Yield for Fire Engineering Assessment Yong S (CFD) Modelling is now widely used by fire safety engineers throughout the world as a tool of the smoke control design as part of the performance based fire safety design in the current industry
Computational Modeling of Chord Fingering for String Instruments
Radicioni, Daniele
Computational Modeling of Chord Fingering for String Instruments Daniele P. Radicioni (radicion for the fingering process with string instruments, based on a constraint satisfaction approach. The model is imple-mechanical aspects of the performer's hand in its interaction with the musical instrument. Introduction Music
Personal Computer-Based Model for Cool Storage Performance Simulation
Kasprowicz, L. M.; Jones, J. W.; Hitzfelder, J.
1990-01-01
A personal computer based hourly simulation model was developed based on the CBS/ICE routines in the DOE-2.1 mainframe building simulation software. The menu driven new model employs more efficient data and information handling than the previous...
Methodology for characterizing modeling and discretization uncertainties in computational simulation
ALVIN,KENNETH F.; OBERKAMPF,WILLIAM L.; RUTHERFORD,BRIAN M.; DIEGERT,KATHLEEN V.
2000-03-01
This research effort focuses on methodology for quantifying the effects of model uncertainty and discretization error on computational modeling and simulation. The work is directed towards developing methodologies which treat model form assumptions within an overall framework for uncertainty quantification, for the purpose of developing estimates of total prediction uncertainty. The present effort consists of work in three areas: framework development for sources of uncertainty and error in the modeling and simulation process which impact model structure; model uncertainty assessment and propagation through Bayesian inference methods; and discretization error estimation within the context of non-deterministic analysis.
Computing Biological Model Parameters by Parallel Statistical Model Checking
Tronci, Enrico
of Treatments for Infertility Related Endocrinological Diseases, 600773). #12;patient-specific model parameters
Models of quantum computation and quantum programming languages
J. A. Miszczak
2011-12-03
The goal of the presented paper is to provide an introduction to the basic computational models used in quantum information theory. We review various models of quantum Turing machine, quantum circuits and quantum random access machine (QRAM) along with their classical counterparts. We also provide an introduction to quantum programming languages, which are developed using the QRAM model. We review the syntax of several existing quantum programming languages and discuss their features and limitations.
Quantum Computers: Noise Propagation and Adversarial Noise Models
Gil Kalai
2009-04-21
In this paper we consider adversarial noise models that will fail quantum error correction and fault-tolerant quantum computation. We describe known results regarding high-rate noise, sequential computation, and reversible noisy computation. We continue by discussing highly correlated noise and the "boundary," in terms of correlation of errors, of the "threshold theorem." Next, we draw a picture of adversarial forms of noise called (collectively) "detrimental noise." Detrimental noise is modeled after familiar properties of noise propagation. However, it can have various causes. We start by pointing out the difference between detrimental noise and standard noise models for two qubits and proceed to a discussion of highly entangled states, the rate of noise, and general noisy quantum systems.
District-heating strategy model: computer programmer's manual
Kuzanek, J.F.
1982-05-01
The US Department of Housing and Urban Development (HUD) and the US Department of Energy (DOE) cosponsor a program aimed at increasing the number of district heating and cooling (DHC) systems. Such systems can reduce the amount and costs of fuels used to heat and cool buildings in a district. Twenty-eight communities have agreed to aid HUD in a national feasibility assessment of DHC systems. The HUD/DOE program entails technical assistance by Argonne National Laboratory and Oak Ridge National Laboratory. The assistance includes a computer program, called the district heating strategy model (DHSM), that performs preliminary calculations to analyze potential DHC systems. This report describes the general capabilities of the DHSM, provides historical background on its development, and explains the computer installation and operation of the model - including the data file structures and the options. Sample problems illustrate the structure of the various input data files, the interactive computer-output listings. The report is written primarily for computer programmers responsible for installing the model on their computer systems, entering data, running the model, and implementing local modifications to the code.
A gas kick model for the personal computer
Miller, Clayton Lowell
1987-01-01
A GAS KICK MODEL FOR THE PERSONAL COMPUTER A Thesis by CLAYTON LOWELL MILLER Submitted to the Graduate College of Texas A6M University in partial fulfillment of the requirements for the degree of MASTER OF SCIENCE May 1987 Major Subject...: Petroleum Engineering A GAS KICK MODEL FOR THE PERSONAL COMPUTER A Thesis by CLAYTON LOWELL MILLER Approved as to style and content by: Wana C. vkam-Wold (Chair f Committee) Robert W. Heine (Member) Tibor G. Rozg yi (Member) W. D. Von Gonten Head...
Concurrent multiscale computational modeling for dense dry granular materials interfacing
Regueiro, Richard A.
of interfacial mechanics between granular soil and tire, tool, or penetrometer, while properly representing far computational modeling of interfacial mechanics between granular materials and deformable solid bodies, agricultural grains (in silo flows), dry soils (sand, silt, gravel), and lunar and martian regolith (soil found
ADVANCED WUFI COMPUTER MODELING WORKSHOP FOR WALL DESIGN AND PERFORMANCE
Oak Ridge National Laboratory
TRANSFER IN BUILDING ENVELOPES) Napa, CA, January 30 - February 1, 2012 WUFI/ORNL1 Program made availableADVANCED WUFI COMPUTER MODELING WORKSHOP FOR WALL DESIGN AND PERFORMANCE (HEAT AND MOISTURE-Institut für Bauphysik (IBP) and co-sponsored by the National Institute of Building Sciences (NIBS)/ Building
Computational model of aortic valve surgical repair using grafted pericardium
1 Computational model of aortic valve surgical repair using grafted pericardium Peter E. Hammer1, aortic valve repair, membrane, surgical planning, leaflet graft, pericardium ABSTRACT Aortic valve leaflets. Difficulty is largely due to the complex geometry and function of the valve and the lower
A Computational Model for Sound Field Absorption by Acoustic Arrays
. We then formulate the acoustic wave equation with the absorption boundary coeÆcient in the frequency the sound absorption property of arrays of micro-acoustic actuators at a control surface. We use the waveA Computational Model for Sound Field Absorption by Acoustic Arrays H. T. Banks #3; D. G. Cole z K
Faceted Models of Blog Feeds Department of Computer
Meng, Weiyi
Faceted Models of Blog Feeds Lifeng Jia Department of Computer Science University of Illinois@cs.binghamton.edu ABSTRACT Faceted blog distillation aims at retrieving the blogs that are not only relevant to a query blogs depict various topics related to the personal experiences of bloggers while official blogs deliver
Open Learner Models at Birmingham Electronic, Electrical and Computer Engineering
Bull, Susan
and Engineering Principles and Methods EE2D2 Introduction to Communications EE2F1 Speech and Audio Technology EE2G EE1B2 Circuits, Devices and Fields EE1E1&2 C Programming EE1F1 Introduction to InformationOpen Learner Models at Birmingham Electronic, Electrical and Computer Engineering University
Modeling and Computational Strategies for Optimal Development Planning of Offshore
Grossmann, Ignacio E.
1 Modeling and Computational Strategies for Optimal Development Planning of Offshore Oilfields for offshore oil and gas fields as a basis to include the generic fiscal rules with ringfencing provisions-integer programming. 1 Introduction Offshore oil and gas field development planning has received significant attention
innovati nNREL Computer Models Integrate Wind Turbines with
innovati nNREL Computer Models Integrate Wind Turbines with Floating Platforms Far off the shores for today's seabed-mounted offshore wind turbines. For the United States to tap into these vast offshore wind energy resources, wind turbines must be mounted on floating platforms to be cost effective
Computational Model for Forced Expiration from Asymmetric Normal Lungs
Lutchen, Kenneth
Computational Model for Forced Expiration from Asymmetric Normal Lungs ADAM G. POLAK 1 losses along the airway branches. Calculations done for succeeding lung volumes result in the semidynamic to the choke points, characteristic differences of lung regional pressures and volumes, and a shape
Computer Modeling of Crystalline Electrolytes Lithium Thiophosphates and Phosphates
Holzwarth, Natalie
migration. I. Introduction During the last 5 years, lithium thiophosphate solid electrolyte materials haveComputer Modeling of Crystalline Electrolytes Lithium Thiophosphates and Phosphates N. D. Lepley properties of (thio)phosphate electrolyte materials, focusing on the "superionic" electrolyte Li7P3S11. We
Pike, J.; Reboul, S.
2015-06-01
SRS High Level Waste Tank Farm personnel rely on conductivity probes for detection of incipient overflow conditions in waste tanks. Minimal information is available concerning the sensitivity that must be achieved such that that liquid detection is assured. Overly sensitive electronics results in numerous nuisance alarms for these safety-related instruments. In order to determine the minimum sensitivity required of the probe, Tank Farm Engineering personnel need adequate conductivity data to improve the existing designs. Little or no measurements of liquid waste conductivity exist; however, the liquid phase of the waste consists of inorganic electrolytes for which the conductivity may be calculated. Savannah River Remediation (SRR) Tank Farm Facility Engineering requested SRNL to determine the conductivity of the supernate resident in SRS waste Tank 40 experimentally as well as computationally. In addition, SRNL was requested to develop a correlation, if possible, that would be generally applicable to liquid waste resident in SRS waste tanks. A waste sample from Tank 40 was analyzed for composition and electrical conductivity as shown in Table 4-6, Table 4-7, and Table 4-9. The conductivity for undiluted Tank 40 sample was 0.087 S/cm. The accuracy of OLI Analyzer™ was determined using available literature data. Overall, 95% of computed estimates of electrical conductivity are within ±15% of literature values for component concentrations from 0 to 15 M and temperatures from 0 to 125 °C. Though the computational results are generally in good agreement with the measured data, a small portion of literature data deviates as much as ±76%. A simplified model was created that can be used readily to estimate electrical conductivity of waste solution in computer spreadsheets. The variability of this simplified approach deviates up to 140% from measured values. Generally, this model can be applied to estimate the conductivity within a factor of two. The comparison of the simplified model to pure component literature data suggests that the simplified model will tend to under estimate the electrical conductivity. Comparison of the computed Tank 40 conductivity with the measured conductivity shows good agreement within the range of deviation identified based on pure component literature data.
Challenges for the CMS computing model in the first year
Fisk, I.; /Fermilab
2009-05-01
CMS is in the process of commissioning a complex detector and a globally distributed computing infrastructure simultaneously. This represents a unique challenge. Even at the beginning there is not sufficient analysis or organized processing resources at CERN alone. In this presentation we discuss the unique computing challenges CMS expects to face during the first year of running and how they influence the baseline computing model decisions. During the early accelerator commissioning periods, CMS will attempt to collect as many events as possible when the beam is on in order to provide adequate early commissioning data. Some of these plans involve overdriving the Tier-0 infrastructure during data collection with recovery when the beam is off. In addition to the larger number of triggered events, there will be pressure in the first year to collect and analyze more complete data formats as the summarized formats mature. The large event formats impact the required storage, bandwidth, and processing capacity across all the computing centers. While the understanding of the detector and the event selections is being improved, there will likely be a larger number of reconstruction passes and skims performed by both central operations and individual users. We discuss how these additional stresses impact the allocation of resources and the changes from the baseline computing model.
A New Perspective for the Calibration of Computational Predictor Models.
Crespo, Luis Guillermo
2014-11-01
This paper presents a framework for calibrating computational models using data from sev- eral and possibly dissimilar validation experiments. The offset between model predictions and observations, which might be caused by measurement noise, model-form uncertainty, and numerical error, drives the process by which uncertainty in the models parameters is characterized. The resulting description of uncertainty along with the computational model constitute a predictor model. Two types of predictor models are studied: Interval Predictor Models (IPMs) and Random Predictor Models (RPMs). IPMs use sets to characterize uncer- tainty, whereas RPMs use random vectors. The propagation of a set through a model makes the response an interval valued function of the state, whereas the propagation of a random vector yields a random process. Optimization-based strategies for calculating both types of predictor models are proposed. Whereas the formulations used to calculate IPMs target solutions leading to the interval value function of minimal spread containing all observations, those for RPMs seek to maximize the models' ability to reproduce the distribution of obser- vations. Regarding RPMs, we choose a structure for the random vector (i.e., the assignment of probability to points in the parameter space) solely dependent on the prediction error. As such, the probabilistic description of uncertainty is not a subjective assignment of belief, nor is it expected to asymptotically converge to a fixed value, but instead it is a description of the model's ability to reproduce the experimental data. This framework enables evaluating the spread and distribution of the predicted response of target applications depending on the same parameters beyond the validation domain (i.e., roll-up and extrapolation).
Wind energy conversion system analysis model (WECSAM) computer program documentation
Downey, W T; Hendrick, P L
1982-07-01
Described is a computer-based wind energy conversion system analysis model (WECSAM) developed to predict the technical and economic performance of wind energy conversion systems (WECS). The model is written in CDC FORTRAN V. The version described accesses a data base containing wind resource data, application loads, WECS performance characteristics, utility rates, state taxes, and state subsidies for a six state region (Minnesota, Michigan, Wisconsin, Illinois, Ohio, and Indiana). The model is designed for analysis at the county level. The computer model includes a technical performance module and an economic evaluation module. The modules can be run separately or together. The model can be run for any single user-selected county within the region or looped automatically through all counties within the region. In addition, the model has a restart capability that allows the user to modify any data-base value written to a scratch file prior to the technical or economic evaluation. Thus, any user-supplied data for WECS performance, application load, utility rates, or wind resource may be entered into the scratch file to override the default data-base value. After the model and the inputs required from the user and derived from the data base are described, the model output and the various output options that can be exercised by the user are detailed. The general operation is set forth and suggestions are made for efficient modes of operation. Sample listings of various input, output, and data-base files are appended. (LEW)
ONSET OF CHAOS IN A MODEL OF QUANTUM COMPUTATION G. BERMAN; ET...
Office of Scientific and Technical Information (OSTI)
OF CHAOS IN A MODEL OF QUANTUM COMPUTATION G. BERMAN; ET AL 71 CLASSICAL AND QUANTUM MECHANICS, GENERAL PHYSICS; 99 GENERAL AND MISCELLANEOUSMATHEMATICS, COMPUTING, AND...
Liu, Huan
Second International Workshop on Social Computing, Behavioral Modeling, and Prediction Phoenix, Arizona March 31 - April 1, 2009 Phoenix, Arizona Proceedings published by Springer Social computing
The origins of computer weather prediction and climate modeling
Lynch, Peter [Meteorology and Climate Centre, School of Mathematical Sciences, University College Dublin, Belfield (Ireland)], E-mail: Peter.Lynch@ucd.ie
2008-03-20
Numerical simulation of an ever-increasing range of geophysical phenomena is adding enormously to our understanding of complex processes in the Earth system. The consequences for mankind of ongoing climate change will be far-reaching. Earth System Models are capable of replicating climate regimes of past millennia and are the best means we have of predicting the future of our climate. The basic ideas of numerical forecasting and climate modeling were developed about a century ago, long before the first electronic computer was constructed. There were several major practical obstacles to be overcome before numerical prediction could be put into practice. A fuller understanding of atmospheric dynamics allowed the development of simplified systems of equations; regular radiosonde observations of the free atmosphere and, later, satellite data, provided the initial conditions; stable finite difference schemes were developed; and powerful electronic computers provided a practical means of carrying out the prodigious calculations required to predict the changes in the weather. Progress in weather forecasting and in climate modeling over the past 50 years has been dramatic. In this presentation, we will trace the history of computer forecasting through the ENIAC integrations to the present day. The useful range of deterministic prediction is increasing by about one day each decade, and our understanding of climate change is growing rapidly as Earth System Models of ever-increasing sophistication are developed.
A General Hippocampal Computational Model Combining Episodic and Spatial Memory in a Spiking Model
Aguiar, Paulo de Castro
The hippocampus, in humans and rats, plays crucial roles in spatial tasks and nonspatial tasks involving episodic-type memory. This thesis presents a novel computational model of the hippocampus (CA1, CA3 and dentate ...
The use of computed radiography plates to determine light and radiation field coincidence
Kerns, James R. [Department of Radiation Physics, The University of Texas MD Anderson Cancer Center, Houston, Texas 77030 and Graduate School of Biomedical Sciences, The University of Texas Health Science Center Houston, Houston, Texas 77030 (United States)] [Department of Radiation Physics, The University of Texas MD Anderson Cancer Center, Houston, Texas 77030 and Graduate School of Biomedical Sciences, The University of Texas Health Science Center Houston, Houston, Texas 77030 (United States); Anand, Aman [Department of Radiation Oncology, Mayo Clinic, Scottsdale, Arizona 85259 (United States)] [Department of Radiation Oncology, Mayo Clinic, Scottsdale, Arizona 85259 (United States)
2013-11-15
Purpose: Photo-stimulable phosphor computed radiography (CR) has characteristics that allow the output to be manipulated by both radiation and optical light. The authors have developed a method that uses these characteristics to carry out radiation field and light field coincidence quality assurance on linear accelerators.Methods: CR detectors from Kodak were used outside their cassettes to measure both radiation and light field edges from a Varian linear accelerator. The CR detector was first exposed to a radiation field and then to a slightly smaller light field. The light impinged on the detector's latent image, removing to an extent the portion exposed to the light field. The detector was then digitally scanned. A MATLAB-based algorithm was developed to automatically analyze the images and determine the edges of the light and radiation fields, the vector between the field centers, and the crosshair center. Radiographic film was also used as a control to confirm the radiation field size.Results: Analysis showed a high degree of repeatability with the proposed method. Results between the proposed method and radiographic film showed excellent agreement of the radiation field. The effect of varying monitor units and light exposure time was tested and found to be very small. Radiation and light field sizes were determined with an uncertainty of less than 1 mm, and light and crosshair centers were determined within 0.1 mm.Conclusions: A new method was developed to digitally determine the radiation and light field size using CR photo-stimulable phosphor plates. The method is quick and reproducible, allowing for the streamlined and robust assessment of light and radiation field coincidence, with no observer interpretation needed.
Computer Modeling of Violent Intent: A Content Analysis Approach
Sanfilippo, Antonio P.; Mcgrath, Liam R.; Bell, Eric B.
2014-01-03
We present a computational approach to modeling the intent of a communication source representing a group or an individual to engage in violent behavior. Our aim is to identify and rank aspects of radical rhetoric that are endogenously related to violent intent to predict the potential for violence as encoded in written or spoken language. We use correlations between contentious rhetoric and the propensity for violent behavior found in documents from radical terrorist and non-terrorist groups and individuals to train and evaluate models of violent intent. We then apply these models to unseen instances of linguistic behavior to detect signs of contention that have a positive correlation with violent intent factors. Of particular interest is the application of violent intent models to social media, such as Twitter, that have proved to serve as effective channels in furthering sociopolitical change.
Comprehensive computer model for magnetron sputtering. II. Charged particle transport
Jimenez, Francisco J., E-mail: fjimenez@ualberta.ca; Dew, Steven K. [Department of Electrical and Computer Engineering, University of Alberta, Edmonton T6G 2V4 (Canada); Field, David J. [Smith and Nephew (Alberta) Inc., Fort Saskatchewan T8L 4K4 (Canada)
2014-11-01
Discharges for magnetron sputter thin film deposition systems involve complex plasmas that are sensitively dependent on magnetic field configuration and strength, working gas species and pressure, chamber geometry, and discharge power. The authors present a numerical formulation for the general solution of these plasmas as a component of a comprehensive simulation capability for planar magnetron sputtering. This is an extensible, fully three-dimensional model supporting realistic magnetic fields and is self-consistently solvable on a desktop computer. The plasma model features a hybrid approach involving a Monte Carlo treatment of energetic electrons and ions, along with a coupled fluid model for thermalized particles. Validation against a well-known one-dimensional system is presented. Various strategies for improving numerical stability are investigated as is the sensitivity of the solution to various model and process parameters. In particular, the effect of magnetic field, argon gas pressure, and discharge power are studied.
Computer Modeling and Simulation of an Active Vision Camera System MingChin Lu
Subbarao, Murali "Rao"
Computer Modeling and Simulation of an Active Vision Camera System MingChin Lu Communications of computer simulation systems. Computer simulation avoids the necessity of building actual camera systems. Based on the proposed model and algorithms, a computer simulation system called Active Vision Simulator
EconoGrid: A detailed Simulation Model of a Standards-based Grid Compute Economy
EconoGrid: A detailed Simulation Model of a Standards-based Grid Compute Economy EconoGrid is a detailed simulation model, implemented in SLX1 , of a grid compute economy that implements selected of users. In a grid compute economy, computing resources are sold to users in a market where price
Nanostructure Determination by Co-Refining Models to Multiple Datasets
Billinge, Simon
2011-05-31
The results of the work are contained in the publications resulting from the grant (which are listed below). Here I summarize the main findings from the last period of the award, 2006-2007: • Published a paper in Science with Igor Levin outlining the “Nanostructure Problem”, our inability to solve structure at the nanoscale. • Published a paper in Nature demonstrating the first ever ab-initio structure determination of a nanoparticle from atomic pair distribution function (PDF) data. • Published one book and 3 overview articles on PDF methods and the nanostructure problem. • Completed a project that sought to find a structural response to the presence of the so-called “intermediate phase” in network glasses which appears close to the rigidity percolation threshold in these systems. The main result was that we did not see convincing evidence for this, which drew into doubt the idea that GexSe1-x glasses were a model system exhibiting rigidity percolation.
A Vast Machine: Computer Models, Climate Data, and the Politics of Global Warming
Maret, Susan
2011-01-01
Review: A Vast Machine: Computer Models, Climate Data, andEdwards, Paul N. A Vast Machine: Computer Models, ClimateEdwards, Paul N. 2004. "A vast machine: standards as social
Models and People: an alternative view of the emergent properties of computational models
Boschetti, Fabio
be used to address scientific questions or support real-world decision making. At these higher levels ecological model may be used i) to address a scientific question and ii) how the scientific insight so gained management problem (Figure 1, top) 11-16 . 2 Computational models and scientific questions Within
Not Available
2012-08-01
Cation degradation insights obtained by computational modeling could result in better performance and longer lifetime for alkaline membrane fuel cells.
FINITE VOLUME METHODS APPLIED TO THE COMPUTATIONAL MODELLING OF WELDING PHENOMENA
Taylor, Gary
1 FINITE VOLUME METHODS APPLIED TO THE COMPUTATIONAL MODELLING OF WELDING PHENOMENA Gareth A.Taylor@brunel.ac.uk ABSTRACT This paper presents the computational modelling of welding phenomena within a versatile numerical) and Computational Solid Mechanics (CSM). With regard to the CFD modelling of the weld pool fluid dynamics, heat
How Computational Models Predict the Behavior of Complex Systems John Symons 1
Boschetti, Fabio
How Computational Models Predict the Behavior of Complex Systems John Symons 1 Fabio Boschetti2,3 1 of prediction in the use of computational models in science. We focus on the consequences of the irreversibility of computational models and on the conditional or ceteris paribus, nature of the kinds of their predictions
Boutchko, R.
2014-01-01
emission tomography systems and computational fluid dynamicsa computational ?uid dynamics (CFD) model of the systemthe computational domain. A Cartesian coordinate system was
Computational fluid dynamics modeling of coal gasification in a pressurized spout-fluid bed
Zhongyi Deng; Rui Xiao; Baosheng Jin; He Huang; Laihong Shen; Qilei Song; Qianjun Li [Southeast University, Nanjing (China). Key Laboratory of Clean Coal Power Generation and Combustion Technology of Ministry of Education
2008-05-15
Computational fluid dynamics (CFD) modeling, which has recently proven to be an effective means of analysis and optimization of energy-conversion processes, has been extended to coal gasification in this paper. A 3D mathematical model has been developed to simulate the coal gasification process in a pressurized spout-fluid bed. This CFD model is composed of gas-solid hydrodynamics, coal pyrolysis, char gasification, and gas phase reaction submodels. The rates of heterogeneous reactions are determined by combining Arrhenius rate and diffusion rate. The homogeneous reactions of gas phase can be treated as secondary reactions. A comparison of the calculated and experimental data shows that most gasification performance parameters can be predicted accurately. This good agreement indicates that CFD modeling can be used for complex fluidized beds coal gasification processes. 37 refs., 7 figs., 5 tabs.
Computing Limb Darkening Coefficients from Stellar Atmosphere Models
David Heyrovsky
2006-10-24
We explore the sensitivity of limb darkening coefficients computed from stellar atmosphere models to different least-squares fitting methods. We demonstrate that conventional methods are strongly biased to fitting the stellar limb. Our suggested method of fitting by minimizing the radially integrated squared residual yields improved fits with better flux conservation. The differences of the obtained coefficients from commonly used values are observationally significant. We show that the new values are in better agreement with solar limb darkening measurements as well as with coefficients reported from analyses of eclipsing binary light curves.
Computing model independent perturbations in dark energy and modified gravity
Battye, Richard A. [Jodrell Bank Centre for Astrophysics, School of Physics and Astronomy, The University of Manchester, Manchester M13 9PL (United Kingdom); Pearson, Jonathan A., E-mail: richard.battye@manchester.ac.uk, E-mail: jonathan.pearson@durham.ac.uk [Department of Mathematical Sciences, Durham University, South Road, Durham, DH1 3LE (United Kingdom)
2014-03-01
We present a methodology for computing model independent perturbations in dark energy and modified gravity. This is done from the Lagrangian for perturbations, by showing how field content, symmetries, and physical principles are often sufficient ingredients for closing the set of perturbed fluid equations. The fluid equations close once ''equations of state for perturbations'' are identified: these are linear combinations of fluid and metric perturbations which construct gauge invariant entropy and anisotropic stress perturbations for broad classes of theories. Our main results are the proof of the equation of state for perturbations presented in a previous paper, and the development of the required calculational tools.
A Model Independent Approach for Determining the Fragmentation Functions
Christova, Ekaterina [Institute for Nuclear research and Nuclear Energy, Sofia (Bulgaria); Leader, Elliot [Imperial College, London (United Kingdom)
2009-08-04
We show that the difference cross sections in unpolarized semi-inclusive deep inelastic scattering (SIDIS)e+N{yields}e+h+X and pp hadron production p+p{yields}h+X determine independently in a model independent way, in any order in Quantum Chromodynamics (QCD), the two FFs: D{sub u}{sup h-h-bar} and D{sub d}{sup h-h-bar}, h = {pi}{sup {+-}}, K{sup {+-}} or a sum over charged hadrons. If both K{sup {+-}} and K{sub 2}{sup 0} are measured, then e{sup +}e{sup -}{yields}K+X, e+N{yields}e+K+X and p+p{yields}K+X present independent measurements of just one FF: D{sub u-d}{sup K{sup +}+K{sup -}}. The above results allow to test the existing parameterizations, obtained with various different assumptions about the FFs, and to test the Q{sup 2} evolution and factorization.
Modeling the Fracture of Ice Sheets on Parallel Computers
Waisman, Haim; Tuminaro, Ray
2013-10-10
The objective of this project was to investigate the complex fracture of ice and understand its role within larger ice sheet simulations and global climate change. This objective was achieved by developing novel physics based models for ice, novel numerical tools to enable the modeling of the physics and by collaboration with the ice community experts. At the present time, ice fracture is not explicitly considered within ice sheet models due in part to large computational costs associated with the accurate modeling of this complex phenomena. However, fracture not only plays an extremely important role in regional behavior but also influences ice dynamics over much larger zones in ways that are currently not well understood. To this end, our research findings through this project offers significant advancement to the field and closes a large gap of knowledge in understanding and modeling the fracture of ice sheets in the polar regions. Thus, we believe that our objective has been achieved and our research accomplishments are significant. This is corroborated through a set of published papers, posters and presentations at technical conferences in the field. In particular significant progress has been made in the mechanics of ice, fracture of ice sheets and ice shelves in polar regions and sophisticated numerical methods that enable the solution of the physics in an efficient way.
A Unified Computational Model for Solar and Stellar Flares
Allred, Joel C; Carlsson, Mats
2015-01-01
We present a unified computational framework which can be used to describe impulsive flares on the Sun and on dMe stars. The models assume that the flare impulsive phase is caused by a beam of charged particles that is accelerated in the corona and propagates downward depositing energy and momentum along the way. This rapidly heats the lower stellar atmosphere causing it to explosively expand and dramatically brighten. Our models consist of flux tubes that extend from the sub-photosphere into the corona. We simulate how flare-accelerated charged particles propagate down one-dimensional flux tubes and heat the stellar atmosphere using the Fokker-Planck kinetic theory. Detailed radiative transfer is included so that model predictions can be directly compared with observations. The flux of flare-accelerated particles drives return currents which additionally heat the stellar atmosphere. These effects are also included in our models. We examine the impact of the flare-accelerated particle beams on model solar and...
Center for Programming Models for Scalable Parallel Computing
John Mellor-Crummey
2008-02-29
Rice University's achievements as part of the Center for Programming Models for Scalable Parallel Computing include: (1) design and implemention of cafc, the first multi-platform CAF compiler for distributed and shared-memory machines, (2) performance studies of the efficiency of programs written using the CAF and UPC programming models, (3) a novel technique to analyze explicitly-parallel SPMD programs that facilitates optimization, (4) design, implementation, and evaluation of new language features for CAF, including communication topologies, multi-version variables, and distributed multithreading to simplify development of high-performance codes in CAF, and (5) a synchronization strength reduction transformation for automatically replacing barrier-based synchronization with more efficient point-to-point synchronization. The prototype Co-array Fortran compiler cafc developed in this project is available as open source software from http://www.hipersoft.rice.edu/caf.
Experimental validation of a kilovoltage x-ray source model for computing imaging dose
Poirier, Yannick, E-mail: yannick.poirier@cancercare.mb.ca [CancerCare Manitoba, 675 McDermot Ave, Winnipeg, Manitoba R3E 0V9 (Canada)] [CancerCare Manitoba, 675 McDermot Ave, Winnipeg, Manitoba R3E 0V9 (Canada); Kouznetsov, Alexei; Koger, Brandon [Department of Physics and Astronomy, University of Calgary, Calgary, Alberta T2N 1N4 (Canada)] [Department of Physics and Astronomy, University of Calgary, Calgary, Alberta T2N 1N4 (Canada); Tambasco, Mauro, E-mail: mtambasco@mail.sdsu.edu [Department of Physics, San Diego State University, San Diego, California 92182-1233 and Department of Physics and Astronomy and Department of Oncology, University of Calgary, Calgary, Alberta T2N 1N4 (Canada)] [Department of Physics, San Diego State University, San Diego, California 92182-1233 and Department of Physics and Astronomy and Department of Oncology, University of Calgary, Calgary, Alberta T2N 1N4 (Canada)
2014-04-15
Purpose: To introduce and validate a kilovoltage (kV) x-ray source model and characterization method to compute absorbed dose accrued from kV x-rays. Methods: The authors propose a simplified virtual point source model and characterization method for a kV x-ray source. The source is modeled by: (1) characterizing the spatial spectral and fluence distributions of the photons at a plane at the isocenter, and (2) creating a virtual point source from which photons are generated to yield the derived spatial spectral and fluence distribution at isocenter of an imaging system. The spatial photon distribution is determined by in-air relative dose measurements along the transverse (x) and radial (y) directions. The spectrum is characterized using transverse axis half-value layer measurements and the nominal peak potential (kVp). This source modeling approach is used to characterize a Varian{sup ®} on-board-imager (OBI{sup ®}) for four default cone-beam CT beam qualities: beams using a half bowtie filter (HBT) with 110 and 125 kVp, and a full bowtie filter (FBT) with 100 and 125 kVp. The source model and characterization method was validated by comparing dose computed by the authors’ inhouse software (kVDoseCalc) to relative dose measurements in a homogeneous and a heterogeneous block phantom comprised of tissue, bone, and lung-equivalent materials. Results: The characterized beam qualities and spatial photon distributions are comparable to reported values in the literature. Agreement between computed and measured percent depth-dose curves is ?2% in the homogeneous block phantom and ?2.5% in the heterogeneous block phantom. Transverse axis profiles taken at depths of 2 and 6 cm in the homogeneous block phantom show an agreement within 4%. All transverse axis dose profiles in water, in bone, and lung-equivalent materials for beams using a HBT, have an agreement within 5%. Measured profiles of FBT beams in bone and lung-equivalent materials were higher than their computed counterparts resulting in an agreement within 2.5%, 5%, and 8% within solid water, bone, and lung, respectively. Conclusions: The proposed virtual point source model and characterization method can be used to compute absorbed dose in both the homogeneous and heterogeneous block phantoms within of 2%–8% of measured values, depending on the phantom and the beam quality. The authors’ results also provide experimental validation for their kV dose computation software, kVDoseCalc.
Computational fluid dynamic modeling of fluidized-bed polymerization reactors
Rokkam, Ram
2012-11-02
Polyethylene is one of the most widely used plastics, and over 60 million tons are produced worldwide every year. Polyethylene is obtained by the catalytic polymerization of ethylene in gas and liquid phase reactors. The gas phase processes are more advantageous, and use fluidized-bed reactors for production of polyethylene. Since they operate so close to the melting point of the polymer, agglomeration is an operational concern in all slurry and gas polymerization processes. Electrostatics and hot spot formation are the main factors that contribute to agglomeration in gas-phase processes. Electrostatic charges in gas phase polymerization fluidized bed reactors are known to influence the bed hydrodynamics, particle elutriation, bubble size, bubble shape etc. Accumulation of electrostatic charges in the fluidized-bed can lead to operational issues. In this work a first-principles electrostatic model is developed and coupled with a multi-fluid computational fluid dynamic (CFD) model to understand the effect of electrostatics on the dynamics of a fluidized-bed. The multi-fluid CFD model for gas-particle flow is based on the kinetic theory of granular flows closures. The electrostatic model is developed based on a fixed, size-dependent charge for each type of particle (catalyst, polymer, polymer fines) phase. The combined CFD model is first verified using simple test cases, validated with experiments and applied to a pilot-scale polymerization fluidized-bed reactor. The CFD model reproduced qualitative trends in particle segregation and entrainment due to electrostatic charges observed in experiments. For the scale up of fluidized bed reactor, filtered models are developed and implemented on pilot scale reactor.
Modeling and Synthesizing Task Placement Constraints in Google Compute Clusters
Cortes, Corinna
characterization for high performance computing and grids focus on task resource require- ments for CPU, memory of compute clusters. Existing workload characteri- zations for high performance computing and grids focus
Computational model for simulation small testing launcher, technical solution
Chelaru, Teodor-Viorel; Cristian, Barbu; Chelaru, Adrian
2014-12-10
The purpose of this paper is to present some aspects regarding the computational model and technical solutions for multistage suborbital launcher for testing (SLT) used to test spatial equipment and scientific measurements. The computational model consists in numerical simulation of SLT evolution for different start conditions. The launcher model presented will be with six degrees of freedom (6DOF) and variable mass. The results analysed will be the flight parameters and ballistic performances. The discussions area will focus around the technical possibility to realize a small multi-stage launcher, by recycling military rocket motors. From technical point of view, the paper is focused on national project 'Suborbital Launcher for Testing' (SLT), which is based on hybrid propulsion and control systems, obtained through an original design. Therefore, while classical suborbital sounding rockets are unguided and they use as propulsion solid fuel motor having an uncontrolled ballistic flight, SLT project is introducing a different approach, by proposing the creation of a guided suborbital launcher, which is basically a satellite launcher at a smaller scale, containing its main subsystems. This is why the project itself can be considered an intermediary step in the development of a wider range of launching systems based on hybrid propulsion technology, which may have a major impact in the future European launchers programs. SLT project, as it is shown in the title, has two major objectives: first, a short term objective, which consists in obtaining a suborbital launching system which will be able to go into service in a predictable period of time, and a long term objective that consists in the development and testing of some unconventional sub-systems which will be integrated later in the satellite launcher as a part of the European space program. This is why the technical content of the project must be carried out beyond the range of the existing suborbital vehicle programs towards the current technological necessities in the space field, especially the European one.
Computational Biology and Bioinformatics 10.10 Models of substitution I: Basic Models A
Goldschmidt, Christina
& stochastic grammars 7.11 RNA structures 9.11 Finding signals in sequences 14.11 Challenges in genome of structure & movements & shapes & grammars 28.11 Integrative genomics: the omics DNA mRNA Protein Metabolite Phenotype 30.11 Integrative genomics: mapping #12;Computational Biology and Bioinformatics 10.10 Models
A computational model of the motivation-learning interface Manish Saggar (mishu@cs.utexas.edu)
Maddox, W. Todd
A computational model of the motivation-learning interface Manish Saggar (mishu the influence of motivation on learning observed by Markman, Baldwin and Maddox (2005). They showed was confirmed. These results constitute a first computational step towards understanding how motivation
A Polarizable QM/MM Explicit Solvent Model for Computational Electrochemistry in Water
Wang, Lee-Ping
We present a quantum mechanical/molecular mechanical (QM/MM) explicit solvent model for the computation of standard reduction potentials E[subscript 0]. The QM/MM model uses density functional theory (DFT) to model the ...
Computation Modeling and Assessment of Nanocoatings for Ultra Supercritical Boilers
J. Shingledecker; D. Gandy; N. Cheruvu; R. Wei; K. Chan
2011-06-21
Forced outages and boiler unavailability of coal-fired fossil plants is most often caused by fire-side corrosion of boiler waterwalls and tubing. Reliable coatings are required for Ultrasupercritical (USC) application to mitigate corrosion since these boilers will operate at a much higher temperatures and pressures than in supercritical (565 C {at} 24 MPa) boilers. Computational modeling efforts have been undertaken to design and assess potential Fe-Cr-Ni-Al systems to produce stable nanocrystalline coatings that form a protective, continuous scale of either Al{sub 2}O{sub 3} or Cr{sub 2}O{sub 3}. The computational modeling results identified a new series of Fe-25Cr-40Ni with or without 10 wt.% Al nanocrystalline coatings that maintain long-term stability by forming a diffusion barrier layer at the coating/substrate interface. The computational modeling predictions of microstructure, formation of continuous Al{sub 2}O{sub 3} scale, inward Al diffusion, grain growth, and sintering behavior were validated with experimental results. Advanced coatings, such as MCrAl (where M is Fe, Ni, or Co) nanocrystalline coatings, have been processed using different magnetron sputtering deposition techniques. Several coating trials were performed and among the processing methods evaluated, the DC pulsed magnetron sputtering technique produced the best quality coating with a minimum number of shallow defects and the results of multiple deposition trials showed that the process is repeatable. scale, inward Al diffusion, grain growth, and sintering behavior were validated with experimental results. The cyclic oxidation test results revealed that the nanocrystalline coatings offer better oxidation resistance, in terms of weight loss, localized oxidation, and formation of mixed oxides in the Al{sub 2}O{sub 3} scale, than widely used MCrAlY coatings. However, the ultra-fine grain structure in these coatings, consistent with the computational model predictions, resulted in accelerated Al diffusion from the coating into the substrate. An effective diffusion barrier interlayer coating was developed to prevent inward Al diffusion. The fire-side corrosion test results showed that the nanocrystalline coatings with a minimum number of defects have a great potential in providing corrosion protection. The coating tested in the most aggressive environment showed no evidence of coating spallation and/or corrosion attack after 1050 hours exposure. In contrast, evidence of coating spallation in isolated areas and corrosion attack of the base metal in the spalled areas were observed after 500 hours. These contrasting results after 500 and 1050 hours exposure suggest that the premature coating spallation in isolated areas may be related to the variation of defects in the coating between the samples. It is suspected that the cauliflower-type defects in the coating were presumably responsible for coating spallation in isolated areas. Thus, a defect free good quality coating is the key for the long-term durability of nanocrystalline coatings in corrosive environments. Thus, additional process optimization work is required to produce defect-free coatings prior to development of a coating application method for production parts.
MOMDIS: a Glauber model computer code for knockout reactions
C. A. Bertulani; A. Gade
2006-04-12
A computer program is described to calculate momentum distributions in stripping and diffraction dissociation reactions. A Glauber model is used with the scattering wavefunctions calculated in the eikonal approximation. The program is appropriate for knockout reactions at intermediate energy collisions (30 MeV $\\leq$ E$_{lab}/$nucleon $\\leq 2000$ MeV). It is particularly useful for reactions involving unstable nuclear beams, or exotic nuclei (e.g. neutron-rich nuclei), and studies of single-particle occupancy probabilities (spectroscopic factors) and other related physical observables. Such studies are an essential part of the scientific program of radioactive beam facilities, as in for instance the proposed RIA (Rare Isotope Accelerator) facility in the US.
GEOCITY: a computer model for systems analysis of geothermal district heating and cooling costs
Fassbender, L.L.; Bloomster, C.H.
1981-06-01
GEOCITY is a computer-simulation model developed to study the economics of district heating/cooling using geothermal energy. GEOCITY calculates the cost of district heating/cooling based on climate, population, resource characteristics, and financing conditions. The basis for our geothermal-energy cost analysis is the unit cost of energy which will recover all the costs of production. The calculation of the unit cost of energy is based on life-cycle costing and discounted-cash-flow analysis. A wide variation can be expected in the range of potential geothermal district heating and cooling costs. The range of costs is determined by the characteristics of the resource, the characteristics of the demand, and the distance separating the resource and the demand. GEOCITY is a useful tool for estimating costs for each of the main parts of the production process and for determining the sensitivity of these costs to several significant parameters under a consistent set of assumptions.
National Nuclear Security Administration (NNSA)
Each successive generation of computing system has provided greater computing power and energy efficiency.
CTS-1 clusters will support NNSA's Life Extension Program and...
Larson, N. M.; Perey, F. G.
1980-11-01
A method is described for determining the parameters of a model from experimental data based upon the utilization of Bayes' theorem. This method has several advantages over the least-squares method as it is commonly used; one important advantage is that the assumptions under which the parameter values have been determined are more clearly evident than in many results based upon least squares. Bayes' method has been used to develop a computer code which can be utilized to analyze neutron cross-section data by means of the R-matrix theory. The required formulae from the R-matrix theory are presented, and the computer implementation of both Bayes' equations and R-matrix theory is described. Details about the computer code and compelte input/output information are given.
Madiraju, Praveen
discovery process lightweight and secure. In this paper we present details of the trust and risk models. We of devices running in a pervasive computing environment [10]. The resource discovery process demands modelsSSRD+: A Privacy-aware Trust and Security Model for Resource Discovery in Pervasive Computing
A Risk-aware Trust Based Secure Resource Discovery (RTSRD) Model for Pervasive Computing
Madiraju, Praveen
security threat to them. Thus, the resource discovery process demands models that ensure the privacyA Risk-aware Trust Based Secure Resource Discovery (RTSRD) Model for Pervasive Computing Sheikh I-hoc network of pervasive computing, a resource discovery model is needed that can resolve security and privacy
Using Parallel MCMC Sampling to Calibrate a Computer Model of a Geothermal Reservoir
Fox, Colin
Using Parallel MCMC Sampling to Calibrate a Computer Model of a Geothermal Reservoir by T. Cui, C. 686 ISSN 1178-360 #12;Using Parallel MCMC Sampling to Calibrate a Computer Model of a Geothermal of a geothermal field to achieve model `calibration' from measured well-test data. We explore three scenarios
Ray tracing computations in the smoothed SEG/EAGE Salt Model
Cerveny, Vlastislav
Ray tracing computations in the smoothed SEG/EAGE Salt Model V#19;aclav Bucha Department to compute rays and synthetic seismograms of refracted and re ected P-waves in the smoothed SEG/EAGE Salt The original 3-D SEG/EAGE Salt Model (Aminzadeh et al. 1997) is very complex model and cannot be used for ray
A Three-Dimensional Computational Model of PEM Fuel Cell with Serpentine Gas Channels
Victoria, University of
A Three-Dimensional Computational Model of PEM Fuel Cell with Serpentine Gas Channels by Phong ABSTRACT A three-dimensional computational fluid dynamics model of a Polymer Electrolyte Membrane (PEM) fuel cell with serpentine gas flow channels is presented in this thesis. This comprehensive model
A network based model for heterogeneous parallel computation
Sathye, Adwait B.
1993-01-01
The computational requirements of science and engineering demand computational resources orders of magnitude of the current day sequential machines. Most of the research effort has been concentrated upon the creation of parallel algorithms...
A New Model for Image-Based Humanities Computing
Brown, Jacob Hohmann
2009-05-15
Image-based humanities computing, the computer-assisted study of digitallyrepresented “objects or artifacts of cultural heritage,” is an increasingly popular yet “established practice” located at the most recent intersections ...
Vladimir P. Gerdt; Vasily M. Severyanov
2005-12-08
A C# package is presented that allows a user for an input quantum circuit to generate a set of multivariate polynomials over the finite field Z_2 whose total number of solutions in Z_2 determines the output of the quantum computation defined by the circuit. The generated polynomial system can further be converted to the canonical Groebner basis form which provides a universal algorithmic tool for counting the number of common roots of the polynomials.
AIR INGRESS ANALYSIS: PART 2 – COMPUTATIONAL FLUID DYNAMIC MODELS
Chang H. Oh; Eung S. Kim; Richard Schultz; Hans Gougar; David Petti; Hyung S. Kang
2011-01-01
The Idaho National Laboratory (INL), under the auspices of the U.S. Department of Energy, is performing research and development that focuses on key phenomena important during potential scenarios that may occur in very high temperature reactors (VHTRs). Phenomena Identification and Ranking Studies to date have ranked an air ingress event, following on the heels of a VHTR depressurization, as important with regard to core safety. Consequently, the development of advanced air ingress-related models and verification and validation data are a very high priority. Following a loss of coolant and system depressurization incident, air will enter the core of the High Temperature Gas Cooled Reactor through the break, possibly causing oxidation of the in-the core and reflector graphite structure. Simple core and plant models indicate that, under certain circumstances, the oxidation may proceed at an elevated rate with additional heat generated from the oxidation reaction itself. Under postulated conditions of fluid flow and temperature, excessive degradation of the lower plenum graphite can lead to a loss of structural support. Excessive oxidation of core graphite can also lead to the release of fission products into the confinement, which could be detrimental to a reactor safety. Computational fluid dynamic model developed in this study will improve our understanding of this phenomenon. This paper presents two-dimensional and three-dimensional CFD results for the quantitative assessment of the air ingress phenomena. A portion of results of the density-driven stratified flow in the inlet pipe will be compared with results of the experimental results.
The impact of the Self-Determined Learning Model of Instruction on student self-determination
Wehmeyer, Michael L.; Shogren, Karrie A.; Palmer, Susan B.; Williams-Diehm, Kendra; Little, Todd D.; Boulton, Aaron Jacob
2012-01-01
possible, attrition of teachers/students at control campuses. After obtaining consent and assent to partici- pate, we collected baseline data, including demo- graphic information and two measures of self-determination. We also collected self-determi- nation... data at the end of Year 1 and of Year 2. As expected, there was attrition in the sample over time. At the end of Year 1, 103 control-group stu- dents had completed baseline and Year 1 posttest scores from the SDS (Wehmeyer & Kelchner, 1995) and 94...
High-Performance Computer Modeling of the Cosmos-Iridium Collision
Olivier, S; Cook, K; Fasenfest, B; Jefferson, D; Jiang, M; Leek, J; Levatin, J; Nikolaev, S; Pertica, A; Phillion, D; Springer, K; De Vries, W
2009-08-28
This paper describes the application of a new, integrated modeling and simulation framework, encompassing the space situational awareness (SSA) enterprise, to the recent Cosmos-Iridium collision. This framework is based on a flexible, scalable architecture to enable efficient simulation of the current SSA enterprise, and to accommodate future advancements in SSA systems. In particular, the code is designed to take advantage of massively parallel, high-performance computer systems available, for example, at Lawrence Livermore National Laboratory. We will describe the application of this framework to the recent collision of the Cosmos and Iridium satellites, including (1) detailed hydrodynamic modeling of the satellite collision and resulting debris generation, (2) orbital propagation of the simulated debris and analysis of the increased risk to other satellites (3) calculation of the radar and optical signatures of the simulated debris and modeling of debris detection with space surveillance radar and optical systems (4) determination of simulated debris orbits from modeled space surveillance observations and analysis of the resulting orbital accuracy, (5) comparison of these modeling and simulation results with Space Surveillance Network observations. We will also discuss the use of this integrated modeling and simulation framework to analyze the risks and consequences of future satellite collisions and to assess strategies for mitigating or avoiding future incidents, including the addition of new sensor systems, used in conjunction with the Space Surveillance Network, for improving space situational awareness.
A model for computer frustration: the role of instrumental and dispositional factors
Shneiderman, Ben
A model for computer frustration: the role of instrumental and dispositional factors on incident 20742, USA Available online 16 April 2004 Abstract Frustration is almost universally accepted
Designing Computing System Architecture and Models for the HL-LHC era
Lothar Bauerdick; Brian Bockelman; Peter Elmer; Stephen Gowdy; Matevz Tadel; Frank Wuerthwein
2015-07-20
This paper describes a programme to study the computing model in CMS after the next long shutdown near the end of the decade.
Designing Computing System Architecture and Models for the HL-LHC era
Bauerdick, Lothar; Elmer, Peter; Gowdy, Stephen; Tadel, Matevz; Wuerthwein, Frank
2015-01-01
This paper describes a programme to study the computing model in CMS after the next long shutdown near the end of the decade.
Climate Determinism Revisited: Multiple Equilibria in a Complex Climate Model
Ferreira, David
Multiple equilibria in a coupled ocean–atmosphere–sea ice general circulation model (GCM) of an aquaplanet with many degrees of freedom are studied. Three different stable states are found for exactly the same set of ...
Computational Modeling and Assessment Of Nanocoatings for Ultra Supercritical Boilers
David W. Gandy; John P. Shingledecker
2011-04-11
Forced outages and boiler unavailability in conventional coal-fired fossil power plants is most often caused by fireside corrosion of boiler waterwalls. Industry-wide, the rate of wall thickness corrosion wastage of fireside waterwalls in fossil-fired boilers has been of concern for many years. It is significant that the introduction of nitrogen oxide (NOx) emission controls with staged burners systems has increased reported waterwall wastage rates to as much as 120 mils (3 mm) per year. Moreover, the reducing environment produced by the low-NOx combustion process is the primary cause of accelerated corrosion rates of waterwall tubes made of carbon and low alloy steels. Improved coatings, such as the MCrAl nanocoatings evaluated here (where M is Fe, Ni, and Co), are needed to reduce/eliminate waterwall damage in subcritical, supercritical, and ultra-supercritical (USC) boilers. The first two tasks of this six-task project-jointly sponsored by EPRI and the U.S. Department of Energy (DE-FC26-07NT43096)-have focused on computational modeling of an advanced MCrAl nanocoating system and evaluation of two nanocrystalline (iron and nickel base) coatings, which will significantly improve the corrosion and erosion performance of tubing used in USC boilers. The computational model results showed that about 40 wt.% is required in Fe based nanocrystalline coatings for long-term durability, leading to a coating composition of Fe-25Cr-40Ni-10 wt.% Al. In addition, the long term thermal exposure test results further showed accelerated inward diffusion of Al from the nanocrystalline coatings into the substrate. In order to enhance the durability of these coatings, it is necessary to develop a diffusion barrier interlayer coating such TiN and/or AlN. The third task 'Process Advanced MCrAl Nanocoating Systems' of the six-task project jointly sponsored by the Electric Power Research Institute, EPRI and the U.S. Department of Energy (DE-FC26-07NT43096)- has focused on processing of advanced nanocrystalline coating systems and development of diffusion barrier interlayer coatings. Among the diffusion interlayer coatings evaluated, the TiN interlayer coating was found to be the optimum one. This report describes the research conducted under the Task 3 workscope.
Rapidly re-computable EEG (electroencephalography) forward models for realistic head shapes
Ermer, J. J.; Mosher, J. C.; Baillet, S.; Leahy, R. M.
2001-01-01
Solution of the EEG source localization (inverse) problem utilizing model-based methods typically requires a significant number of forward model evaluations. For subspace based inverse methods like MUSIC [6], the total number of forward model evaluations can often approach an order of 10{sup 3} or 10{sup 4}. Techniques based on least-squares minimization may require significantly more evaluations. The observed set of measurements over an M-sensor array is often expressed as a linear forward spatio-temporal model of the form: F = GQ + N (1) where the observed forward field F (M-sensors x N-time samples) can be expressed in terms of the forward model G, a set of dipole moment(s) Q (3xP-dipoles x N-time samples) and additive noise N. Because of their simplicity, ease of computation, and relatively good accuracy, multi-layer spherical models [7] (or fast approximations described in [1], [7]) have traditionally been the 'forward model of choice' for approximating the human head. However, approximation of the human head via a spherical model does have several key drawbacks. By its very shape, the use of a spherical model distorts the true distribution of passive currents in the skull cavity. Spherical models also require that the sensor positions be projected onto the fitted sphere (Fig. 1), resulting in a distortion of the true sensor-dipole spatial geometry (and ultimately the computed surface potential). The use of a single 'best-fitted' sphere has the added drawback of incomplete coverage of the inner skull region, often ignoring areas such as the frontal cortex. In practice, this problem is typically countered by fitting additional sphere(s) to those region(s) not covered by the primary sphere. The use of these additional spheres results in added complication to the forward model. Using high-resolution spatial information obtained via X-ray CT or MR imaging, a realistic head model can be formed by tessellating the head into a set of contiguous regions (typically the scalp, outer skull, and inner skull surfaces). Since accurate in vivo determination of internal conductivities is currently not currently possible, the head is typically assumed to consist of a set of contiguous isotropic regions, each with constant conductivity.
A Mathematical Model for Virus Infection in a System of Interacting Computers
Cipolatti, Rolci
A Mathematical Model for Virus Infection in a System of Interacting Computers J. L´opez Gondar & R are explored and enlightened in this paper. 1. Introduction The infection of computers by virtual viruses of virtual viruses in a system of interacting computers could be compared with a disease transmitted
Modeling Computational Security in Long-Lived Systems, Version 2 Ran Canetti1,2
International Association for Cryptologic Research (IACR)
Modeling Computational Security in Long-Lived Systems, Version 2 Ran Canetti1,2 , Ling Cheung2 Introduction Computational security in long-lived systems: Security properties of cryptographic protocols computational power. This type of security degrades progressively over the lifetime of a protocol. However, some
Modeling Computational Security in Long-Lived Systems Ran Canetti1,2
International Association for Cryptologic Research (IACR)
Modeling Computational Security in Long-Lived Systems Ran Canetti1,2 , Ling Cheung2 , Dilsun Kaynar Introduction Computational security in long-lived systems: Security properties of cryptographic protocols protocols, security relies on the assumption that adversarial entities have lim- ited computational power
Novel properties generated by interacting computational systems: A minimal model Fabio Boschetti1,2
Boschetti, Fabio
Novel properties generated by interacting computational systems: A minimal model Fabio questions: First, what is the smallest number of components a computational system needs in order such as selforganisation and emergence have been discussed in computational terms within Complex System Science
Pedram, Massoud
Trace-Based Analysis and Prediction of Cloud Computing User Behavior Using the Fractal Modeling and technology. In this paper, we investigate the characteristics of the cloud computing requests received the alpha- stable distribution. Keywords- cloud computing; alpha-stable distribution; fractional order
CPT: An Energy-Efficiency Model for Multi-core Computer Systems
Shi, Weisong
CPT: An Energy-Efficiency Model for Multi-core Computer Systems Weisong Shi, Shinan Wang and Bing efficiency of computer systems. These techniques affect the energy efficiency across different layers metric that represents the energy efficiency of a computer system, for a specific configuration, given
Baer, Ferdinand
Optimizing Computations in Weather and Climate Prediction Models* F. BAER, BANGLIN ZHANG, AND BING scenarios for many time scales, more computer power than is currently available will be needed. One and sometimes with a biosphere included, are very complex and require so much computing power on available
Computer modeling of the spatial resolution properties of a dedicated breast CT system
Yang Kai; Kwan, Alexander L. C.; Boone, John M. [Department of Radiology, University of California, Davis Medical Center, 4860 Y Street, Suite 3100 Ellison Building, Sacramento, California 95817 (United States) and Department of Biomedical Engineering, University of California, Davis, California, 95616 (United States); Department of Radiology, University of California, Davis Medical Center, 4860 Y Street, Suite 3100 Ellison Building, Sacramento, California 95817 (United States); Department of Radiology, University of California, Davis Medical Center, 4860 Y Street, Suite 3100 Ellison Building, Sacramento, California 95817 (United States) and Department of Biomedical Engineering, University of California, Davis, California 95616 (United States)
2007-06-15
Computer simulation methods were used to evaluate the spatial resolution properties of a dedicated cone-beam breast CT system. x-ray projection data of a 70 {mu}m nickel-chromium wire were simulated. The modulation transfer function (MTF) was calculated from the reconstructed axial images at different radial positions from the isocenter to study the spatial dependency of the spatial resolution of the breast CT scanner. The MTF was also calculated in both the radial and azimuthal directions. Subcomponents of the cone beam CT system that affect the MTF were modeled in the computer simulation in a serial manner, including the x-ray focal spot distribution, gantry rotation under the condition of continuous fluoroscopy, detector lag, and detector spatial resolution. Comparison between the computer simulated and physically measured MTF values demonstrates reasonable accuracy in the simulation process, with a small systematic difference ({approx}9.5{+-}6.4% difference, due to unavoidable uncertainties from physical measurement and system calibration). The intrinsic resolution in the radial direction determined by simulation was about 2.0 mm{sup -1} uniformly through the field of view. The intrinsic resolution in the azimuthal direction degrades from 2.0 mm{sup -1} at the isocenter to 1.0 mm{sup -1} at the periphery with 76.9 mm from the isocenter. The results elucidate the intrinsic spatial resolution properties of the prototype breast CT system, and suggest ways in which spatial resolution can be improved with system modification.
OptCDR: a general computational method for the design of antibody complementarity determining
Maranas, Costas
regions for targeted epitope binding R.J. Pantazes and C.D. Maranas1 Department of Chemical Engineering for each position. OptCDR is applied to three computational test cases: a peptide from the capsid and affinity. They are composed of pairs of heavy and light chains folded into the well-known `Y' shape. Two
A Calibrated Computer Model for the Thermal Simulation of Courtyard Microclimates
Bagneid, A.; Haberl, J.
2006-01-01
This paper describes a calibrated stand-alone courtyard microclimate model. This model is considered to be the fIrst calibrated computer program for the simulation of courtyard microclimates. In order to accomplish this a calibrated simplif...
Cielo Computational Environment Usage Model With Mappings to...
Office of Scientific and Technical Information (OSTI)
Environment (ACE) Version 8.0 requirements. The ACE requirements reflect the high performance computing requirements for the Production Readiness Milestone user environment...
Determining Interactions in PSA models: Application to a Space PSA
C. Smith; E. Borgonovo
2010-06-01
This paper addresses use of an importance measure interaction study of a probabilistic risk analysis (PSA) performed for a hypothetical aerospace lunar mission. The PSA methods used in this study follow the general guidance provided in the NASA Probabilistic Risk Assessment Procedures Guide for NASA Managers and Practitioners. For the PSA portion, we used phased-based event tree and fault tree logic structures are used to model a lunar mission, including multiple phases (from launch to return to the Earth surface) and multiple critical systems. Details of the analysis results are not provided in this paper – instead specific basic events are denoted by number (e.g., the first event is 1, the second is 2, and so on). However, in the model, we used approximately 150 fault trees and over 800 basic events. Following analysis and truncation of cut sets, we were left with about 400 basic events to evaluate. We used this model to explore interactions between different basic events and systems. These sensitivity studies provide high-level insights into features of the PSA for the hypothetical lunar mission.
Mathematical and Computer Modelling 53 (2011) 716730 Contents lists available at ScienceDirect
Berges, John A.
2011-01-01
Direct Mathematical and Computer Modelling journal homepage: www.elsevier.com/locate/mcm Dynamics of a virusMathematical and Computer Modelling 53 (2011) 716730 Contents lists available at Science: Received 19 January 2010 Accepted 17 October 2010 Keywords: Virushost dynamics Quota Bacteriophage
8/30/2001 Parallel Programming -Fall 2001 1 Models of Parallel Computation
Browne, James C.
8/30/2001 Parallel Programming - Fall 2001 1 Models of Parallel Computation Philosophy Parallel of parallel programming. #12;8/30/2001 Parallel Programming - Fall 2001 2 Models of Parallel Computation will discuss parallelism from the viewpoint of programming but with connections to other domains. #12;8/30/2001
Computing Estimates in the Proportional Odds Model David R. Hunter1
Hunter, David
Computing Estimates in the Proportional Odds Model David R. Hunter1 Kenneth Lange2 Department running head: Computing MLE for proportional odds Submitted to Annals of the Institute of Statistical The semiparametric proportional odds model for survival data is useful when mortality rates of different groups
Dessouky, Maged
A Hierarchical Task Model for Dispatching in Computer- Assisted Demand-Responsive Paratransit Model for Dispatching in Computer-Assisted Demand-Responsive Paratransit Operation ABSTRACT, Dispatch Training #12;1 INTRODUCTION Demand-responsive paratransit service is on the rise. For example
Using error correction to determine the noise model
M. Laforest; D. Simon; J. -C. Boileau; J. Baugh; M. Ditty; R. Laflamme
2007-01-25
Quantum error correcting codes have been shown to have the ability of making quantum information resilient against noise. Here we show that we can use quantum error correcting codes as diagnostics to characterise noise. The experiment is based on a three-bit quantum error correcting code carried out on a three-qubit nuclear magnetic resonance (NMR) quantum information processor. Utilizing both engineered and natural noise, the degree of correlations present in the noise affecting a two-qubit subsystem was determined. We measured a correlation factor of c=0.5+/-0.2 using the error correction protocol, and c=0.3+/-0.2 using a standard NMR technique based on coherence pathway selection. Although the error correction method demands precise control, the results demonstrate that the required precision is achievable in the liquid-state NMR setting.
Burk, David Edward
2005-08-29
of the RDD debris. The objective of this research was to benchmark a forward model methodology for predicting isotopic composition of spent nuclear fuel used in an RDD while at the same time optimizing the fidelity of the model to reduce computational time...
Tarragon : a programming model for latency-hiding scientific computations
Cicotti, Pietro
2011-01-01
Chapter 2 Programming Model . . . . . . . . . 2.1vi Chapter 6 Dynamic programming . . . . . . . . . . 6.1 Theof related programming model implementations. . . . . .
Scalar products in GL(3)-based models with trigonometric R-matrix. Determinant representation
N. A. Slavnov
2015-03-20
We study quantum integrable GL(3)-based models with a trigonometric R-matrix solvable by the nested algebraic Bethe ansatz. We derive a determinant representation for a special case of scalar products of Bethe vectors. This representation allows one to find a determinant formula for the form factor of one of the monodromy matrix entries. We also point out essential difference between form factors in the models with the trigonometric R-matrix and their analogs in GL(3)-invariant models.
Towards a Model for Computing in European Astroparticle Physics
T. Berghöfer; I. Agrafioti; B. Allen; V. Beckmann; T. Chiarusi; M. Delfino; S. Hesping; J. Chudoba; L. Dell'Agnello; S. Katsanevas; G. Lamanna; R. Lemrani; A. Margiotta; G. Maron; C. Palomba; G. Russo; P. Wegner
2015-12-03
Current and future astroparticle physics experiments are operated or are being built to observe highly energetic particles, high energy electromagnetic radiation and gravitational waves originating from all kinds of cosmic sources. The data volumes taken by the experiments are large and expected to grow significantly during the coming years. This is a result of advanced research possibilities and improved detector technology. To cope with the substantially increasing data volumes of astroparticle physics projects it is important to understand the future needs for computing resources in this field. Providing these resources constitutes a larger fraction of the overall running costs of future infrastructures. This document presents the results of a survey made by APPEC with the help of computing experts of major projects and future initiatives in astroparticle physics, representatives of current Tier-1 and Tier-2 LHC computing centers, as well as specifically astroparticle physics computing centers, e.g. the Albert Einstein Institute for gravitational waves analysis in Hanover. In summary, the overall CPU usage and short-term disk and long-term (tape) storage space currently available for astroparticle physics projects' computing services is of the order of one third of the central computing available for LHC data at the Tier-0 center at CERN. Till the end of the decade the requirements for computing resources are estimated to increase by a factor of 10. Furthermore, this document shall describe the diversity of astroparticle physics data handling and serve as a basis to estimate a distribution of computing and storage tasks among the major computing centers. (Abridged)
Cummings, Peter [Vanderbilt University] [Vanderbilt University
2009-11-15
The document is the final report of the DOE Computational Nanoscience Project DE-FG02-03ER46096: Integrated Multiscale Modeling of Molecular Computing Devices. It included references to 62 publications that were supported by the grant.
Math 484: Mathematical & Computational Modeling Course Information and Syllabus Spring 2013
Math 484: Mathematical & Computational Modeling Course Information and Syllabus Spring 2013 students how to build and analyze mathematical models. Most concepts presented in the context of physical students how to build mathematical models to physical problems. 2. Be able to solve modeling problems
Math 484: Mathematical & Computational Modeling Course Information and Syllabus Spring 2012
Math 484: Mathematical & Computational Modeling Course Information and Syllabus Spring 2012 students how to build and analyze mathematical models. Most concepts presented in the context of physical students how to build mathematical models to physical problems. 2. Be able to solve modeling problems
JACKSON VL
2011-08-31
The primary purpose of the tank mixing and sampling demonstration program is to mitigate the technical risks associated with the ability of the Hanford tank farm delivery and celtification systems to measure and deliver a uniformly mixed high-level waste (HLW) feed to the Waste Treatment and Immobilization Plant (WTP) Uniform feed to the WTP is a requirement of 24590-WTP-ICD-MG-01-019, ICD-19 - Interface Control Document for Waste Feed, although the exact definition of uniform is evolving in this context. Computational Fluid Dynamics (CFD) modeling has been used to assist in evaluating scaleup issues, study operational parameters, and predict mixing performance at full-scale.
A Fire Model for 2-D Computer Animation
Yu, J-H.; Patterson, J.W.
Yu,J-H. Patterson,J.W. Proceedings of the EUROGRAPHICS Workshop on Computer Animation '96, Poitiers, France. Published in Eurographics Series, (Boulic R. and Hegron, G., Eds.). pp 49-60 Springer
Edinburgh Research Explorer Computational models in systems biology
Millar, Andrew J.
called process algebras and Petri nets offer alternative ways of constructing computational systems networks in an abstract way that is independent of particular mathe- matical techniques of analysis. Access
The Need for Biological Computation System Models | GE Global...
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
2012.10.09 Hello everyone, I'm Maria Zavodszky and I work in the Computational Biology and Biostatistics Lab at GE Global Research in Niskayuna, New York. This being our...
Computational Approaches for Modelling Elastohydrodynamic Lubrication Using Multiphysics Software
Jimack, Peter
the Integral Approach (IA). The computational cost of performing this calculation is , for N points over developing highly specialised, bespoke software are highlighted. In order to calculate the complex nonlinear partial differential equation system representing the full problem. This includes
Multiscale Computational Modeling of Multiphase Composites with Damage
Cheng, Feifei
2013-11-01
A multiscale computational framework for multiphase composites considering damage is developed in this research. In micro-scale, micromechanics based homogenization methods are used to estimate effective elastic moduli of graded Ti_(2)Al...
Automatic Symmetry Detection for Model Checking Using Computational Group Theory
Donaldson, A.F.; Miller, A.
Donaldson,A.F. Miller,A. Proceedings of the 13th International Symposium on Formal Methods Europe (FME 2005). Lecture Notes in Computing Science volume 3582. pp 481--496 Springer
Applications of Computer Modelling to Fire Safety Design
Torero, Jose L; Steinhaus, Thomas
Tools in support of fire safety engineering design have proliferated in the last few years due to the increased performance of computers. These tools are currently being used in a generalized manner in areas such as egress, ...
Cloud computing adoption model for governments and large enterprises
Trivedi, Hrishikesh
2013-01-01
Cloud Computing has held organizations across the globe spell bound with its promise. As it moves from being a buzz word and hype into adoption, organizations are faced with question of how to best adopt cloud. Existing ...
Empirical model determines energy required to clean sand from well bore
Appah, D.; Ichara, M. (Univ. of Port Harcourt (Nigeria))
1994-02-28
An empirical hydraulic model has been developed for determining the energy required for cleaning a vertical and nearly vertical well bore plugged with sand particles. The model considers pressure losses and cleanout time and compares sand cleanout time during direct and reverse circulation of water. Good agreement was obtained between the model and experimental results.
Effective Design And Use Of Computer Decision Models.
Fuerst, William L.
1984-03-01
of decision models such as simulation have not been demonstrated. This paper looks at recent literature regarding decision model deficiencies, evaluates selected financial simulation model packages, and suggests design needs for expanding the use of decision...
Modeling-Computer Simulations At Dixie Valley Geothermal Area...
W. Wisian, David D. Blackwell (2004) Numerical Modeling Of Basin And Range Geothermal Systems Additional References Retrieved from "http:en.openei.orgwindex.php?titleModel...
Bauer, Roman; Stoll, Elizabeth
2015-01-01
Glioma is the most common form of primary brain tumor. Demographically, the risk of occurrence increases until old age. Here we present a novel computational model to reproduce the probability of glioma incidence across the lifespan. Previous mathematical models explaining glioma incidence are framed in a rather abstract way, and do not directly relate to empirical findings. To decrease this gap between theory and experimental observations, we incorporate recent data on cellular and molecular factors underlying gliomagenesis. Since evidence implicates the adult neural stem cell as the likely cell-of-origin of glioma, we have incorporated empirically-determined estimates of neural stem cell number, cell division rate, mutation rate and oncogenic potential into our model. We demonstrate that our model yields results which match actual demographic data in the human population. In particular, this model accounts for the observed peak incidence of glioma at approximately 80 years of age, without the need to assert...
Compensator models for fluence field modulated computed tomography
Bartolac, Steven; Jaffray, David; Radiation Medicine Program, Princess Margaret Hospital Department of Radiation Oncology, University of Toronto, Toronto, Ontario M5G 2M9
2013-12-15
Purpose: Fluence field modulated computed tomography (FFMCT) presents a novel approach for acquiring CT images, whereby a patient model guides dynamically changing fluence patterns in an attempt to achieve task-based, user-prescribed, regional variations in image quality, while also controlling dose to the patient. This work aims to compare the relative effectiveness of FFMCT applied to different thoracic imaging tasks (routine diagnostic CT, lung cancer screening, and cardiac CT) when the modulator is subject to limiting constraints, such as might be present in realistic implementations.Methods: An image quality plan was defined for a simulated anthropomorphic chest slice, including regions of high and low image quality, for each of the thoracic imaging tasks. Modulated fluence patterns were generated using a simulated annealing optimization script, which attempts to achieve the image quality plan under a global dosimetric constraint. Optimization was repeated under different types of modulation constraints (e.g., fixed or gantry angle dependent patterns, continuous or comprised of discrete apertures) with the most limiting case being a fixed conventional bowtie filter. For each thoracic imaging task, an image quality map (IQM{sub sd}) representing the regionally varying standard deviation is predicted for each modulation method and compared to the prescribed image quality plan as well as against results from uniform fluence fields. Relative integral dose measures were also compared.Results: Each IQM{sub sd} resulting from FFMCT showed improved agreement with planned objectives compared to those from uniform fluence fields for all cases. Dynamically changing modulation patterns yielded better uniformity, improved image quality, and lower dose compared to fixed filter patterns with optimized tube current. For the latter fixed filter cases, the optimal choice of tube current modulation was found to depend heavily on the task. Average integral dose reduction compared to a uniform fluence field ranged from 10% using a bowtie filter to 40% or greater using an idealized modulator.Conclusions: The results support that FFMCT may achieve regionally varying image quality distributions in good agreement with user-prescribed values, while limiting dose. The imposition of constraints inhibits dose reduction capacity and agreement with image quality plans but still yields significant improvement over what is afforded by conventional dose minimization techniques. These results suggest that FFMCT can be implemented effectively even when the modulator has limited modulation capabilities.
Center for Programming Models for Scalable Parallel Computing: Future Programming Models
Gao, Guang, R.
2008-07-24
The mission of the pmodel center project is to develop software technology to support scalable parallel programming models for terascale systems. The goal of the specific UD subproject is in the context developing an efficient and robust methodology and tools for HPC programming. More specifically, the focus is on developing new programming models which facilitate programmers in porting their application onto parallel high performance computing systems. During the course of the research in the past 5 years, the landscape of microprocessor chip architecture has witnessed a fundamental change – the emergence of multi-core/many-core chip architecture appear to become the mainstream technology and will have a major impact to for future generation parallel machines. The programming model for shared-address space machines is becoming critical to such multi-core architectures. Our research highlight is the in-depth study of proposed fine-grain parallelism/multithreading support on such future generation multi-core architectures. Our research has demonstrated the significant impact such fine-grain multithreading model can have on the productivity of parallel programming models and their efficient implementation.
Computer vision determination of the stem/root joint on processing carrots
Batchelor, Matthew McMahon
1987-01-01
, a linear photo-diode array camera, and a microprocessor to control an orientating mechanism. The photo- diode array measured the diameter of ears of corn as they were transported by a conveyor belt. Because the corn was being transported in a... direction perpendicular to the plane of the photo- diode array, the measure produced by the array was the width or diameter of the ear on the belt. This diameter information described the shape of the ear. Ear shape was analyzed to determine whether...
Protein translocation without specific quality control in a computational model of the Tat system
Chitra R. Nayak; Aidan I. Brown; Andrew D. Rutenberg
2014-08-20
The twin-arginine translocation (Tat) system transports folded proteins of various sizes across both bacterial and plant thylakoid membranes. The membrane-associated TatA protein is an essential component of the Tat translocon, and a broad distribution of different sized TatA-clusters is observed in bacterial membranes. We assume that the size dynamics of TatA clusters are affected by substrate binding, unbinding, and translocation to associated TatBC clusters, where clusters with bound translocation substrates favour growth and those without associated substrates favour shrinkage. With a stochastic model of substrate binding and cluster dynamics, we numerically determine the TatA cluster size distribution. We include a proportion of targeted but non-translocatable (NT) substrates, with the simplifying hypothesis that the substrate translocatability does not directly affect cluster dynamical rate constants or substrate binding or unbinding rates. This amounts to a translocation model without specific quality control. Nevertheless, NT substrates will remain associated with TatA clusters until unbound and so will affect cluster sizes and translocation rates. We find that the number of larger TatA clusters depends on the NT fraction $f$. The translocation rate can be optimized by tuning the rate of spontaneous substrate unbinding, $\\Gamma_U$. We present an analytically solvable three-state model of substrate translocation without cluster size dynamics that follows our computed translocation rates, and that is consistent with {\\em in vitro} Tat-translocation data in the presence of NT substrates.
Tesfatsion, Leigh
Bounded Computing Capacity ·· Explicit SpaceExplicit Space ·· Local InteractionsLocal Interactions ·· Non explanatory notion.notion. #12;4 SugarscapeSugarscape ·· Events unfold on a landscape of renewableEvents
A.24-1 A.24 ENHANCING THE CAPABILITY OF COMPUTATIONAL EARTH SYSTEM MODELS AND NASA DATA) computational support of Earth system modeling. #12;A.24-2 2.1 Acceleration of Operational Use of Research Data
Computer modeling reveals how surprisingly potent hepatitis C drug works
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
AFDC Printable Version Share this resource Send a link to EERE: Alternative Fuels Data Center Home Page to someone by E-mail Share EERE: Alternative Fuels Data Center Home Page on Facebook Tweet about EERE: Alternative Fuels Data Center Home Page on Twitter Bookmark EERE: Alternative Fuels Data Center Homesum_a_epg0_fpd_mmcf_m.xls" ,"Available from WebQuantityBonneville Power Administration would like submit the following commentsMethodsCompositional6 ComputationalComputerHepatitis C
Computer-Aided Construction of Combustion Chemistry Models
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
AFDC Printable Version Share this resource Send a link to EERE: Alternative Fuels Data Center Home Page to someone by E-mail Share EERE: Alternative Fuels Data Center Home Page on Facebook Tweet about EERE: Alternative Fuels Data Center Home Page on Twitter Bookmark EERE: Alternative Fuels Data Center Homesum_a_epg0_fpd_mmcf_m.xls" ,"Available from WebQuantityBonneville Power Administration would like submit the following commentsMethodsCompositional6 ComputationalComputerHepatitis
The U-Machine: A Model of Generalized Computation
MacLennan, Bruce
classification: 68Q05. This report may be used for any non-profit purpose provided that the source is credited in the broad- est sense, including analog and digital, and provides a framework for computation using novel known, the end of Moore's Law is in sight, and we are approaching the limits of digital electronics
WUFI COMPUTER MODELING WORKSHOP FOR WALL DESIGN AND PERFORMANCE
Oak Ridge National Laboratory
from rain, solar radiation and other crucial weather events on an hourly basis. Both vapor and liquid which can easily be selected from a map. The temporal behavior of the computed quantities (temperatures tell me performance predicting · Mold growth predictions new post-processing modules · Development
Modeling Living Systems for Computer Vision Demetri Terzopoulos
Terzopoulos, Demetri
, for computer vision. First, I present a new breed of artificial animals in a physicsbased virtual marine world vision. 1 These advances center around the idea of artificial animals, or ``animats'' a term coined and 3 present our work on artificial fishes and animat vision. The basic idea in a nutshell
Modelling Photochemical Pollution using Parallel and Distributed Computing Platforms
Abramson, David
of photochemical air pollution (smog) in industrialised cities. However, computational hardware demands can that have been used as part of an air pollution study being conducted in Melbourne, Australia. We also necessary to perform real air pollution studies. The system is used as part of the Melbourne Airshed study
Modeling-Computer Simulations At Kilauea East Rift Geothermal...
East Rift Zone Notes Three models were made from data collected in the exploratory well HGP-A. The models simulated constant heat sources from a vertical dike, a magma chamber,...
An Axiomatisation of Computationally Adequate Domain Theoretic Models of FPC
Fiore, Marcelo P; Plotkin, Gordon
1994-01-01
Categorical models of the metalanguage FPC (a type theory with sums, products, exponentials and recursive types) are defined. Then, domain-theoretic models of FPC are axiomatised and a wide subclass of them —the ...
Experimental Evaluations of Expert and Non-expert Computer Users' Mental Models of Security Risks
Camp, L. Jean
Experimental Evaluations of Expert and Non-expert Computer Users' Mental Models of Security Risks risks and thereby enable informed decisions by naive users. Yet computer security has not been en- gaged with the scholarship of risk communication. While the existence of malicious actors may appear at first to distinguish
Creative Commons Copyright 2013 Some Rights Reserved CMC: A Model Computer Science Curriculum
Iyer, Sridhar
for K-12 Schools 3rd Edition, Released June 2013 Technical Report: TR-CSE-2013-52 Department of Computer-12 Schools Creative Commons Copyright © 2013 Some Rights Reserved 2 CMC: A Model Computer Science Curriculum for K-12 Schools 3rd Edition June, 2013 Authors Sridhar Iyer*, Farida Khan, Sahana Murthy
An Exact Modeling of Signal Statistics in Energy-integrating X-ray Computed Tomography
used by modern computed tomography (CT) scanners and has been an interesting research topicAn Exact Modeling of Signal Statistics in Energy-integrating X-ray Computed Tomography Yi Fan1.i.d.), such as Gamma, Gaussian, etc, would be valid. A comparison study was performed to estimate the introduced errors
Internship Parallel Computer Evaluation Parallelization of a Lagrangian Particle Diffusion Model
use case is a nuclear accident like a core meltdown at a atomic power plant, where atomic radiation emits in the air. The Lagrangian model can predict how the nuclear cloud spreads under different that will be computed. Particle: One single molecule floating in the wind field. Compute unit: One unit that runs
A computational model for nanoscale adhesion between deformable solids and its application to gecko adhesion Roger A. Sauer 1 Aachen Institute for Advanced Study in Computational Engineering Science (AICES), RWTH Aachen University, Templergaben 55, 52056 Aachen, Germany Published2 in the Journal of Adhesion
Author's personal copy Calibration procedures for a computational model of ductile fracture
Hutchinson, John W.
Author's personal copy Calibration procedures for a computational model of ductile fracture Z. Xue fracture Computational fracture Shear fracture Damage parameters a b s t r a c t A recent extension of the cup-cone fracture mode in the neck of a round tensile bar. Ductility of a notched round bar provides
A Vast Machine Computer Models, Climate Data, and the Politics of Global Warming
Edwards, Paul N.
A Vast Machine Computer Models, Climate Data, and the Politics of Global Warming Paul N. Edwards models, climate data, and the politics of global warming / Paul N. Edwards. p. cm. Includes this: Global warming is a myth. It's all model predictions, nothing but simulations. Before you believe
A Computational Model of Knowledge-Intensive Learning and Problem Solving1
Aamodt, Agnar
1 A Computational Model of Knowledge-Intensive Learning and Problem Solving1 Agnar Aamodt Knowledge. If knowledge-based systems are to become more competent and robust in solving real world problems, they need model - a framework -for knowledge-intensive problem solving and learning from experience. The model has
COMPUTERS AND BIOMEDICAL RESEARCH 17, 580-589 ( 1984) Multicompartment Model of Lung Dynamics*
Longtin, André
COMPUTERS AND BIOMEDICAL RESEARCH 17, 580-589 ( 1984) Multicompartment Model of Lung Dynamics* B of the lungs. The lungs are represented by 24 compartments each corresponding to a generation of the Weibel model A. In the model it is assumed that gases are transported in the lungs by convection and diffusion
Lecture Notes in Computer Science 1 Data Reduction Using Multiple Models Integration
Obradovic, Zoran
Lecture Notes in Computer Science 1 Data Reduction Using Multiple Models Integration Aleksandar the models constructed on previously considered data samples. In addition to random sampling, controllable sampling based on the boosting algorithm is proposed, where the models are combined using a weighted voting
Computational Fluid Dynamics Modeling of a Lithium/Thionyl Chloride Battery with Electrolyte Flow
Wang, Chao-Yang
Computational Fluid Dynamics Modeling of a Lithium/Thionyl Chloride Battery with Electrolyte Flow W-dimensional model is developed to simulate discharge of a primary lithium/thionyl chloride battery. The model to the first task with important examples of lead-acid,1-3 nickel-metal hydride,4-8 and lithium-based batteries
DEVELOPMENT OF A COMPUTER SIMULATION MODEL FOR BLOWING GLASS CONTAINERS C. G. Giannopapa
Eindhoven, Technische Universiteit
1 DEVELOPMENT OF A COMPUTER SIMULATION MODEL FOR BLOWING GLASS CONTAINERS C. G. Giannopapa Dept to be used for industrial purposes that accurately captures the blowing step of glass containers. The model,2]. This paper concentrates on modeling the blowing stage of the forming process of glass containers
Ortiz, Michael
mechanical re- sponse of a ceramic matrix composite is simulated by a numerical model for a ®ber-matrix unitComputational modeling of damage evolution in unidirectional fiber reinforced ceramic matrix evolution in brittle matrix composites was developed. This modeling is based on an axisymmetric unit cell
Gedeon, Tomas
, from those appearing in physiology and ecology to Earth systems modeling, often experience critical
Computational Modeling of Conventionally Reinforced Concrete Coupling Beams
Shastri, Ajay Seshadri
2012-02-14
. The model is developed in the finite element analysis software ABAQUS. The concrete damaged plasticity model was used to simulate the behavior of concrete. A calibration model using a cantilever beam was produced to generate key parameters in the model... Stress (ABAQUS 2008)??????.????...61 Fig. 3.9. CPS8 Element Used for Modeling Concrete (ABAQUS 2008)?????64 Fig. 4.1. Elevation and Cross-Section of the Cantilever Beam?????????66 Fig. 4.2. Compressive Stress-Strain Behavior of Concrete...
Seagraves, Andrew Nathan
2010-01-01
In this thesis a new parallel computational method is proposed for modeling threedimensional dynamic fracture of brittle solids. The method is based on a combination of the discontinuous Galerkin (DG) formulation of the ...
Curry, Benjamin David
A goal of Artificial Intelligence is to develop computational models of what would be considered intelligent behaviour in a human. One such task is that of musical performance. This research specifically focuses on aspects ...
A Computational Market Model for Distributed Configuration Design Michael P. Wellman
Wellman, Michael P.
economies" constitutes the market solution to the original problem. After defining the configuration design. Consider a hyper-simplified scenario in aircraft design. (We choose this not as a serious exemplar ¡ ¢ £ £ ¤ ¢ ¥ ¦ § ¨ © ¨ A Computational Market Model for Distributed Configuration
THEORETICAL MODELING AND COMPUTATIONAL SIMULATION OF ROBUST CONTROL FOR MARS AIRCRAFT
Oh, Seyool
2014-05-31
The focus of this dissertation is the development of control system design algorithms for autonomous operation of an aircraft in the Martian atmosphere. This research will show theoretical modeling and computational ...
Ma, Yongting
2011-01-11
This thesis presents development of mathematical models for multi-media interaction process using Eulerian description and associated computational infrastructure to obtain numerical solution of the initial value problems ...
Liquefied Natural Gas (LNG) Vapor Dispersion Modeling with Computational Fluid Dynamics Codes
Qi, Ruifeng
2012-10-19
Federal regulation 49 CFR 193 and standard NFPA 59A require the use of validated consequence models to determine the vapor cloud dispersion exclusion zones for accidental liquefied natural gas (LNG) releases. For modeling purposes, the physical...
Daigle, Matthew
, and availability. Prognos- tics deals with determining the health state of compo- nents, and projecting) predictions. Model-based prognos- tics approaches perform these tasks with the aid of a model that captures
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
AFDC Printable Version Share this resource Send a link to EERE: Alternative Fuels Data Center Home Page to someone by E-mail Share EERE: Alternative Fuels Data Center Home Page on Facebook Tweet about EERE: Alternative Fuels Data Center Home Page on Twitter Bookmark EERE: Alternative Fuels Data Center Homesum_a_epg0_fpd_mmcf_m.xls" ,"Available from WebQuantityBonneville Power Administration would like submit the following commentsMethodsCompositional6Energy Computers,Computing and
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
AFDC Printable Version Share this resource Send a link to EERE: Alternative Fuels Data Center Home Page to someone by E-mail Share EERE: Alternative Fuels Data Center Home Page on Facebook Tweet about EERE: Alternative Fuels Data Center Home Page on Twitter Bookmark EERE: Alternative Fuels Data Center Homesum_a_epg0_fpd_mmcf_m.xls" ,"Available from WebQuantityBonneville Power Administration would like submit the following commentsMethodsCompositional6Energy Computers,Computing
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
AFDC Printable Version Share this resource Send a link to EERE: Alternative Fuels Data Center Home Page to someone by E-mail Share EERE: Alternative Fuels Data Center Home Page on Facebook Tweet about EERE: Alternative Fuels Data Center Home Page on Twitter Bookmark EERE: Alternative Fuels Data Center Homesum_a_epg0_fpd_mmcf_m.xls" ,"Available from WebQuantity ofkandz-cm11 Outreach Home Room News PublicationsAuditsCluster Compatibilitydefault Changes TheComputeComputing
Formal computational models and non-standard finiteness
Ayala-RincÃ³n, Mauricio
(f) #12;Finiteness in Computation ^ without NNO T : Pos - f : P(Q Ã? Pos Ã? Tfin) - P(Q Ã? Posfin Ã? Q), such that Tfin = {h/fin({p Pos : h(p) = b})} X = P(Q Ã? Pos Ã? Tfin), St(X) = {S X : f(S) S} i : St(X) P(X) has = { x , y /Pos , yT > ^ Pos , xT > and final(q)} #12;Finitenes
Final Report for Integrated Multiscale Modeling of Molecular Computing Devices
Glotzer, Sharon C.
2013-08-28
In collaboration with researchers at Vanderbilt University, North Carolina State University, Princeton and Oakridge National Laboratory we developed multiscale modeling and simulation methods capable of modeling the synthesis, assembly, and operation of molecular electronics devices. Our role in this project included the development of coarse-grained molecular and mesoscale models and simulation methods capable of simulating the assembly of millions of organic conducting molecules and other molecular components into nanowires, crossbars, and other organized patterns.
Modeling-Computer Simulations At Long Valley Caldera Geothermal...
surrounding a vertically dipping prolate spheroid source during an active period of time-dependent deformation between 1995 and 2000 at Long Valley caldera. We model a rapid...
Modeling-Computer Simulations At Long Valley Caldera Geothermal...
and seismic data was conducted in 2003 to investigate the cause of recent uplift of the resurgent dome. Notes Modeling of deformation and microgravity data suggests...
Advanced Computing Tools and Models for Accelerator Physics
Ryne, Robert D.
2008-01-01
MODELS FOR ACCELERATOR PHYSICS * Robert D. Ryne, Lawrencetools for accelerator physics. Following an introduction Icomputing in accelerator physics. INTRODUCTION To begin I
Mathematical modeling and computer simulation of processes in energy systems
Hanjalic, K.C. )
1990-01-01
This book is divided into the following chapters. Modeling techniques and tools (fundamental concepts of modeling); 2. Fluid flow, heat and mass transfer, chemical reactions, and combustion; 3. Processes in energy equipment and plant components (boilers, steam and gas turbines, IC engines, heat exchangers, pumps and compressors, nuclear reactors, steam generators and separators, energy transport equipment, energy convertors, etc.); 4. New thermal energy conversion technologies (MHD, coal gasification and liquefaction fluidized-bed combustion, pulse-combustors, multistage combustion, etc.); 5. Combined cycles and plants, cogeneration; 6. Dynamics of energy systems and their components; 7. Integrated approach to energy systems modeling, and 8. Application of modeling in energy expert systems.
Modeling-Computer Simulations At Long Valley Caldera Geothermal...
LVEW are best matched using modeled solutions for a flow system consisting of a rock matrix with finite hydraulic conductivity cut by a steeply dipping fracture with infinite...
Vassilis Geroyannis; Georgios Kleftogiannis
2014-06-14
We revisit the problem of radial pulsations of neutron stars by computing four general-relativistic polytropic models, in which "density" and "adiabatic index" are involved with their discrete meanings: (i) "rest-mass density" or (ii) "mass-energy density" regarding the density, and (i) "constant" or (ii) "variable" regarding the adiabatic index. Considering the resulting four discrete combinations, we construct corresponding models and compute for each model the frequencies of the lowest three radial modes. Comparisons with previous results are made. The deviations of respective frequencies of the resolved models seem to exhibit a systematic behavior, an issue discussed here in detail.
Gaussian Process Modeling and Computation in Engineering Applications
Pourhabib, Arash
2014-07-08
; and predictive modeling for large datasets. First, we develop a spatial-temporal model for local wind fields in a wind farm with more than 200 wind turbines. Our framework utilizes the correlation among the derivatives of wind speeds to find a neighborhood...
Modeling civil violence: An agent-based computational approach
Tesfatsion, Leigh
," I do so advisedly, recognizing that no political or social order is represented in the model of revolutions properly speaking. The dynamics of decentralized upheaval, rather than its political substance Against Central Authority This model involves two categories of actors. ``Agents'' are members
Huang, Su-Yun
factors · Simulation codes with calibration parameters 8 #12;Example: Designing Cellular Heat Exchangers in Qian et al. (2006, ASME) Related to the autoregressive model in Kennedy and O'Hagan (2000) · x = (x1
Didactyl: Toward a Useful Computational Model of Piano Fingering
, ergonomics, cognition, and habit. Theoretically, the search space for an ideal) to define evaluation corpora to allow comparisons between the models, and (3 and evaluation corpora to reduce opportunity costs for future researchers working
Computational tools for modeling and measuring chromosome structure
Ross, Brian Christopher
2012-01-01
DNA conformation within cells has many important biological implications, but there are challenges both in modeling DNA due to the need for specialized techniques, and experimentally since tracing out in vivo conformations ...
When Model Checking Met Deduction Computer Science Laboratory
Clarke, Edmund M.
Park, CA Sep 19, 2014 #12;Alan Turing It is of course important that some efforts be made to verify hold in each case. Alan Turing (quoted by D. MacKenzie in Risk and Reason) N. Shankar Model checking
Scalable computational architecture for integrating biological pathway models
Shiva, V. A
2007-01-01
A grand challenge of systems biology is to model the cell. The cell is an integrated network of cellular functions. Each cellular function, such as immune response, cell division, metabolism or apoptosis, is defined by an ...
A Computational Model of How the Basal Ganglia Produce Sequences
Berns, Gregory S.
closely on known anatomy and physiology. First, we assume that the thalamic targets, which relay ascend the external globus pallidus (GPe) and the subthalamic nucleus (STN). As a test of the model, the system
Computational Modeling of Combined Steam Pyrolysis and Hydrogasification of Ethanol
Singh, S; Park, C S; Norbeck, J N
2005-01-01
Approximate modelling of coal pyrolysis. Fuel, 78(7), 825-9. Gonzalez, JF. (2003). Pyrolysis of cherry stones: energyof analytical and applied pyrolysis, 67(1), 165-190. 10.
Computer support to run models of the atmosphere. Final report
Fung, I.
1996-08-30
This research is focused on a better quantification of the variations in CO{sub 2} exchanges between the atmosphere and biosphere and the factors responsible for these exchangers. The principal approach is to infer the variations in the exchanges from variations in the atmospheric CO{sub 2} distribution. The principal tool involves using a global three-dimensional tracer transport model to advect and convect CO{sub 2} in the atmosphere. The tracer model the authors used was developed at the Goddard institute for Space Studies (GISS) and is derived from the GISS atmospheric general circulation model. A special run of the GCM is made to save high-frequency winds and mixing statistics for the tracer model.
Modeling-Computer Simulations At White Mountains Area (Goff ...
useful DOE-funding Unknown Notes Review and identification of 24 potential sites for EGS development across the U.S., as well as modeling of the representative geologic systems...
Modeling-Computer Simulations At Fenton Hill HDR Geothermal Area...
useful DOE-funding Unknown Notes Review and identification of 24 potential sites for EGS development across the U.S., as well as modeling of the representative geologic systems...
Continuum-based Multiscale Computational Damage Modeling of Cementitous Composites
Kim, Sun-Myung
2011-08-08
, aggregates, and interfacial transition zone (ITZ) and interaction among components at meso-scale, and the interaction between reinforcements, such as fiber and carbon nanotubes (CNTs) and mortar matrix or the ITZ at nano scale in order to predict more... of Advisory Committee: Dr. Rashid K. Abu Al-Rub Based on continuum damage mechanics (CDM), an isotropic and anisotropic damage model coupled with a novel plasticity model for plain concrete is proposed in this research. Two different damage evolution laws...
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
AFDC Printable Version Share this resource Send a link to EERE: Alternative Fuels Data Center Home Page to someone by E-mail Share EERE: Alternative Fuels Data Center Home Page on Facebook Tweet about EERE: Alternative Fuels Data Center Home Page on Twitter Bookmark EERE: Alternative Fuels Data Center Homesum_a_epg0_fpd_mmcf_m.xls" ,"Available from WebQuantityBonneville Power Administration would like submit the following commentsMethodsCompositional6 Computational CSEEnergy
Slepoy, Alexander; Mitchell, Scott A.; Backus, George A.; McNamara, Laura A.; Trucano, Timothy Guy
2008-09-01
Sandia National Laboratories is investing in projects that aim to develop computational modeling and simulation applications that explore human cognitive and social phenomena. While some of these modeling and simulation projects are explicitly research oriented, others are intended to support or provide insight for people involved in high consequence decision-making. This raises the issue of how to evaluate computational modeling and simulation applications in both research and applied settings where human behavior is the focus of the model: when is a simulation 'good enough' for the goals its designers want to achieve? In this report, we discuss two years' worth of review and assessment of the ASC program's approach to computational model verification and validation, uncertainty quantification, and decision making. We present a framework that extends the principles of the ASC approach into the area of computational social and cognitive modeling and simulation. In doing so, we argue that the potential for evaluation is a function of how the modeling and simulation software will be used in a particular setting. In making this argument, we move from strict, engineering and physics oriented approaches to V&V to a broader project of model evaluation, which asserts that the systematic, rigorous, and transparent accumulation of evidence about a model's performance under conditions of uncertainty is a reasonable and necessary goal for model evaluation, regardless of discipline. How to achieve the accumulation of evidence in areas outside physics and engineering is a significant research challenge, but one that requires addressing as modeling and simulation tools move out of research laboratories and into the hands of decision makers. This report provides an assessment of our thinking on ASC Verification and Validation, and argues for further extending V&V research in the physical and engineering sciences toward a broader program of model evaluation in situations of high consequence decision-making.
Theoretical and computer models of detonation in solid explosives
Tarver, C.M.; Urtiew, P.A.
1997-10-01
Recent experimental and theoretical advances in understanding energy transfer and chemical kinetics have led to improved models of detonation waves in solid explosives. The Nonequilibrium Zeldovich - von Neumann - Doring (NEZND) model is supported by picosecond laser experiments and molecular dynamics simulations of the multiphonon up-pumping and internal vibrational energy redistribution (IVR) processes by which the unreacted explosive molecules are excited to the transition state(s) preceding reaction behind the leading shock front(s). High temperature, high density transition state theory calculates the induction times measured by laser interferometric techniques. Exothermic chain reactions form product gases in highly excited vibrational states, which have been demonstrated to rapidly equilibrate via supercollisions. Embedded gauge and Fabry-Perot techniques measure the rates of reaction product expansion as thermal and chemical equilibrium is approached. Detonation reaction zone lengths in carbon-rich condensed phase explosives depend on the relatively slow formation of solid graphite or diamond. The Ignition and Growth reactive flow model based on pressure dependent reaction rates and Jones-Wilkins-Lee (JWL) equations of state has reproduced this nanosecond time resolved experimental data and thus has yielded accurate average reaction zone descriptions in one-, two- and three- dimensional hydrodynamic code calculations. The next generation reactive flow model requires improved equations of state and temperature dependent chemical kinetics. Such a model is being developed for the ALE3D hydrodynamic code, in which heat transfer and Arrhenius kinetics are intimately linked to the hydrodynamics.
A Vast Machine Computer Models, Climate Data, and the Politics of Global Warming
Edwards, Paul N.
A Vast Machine Computer Models, Climate Data, and the Politics of Global Warming Paul N. Edwards models, climate data, and the politics of global warming / Paul N. Edwards. p. cm. Includes. Climatology--History. 3. Meteorology--History. 4. Climatology--Technological innovation. 5. Global temperature
Computational modeling of thermal conductivity of single walled carbon nanotube polymer composites
Maruyama, Shigeo
was developed to study the thermal conductivity of single walled carbon nanotube (SWNT)-polymer composites1 Computational modeling of thermal conductivity of single walled carbon nanotube polymer resistance on effective conductivity of composites were quantified. The present model is a useful tool
The Application of L-systems and Developmental Models to Computer Art,
McCormack, Jon
The Application of L-systems and Developmental Models to Computer Art, Animation, and Music Synthesis Jon McCormack, B.Sc.(Hons), Grad Dip. Art (Film & Television) A Thesis Submitted for the Degree.3.1 Generative Modelling Systems ................................................ 8 1.3.2 Critical and Art
Toward Cost-Sensitive Modeling for Intrusion Detection Computer Science Department
Toward Cost-Sensitive Modeling for Intrusion Detection Wenke Lee Computer Science Department North,ezk,weaselg@cs.columbia.edu Abstract Intrusion detection systems need to maximize security while minimizing costs. In this paper, we study the problem of building cost-sensitive intrusion detection models. We examine the major cost
STATISTICAL MODELING OF THE LUNG NODULES IN LOW DOSE COMPUTED TOMOGRAPHY SCANS OF THE CHEST
Louisville, University of
STATISTICAL MODELING OF THE LUNG NODULES IN LOW DOSE COMPUTED TOMOGRAPHY SCANS OF THE CHEST Amal in automatic detection of the lung nodules and is compared with respect to parametric nodule models in terms appearing in low dose CT (LDCT) scans of the human chest. Four types of common lung nodules are analyzed
A two-layer granular landslide model for tsunami wave generation: Theory and computation
Kirby, James T.
A two-layer granular landslide model for tsunami wave generation: Theory and computation Gangfeng for granular landslide motion and tsunami wave generation. The landslide, either submarine or subaerial experiments on impulsive wave generation by subaerial granular landslides. Model results illustrate a complex
Stallard, Alisha M.
2011-08-08
. This prompted a reconstruction of the dose to the worker’s hands. The computer code MCNP was chosen to model the tasks that the worker performed to evaluate the potential nonuniform hand dose distribution. A model was constructed similar to the worker’s hands...
A computational model for predicting damage evolution in laminated composite plates
Phillips, Mark Lane
1999-01-01
computationally tenable is shown herein. Due to the complicated nature of the many cracks and their interactions, a multi-scale micro-meso-local-global methodology is employed in order to model damage modes. Interface degradation is first modeled analytically...
Modeling and Design of RF MEMS Structures Using Computationally Efficient Numerical Techniques
Tentzeris, Manos
Modeling and Design of RF MEMS Structures Using Computationally Efficient Numerical Techniques N. A Abstract The modeling of MEMS structures using MRTD is presented. Many complex RF structures have been communication systems efficiently and accurately. Specifically, micromachined structures such as MEMS
Tentzeris, Manos
Modeling and Optimization of RF-MEMS Reconfigurable Tuners with Computationally Efficient Time of Technology, Atlanta, GA 30332 2 Raytheon Company, Tucson AZ, 85734 Abstract -- Modern RF-MEMS device design methods in which the FDTD technique can be used to model a reconfigurable RF-MEMS tuner. A new method
An efficient computational model for macroscale simulations of moving contact lines
Boyer, Edmond
with CO2, for example). A major challenge in numerical simulations of moving contact linesAn efficient computational model for macroscale simulations of moving contact lines Y. Sui1 simulation of moving contact lines. The main purpose is to formulate and test a model wherein the macroscale
Coupling remote sensing with computational fluid dynamics modelling to estimate lake chlorophyll form 17 October 2000; accepted 1 June 2001 Abstract A remotely sensed image of Loch Leven, a shallow in the remotely sensed image. It is proposed that CFD modelling benefits the interpretation of remotely sensed
GWU Department of Mathematics Topics in Model Theory: Classical and Computable
Harizanov, Valentina S.
framework for the notions of language, meaning, and truth. A model, a concept used in all of sciences course will be, in some sense, self-contained. We will start by reviewing the fundamental conceptsÂ194 (survey chapter without proofs). (3) V. Harizanov, "Pure computable model theory," in the volume: Handbook
HIGH RESOLUTION FORWARD AND INVERSE EARTHQUAKE MODELING ON TERASCALE COMPUTERS
Shewchuk, Jonathan
highly populated seismic region in the U.S., it has well- characterized geological structures (including in characterizing earthquake source and basin material properties, a critical remaining challenge is to invert basin geology and earthquake sources, and to use this capability to model and forecast strong ground
CASE FOR SUPPORT: Computational Modeling of Salience Sensitive Control
Heinke, Dietmar
in neural network modeling, machine learning, adaptive systems in general and self-organising systems] and verification of real-time systems [6]. A large amount of this research has been performed using the CADP verification environment, which is one of the most powerful tool suites available, boasting a spectrum
A quantitative model of computation Dan R. Ghica
Ghica, Dan
languages. We define a Hyland-Ong-style games framework called slot games, which consists of HO games be nevertheless difficult, if not impossible, to handle using known operational techniques. Categories and Subject]: Modeling techniques. General Terms: Languages, Performance, Theory, Verifi- cation Keywords: Game semantics
Corley, Richard A.; Kabilan, Senthil; Kuprat, Andrew P.; Carson, James P.; Minard, Kevin R.; Jacob, Rick E.; Timchalk, Charles; Glenny, Robb W.; Pipavath, Sudhaker; Cox, Timothy C.; Wallis, Chris; Larson, Richard; Fanucchi, M.; Postlewait, Ed; Einstein, Daniel R.
2012-07-01
Coupling computational fluid dynamics (CFD) with physiologically based pharmacokinetic (PBPK) models is useful for predicting site-specific dosimetry of airborne materials in the respiratory tract and elucidating the importance of species differences in anatomy, physiology, and breathing patterns. Historically, these models were limited to discrete regions of the respiratory system. CFD/PBPK models have now been developed for the rat, monkey, and human that encompass airways from the nose or mouth to the lung. A PBPK model previously developed to describe acrolein uptake in nasal tissues was adapted to the extended airway models as an example application. Model parameters for each anatomic region were obtained from the literature, measured directly, or estimated from published data. Airflow and site-specific acrolein uptake patterns were determined under steadystate inhalation conditions to provide direct comparisons with prior data and nasalonly simulations. Results confirmed that regional uptake was dependent upon airflow rates and acrolein concentrations with nasal extraction efficiencies predicted to be greatest in the rat, followed by the monkey, then the human. For human oral-breathing simulations, acrolein uptake rates in oropharyngeal and laryngeal tissues were comparable to nasal tissues following nasal breathing under the same exposure conditions. For both breathing modes, higher uptake rates were predicted for lower tracheo-bronchial tissues of humans than either the rat or monkey. These extended airway models provide a unique foundation for comparing dosimetry across a significantly more extensive range of conducting airways in the rat, monkey, and human than prior CFD models.
CASTING DEFECT MODELING IN AN INTEGRATED COMPUTATIONAL MATERIALS ENGINEERING APPROACH
Sabau, Adrian S [ORNL
2015-01-01
To accelerate the introduction of new cast alloys, the simultaneous modeling and simulation of multiphysical phenomena needs to be considered in the design and optimization of mechanical properties of cast components. The required models related to casting defects, such as microporosity and hot tears, are reviewed. Three aluminum alloys are considered A356, 356 and 319. The data on calculated solidification shrinkage is presented and its effects on microporosity levels discussed. Examples are given for predicting microporosity defects and microstructure distribution for a plate casting. Models to predict fatigue life and yield stress are briefly highlighted here for the sake of completion and to illustrate how the length scales of the microstructure features as well as porosity defects are taken into account for modeling the mechanical properties. Thus, the data on casting defects, including microstructure features, is crucial for evaluating the final performance-related properties of the component. ACKNOWLEDGEMENTS This work was performed under a Cooperative Research and Development Agreement (CRADA) with the Nemak Inc., and Chrysler Co. for the project "High Performance Cast Aluminum Alloys for Next Generation Passenger Vehicle Engines. The author would also like to thank Amit Shyam for reviewing the paper and Andres Rodriguez of Nemak Inc. Research sponsored by the U. S. Department of Energy, Office of Energy Efficiency and Renewable Energy, Vehicle Technologies Office, as part of the Propulsion Materials Program under contract DE-AC05-00OR22725 with UT-Battelle, LLC. Part of this research was conducted through the Oak Ridge National Laboratory's High Temperature Materials Laboratory User Program, which is sponsored by the U. S. Department of Energy, Office of Energy Efficiency and Renewable Energy, Vehicle Technologies Program.
Computational models for the berry phase in semiconductor quantum dots
Prabhakar, S. Melnik, R. V. N.; Sebetci, A.
2014-10-06
By developing a new model and its finite element implementation, we analyze the Berry phase low-dimensional semiconductor nanostructures, focusing on quantum dots (QDs). In particular, we solve the Schrödinger equation and investigate the evolution of the spin dynamics during the adiabatic transport of the QDs in the 2D plane along circular trajectory. Based on this study, we reveal that the Berry phase is highly sensitive to the Rashba and Dresselhaus spin-orbit lengths.
A Variable Refrigerant Flow Heat Pump Computer Model in EnergyPlus
Raustad, Richard A. [Florida Solar Energy Center
2013-01-01
This paper provides an overview of the variable refrigerant flow heat pump computer model included with the Department of Energy's EnergyPlusTM whole-building energy simulation software. The mathematical model for a variable refrigerant flow heat pump operating in cooling or heating mode, and a detailed model for the variable refrigerant flow direct-expansion (DX) cooling coil are described in detail.
Determination and modeling of the 3-D gradient refractive indices in crystalline lenses
Chan, Derek Y C
Determination and modeling of the 3-D gradient refractive indices in crystalline lenses Derek Y. C gradient refractive-index profiles in crystalline lenses is proposed. The input data are derivedfrom 2 crystalline lenses.2 This method is based on measuring the total refraction suffered by a light beam
Verification of a VRF Heat Pump Computer Model in EnergyPlus
Nigusse, Bereket; Raustad, Richard
2013-06-01
This paper provides verification results of the EnergyPlus variable refrigerant flow (VRF) heat pump computer model using manufacturer's performance data. The paper provides an overview of the VRF model, presents the verification methodology, and discusses the results. The verification provides quantitative comparison of full and part-load performance to manufacturer's data in cooling-only and heating-only modes of operation. The VRF heat pump computer model uses dual range bi-quadratic performance curves to represent capacity and Energy Input Ratio (EIR) as a function of indoor and outdoor air temperatures, and dual range quadratic performance curves as a function of part-load-ratio for modeling part-load performance. These performance curves are generated directly from manufacturer's published performance data. The verification compared the simulation output directly to manufacturer's performance data, and found that the dual range equation fit VRF heat pump computer model predicts the manufacturer's performance data very well over a wide range of indoor and outdoor temperatures and part-load conditions. The predicted capacity and electric power deviations are comparbale to equation-fit HVAC computer models commonly used for packaged and split unitary HVAC equipment.
Computational Human Performance Modeling For Alarm System Design
Jacques Hugo
2012-07-01
The introduction of new technologies like adaptive automation systems and advanced alarms processing and presentation techniques in nuclear power plants is already having an impact on the safety and effectiveness of plant operations and also the role of the control room operator. This impact is expected to escalate dramatically as more and more nuclear power utilities embark on upgrade projects in order to extend the lifetime of their plants. One of the most visible impacts in control rooms will be the need to replace aging alarm systems. Because most of these alarm systems use obsolete technologies, the methods, techniques and tools that were used to design the previous generation of alarm system designs are no longer effective and need to be updated. The same applies to the need to analyze and redefine operators’ alarm handling tasks. In the past, methods for analyzing human tasks and workload have relied on crude, paper-based methods that often lacked traceability. New approaches are needed to allow analysts to model and represent the new concepts of alarm operation and human-system interaction. State-of-the-art task simulation tools are now available that offer a cost-effective and efficient method for examining the effect of operator performance in different conditions and operational scenarios. A discrete event simulation system was used by human factors researchers at the Idaho National Laboratory to develop a generic alarm handling model to examine the effect of operator performance with simulated modern alarm system. It allowed analysts to evaluate alarm generation patterns as well as critical task times and human workload predicted by the system.
Wilson, John 1988-
2012-11-15
physiologic geometries obtained from confocal images of a rat mesenteric lymphatic vessel to determine the characteristics of NO transport in the lymphatic flow regime. Both steady and unsteady analyses were performed. Steady models were simulated...
Mohaghegh, Shahab
Determination From Well Log Data: Part 1- A Comparative Study, Model Development Balan, B., Mohaghegh, S., Ameri use of empirically determined models, multiple variable regression, and artificial neural networks present only the model develop- ment phase in which we are testing the capability of each method to match
Modeling of BWR core meltdown accidents - for application in the MELRPI. MOD2 computer code
Koh, B R; Kim, S H; Taleyarkhan, R P; Podowski, M Z; Lahey, Jr, R T
1985-04-01
This report summarizes improvements and modifications made in the MELRPI computer code. A major difference between this new, updated version of the code, called MELRPI.MOD2, and the one reported previously, concerns the inclusion of a model for the BWR emergency core cooling systems (ECCS). This model and its computer implementation, the ECCRPI subroutine, account for various emergency injection modes, for both intact and rubblized geometries. Other changes to MELRPI deal with an improved model for canister wall oxidation, rubble bed modeling, and numerical integration of system equations. A complete documentation of the entire MELRPI.MOD2 code is also given, including an input guide, list of subroutines, sample input/output and program listing.
Rakowski, Cynthia L.; Serkowski, John A.; Richmond, Marshall C.; Perkins, William A.
2010-12-01
Although fisheries biology studies are frequently performed at US Army Corps of Engineers (USACE) projects along the Columbia and Snake Rivers, there is currently no consistent definition of the ``forebay'' and ``tailrace'' regions for these studies. At this time, each study may use somewhat arbitrary lines (e.g., the Boat Restriction Zone) to define the upstream and downstream limits of the study, which may be significantly different at each project. Fisheries researchers are interested in establishing a consistent definition of project forebay and tailrace regions for the hydroelectric projects on the lower Columbia and Snake rivers. The Hydraulic Extent of a project was defined by USACE (Brad Eppard, USACE-CENWP) as follows: The river reach directly upstream (forebay) and downstream (tailrace) of a project that is influenced by the normal range of dam operations. Outside this reach, for a particular river discharge, changes in dam operations cannot be detected by hydraulic measurement. The purpose of this study was to, in consultation with USACE and regional representatives, develop and apply a consistent set of criteria for determining the hydraulic extent of each of the projects in the lower Columbia and Snake rivers. A 2D depth-averaged river model, MASS2, was applied to the Snake and Columbia Rivers. New computational meshes were developed most reaches and the underlying bathymetric data updated to the most current survey data. The computational meshes resolved each spillway bay and turbine unit at each project and extended from project to project. MASS2 was run for a range of total river flows and each flow for a range of project operations at each project. The modeled flow was analyzed to determine the range of velocity magnitude differences and the range of flow direction differences at each location in the computational mesh for each total river flow. Maps of the differences in flow direction and velocity magnitude were created. USACE fishery biologists requested data analysis to determine the project hydraulic extent based on the following criteria: 1) For areas where the mean velocities are less than 4 ft/s, the water velocity differences between operations are not greater than 0.5 ft/sec and /or the differences in water flow direction are not greater than 10 degrees, 2) If mean water velocity is 4.0 ft/second or greater the boundary is determined using the differences in water flow direction (i.e., not greater than 10 degrees). Based on these criteria, and excluding areas with a mean velocity of less than 0.1 ft/s (within the error of the model), a final set of graphics were developed that included data from all flows and all operations. Although each hydroelectric project has a different physical setting, there were some common results. The downstream hydraulic extent tended to be greater than the hydraulic extent in the forebay. The hydraulic extent of the projects tended to be larger at the mid-range flows. At higher flows, the channel geometry tends to reduce the impact of project operations.
Medical Nuclear Supply Chain Design: A Tractable Network Model and Computational
Nagurney, Anna
importance of considering waste management. Anna Nagurney and Ladimer S. Nagurney Medical Nuclear SupplyMedical Nuclear Supply Chain Design: A Tractable Network Model and Computational Approach Anna Nagurney1 and Ladimer S. Nagurney2 1John F. Smith Memorial Professor - Isenberg School of Management
Medical Nuclear Supply Chain Design: A Tractable Network Model and Computational Approach
Nagurney, Anna
operational cost minimization, and the minimization of cost associated with nuclear waste discarding, coupled, the underlying physics of radioactive decay, and the inclusion of waste management. We focus on Molybdenum-99 dueMedical Nuclear Supply Chain Design: A Tractable Network Model and Computational Approach Anna
Many Task Computing for Modeling the Fate of Oil Discharged from the Deep Water Horizon Well
Many Task Computing for Modeling the Fate of Oil Discharged from the Deep Water Horizon Well@jhu.edu Abstract--The Deep Water Horizon well blowout on April 20th 2010 discharged between 40,000 - 1.2 million@med.miami.edu O. M. Knio Dept of Mechanical Engineering Johns Hopkins University Baltimore, MD knio
Computational Modeling of Neural Plasticity for Self-Organization of Neural Networks
Jin, Yaochu
Computational Modeling of Neural Plasticity for Self-Organization of Neural Networks Joseph Chrol on the learning per- formance of neural networks for accomplishing machine learning tasks such as classication, dynamics and learning per- formance of neural networks remains elusive. The purpose of this article
In-Vehicle Testing and Computer Modeling of Electric Vehicle Batteries
Wang, Chao-Yang
In-Vehicle Testing and Computer Modeling of Electric Vehicle Batteries B. Thomas, W.B. Gu, J.edu Abstract A combined simulation and testing approach has been developed to evaluate battery packs in real accelerates battery development cycle, and enables innovative battery design and optimization. Several
Computational Modeling of Electrolyte/Cathode Interfaces in Proton Exchange Membrane Fuel Cells
Bjørnstad, Ottar Nordal
Computational Modeling of Electrolyte/Cathode Interfaces in Proton Exchange Membrane Fuel Cells Dr Proton exchange membrane fuel cells (PEMFCs) are alternative energy conversion devices that efficiently. The fundamental relationship between operating conditions and device performance will help to optimize the device
A computational strategy for multiscale systems with applications to Lorenz 96 model
Van Den Eijnden, Eric
A computational strategy for multiscale systems with applications to Lorenz 96 model Ibrahim 2004 Available online Abstract Numerical schemes for systems with multiple spatio-temporal scales are investigated. The multiscale schemes use asymptotic results for this type of systems which guarantee
P&P: a Combined Push-Pull Model for Resource Monitoring in Cloud Computing Environment
Wang, Liqiang
P&P: a Combined Push-Pull Model for Resource Monitoring in Cloud Computing Environment He Huang, various platforms and software. Resource monitoring involves collecting information of system resources to facilitate decision making by other components in Cloud environ- ment. It is the foundation of many major
COMPUTATIONAL CHALLENGES IN THE NUMERICAL TREATMENT OF LARGE AIR POLLUTION MODELS
Ostromsky, Tzvetan
COMPUTATIONAL CHALLENGES IN THE NUMERICAL TREATMENT OF LARGE AIR POLLUTION MODELS I. DIMOV , K. GEORGIEVy, TZ. OSTROMSKY , R. J. VAN DER PASz, AND Z. ZLATEVx Abstract. The air pollution, and especially the reduction of the air pollution to some acceptable levels, is an important environmental problem, which
MATHEMATICAL PERGAMON Mathematical and Computer Modelling 35 (2002) 1371-1375
Gorban, Alexander N.
2002-01-01
Application to the Efficiency of Free Flow Turbines A. GORBAN' Institute of Computational Modeling, Russian obstacle is considered. Its application to estimating the efficiency of free flow turbines is discussed hydraulic turbines, i.e., the turbines that work without dams [l]. For this kind of turbine, the term
A Computational Model of Aging and Calcification in the Aortic Heart Valve
Mofrad, Mohammad R. K.
A Computational Model of Aging and Calcification in the Aortic Heart Valve Eli J. Weinberg1 of America Abstract The aortic heart valve undergoes geometric and mechanical changes over time. The cusps of a normal, healthy valve thicken and become less extensible over time. In the disease calcific aortic
Computational Modeling and the Experimental Plasma Research Program A White Paper Submitted of the fusion energy program. The experimental plasma research (EPR) program is well positioned to make major in fusion development and promote scientific discovery. Experimental plasma research projects explore
PhD Position Available: integrative biomechanics, computational modeling, nonlinear dynamics
Clewley, Robert
PhD Position Available: integrative biomechanics, computational modeling, nonlinear dynamics and mathematical analysis of biomechanical and neural control systems. We are looking for an excellent and highly.edu/~biodhe/#Research). These are being used to study the Crayfish swim escape mechanism as a case study in integrative biomechanical
Computational framework for modeling the dynamic evolution of large-scale multi-agent organizations
Lazar, Alina
disciplines as well. This is because the emergence of such a social structure can have a profound impact 48202 ABSTRACT The process by which complex social entities such as the state emerged from lower level on the societies' physical and social environment. However, the task of developing realistic computational models
Broader source: Energy.gov [DOE]
The objective of this Funding Opportunity Announcement (FOA) is to leverage scientific advancements in mathematics and computation for application to power system models and software tools, with the long-term goal of enabling real-time protection and control based on wide-area sensor measurements.
Buyya, Rajkumar
infrastructure (hardware, software, services) for different application and service models under varying load problem to tackle. To simplify this process, in this paper we propose CloudSim: a new generalized Cloud computing infrastructures and management services. The simulation framework has the following
Margaliot, Michael
1 Entrainment to Periodic Initiation and Transition Rates in a Computational Model for Gene, the biological system must entrain or phase-lock to the periodic excitation. Entrainment is also important in synthetic biology. For example, connecting several artificial biological systems that entrain to a common
Sontag, Eduardo
Entrainment to Periodic Initiation and Transition Rates in a Computational Model for Gene to the solar day. In the terminology of systems theory, the biological system must entrain or phase-lock to the periodic excitation. Entrainment is also important in synthetic biology. For example, connecting several
3D Bone Microarchitecture Modeling and Fracture Risk Department of Computer
Buffalo, State University of New York
3D Bone Microarchitecture Modeling and Fracture Risk Prediction Hui Li Department of Computer will also rise. It calls for innovative research on understanding of osteoporo- sis and fracture mechanisms-of-the-art probabilistic approach to analyze bone fracture risk factors including demographic attributes and life styles
Baker, Jack W.
Conditional Spectrum Computation Incorporating Multiple Causal Earthquakes and Ground-Motion Prediction Models by Ting Lin, Stephen C. Harmsen, Jack W. Baker, and Nicolas Luco Abstract The conditional uncertainties in all earthquake scenarios and resulting ground motions, as well as the epistemic uncertainties
A Computational Model for Understanding Stem Cell, Trophectoderm and Endoderm Lineage Determination
Peterson, Carsten
under the terms of the Creative Commons Attribution License, which permits unrestricted use Grant and the National Science Foundation-FIBR Award EF-0330786. The funders had no role in study design
International Association for Cryptologic Research (IACR)
Modeling Computational Security in LongLived Systems # ## Ran Canetti 1,2 , Ling Cheung 2 , Dilsun Introduction Computational security in longlived systems: Security properties of cryptographic protocols computational power. This type of security degrades progressively over the lifetime of a protocol. However, some
Buyya, Rajkumar
CloudAnalyst: A CloudSim-based Visual Modeller for Analysing Cloud Computing Environments and Applications Bhathiya Wickremasinghe1 , Rodrigo N. Calheiros2 , and Rajkumar Buyya1 1 The Cloud Computing and Distributed Systems (CLOUDS) Laboratory Department of Computer Science and Software Engineering The University
Technical Review of the CENWP Computational Fluid Dynamics Model of the John Day Dam Forebay
Rakowski, Cynthia L.; Serkowski, John A.; Richmond, Marshall C.
2010-12-01
The US Army Corps of Engineers Portland District (CENWP) has developed a computational fluid dynamics (CFD) model of the John Day forebay on the Columbia River to aid in the development and design of alternatives to improve juvenile salmon passage at the John Day Project. At the request of CENWP, Pacific Northwest National Laboratory (PNNL) Hydrology Group has conducted a technical review of CENWP's CFD model run in CFD solver software, STAR-CD. PNNL has extensive experience developing and applying 3D CFD models run in STAR-CD for Columbia River hydroelectric projects. The John Day forebay model developed by CENWP is adequately configured and validated. The model is ready for use simulating forebay hydraulics for structural and operational alternatives. The approach and method are sound, however CENWP has identified some improvements that need to be made for future models and for modifications to this existing model.
Computer modeling approach for microsphere-packed bone scaffold Pallavi Lal, Wei Sun*
Sun, Wei
bone graft [5,6], for structural and human cellular assessment of scaffolds for bone repair [7 modeling approach for constructing a three-dimensional microsphere-packed bone graft structure is presented packing model to determine the number of microspheres packed in a synthesized bone graft. The pore size
Yock, Adam D. Kudchadker, Rajat J.; Rao, Arvind; Dong, Lei; Beadle, Beth M.; Garden, Adam S.; Court, Laurence E.
2014-05-15
Purpose: The purpose of this work was to develop and evaluate the accuracy of several predictive models of variation in tumor volume throughout the course of radiation therapy. Methods: Nineteen patients with oropharyngeal cancers were imaged daily with CT-on-rails for image-guided alignment per an institutional protocol. The daily volumes of 35 tumors in these 19 patients were determined and used to generate (1) a linear model in which tumor volume changed at a constant rate, (2) a general linear model that utilized the power fit relationship between the daily and initial tumor volumes, and (3) a functional general linear model that identified and exploited the primary modes of variation between time series describing the changing tumor volumes. Primary and nodal tumor volumes were examined separately. The accuracy of these models in predicting daily tumor volumes were compared with those of static and linear reference models using leave-one-out cross-validation. Results: In predicting the daily volume of primary tumors, the general linear model and the functional general linear model were more accurate than the static reference model by 9.9% (range: ?11.6%–23.8%) and 14.6% (range: ?7.3%–27.5%), respectively, and were more accurate than the linear reference model by 14.2% (range: ?6.8%–40.3%) and 13.1% (range: ?1.5%–52.5%), respectively. In predicting the daily volume of nodal tumors, only the 14.4% (range: ?11.1%–20.5%) improvement in accuracy of the functional general linear model compared to the static reference model was statistically significant. Conclusions: A general linear model and a functional general linear model trained on data from a small population of patients can predict the primary tumor volume throughout the course of radiation therapy with greater accuracy than standard reference models. These more accurate models may increase the prognostic value of information about the tumor garnered from pretreatment computed tomography images and facilitate improved treatment management.
Computer modeling of electromagnetic edge containment in twin-roll casting
Chang, F.C.; Turner, L.R.; Hull, J.R.; Wang, Y.H.; Blazek, K.E.
1998-07-01
This paper presents modeling studies of magnetohydrodynamics (MHD) analysis in twin-roll casting. Argonne National Laboratory (ANL) and Inland Steel Company have worked together to develop a 3-D computer model that can predict eddy currents, fluid flows, and liquid metal containment for an electromagnetic (EM) edge containment device. This mathematical model can greatly shorten casting research on the use of EM fields for liquid metal containment and control. It can also optimize the existing casting processes and minimize expensive, time-consuming full-scale testing. The model was verified by comparing predictions with experimental results of liquid-metal containment and fluid flow in EM edge dams designed at Inland Steel for twin-roll casting. Numerical simulation was performed by coupling a three-dimensional (3-D) finite-element EM code (ELEKTRA) and a 3-D finite-difference fluids code (CaPS-EM) to solve Maxwell`s equations, Ohm`s law, Navier-Stokes equations, and transport equations of turbulence flow in a casting process that uses EM fields. ELEKTRA is able to predict the eddy-current distribution and electromagnetic forces in complex geometry. CaPS-EM is capable of modeling fluid flows with free-surfaces and dynamic rollers. The computed 3-D magnetic fields and induced eddy currents in ELEKTRA are used as input to flow-field computations in CaPS-EM. Results of the numerical simulation compared well with measurements obtained from both static and dynamic tests.
High-Performance Computing Modeling Advances Accelerator Science for High-Energy Physics
DOE Public Access Gateway for Energy & Science Beta (PAGES Beta)
Amundson, James; Macridin, Alexandru; Spentzouris, Panagiotis
2014-11-01
The development and optimization of particle accelerators are essential for advancing our understanding of the properties of matter, energy, space and time. Particle accelerators are complex devices whose behavior involves many physical effects on multiple scales. Therefore, advanced computational tools utilizing high-performance computing (HPC) are essential for accurately modeling them. In the past decade, the DOE SciDAC program has produced such accelerator-modeling tools, which have beem employed to tackle some of the most difficult accelerator science problems. In this article we discuss the Synergia beam-dynamics framework and its applications to high-intensity particle accelerator physics. Synergia is an accelerator simulation packagemore »capable of handling the entire spectrum of beam dynamics simulations. We present the design principles, key physical and numerical models in Synergia and its performance on HPC platforms. Finally, we present the results of Synergia applications for the Fermilab proton source upgrade, known as the Proton Improvement Plan (PIP).« less
High-Performance Computing Modeling Advances Accelerator Science for High-Energy Physics
Amundson, James; Macridin, Alexandru; Spentzouris, Panagiotis
2014-11-01
The development and optimization of particle accelerators are essential for advancing our understanding of the properties of matter, energy, space and time. Particle accelerators are complex devices whose behavior involves many physical effects on multiple scales. Therefore, advanced computational tools utilizing high-performance computing (HPC) are essential for accurately modeling them. In the past decade, the DOE SciDAC program has produced such accelerator-modeling tools, which have beem employed to tackle some of the most difficult accelerator science problems. In this article we discuss the Synergia beam-dynamics framework and its applications to high-intensity particle accelerator physics. Synergia is an accelerator simulation package capable of handling the entire spectrum of beam dynamics simulations. We present the design principles, key physical and numerical models in Synergia and its performance on HPC platforms. Finally, we present the results of Synergia applications for the Fermilab proton source upgrade, known as the Proton Improvement Plan (PIP).
Gabriel Álvarez; Luis Martínez Alonso; Elena Medina
2011-01-14
We present a method to compute the genus expansion of the free energy of Hermitian matrix models from the large N expansion of the recurrence coefficients of the associated family of orthogonal polynomials. The method is based on the Bleher-Its deformation of the model, on its associated integral representation of the free energy, and on a method for solving the string equation which uses the resolvent of the Lax operator of the underlying Toda hierarchy. As a byproduct we obtain an efficient algorithm to compute generating functions for the enumeration of labeled k-maps which does not require the explicit expressions of the coefficients of the topological expansion. Finally we discuss the regularization of singular one-cut models within this approach.
Vandersall, Jennifer A.; Gardner, Shea N.; Clague, David S.
2010-05-04
A computational method and computer-based system of modeling DNA synthesis for the design and interpretation of PCR amplification, parallel DNA synthesis, and microarray chip analysis. The method and system include modules that address the bioinformatics, kinetics, and thermodynamics of DNA amplification and synthesis. Specifically, the steps of DNA selection, as well as the kinetics and thermodynamics of DNA hybridization and extensions, are addressed, which enable the optimization of the processing and the prediction of the products as a function of DNA sequence, mixing protocol, time, temperature and concentration of species.
Carbajo, Juan; Jeong, Hae-Yong; Wigeland, Roald; Corradini, Michael; Schmidt, Rodney Cannon; Thomas, Justin; Wei, Tom; Sofu, Tanju; Ludewig, Hans; Tobita, Yoshiharu; Ohshima, Hiroyuki; Serre, Frederic
2011-06-01
This report summarizes the results of an expert-opinion elicitation activity designed to qualitatively assess the status and capabilities of currently available computer codes and models for accident analysis and reactor safety calculations of advanced sodium fast reactors, and identify important gaps. The twelve-member panel consisted of representatives from five U.S. National Laboratories (SNL, ANL, INL, ORNL, and BNL), the University of Wisconsin, the KAERI, the JAEA, and the CEA. The major portion of this elicitation activity occurred during a two-day meeting held on Aug. 10-11, 2010 at Argonne National Laboratory. There were two primary objectives of this work: (1) Identify computer codes currently available for SFR accident analysis and reactor safety calculations; and (2) Assess the status and capability of current US computer codes to adequately model the required accident scenarios and associated phenomena, and identify important gaps. During the review, panel members identified over 60 computer codes that are currently available in the international community to perform different aspects of SFR safety analysis for various event scenarios and accident categories. A brief description of each of these codes together with references (when available) is provided. An adaptation of the Predictive Capability Maturity Model (PCMM) for computational modeling and simulation is described for use in this work. The panel's assessment of the available US codes is presented in the form of nine tables, organized into groups of three for each of three risk categories considered: anticipated operational occurrences (AOOs), design basis accidents (DBA), and beyond design basis accidents (BDBA). A set of summary conclusions are drawn from the results obtained. At the highest level, the panel judged that current US code capabilities are adequate for licensing given reasonable margins, but expressed concern that US code development activities had stagnated and that the experienced user-base and the experimental validation base was decaying away quickly.
Computational Modeling of Blood Flow in the TrapEase Inferior Vena Cava Filter
Singer, M A; Henshaw, W D; Wang, S L
2008-02-04
To evaluate the flow hemodynamics of the TrapEase vena cava filter using three dimensional computational fluid dynamics, including simulated thrombi of multiple shapes, sizes, and trapping positions. The study was performed to identify potential areas of recirculation and stagnation and areas in which trapped thrombi may influence intrafilter thrombosis. Computer models of the TrapEase filter, thrombi (volumes ranging from 0.25mL to 2mL, 3 different shapes), and a 23mm diameter cava were constructed. The hemodynamics of steady-state flow at Reynolds number 600 was examined for the unoccluded and partially occluded filter. Axial velocity contours and wall shear stresses were computed. Flow in the unoccluded TrapEase filter experienced minimal disruption, except near the superior and inferior tips where low velocity flow was observed. For spherical thrombi in the superior trapping position, stagnant and recirculating flow was observed downstream of the thrombus; the volume of stagnant flow and the peak wall shear stress increased monotonically with thrombus volume. For inferiorly trapped spherical thrombi, marked disruption to the flow was observed along the cava wall ipsilateral to the thrombus and in the interior of the filter. Spherically shaped thrombus produced a lower peak wall shear stress than conically shaped thrombus and a larger peak stress than ellipsoidal thrombus. We have designed and constructed a computer model of the flow hemodynamics of the TrapEase IVC filter with varying shapes, sizes, and positions of thrombi. The computer model offers several advantages over in vitro techniques including: improved resolution, ease of evaluating different thrombus sizes and shapes, and easy adaptation for new filter designs and flow parameters. Results from the model also support a previously reported finding from photochromic experiments that suggest the inferior trapping position of the TrapEase IVC filter leads to an intra-filter region of recirculating/stagnant flow with very low shear stress that may be thrombogenic.
Victoria, University of
On the Use of Computational Models for Wave Climate Assessment in Support of the Wave Energy On the Use of Computational Models for Wave Climate Assessment in Support of the Wave Energy Industry Effective, economic extraction of ocean wave energy requires an intimate under- standing of the ocean wave
DualTrust: A Trust Management Model for Swarm-Based Autonomic Computing Systems
Maiden, Wendy M.
2010-05-01
Trust management techniques must be adapted to the unique needs of the application architectures and problem domains to which they are applied. For autonomic computing systems that utilize mobile agents and ant colony algorithms for their sensor layer, certain characteristics of the mobile agent ant swarm -- their lightweight, ephemeral nature and indirect communication -- make this adaptation especially challenging. This thesis looks at the trust issues and opportunities in swarm-based autonomic computing systems and finds that by monitoring the trustworthiness of the autonomic managers rather than the swarming sensors, the trust management problem becomes much more scalable and still serves to protect the swarm. After analyzing the applicability of trust management research as it has been applied to architectures with similar characteristics, this thesis specifies the required characteristics for trust management mechanisms used to monitor the trustworthiness of entities in a swarm-based autonomic computing system and describes a trust model that meets these requirements.
Compare Energy Use in Variable Refrigerant Flow Heat Pumps Field Demonstration and Computer Model
Sharma, Chandan; Raustad, Richard
2013-06-01
Variable Refrigerant Flow (VRF) heat pumps are often regarded as energy efficient air-conditioning systems which offer electricity savings as well as reduction in peak electric demand while providing improved individual zone setpoint control. One of the key advantages of VRF systems is minimal duct losses which provide significant reduction in energy use and duct space. However, there is limited data available to show their actual performance in the field. Since VRF systems are increasingly gaining market share in the US, it is highly desirable to have more actual field performance data of these systems. An effort was made in this direction to monitor VRF system performance over an extended period of time in a US national lab test facility. Due to increasing demand by the energy modeling community, an empirical model to simulate VRF systems was implemented in the building simulation program EnergyPlus. This paper presents the comparison of energy consumption as measured in the national lab and as predicted by the program. For increased accuracy in the comparison, a customized weather file was created by using measured outdoor temperature and relative humidity at the test facility. Other inputs to the model included building construction, VRF system model based on lab measured performance, occupancy of the building, lighting/plug loads, and thermostat set-points etc. Infiltration model inputs were adjusted in the beginning to tune the computer model and then subsequent field measurements were compared to the simulation results. Differences between the computer model results and actual field measurements are discussed. The computer generated VRF performance closely resembled the field measurements.
James P. Crutchfield; Christopher J. Ellison; Ryan G. James; John R. Mahoney
2010-07-30
We adapt tools from information theory to analyze how an observer comes to synchronize with the hidden states of a finitary, stationary stochastic process. We show that synchronization is determined by both the process's internal organization and by an observer's model of it. We analyze these components using the convergence of state-block and block-state entropies, comparing them to the previously known convergence properties of the Shannon block entropy. Along the way, we introduce a hierarchy of information quantifiers as derivatives and integrals of these entropies, which parallels a similar hierarchy introduced for block entropy. We also draw out the duality between synchronization properties and a process's controllability. The tools lead to a new classification of a process's alternative representations in terms of minimality, synchronizability, and unifilarity.
Abundance determinations in HII regions: model fitting versus Te-method
L. S. Pilyugin
2002-11-14
The discrepancy between the oxygen abundances in high-metallicity HII regions determined through the Te-method (and/or through the corresponding "strong lines - oxygen abundance" calibration) and that determined through the model fitting (and/or through the corresponding "strong lines - oxygen abundance" calibration) is discussed. It is suggested to use the interstellar oxygen abundance in the solar vicinity, derived with very high precision from the high-resolution observations of the weak interstellar absorption lines towards the stars, as a "Rosetta stone" to verify the validity of the oxygen abundances derived in HII regions with the Te-method at high abundances. The agreement between the value of the oxygen abundance at the solar galactocentric distance traced by the abundances derived in HII regions through the Te-method and that derived from the interstellar absorption lines towards the stars is strong evidence in favor of that i) the two-zone model for Te seems to be a realistic interpretation of the temperature structure within HII regions, and ii) the classic Te-method provides accurate oxygen abundances in HII regions. It has been concluded that the "strong lines - oxygen abundance" calibrations must be based on the HII regions with the oxygen abundances derived with the Te-method but not on the existing grids of the models for HII regions.
Lambert, Michael Andrew
1990-01-01
characterization of nuclear pressure boundary materials. Substituting typical values for upper shelf fracture toughness and yield suength, 200 ksi~in. (220 MPaWm) and 50 ksi (345 MPa), respectively, for nudear pressure vessel steels into Bquatton B3 results in a... Carlo Computer Simulation of the Drop-Weight Test for the Determination of Nil-Ductility Transition Temperature. (August 1990) Michael A. Lambert, B. S. , Texas AdcM University Chair of Advisory Committee: Dr. ' Ted L. Anderson Material inhomogeneity...
Parametric Studies and Optimization of Eddy Current Techniques through Computer Modeling
Todorov, E. I. [EWI, Engineering and NDE, 1250 Arthur E. Adams Dr., Columbus, OH 43221-3585 (United States)
2007-03-21
The paper demonstrates the use of computer models for parametric studies and optimization of surface and subsurface eddy current techniques. The study with high-frequency probe investigates the effect of eddy current frequency and probe shape on the detectability of flaws in the steel substrate. The low-frequency sliding probe study addresses the effect of conductivity between the fastener and the hole, frequency and coil separation distance on detectability of flaws in subsurface layers.
NREL Computer Models Integrate Wind Turbines with Floating Platforms (Fact Sheet)
Not Available
2011-07-01
Far off the shores of energy-hungry coastal cities, powerful winds blow over the open ocean, where the water is too deep for today's seabed-mounted offshore wind turbines. For the United States to tap into these vast offshore wind energy resources, wind turbines must be mounted on floating platforms to be cost effective. Researchers at the National Renewable Energy Laboratory (NREL) are supporting that development with computer models that allow detailed analyses of such floating wind turbines.
An Additive Bivariate Hierarchical Model for Functional Data and Related Computations
Redd, Andrew Middleton
2011-10-21
the penalties 11 are given is computationally intensive. In addition the space for nding the penalty parameters is four dimensional, two for each predictor variable, one each for the mean and principal component functions. Optimizing over a four dimensional... ed by the structure of the formula used. I implement two new formula operators that only work with the pfda package; %&% bind together variables, on the left side of the formula indicates the paired model, on the right an additive variable...
Path2Models: large-scale generation of computational models from biochemical pathway maps
2013-01-01
C: Petri net modelling of biological networks. Brief Bioinfonetworks [14-16] to discrete algebra [17] and dif- ferential equations [18], Petri
Test of the notch technique for determining the radial sensitivity of the optical model potential
Lei Yang; Cheng-Jian Lin; Hui-ming Jia; Xin-Xing Xu; Nan-Ru Ma; Li-Jie Sun; Feng Yang; Huan-Qiao Zhang; Zu-Hua Li; Dong-Xi Wang
2015-08-10
Detailed investigations on the notch technique are performed on the ideal data generated by the optical model potential parameters extracted from the 16O+208Pb system at the laboratory energy of 129.5 MeV, to study the sensitivities of this technique on the model parameters as well as the experimental data. It is found that, for the perturbation parameters, a sufficient large reduced fraction and an appropriate small perturbation width are necessary to determine the accurate radial sensitivity; while for the potential parameters, almost no dependence was observed. For the experimental measurements, the number of data points has little influence for the heavy target system, and the relative inner information of the nuclear potential can be derived when the measurement extended to a lower cross section.
Broader source: Energy.gov [DOE]
Presentation given by NREL at 2014 DOE Hydrogen and Fuel Cells Program and Vehicle Technologies Office Annual Merit Review and Peer Evaluation Meeting about significant enhancement of computational...
Efficient Computation of Info-Gap Robustness for Finite Element Models
Stull, Christopher J.; Hemez, Francois M.; Williams, Brian J.
2012-07-05
A recent research effort at LANL proposed info-gap decision theory as a framework by which to measure the predictive maturity of numerical models. Info-gap theory explores the trade-offs between accuracy, that is, the extent to which predictions reproduce the physical measurements, and robustness, that is, the extent to which predictions are insensitive to modeling assumptions. Both accuracy and robustness are necessary to demonstrate predictive maturity. However, conducting an info-gap analysis can present a formidable challenge, from the standpoint of the required computational resources. This is because a robustness function requires the resolution of multiple optimization problems. This report offers an alternative, adjoint methodology to assess the info-gap robustness of Ax = b-like numerical models solved for a solution x. Two situations that can arise in structural analysis and design are briefly described and contextualized within the info-gap decision theory framework. The treatments of the info-gap problems, using the adjoint methodology are outlined in detail, and the latter problem is solved for four separate finite element models. As compared to statistical sampling, the proposed methodology offers highly accurate approximations of info-gap robustness functions for the finite element models considered in the report, at a small fraction of the computational cost. It is noted that this report considers only linear systems; a natural follow-on study would extend the methodologies described herein to include nonlinear systems.
Chang, F.C.; Hull, J.R.; Wang, Y.H.; Blazek, K.E.
1996-02-01
A computer model was developed to predict eddy currents and fluid flows in molten steel. The model was verified by comparing predictions with experimental results of liquid-metal containment and fluid flow in electromagnetic (EM) edge dams (EMDs) designed at Inland Steel for twin-roll casting. The model can optimize the EMD design so it is suitable for application, and minimize expensive, time-consuming full-scale testing. Numerical simulation was performed by coupling a three-dimensional (3-D) finite-element EM code (ELEKTRA) and a 3-D finite-difference fluids code (CaPS-EM) to solve heat transfer, fluid flow, and turbulence transport in a casting process that involves EM fields. ELEKTRA is able to predict the eddy- current distribution and the electromagnetic forces in complex geometries. CaPS-EM is capable of modeling fluid flows with free surfaces. Results of the numerical simulation compared well with measurements obtained from a static test.
Leishear, Robert A.; Lee, Si Y.; Poirier, Michael R.; Steeper, Timothy J.; Ervin, Robert C.; Giddings, Billy J.; Stefanko, David B.; Harp, Keith D.; Fowley, Mark D.; Van Pelt, William B.
2012-10-07
Computational fluid dynamics (CFD) is recognized as a powerful engineering tool. That is, CFD has advanced over the years to the point where it can now give us deep insight into the analysis of very complex processes. There is a danger, though, that an engineer can place too much confidence in a simulation. If a user is not careful, it is easy to believe that if you plug in the numbers, the answer comes out, and you are done. This assumption can lead to significant errors. As we discovered in the course of a study on behalf of the Department of Energy's Savannah River Site in South Carolina, CFD models fail to capture some of the large variations inherent in complex processes. These variations, or scatter, in experimental data emerge from physical tests and are inadequately captured or expressed by calculated mean values for a process. This anomaly between experiment and theory can lead to serious errors in engineering analysis and design unless a correction factor, or safety factor, is experimentally validated. For this study, blending times for the mixing of salt solutions in large storage tanks were the process of concern under investigation. This study focused on the blending processes needed to mix salt solutions to ensure homogeneity within waste tanks, where homogeneity is required to control radioactivity levels during subsequent processing. Two of the requirements for this task were to determine the minimum number of submerged, centrifugal pumps required to blend the salt mixtures in a full-scale tank in half a day or less, and to recommend reasonable blending times to achieve nearly homogeneous salt mixtures. A full-scale, low-flow pump with a total discharge flow rate of 500 to 800 gpm was recommended with two opposing 2.27-inch diameter nozzles. To make this recommendation, both experimental and CFD modeling were performed. Lab researchers found that, although CFD provided good estimates of an average blending time, experimental blending times varied significantly from the average.
Quantifying the role of chaperones in protein translocation by computational modelling
Salvatore Assenza; Paolo De Los Rios; Alessandro Barducci
2014-11-25
The molecular chaperone Hsp70 plays a central role in the import of cytoplasmic proteins into organelles, driving their translocation by binding them from the organellar interior. Starting from the experimentally-determined structure of the \\textit{E. coli} Hsp70, we computed, by means of molecular simulations, the effective free-energy profile for substrate translocation upon chaperone binding. We then used the resulting free energy to quantitatively characterize the kinetics of the import process, whose comparison with unassisted translocation highlights the essential role played by Hsp70 in importing cytoplasmic proteins.
Steinhaus, Thomas
2010-01-01
Computational Fluid Dynamics (CFD) codes are being increasingly used in the field of fire safety engineering. They provide, amongst other things, velocity, species and heat flux distributions throughout the computational ...
Ellis, Matthew
2015-01-01
efficiency. . . . . . . . . . . . . . . . . . . . . . . .Performance and Computational Efficiency . . 4.4.1 Class ofestimation and computational efficiency. Journal of Process
Stack and cell modelling with SOFC3D: a computer program for the 3D simulations of
Herbin, Raphaèle
Stack and cell modelling with SOFC3D: a computer program for the 3D simulations of solid oxide fuel, France 1 Introduction SOFC3D is a computer program, which simulates the behaviour of a solid oxide fuel or the channels, the electrical potential \\Phi at any point of the solid part of the SOFC, and the molar fractions
Dr. Chenn Zhou
2008-10-15
Pulverized coal injection (PCI) into the blast furnace (BF) has been recognized as an effective way to decrease the coke and total energy consumption along with minimization of environmental impacts. However, increasing the amount of coal injected into the BF is currently limited by the lack of knowledge of some issues related to the process. It is therefore important to understand the complex physical and chemical phenomena in the PCI process. Due to the difficulty in attaining trus BF measurements, Computational fluid dynamics (CFD) modeling has been identified as a useful technology to provide such knowledge. CFD simulation is powerful for providing detailed information on flow properties and performing parametric studies for process design and optimization. In this project, comprehensive 3-D CFD models have been developed to simulate the PCI process under actual furnace conditions. These models provide raceway size and flow property distributions. The results have provided guidance for optimizing the PCI process.
Go-Smart: Web-based Computational Modeling of Minimally Invasive Cancer Treatments
Weir, Phil; Ellerweg, Roland; Alhonnoro, Tuomas; Pollari, Mika; Voglreiter, Philip; Mariappan, Panchatcharam; Flanagan, Ronan; Park, Chang Sub; Payne, Stephen; Staerk, Elmar; Voigt, Peter; Moche, Michael; Kolesnik, Marina
2015-01-01
The web-based Go-Smart environment is a scalable system that allows the prediction of minimally invasive cancer treatment. Interventional radiologists create a patient-specific 3D model by semi-automatic segmentation and registration of pre-interventional CT (Computed Tomography) and/or MRI (Magnetic Resonance Imaging) images in a 2D/3D browser environment. This model is used to compare patient-specific treatment plans and device performance via built-in simulation tools. Go-Smart includes evaluation techniques for comparing simulated treatment with real ablation lesions segmented from follow-up scans. The framework is highly extensible, allowing manufacturers and researchers to incorporate new ablation devices, mathematical models and physical parameters.
Briceno, Luis Diego [Colorado State University, Fort Collins; Khemka, Bhavesh [Colorado State University, Fort Collins; Siegel, Howard Jay [Colorado State University, Fort Collins; Maciejewski, Anthony A [ORNL; Groer, Christopher S [ORNL; Koenig, Gregory A [ORNL; Okonski, Gene D [ORNL; Poole, Stephen W [ORNL
2011-01-01
This study considers a heterogeneous computing system and corresponding workload being investigated by the Extreme Scale Systems Center (ESSC) at Oak Ridge National Laboratory (ORNL). The ESSC is part of a collaborative effort between the Department of Energy (DOE) and the Department of Defense (DoD) to deliver research, tools, software, and technologies that can be integrated, deployed, and used in both DOE and DoD environments. The heterogeneous system and workload described here are representative of a prototypical computing environment being studied as part of this collaboration. Each task can exhibit a time-varying importance or utility to the overall enterprise. In this system, an arriving task has an associated priority and precedence. The priority is used to describe the importance of a task, and precedence is used to describe how soon the task must be executed. These two metrics are combined to create a utility function curve that indicates how valuable it is for the system to complete a task at any given moment. This research focuses on using time-utility functions to generate a metric that can be used to compare the performance of different resource schedulers in a heterogeneous computing system. The contributions of this paper are: (a) a mathematical model of a heterogeneous computing system where tasks arrive dynamically and need to be assigned based on their priority, precedence, utility characteristic class, and task execution type, (b) the use of priority and precedence to generate time-utility functions that describe the value a task has at any given time, (c) the derivation of a metric based on the total utility gained from completing tasks to measure the performance of the computing environment, and (d) a comparison of the performance of resource allocation heuristics in this environment.
IEEE TRANSACTIONS ON NANOBIOSCIENCE, VOL. 5, NO. 1, MARCH 2006 41 Computational Modeling of a New Truong, Member, IEEE Abstract--The class of fluorescence resonance energy transfer (FRET) protein
Zhou, Shujia
2009-01-01
Acceleration of Numerical Weather Prediction,” ProceedingsComputer Systems for Climate and Weather Models Shujia Zhouprocesses in climate and weather models demands a continual
Enabling a Highly-Scalable Global Address Space Model for Petascale Computing
Apra, Edoardo; Vetter, Jeffrey S; Yu, Weikuan
2010-01-01
Over the past decade, the trajectory to the petascale has been built on increased complexity and scale of the underlying parallel architectures. Meanwhile, software de- velopers have struggled to provide tools that maintain the productivity of computational science teams using these new systems. In this regard, Global Address Space (GAS) programming models provide a straightforward and easy to use addressing model, which can lead to improved produc- tivity. However, the scalability of GAS depends directly on the design and implementation of the runtime system on the target petascale distributed-memory architecture. In this paper, we describe the design, implementation, and optimization of the Aggregate Remote Memory Copy Interface (ARMCI) runtime library on the Cray XT5 2.3 PetaFLOPs computer at Oak Ridge National Laboratory. We optimized our implementation with the flow intimation technique that we have introduced in this paper. Our optimized ARMCI implementation improves scalability of both the Global Arrays (GA) programming model and a real-world chemistry application NWChem from small jobs up through 180,000 cores.
Computational Model of Population Dynamics Based on the Cell Cycle and Local Interactions
Oprisan, Sorinel Adrian; Oprisan, Ana
2005-03-31
Our study bridges cellular (mesoscopic) level interactions and global population (macroscopic) dynamics of carcinoma. The morphological differences and transitions between well and smooth defined benign tumors and tentacular malignat tumors suggest a theoretical analysis of tumor invasion based on the development of mathematical models exhibiting bifurcations of spatial patterns in the density of tumor cells. Our computational model views the most representative and clinically relevant features of oncogenesis as a fight between two distinct sub-systems: the immune system of the host and the neoplastic system. We implemented the neoplastic sub-system using a three-stage cell cycle: active, dormant, and necrosis. The second considered sub-system consists of cytotoxic active (effector) cells -- EC, with a very broad phenotype ranging from NK cells to CTL cells, macrophages, etc. Based on extensive numerical simulations, we correlated the fractal dimensions for carcinoma, which could be obtained from tumor imaging, with the malignat stage. Our computational model was able to also simulate the effects of surgical, chemotherapeutical, and radiotherapeutical treatments.
Swelling in light water reactor internal components: Insights from computational modeling
Stoller, Roger E.; Barashev, Alexander V.; Golubov, Stanislav I.
2015-08-01
A modern cluster dynamics model has been used to investigate the materials and irradiation parameters that control microstructural evolution under the relatively low-temperature exposure conditions that are representative of the operating environment for in-core light water reactor components. The focus is on components fabricated from austenitic stainless steel. The model accounts for the synergistic interaction between radiation-produced vacancies and the helium that is produced by nuclear transmutation reactions. Cavity nucleation rates are shown to be relatively high in this temperature regime (275 to 325°C), but are sensitive to assumptions about the fine scale microstructure produced under low-temperature irradiation. The cavity nucleation rates observed run counter to the expectation that void swelling would not occur under these conditions. This expectation was based on previous research on void swelling in austenitic steels in fast reactors. This misleading impression arose primarily from an absence of relevant data. The results of the computational modeling are generally consistent with recent data obtained by examining ex-service components. However, it has been shown that the sensitivity of the model s predictions of low-temperature swelling behavior to assumptions about the primary damage source term and specification of the mean-field sink strengths is somewhat greater that that observed at higher temperatures. Further assessment of the mathematical model is underway to meet the long-term objective of this research, which is to provide a predictive model of void swelling at relevant lifetime exposures to support extended reactor operations.
Use of model calibration to achieve high accuracy in analysis of computer networks
Frogner, Bjorn; Guarro, Sergio; Scharf, Guy
2004-05-11
A system and method are provided for creating a network performance prediction model, and calibrating the prediction model, through application of network load statistical analyses. The method includes characterizing the measured load on the network, which may include background load data obtained over time, and may further include directed load data representative of a transaction-level event. Probabilistic representations of load data are derived to characterize the statistical persistence of the network performance variability and to determine delays throughout the network. The probabilistic representations are applied to the network performance prediction model to adapt the model for accurate prediction of network performance. Certain embodiments of the method and system may be used for analysis of the performance of a distributed application characterized as data packet streams.
Mork, B; Nelson, R; Kirkendall, B; Stenvig, N
2009-11-30
Application of BPL technologies to existing overhead high-voltage power lines would benefit greatly from improved simulation tools capable of predicting performance - such as the electromagnetic fields radiated from such lines. Existing EMTP-based frequency-dependent line models are attractive since their parameters are derived from physical design dimensions which are easily obtained. However, to calculate the radiated electromagnetic fields, detailed current distributions need to be determined. This paper presents a method of using EMTP line models to determine the current distribution on the lines, as well as a technique for using these current distributions to determine the radiated electromagnetic fields.
Brown, Matthew Alan
2013-12-31
of the cranked wing approach, most planforms can be modeled including nearly all practical lifting surfaces with some notable exceptions. The methods are extremely accurate for elliptic wings and rectangular wings with some notable difficulty attributed to swept...
and NCAR in the development of a comprehensive, earth systems model. This model incorporates the most-performance climate models. Through the addition of relevant physical processes, we are developing an earth systems modeling capability as well. Our collaborators in climate research include the National Center
Lunds Universitet
Alignments for the IBM-3 Translation Model Thomas Schoenemann Centre for Mathematical Sciences Lund University, Sweden Abstract Prior work on training the IBM-3 transla- tion model is based on suboptimal meth from a statistical viewpoint and introduced five probability models, known as IBM 1-5. Their models
G. R. Odette; G. E. Lucas
2005-11-15
This final report on "In-Service Design & Performance Prediction of Advanced Fusion Material Systems by Computational Modeling and Simulation" (DE-FG03-01ER54632) consists of a series of summaries of work that has been published, or presented at meetings, or both. It briefly describes results on the following topics: 1) A Transport and Fate Model for Helium and Helium Management; 2) Atomistic Studies of Point Defect Energetics, Dynamics and Interactions; 3) Multiscale Modeling of Fracture consisting of: 3a) A Micromechanical Model of the Master Curve (MC) Universal Fracture Toughness-Temperature Curve Relation, KJc(T - To), 3b) An Embrittlement DTo Prediction Model for the Irradiation Hardening Dominated Regime, 3c) Non-hardening Irradiation Assisted Thermal and Helium Embrittlement of 8Cr Tempered Martensitic Steels: Compilation and Analysis of Existing Data, 3d) A Model for the KJc(T) of a High Strength NFA MA957, 3e) Cracked Body Size and Geometry Effects of Measured and Effective Fracture Toughness-Model Based MC and To Evaluations of F82H and Eurofer 97, 3-f) Size and Geometry Effects on the Effective Toughness of Cracked Fusion Structures; 4) Modeling the Multiscale Mechanics of Flow Localization-Ductility Loss in Irradiation Damaged BCC Alloys; and 5) A Universal Relation Between Indentation Hardness and True Stress-Strain Constitutive Behavior. Further details can be found in the cited references or presentations that generally can be accessed on the internet, or provided upon request to the authors. Finally, it is noted that this effort was integrated with our base program in fusion materials, also funded by the DOE OFES.
Validation of the thermospheric vector spherical harmonic (VSH) computer model. Master's thesis
Davis, J.L.
1991-01-01
A semi-empirical computer model of the lower thermosphere has been developed that provides a description of the composition and dynamics of the thermosphere (Killeen et al., 1992). Input variables needed to run the VSH model include time, space and geophysical conditions. One of the output variables the model provides, neutral density, is of particular interest to the U.S. Air Force. Neutral densities vary both as a result of change in solar flux (eg. the solar cycle) and as a result of changes in the magnetosphere (eg. large changes occur in neutral density during geomagnetic storms). Satellites in earth orbit experience aerodynamic drag due to the atmospheric density of the thermosphere. Variability in the neutral density described above affects the drag a satellite experiences and as a result can change the orbital characteristics of the satellite. These changes make it difficult to track the satellite's position. Therefore, it is particularly important to insure that the accuracy of the model's neutral density is optimized for all input parameters. To accomplish this, a validation program was developed to evaluate the strengths and weaknesses of the model's density output by comparing it to SETA-2 (satellite electrostatic accelerometer) total mass density measurements.
Blough, E.; Russell, W.; Leach, J.W.
1990-08-01
Computer models have been developed for evaluating conceptual designs of integrated coal gasification combined cycle power plants. An overall system model was developed for performing thermodynamic cycle analyses, and detailed models were developed for predicting performance characteristics of fixed bed coal gasifiers and hot gas clean up subsystem components. The overall system model performs mass and energy balances and does chemical equilibrium analyses to determine the effects of changes in operating conditions, or to evaluate proposed design changes. An existing plug flow model for fixed bed gasifiers known as the Wen II model was revised and updated. Also, a spread sheet model of zinc ferrite sulfur sorbent regeneration subsystem was developed. Parametric analyses were performed to determine how performance depends on variables in the system design. The work was done to support CRS Sirrine Incorporated in their study of standardized air blown coal gasifier gas turbine concepts.
Shimizu, Y.
2001-01-11
This report provides computational results of an extensive study to examine the following: (1) infinite media neutron-multiplication factors; (2) material bucklings; (3) bounding infinite media critical concentrations; (4) bounding finite critical dimensions of water-reflected and homogeneously water-moderated one-dimensional systems (i.e., spheres, cylinders of infinite length, and slabs that are infinite in two dimensions) that were comprised of various proportions and densities of plutonium oxides and uranium oxides, each having various isotopic compositions; and (5) sensitivity coefficients of delta k-eff with respect to critical geometry delta dimensions were determined for each of the three geometries that were studied. The study was undertaken to support the development of a standard that is sponsored by the International Standards Organization (ISO) under Technical Committee 85, Nuclear Energy (TC 85)--Subcommittee 5, Nuclear Fuel Technology (SC 5)--Working Group 8, Standardization of Calculations, Procedures and Practices Related to Criticality Safety (WG 8). The designation and title of the ISO TC 85/SC 5/WG 8 standard working draft is WD 14941, ''Nuclear energy--Fissile materials--Nuclear criticality control and safety of plutonium-uranium oxide fuel mixtures outside of reactors.'' Various ISO member participants performed similar computational studies using their indigenous computational codes to provide comparative results for analysis in the development of the standard.
Leishear, R.; Poirier, M.; Fowley, M.
2011-05-26
The Salt Disposition Integration (SDI) portfolio of projects provides the infrastructure within existing Liquid Waste facilities to support the startup and long term operation of the Salt Waste Processing Facility (SWPF). Within SDI, the Blend and Feed Project will equip existing waste tanks in the Tank Farms to serve as Blend Tanks where 300,000-800,000 gallons of salt solution will be blended in 1.3 million gallon tanks and qualified for use as feedstock for SWPF. Blending requires the miscible salt solutions from potentially multiple source tanks per batch to be well mixed without disturbing settled sludge solids that may be present in a Blend Tank. Disturbing solids may be problematic both from a feed quality perspective as well as from a process safety perspective where hydrogen release from the sludge is a potential flammability concern. To develop the necessary technical basis for the design and operation of blending equipment, Savannah River National Laboratory (SRNL) completed scaled blending and transfer pump tests and computational fluid dynamics (CFD) modeling. A 94 inch diameter pilot-scale blending tank, including tank internals such as the blending pump, transfer pump, removable cooling coils, and center column, were used in this research. The test tank represents a 1/10.85 scaled version of an 85 foot diameter, Type IIIA, nuclear waste tank that may be typical of Blend Tanks used in SDI. Specifically, Tank 50 was selected as the tank to be modeled per the SRR, Project Engineering Manager. SRNL blending tests investigated various fixed position, non-rotating, dual nozzle pump designs, including a blending pump model provided by the blend pump vendor, Curtiss Wright (CW). Primary research goals were to assess blending times and to evaluate incipient sludge disturbance for waste tanks. Incipient sludge disturbance was defined by SRR and SRNL as minor blending of settled sludge from the tank bottom into suspension due to blending pump operation, where the sludge level was shown to remain constant. To experimentally model the sludge layer, a very thin, pourable, sludge simulant was conservatively used for all testing. To experimentally model the liquid, supernate layer above the sludge in waste tanks, two salt solution simulants were used, which provided a bounding range of supernate properties. One solution was water (H{sub 2}O + NaOH), and the other was an inhibited, more viscous salt solution. The research performed and data obtained significantly advances the understanding of fluid mechanics, mixing theory and CFD modeling for nuclear waste tanks by benchmarking CFD results to actual experimental data. This research significantly bridges the gap between previous CFD models and actual field experiences in real waste tanks. A finding of the 2009, DOE, Slurry Retrieval, Pipeline Transport and Plugging, and Mixing Workshop was that CFD models were inadequate to assess blending processes in nuclear waste tanks. One recommendation from that Workshop was that a validation, or bench marking program be performed for CFD modeling versus experiment. This research provided experimental data to validate and correct CFD models as they apply to mixing and blending in nuclear waste tanks. Extensive SDI research was a significant step toward bench marking and applying CFD modeling. This research showed that CFD models not only agreed with experiment, but demonstrated that the large variance in actual experimental data accounts for misunderstood discrepancies between CFD models and experiments. Having documented this finding, SRNL was able to provide correction factors to be used with CFD models to statistically bound full scale CFD results. Through the use of pilot scale tests performed for both types of pumps and available engineering literature, SRNL demonstrated how to effectively apply CFD results to salt batch mixing in full scale waste tanks. In other words, CFD models were in error prior to development of experimental correction factors determined during this research, which provided a technique to use CFD models fo
Quantum Chaos & Quantum Computers
D. L. Shepelyansky
2000-06-15
The standard generic quantum computer model is studied analytically and numerically and the border for emergence of quantum chaos, induced by imperfections and residual inter-qubit couplings, is determined. This phenomenon appears in an isolated quantum computer without any external decoherence. The onset of quantum chaos leads to quantum computer hardware melting, strong quantum entropy growth and destruction of computer operability. The time scales for development of quantum chaos and ergodicity are determined. In spite the fact that this phenomenon is rather dangerous for quantum computing it is shown that the quantum chaos border for inter-qubit coupling is exponentially larger than the energy level spacing between quantum computer eigenstates and drops only linearly with the number of qubits n. As a result the ideal multi-qubit structure of the computer remains rather robust against imperfections. This opens a broad parameter region for a possible realization of quantum computer. The obtained results are related to the recent studies of quantum chaos in such many-body systems as nuclei, complex atoms and molecules, finite Fermi systems and quantum spin glass shards which are also reviewed in the paper.
Ellis, Matthew
2015-01-01
integration. Computers & Chemical Engineering, 20:315–323,control. Computers & Chemical Engineering, 58:334–343,reactors: a review. Chemical Engineering Communications, 1:
2004-01-01
Mathematics and Computers in Simulation 65 (2004) 557577 Parallel runs of a large air pollution 20 January 2004; accepted 21 January 2004 Abstract Large-scale air pollution models can successfully. The mathematical description of a large-scale air pollution model will be discussed in this paper. The principles
Shepelyansky, Dima
Quantum computing of quantum chaos in the kicked rotator model B. LeÂ´vi, B. Georgeot, and D. L model, a system that displays rich physical properties and enables to study problems of quantum chaos are robust in presence of imperfections. This implies that the algorithm can simulate the dynamics of quantum
Hawick, Ken
0 Computational Science Technical Note CSTN-163 Transients in a Forest-Fire Simulation Model with Varying Combustion Neighbourhoods and Watercourse Firebreaks K. A. Hawick 2012 Forest and bush fires-Schwabl Forest-fire model is investigated with various localised combustion neighbourhoods. The transient
Cerveny, Vlastislav
Ray tracing computations in the smoothed SEG/EAGE Salt Model V. Bucha*, Department of Geophysics in comparison with precise methods, e.g., finite differences or finite elements, are the speed of computation, because ray tracing computations need smooth velocity macro models (Bucha, Bulant & Klimes 2003
Economou, Tassos
understanding how it functions; correspondingly, one of the holy grails of computational struc- tural biology
Michael V. Glazoff; Piyush Sabharwall; Akira Tokuhiro
2014-09-01
An evaluation of thermodynamic aspects of hot corrosion of the superalloys Haynes 242 and HastelloyTM N in the eutectic mixtures of KF and ZrF4 is carried out for development of Advanced High Temperature Reactor (AHTR). This work models the behavior of several superalloys, potential candidates for the AHTR, using computational thermodynamics tool (ThermoCalc), leading to the development of thermodynamic description of the molten salt eutectic mixtures, and on that basis, mechanistic prediction of hot corrosion. The results from these studies indicated that the principal mechanism of hot corrosion was associated with chromium leaching for all of the superalloys described above. However, HastelloyTM N displayed the best hot corrosion performance. This was not surprising given it was developed originally to withstand the harsh conditions of molten salt environment. However, the results obtained in this study provided confidence in the employed methods of computational thermodynamics and could be further used for future alloy design efforts. Finally, several potential solutions to mitigate hot corrosion were proposed for further exploration, including coating development and controlled scaling of intermediate compounds in the KF-ZrF4 system.
Emergency Response Equipment and Related Training: Airborne Radiological Computer System (Model II)
David P. Colton
2007-02-28
The materials included in the Airborne Radiological Computer System, Model-II (ARCS-II) were assembled with several considerations in mind. First, the system was designed to measure and record the airborne gamma radiation levels and the corresponding latitude and longitude coordinates, and to provide a first overview look of the extent and severity of an accident's impact. Second, the portable system had to be light enough and durable enough that it could be mounted in an aircraft, ground vehicle, or watercraft. Third, the system must control the collection and storage of the data, as well as provide a real-time display of the data collection results to the operator. The notebook computer and color graphics printer components of the system would only be used for analyzing and plotting the data. In essence, the provided equipment is composed of an acquisition system and an analysis system. The data can be transferred from the acquisition system to the analysis system at the end of the data collection or at some other agreeable time.
Ghosh, Somnath
. New research initiatives like the Materials Genome Initiative (MGI) and the Integrated Computational Materials Science & Engineering (ICMSE) are creating unprecedented opportunities for unraveling new1 PREFACE The recent times have seen a surge in computational modeling of materials and processes
Lee, Suk-Hyang; Palmer, Susan B.; Turnbull, Ann P.; Wehmeyer, Michael L.
2006-01-01
Curriculum Modifications and Accommodations t s a A Model for Parent-Teacher Collaboration to Promote SeK-Determination in Young Children With Disabilities Suk-Hyang Lee Susan B. Palmer Ann P. TXirnbull Michael L. Wehmeyer Young's mother says she...Curriculum Modifications and Accommodations t s a A Model for Parent-Teacher Collaboration to Promote SeK-Determination in Young Children With Disabilities Suk-Hyang Lee Susan B. Palmer Ann P. TXirnbull Michael L. Wehmeyer Young's mother says she...
Kaper, Tasso J. Kramer, Mark A.; Rotstein, Horacio G.
2013-12-15
Rhythmic neuronal oscillations across a broad range of frequencies, as well as spatiotemporal phenomena, such as waves and bumps, have been observed in various areas of the brain and proposed as critical to brain function. While there is a long and distinguished history of studying rhythms in nerve cells and neuronal networks in healthy organisms, the association and analysis of rhythms to diseases are more recent developments. Indeed, it is now thought that certain aspects of diseases of the nervous system, such as epilepsy, schizophrenia, Parkinson's, and sleep disorders, are associated with transitions or disruptions of neurological rhythms. This focus issue brings together articles presenting modeling, computational, analytical, and experimental perspectives about rhythms and dynamic transitions between them that are associated to various diseases.
Sasaki, Y.; Kitatsuji, Y.; Hirata, M.; Kimura, T.; Yoshizuka, K.
2008-07-01
Multi-dentate diamides have been synthesized and examined for actinide (An) extractions. Bi- and tridentate extractants are the focus in this work. The extraction of actinides was performed from 0.1-6 M HNO{sub 3} to organic solvents. It was obvious that N,N,N',N'-tetra-alkyl-diglycolamide (DGA) derivatives, 2,2'-(methylimino)bis(N,N-dioctyl-acetamide) (MIDOA), and N,N'-dimethyl-N,N'-dioctyl-2-(3-oxa-pentadecane)-malonamide (DMDOOPDMA) have relatively high D values (D(Pu) > 70). The following notable results using DGA extractants were obtained: (1) DGAs with short alkyl chains give higher D values than those with long alkyl chain, (2) DGAs with long alkyl chain have high solubility in n-dodecane. Computational molecular modeling was also used to elucidate the effects of structural and electronic properties of the reagents on their different extractabilities. (authors)
Anderson, C.A.
1981-01-01
Six years ago the Reactor Safety Research Division of the Nuclear Regulatory Commission (NRC) approached the Los Alamos National Laboratory to develop a comprehensive concrete structural analysis code to predict the static and dynamic behavior of Prestressed Concrete Reactor Vessels (PCRVs) that serve as the containment structure of a High-Temperature Gas-Cooled Reactor. The PCRV is a complex concrete structure that must be modeled in three dimensions and posseses other complicating features such as a steel liner for the reactor cavity and woven cables embedded vertically in the PCRV and wound circumferentially on the outside of the PCRV. The cables, or tendons, are used for prestressing the reactor vessel. In addition to developing the computational capability to predict inelastic three dimensional concrete structural behavior, the code response was verified against documented experiments on concrete structural behavior. This code development/verification effort is described.
Pierre-Luc Dallaire-Demers; Frank K. Wilhelm
2015-08-18
Many phenomena of strongly correlated materials are encapsulated in the Fermi-Hubbard model whose thermodynamical properties can be computed from its grand canonical potential according to standard procedures. In general, there is no closed form solution for lattices of more than one spatial dimension, but solutions can be approximated with cluster perturbation theory. To model long-range effects such as order parameters, a powerful method to compute the cluster's Green's function consists in finding its self-energy through a variational principle of the grand canonical potential. This opens the possibility of studying various phase transitions at finite temperature in the Fermi-Hubbard model. However, a classical cluster solver quickly hits an exponential wall in the memory (or computation time) required to store the computation variables. Here it is shown theoretically that that the cluster solver can be mapped to a subroutine on a quantum computer whose quantum memory scales as the number of orbitals in the simulated cluster. A quantum computer with a few tens of qubits could therefore simulate the thermodynamical properties of complex fermionic lattices inaccessible to classical supercomputers.
Pierre-Luc Dallaire-Demers; Frank K. Wilhelm
2015-11-27
Many phenomena of strongly correlated materials are encapsulated in the Fermi-Hubbard model whose thermodynamical properties can be computed from its grand canonical potential according to standard procedures. In general, there is no closed form solution for lattices of more than one spatial dimension, but solutions can be approximated with cluster perturbation theory. To model long-range effects such as order parameters, a powerful method to compute the cluster's Green's function consists in finding its self-energy through a variational principle of the grand canonical potential. This opens the possibility of studying various phase transitions at finite temperature in the Fermi-Hubbard model. However, a classical cluster solver quickly hits an exponential wall in the memory (or computation time) required to store the computation variables. Here it is shown theoretically that that the cluster solver can be mapped to a subroutine on a quantum computer whose quantum memory scales as the number of orbitals in the simulated cluster. A quantum computer with a few tens of qubits could therefore simulate the thermodynamical properties of complex fermionic lattices inaccessible to classical supercomputers.
Reznik, Alexander N.
2014-08-25
An electrodynamic model is proposed for the tunneling microwave microscope with subnanometer space resolution as developed by Lee et al. [Appl. Phys. Lett. 97, 183111 (2010)]. Tip-sample impedance Z{sub a} was introduced and studied in the tunneling and non-tunneling regimes. At tunneling breakdown, the microwave current between probe and sample flows along two parallel channels characterized by impedances Z{sub p} and Z{sub t} that add up to form overall impedance Z{sub a}. Quantity Z{sub p} is the capacitive impedance determined by the near field of the probe and Z{sub t} is the impedance of the tunnel junction. By taking into account the distance dependences of effective tip radius r{sub 0}(z) and tunnel resistance R{sub t}(z)?=?Re[Z{sub t}(z)], we were able to explain the experimentally observed dependences of resonance frequency f{sub r}(z) and quality factor Q{sub L}(z) of the microscope. The obtained microwave resistance R{sub t}(z) and direct current tunnel resistance R{sub t}{sup dc}(z) exhibit qualitatively similar behavior, although being largely different in both magnitude and the characteristic scale of height dependence. Interpretation of the microwave images of the atomic structure of test samples proved possible by taking into account the inductive component of tunnel impedance ImZ{sub t}?=??L{sub t}. Relation ?L{sub t}/R{sub t}???0.235 was obtained.
The Use Of Computational Human Performance Modeling As Task Analysis Tool
Jacuqes Hugo; David Gertman
2012-07-01
During a review of the Advanced Test Reactor safety basis at the Idaho National Laboratory, human factors engineers identified ergonomic and human reliability risks involving the inadvertent exposure of a fuel element to the air during manual fuel movement and inspection in the canal. There were clear indications that these risks increased the probability of human error and possible severe physical outcomes to the operator. In response to this concern, a detailed study was conducted to determine the probability of the inadvertent exposure of a fuel element. Due to practical and safety constraints, the task network analysis technique was employed to study the work procedures at the canal. Discrete-event simulation software was used to model the entire procedure as well as the salient physical attributes of the task environment, such as distances walked, the effect of dropped tools, the effect of hazardous body postures, and physical exertion due to strenuous tool handling. The model also allowed analysis of the effect of cognitive processes such as visual perception demands, auditory information and verbal communication. The model made it possible to obtain reliable predictions of operator performance and workload estimates. It was also found that operator workload as well as the probability of human error in the fuel inspection and transfer task were influenced by the concurrent nature of certain phases of the task and the associated demand on cognitive and physical resources. More importantly, it was possible to determine with reasonable accuracy the stages as well as physical locations in the fuel handling task where operators would be most at risk of losing their balance and falling into the canal. The model also provided sufficient information for a human reliability analysis that indicated that the postulated fuel exposure accident was less than credible.
Computer Modeling VRF Heat Pumps in Commercial Buildings using EnergyPlus
Raustad, Richard
2013-06-01
Variable Refrigerant Flow (VRF) heat pumps are increasingly used in commercial buildings in the United States. Monitored energy use of field installations have shown, in some cases, savings exceeding 30% compared to conventional heating, ventilating, and air-conditioning (HVAC) systems. A simulation study was conducted to identify the installation or operational characteristics that lead to energy savings for VRF systems. The study used the Department of Energy EnergyPlus? building simulation software and four reference building models. Computer simulations were performed in eight U.S. climate zones. The baseline reference HVAC system incorporated packaged single-zone direct-expansion cooling with gas heating (PSZ-AC) or variable-air-volume systems (VAV with reheat). An alternate baseline HVAC system using a heat pump (PSZ-HP) was included for some buildings to directly compare gas and electric heating results. These baseline systems were compared to a VRF heat pump model to identify differences in energy use. VRF systems combine multiple indoor units with one or more outdoor unit(s). These systems move refrigerant between the outdoor and indoor units which eliminates the need for duct work in most cases. Since many applications install duct work in unconditioned spaces, this leads to installation differences between VRF systems and conventional HVAC systems. To characterize installation differences, a duct heat gain model was included to identify the energy impacts of installing ducts in unconditioned spaces. The configuration of variable refrigerant flow heat pumps will ultimately eliminate or significantly reduce energy use due to duct heat transfer. Fan energy is also studied to identify savings associated with non-ducted VRF terminal units. VRF systems incorporate a variable-speed compressor which may lead to operational differences compared to single-speed compression systems. To characterize operational differences, the computer model performance curves used to simulate cooling operation are also evaluated. The information in this paper is intended to provide a relative difference in system energy use and compare various installation practices that can impact performance. Comparative results of VRF versus conventional HVAC systems include energy use differences due to duct location, differences in fan energy when ducts are eliminated, and differences associated with electric versus fossil fuel type heating systems.
Donna Post Guillen; Tami Grimmett; Anastasia M. Gribik; Steven P. Antal
2010-09-01
The Hybrid Energy Systems Testing (HYTEST) Laboratory is being established at the Idaho National Laboratory to develop and test hybrid energy systems with the principal objective to safeguard U.S. Energy Security by reducing dependence on foreign petroleum. A central component of the HYTEST is the slurry bubble column reactor (SBCR) in which the gas-to-liquid reactions will be performed to synthesize transportation fuels using the Fischer Tropsch (FT) process. SBCRs are cylindrical vessels in which gaseous reactants (for example, synthesis gas or syngas) is sparged into a slurry of liquid reaction products and finely dispersed catalyst particles. The catalyst particles are suspended in the slurry by the rising gas bubbles and serve to promote the chemical reaction that converts syngas to a spectrum of longer chain hydrocarbon products, which can be upgraded to gasoline, diesel or jet fuel. These SBCRs operate in the churn-turbulent flow regime which is characterized by complex hydrodynamics, coupled with reacting flow chemistry and heat transfer, that effect reactor performance. The purpose of this work is to develop a computational multiphase fluid dynamic (CMFD) model to aid in understanding the physico-chemical processes occurring in the SBCR. Our team is developing a robust methodology to couple reaction kinetics and mass transfer into a four-field model (consisting of the bulk liquid, small bubbles, large bubbles and solid catalyst particles) that includes twelve species: (1) CO reactant, (2) H2 reactant, (3) hydrocarbon product, and (4) H2O product in small bubbles, large bubbles, and the bulk fluid. Properties of the hydrocarbon product were specified by vapor liquid equilibrium calculations. The absorption and kinetic models, specifically changes in species concentrations, have been incorporated into the mass continuity equation. The reaction rate is determined based on the macrokinetic model for a cobalt catalyst developed by Yates and Satterfield [1]. The model includes heat generation due to the exothermic chemical reaction, as well as heat removal from a constant temperature heat exchanger. Results of the CMFD simulations (similar to those shown in Figure 1) will be presented.
Johnson, J. D.; Oberkampf, William Louis; Helton, Jon Craig (Arizona State University, Tempe, AZ); Storlie, Curtis B. (North Carolina State University, Raleigh, NC)
2006-10-01
Evidence theory provides an alternative to probability theory for the representation of epistemic uncertainty in model predictions that derives from epistemic uncertainty in model inputs, where the descriptor epistemic is used to indicate uncertainty that derives from a lack of knowledge with respect to the appropriate values to use for various inputs to the model. The potential benefit, and hence appeal, of evidence theory is that it allows a less restrictive specification of uncertainty than is possible within the axiomatic structure on which probability theory is based. Unfortunately, the propagation of an evidence theory representation for uncertainty through a model is more computationally demanding than the propagation of a probabilistic representation for uncertainty, with this difficulty constituting a serious obstacle to the use of evidence theory in the representation of uncertainty in predictions obtained from computationally intensive models. This presentation describes and illustrates a sampling-based computational strategy for the representation of epistemic uncertainty in model predictions with evidence theory. Preliminary trials indicate that the presented strategy can be used to propagate uncertainty representations based on evidence theory in analysis situations where naive sampling-based (i.e., unsophisticated Monte Carlo) procedures are impracticable due to computational cost.
Nair, S.K.; Chambers, D.B.; Park, S.H.; Hoffman, F.O. [Senes Oak Ridge, Inc., TN (United States). Center for Risk Analysis
1997-11-01
The objective of this study is to examine the usefulness and effectiveness of currently existing models that simulate the release of uranium hexafluoride from UF{sub 6}-handling facilities, subsequent reactions of UF{sub 6} with atmospheric moisture, and the dispersion of UF{sub 6} and reaction products in the atmosphere. The study evaluates screening-level and detailed public-domain models that were specifically developed for UF{sub 6} and models that were originally developed for the treatment of dense gases but are applicable to UF{sub 6} release, reaction, and dispersion. The model evaluation process is divided into three specific tasks: model-component evaluation; applicability evaluation; and user interface and quality assurance and quality control (QA/QC) evaluation. Within the model-component evaluation process, a model`s treatment of source term, thermodynamics, and atmospheric dispersion are considered and model predictions are compared with actual observations. Within the applicability evaluation process, a model`s applicability to Integrated Safety Analysis, Emergency Response Planning, and Post-Accident Analysis, and to site-specific considerations are assessed. Finally, within the user interface and QA/QC evaluation process, a model`s user-friendliness, presence and clarity of documentation, ease of use, etc. are assessed, along with its handling of QA/QC. This document presents the complete methodology used in the evaluation process.
Harting, Jens
Steering in computational science: mesoscale modelling and simulation J. CHIN{, J. HARTING{, S. JHA steering for high performance computing applications. Lattice-Boltzmann mesoscale fluid simulations, there is currently considerable interest in mesoscale models. These models coarse grain most of the atomic
Wind Turbine Modeling for Computational Fluid Dynamics: December 2010 - December 2012
Tossas, L. A. M.; Leonardi, S.
2013-07-01
With the shortage of fossil fuel and the increasing environmental awareness, wind energy is becoming more and more important. As the market for wind energy grows, wind turbines and wind farms are becoming larger. Current utility-scale turbines extend a significant distance into the atmospheric boundary layer. Therefore, the interaction between the atmospheric boundary layer and the turbines and their wakes needs to be better understood. The turbulent wakes of upstream turbines affect the flow field of the turbines behind them, decreasing power production and increasing mechanical loading. With a better understanding of this type of flow, wind farm developers could plan better-performing, less maintenance-intensive wind farms. Simulating this flow using computational fluid dynamics is one important way to gain a better understanding of wind farm flows. In this study, we compare the performance of actuator disc and actuator line models in producing wind turbine wakes and the wake-turbine interaction between multiple turbines. We also examine parameters that affect the performance of these models, such as grid resolution, the use of a tip-loss correction, and the way in which the turbine force is projected onto the flow field.
Computer modelling of the reduction of rare earth dopants in barium aluminate
Rezende, Marcos V. dos S; Valerio, Mario E.G. [Department of Physics, Federal University of Sergipe, 49100-000 Sao Cristovao, SE (Brazil); Jackson, Robert A., E-mail: r.a.jackson@chem.keele.ac.uk [School of Physical and Geographical Sciences, Keele University, Keele, Staffordshire ST5 5BG (United Kingdom)
2011-08-15
Long lasting phosphorescence in barium aluminates can be achieved by doping with rare earth ions in divalent charge states. The rare earth ions are initially in a trivalent charge state, but are reduced to a divalent charge state before being doped into the material. In this paper, the reduction of trivalent rare earth ions in the BaAl{sub 2}O{sub 4} lattice is studied by computer simulation, with the energetics of the whole reduction and doping process being modelled by two methods, one based on single ion doping and one which allows dopant concentrations to be taken into account. A range of different reduction schemes are considered and the most energetically favourable schemes identified. - Graphical abstract: The doping and subsequent reduction of a rare earth ion into the barium aluminate lattice. Highlights: > The doping of barium aluminate with rare earth ions reduced in a range of atmospheres has been modelled. > The overall solution energy for the doping process for each ion in each reducing atmosphere is calculated using two methods. > The lowest energy reduction process is predicted and compared with experimental results.
Development of Computational Tools for Metabolic Model Curation, Flux Elucidation and Strain Design
Maranas, Costas D
2012-05-21
An overarching goal of the Department of EnergyÂ? mission is the efficient deployment and engineering of microbial and plant systems to enable biomass conversion in pursuit of high energy density liquid biofuels. This has spurred the pace at which new organisms are sequenced and annotated. This torrent of genomic information has opened the door to understanding metabolism in not just skeletal pathways and a handful of microorganisms but for truly genome-scale reconstructions derived for hundreds of microbes and plants. Understanding and redirecting metabolism is crucial because metabolic fluxes are unique descriptors of cellular physiology that directly assess the current cellular state and quantify the effect of genetic engineering interventions. At the same time, however, trying to keep pace with the rate of genomic data generation has ushered in a number of modeling and computational challenges related to (i) the automated assembly, testing and correction of genome-scale metabolic models, (ii) metabolic flux elucidation using labeled isotopes, and (iii) comprehensive identification of engineering interventions leading to the desired metabolism redirection.
Modeling a Reconfigurable System for Computing the FFT in Place via RewritingLogic #
Ayala-Rincón, Mauricio
hardware implementation of the Fast Fourier Transform -- FFT using rewritinglogic. It is shown on general purpose processors, reconfigurable computing delivers more processing power due electronic market. There are several taxonomies applied to reconfigurable computing. Concerning the specific
Huang, Yongxin
2010-01-16
ferromagnets, there exist two main challenges: the complicated microelasticity due to the magnetostrictive strain, and very expensive computation mainly caused by the calculation of long-range magnetostatic and elastic interactions. A parallel computing...
Study of behavior and determination of customer lifetime value(CLV) using Markov chain model
Permana, Dony, E-mail: donypermana@students.itb.ac.id [Statistics Research Division, Faculty of Mathematics and Natural Science, Bandung Institute of Technology, Indonesia and Statistics Study Program, Faculty of Mathematics and Natural Sciences, Padang State University (Indonesia); Indratno, Sapto Wahyu; Pasaribu, Udjianna S. [Statistics Research Division, Faculty of Mathematics and Natural Science, Bandung Institute of Technology (Indonesia)
2014-03-24
Customer Lifetime Value or CLV is a restriction on interactive marketing to help a company in arranging financial for the marketing of new customer acquisition and customer retention. Additionally CLV can be able to segment customers for financial arrangements. Stochastic models for the fairly new CLV used a Markov chain. In this model customer retention probability and new customer acquisition probability play an important role. This model is originally introduced by Pfeifer and Carraway in 2000 [1]. They introduced several CLV models, one of them only involves customer and former customer. In this paper we expand the model by adding the assumption of the transition from former customer to customer. In the proposed model, the CLV value is higher than the CLV value obtained by Pfeifer and Caraway model. But our model still requires a longer convergence time.
Determination of Retrofit Savings Using a Calibrated Building Energy Simulation Model
Reddy, S. N.; Hunn, B. D.; Hood, D. B.
1994-01-01
for whole-building electric, cooling, and heating energy use, and were compared with savings calculated using a regression model developed under the LoanSTAR program. Finally, to validate the model, postretrofit DOE-2 results were compared with measured...
Lyon, Richard Harry, 1981-
2004-01-01
Correct modeling of the space environment, including radiative forces, is an important aspect of space situational awareness for geostationary (GEO) spacecraft. Solar radiation pressure has traditionally been modeled using ...
A Computational Model which Learns to Selectively Attend in Category Learning
Cottrell, Garrison W.
Garrison W. Cottrell lingyun,gary@cs.ucsd.edu UCSD Computer Science and Engineering 9500 Gilman Dr., La
DOE Public Access Gateway for Energy & Science Beta (PAGES Beta)
Kilina, Svetlana; Yarotski, Dzmitry A.; Talin, A. Alec; Tretiak, Sergei; Taylor, Antoinette J.; Balatsky, Alexander V.
2011-01-01
We present a combined approach that relies on computational simulations and scanning tunneling microscopy (STM) measurements to reveal morphological properties and stability criteria of carbon nanotube-DNA (CNT-DNA) constructs. Application of STM allows direct observation of very stable CNT-DNA hybrid structures with the well-defined DNA wrapping angle of 63.4 ° and a coiling period of 3.3?nm. Using force field simulations, we determine how the DNA-CNT binding energy depends on the sequence and binding geometry of a single strand DNA. This dependence allows us to quantitatively characterize the stability of a hybrid structure with an optimal ?-stacking between DNA nucleotides andmore »the tube surface and better interpret STM data. Our simulations clearly demonstrate the existence of a very stable DNA binding geometry for (6,5) CNT as evidenced by the presence of a well-defined minimum in the binding energy as a function of an angle between DNA strand and the nanotube chiral vector. This novel approach demonstrates the feasibility of CNT-DNA geometry studies with subnanometer resolution and paves the way towards complete characterization of the structural and electronic properties of drug-delivering systems based on DNA-CNT hybrids as a function of DNA sequence and a nanotube chirality. « less
Stockman, Mark [Georgia State University Research Foundation] [Georgia State University Research Foundation; Gray, Steven [Argon National Laboratory] [Argon National Laboratory
2014-02-21
The program is directed toward development of new computational approaches to photoprocesses in nanostructures whose geometry and composition are tailored to obtain desirable optical responses. The emphasis of this specific program is on the development of computational methods and prediction and computational theory of new phenomena of optical energy transfer and transformation on the extreme nanoscale (down to a few nanometers).
Bergman, Keren
at the scale of high performance computer clusters and warehouse scale data centers, system level simulations and results for rack scale photonic interconnection networks for high performance computing. Keywords: optical to the newsworthy power consumption [3], latency [4] and bandwidth challenges [5] of high performance computing (HPC
Robinson, Mark R. (Albuquerque, NM); Ward, Kenneth J. (Albuquerque, NM); Eaton, Robert P. (Albuquerque, NM); Haaland, David M. (Albuquerque, NM)
1990-01-01
The characteristics of a biological fluid sample having an analyte are determined from a model constructed from plural known biological fluid samples. The model is a function of the concentration of materials in the known fluid samples as a function of absorption of wideband infrared energy. The wideband infrared energy is coupled to the analyte containing sample so there is differential absorption of the infrared energy as a function of the wavelength of the wideband infrared energy incident on the analyte containing sample. The differential absorption causes intensity variations of the infrared energy incident on the analyte containing sample as a function of sample wavelength of the energy, and concentration of the unknown analyte is determined from the thus-derived intensity variations of the infrared energy as a function of wavelength from the model absorption versus wavelength function.
Nair, S.K.; Chambers, D.B.; Park, S.H.; Radonjic, Z.R.; Coutts, P.T.; Lewis, C.J.; Hammonds, J.S.; Hoffman, F.O. [Senes Oak Ridge, Inc., TN (United States). Center for Risk Analysis
1997-11-01
Three uranium hexafluoride-(UF{sub 6}-) specific models--HGSYSTEM/UF{sub 6}, Science Application International Corporation, and RTM-96; three dense-gas models--DEGADIS, SLAB, and the Chlorine Institute methodology; and one toxic chemical model--AFTOX--are evaluated on their capabilities to simulate the chemical reactions, thermodynamics, and atmospheric dispersion of UF{sub 6} released from accidents at nuclear fuel-cycle facilities, to support Integrated Safety Analysis, Emergency Response Planning, and Post-Accident Analysis. These models are also evaluated for user-friendliness and for quality assurance and quality control features, to ensure the validity and credibility of the results. Model performance evaluations are conducted for the three UF{sub 6}-specific models, using field data on releases of UF{sub 6} and other heavy gases. Predictions from the HGSYSTEM/UF{sub 6} and SAIC models are within an order of magnitude of the field data, but the SAIC model overpredicts beyond an order of magnitude for a few UF{sub 6}-specific data points. The RTM-96 model provides overpredictions within a factor of 3 for all data points beyond 400 m from the source. For one data set, however, the RTM-96 model severely underpredicts the observations within 200 m of the source. Outputs of the models are most sensitive to the meteorological parameters at large distances from the source and to certain source-specific and meteorological parameters at distances close to the source. Specific recommendations are being made to improve the applicability and usefulness of the three models and to choose a specific model to support the intended analyses. Guidance is also provided on the choice of input parameters for initial dilution, building wake effects, and distance to completion of UF{sub 6} reaction with water.
Okuma, Tomohisa, E-mail: o-kuma@msic.med.osaka-cu.ac.jp; Matsuoka, Toshiyuki; Yamamoto, Akira; Oyama, Yoshimasa; Hamamoto, Shinichi [Osaka City University Graduate School of Medicine, Department of Radiology (Japan); Toyoshima, Masami [Kobe City Medical Center West Hospital, Department of Radiology (Japan); Nakamura, Kenji; Miki, Yukio [Osaka City University Graduate School of Medicine, Department of Radiology (Japan)
2010-08-15
The purpose of this study was to retrospectively determine the local control rate and contributing factors to local progression after computed tomography (CT)-guided radiofrequency ablation (RFA) for unresectable lung tumor. This study included 138 lung tumors in 72 patients (56 men and 16 women; age 70.0 {+-} 11.6 years (range 31-94); mean tumor size 2.1 {+-} 1.2 cm [range 0.2-9]) who underwent lung RFA between June 2000 and May 2009. Mean follow-up periods for patients and tumors were 14 and 12 months, respectively. The local progression-free rate and survival rate were calculated to determine the contributing factors to local progression. During follow-up, 44 of 138 (32%) lung tumors showed local progression. The 1-, 2-, 3-, and 5-year overall local control rates were 61, 57, 57, and 38%, respectively. The risk factors for local progression were age ({>=}70 years), tumor size ({>=}2 cm), sex (male), and no achievement of roll-off during RFA (P < 0.05). Multivariate analysis identified tumor size {>=}2 cm as the only independent factor for local progression (P = 0.003). For tumors <2 cm, 17 of 68 (25%) showed local progression, and the 1-, 2-, and 3-year overall local control rates were 77, 73, and 73%, respectively. Multivariate analysis identified that age {>=}70 years was an independent determinant of local progression for tumors <2 cm in diameter (P = 0.011). The present study showed that 32% of lung tumors developed local progression after CT-guided RFA. The significant risk factor for local progression after RFA for lung tumors was tumor size {>=}2 cm.
Rondon, Nolys Javier
2009-05-15
Knowledge of viscosity of flow streams is essential for the design and operation of production facilities, drilling operations and reservoir engineering calculations. The determination of the viscosity of a reservoir fluid ...
Kramer, Sharlotte Lorraine Bolyard; Scherzinger, William M.
2014-09-01
The Virtual Fields Method (VFM) is an inverse method for constitutive model parameter identication that relies on full-eld experimental measurements of displacements. VFM is an alternative to standard approaches that require several experiments of simple geometries to calibrate a constitutive model. VFM is one of several techniques that use full-eld exper- imental data, including Finite Element Method Updating (FEMU) techniques, but VFM is computationally fast, not requiring iterative FEM analyses. This report describes the im- plementation and evaluation of VFM primarily for nite-deformation plasticity constitutive models. VFM was successfully implemented in MATLAB and evaluated using simulated FEM data that included representative experimental noise found in the Digital Image Cor- relation (DIC) optical technique that provides full-eld displacement measurements. VFM was able to identify constitutive model parameters for the BCJ plasticity model even in the presence of simulated DIC noise, demonstrating VFM as a viable alternative inverse method. Further research is required before VFM can be adopted as a standard method for constitu- tive model parameter identication, but this study is a foundation for ongoing research at Sandia for improving constitutive model calibration.
Gas-Phase Lubrication of ta-C by Glycerol and Hydrogen Peroxide. Experimental and Computer Modeling
Goddard III, William A.
Gas-Phase Lubrication of ta-C by Glycerol and Hydrogen Peroxide. Experimental and Computer Modeling lubrication conditions at 80 Â°C in presence of OH-containing molecules. To understand the mechanism of ultralow friction, we performed gas-phase lubrication experiments followed by time-of-flight secondary ion
Geller, Michael R.
2013-01-01
code. Section III maps the calculation of the ground-state energy for the TIM onto a quantum phase in the transverse Ising model (TIM) [12], there is a large number of physical qubits and lengthy computational time]. Here we investigate the quantum simulation of the TIM ground-state energy on a surface code quantum
Computer determination of bacterial volume
Griffis, David William
1978-01-01
51-I!. 55 (1'j ~3) 17. M. J. Tnornton, The Applicaticn of the Coul Counter to Problems in the Siz "-. nalys's cf Insoluble D u"s. Jour. Pharm. =nd Phar. . ecol. (11)~ 742"752 i" 963~ 18 . G. F. Bahr, G . R. Rer'caner, and 8 . , G' as, A...
Nishino, Takafumi
2012-01-01
Modelling of turbine blade-induced turbulence (BIT) is discussed within the framework of three-dimensional Reynolds-averaged Navier-Stokes (RANS) actuator disk computations. We first propose a generic (baseline) BIT model, which is applied only to the actuator disk surface, does not include any model coefficients (other than those used in the original RANS turbulence model) and is expected to be valid in the limiting case where BIT is fully isotropic and in energy equilibrium. The baseline model is then combined with correction functions applied to the region behind the disk to account for the effect of rotor tip vortices causing a mismatch of Reynolds shear stress between short- and long-time averaged flow fields. Results are compared with wake measurements of a two-bladed wind turbine model of Medici and Alfredsson [Wind Energy, Vol. 9, 2006, pp. 219-236] to demonstrate the capability of the new model.
Aalborg Universitet Determining HTPEM electrode parameters using a mechanistic impedance model
Berning, Torsten
International Conference on Medium and High Temperature Proton Exchange Membrane Fuel Cells, Copenhagen, Denmark-realistic parameters. Model The fuel cell model is implemented using a 2D finite volume approach, taking into account at the School of Chemical Engineering and Advanced Materials, Newcastle University. The membrane was PBI doped
Musial, W.; Lawson, M.; Rooney, S.
2013-02-01
The Marine and Hydrokinetic Technology (MHK) Instrumentation, Measurement, and Computer Modeling Workshop was hosted by the National Renewable Energy Laboratory (NREL) in Broomfield, Colorado, July 9-10, 2012. The workshop brought together over 60 experts in marine energy technologies to disseminate technical information to the marine energy community and collect information to help identify ways in which the development of a commercially viable marine energy industry can be accelerated. The workshop was comprised of plenary sessions that reviewed the state of the marine energy industry and technical sessions that covered specific topics of relevance. Each session consisted of presentations, followed by facilitated discussions. During the facilitated discussions, the session chairs posed several prepared questions to the presenters and audience to encourage communication and the exchange of ideas between technical experts. Following the workshop, attendees were asked to provide written feedback on their takeaways and their best ideas on how to accelerate the pace of marine energy technology development. The first four sections of this document give a general overview of the workshop format, provide presentation abstracts and discussion session notes, and list responses to the post-workshop questions. The final section presents key findings and conclusions from the workshop that suggest how the U.S. Department of Energy and national laboratory resources can be utilized to most effectively assist the marine energy industry.
Musial, W.; Lawson, M.; Rooney, S.
2013-02-01
The Marine and Hydrokinetic Technology (MHK) Instrumentation, Measurement, and Computer Modeling Workshop was hosted by the National Renewable Energy Laboratory (NREL) in Broomfield, Colorado, July 9–10, 2012. The workshop brought together over 60 experts in marine energy technologies to disseminate technical information to the marine energy community, and to collect information to help identify ways in which the development of a commercially viable marine energy industry can be accelerated. The workshop was comprised of plenary sessions that reviewed the state of the marine energy industry and technical sessions that covered specific topics of relevance. Each session consisted of presentations, followed by facilitated discussions. During the facilitated discussions, the session chairs posed several prepared questions to the presenters and audience to encourage communication and the exchange of ideas between technical experts. Following the workshop, attendees were asked to provide written feedback on their takeaways from the workshop and their best ideas on how to accelerate the pace of marine energy technology development. The first four sections of this document give a general overview of the workshop format, provide presentation abstracts, supply discussion session notes, and list responses to the post-workshop questions. The final section presents key findings and conclusions from the workshop that suggest what the most pressing MHK technology needs are and how the U.S. Department of Energy (DOE) and national laboratory resources can be utilized to assist the marine energy industry in the most effective manner.
Shadmehr, Reza
of Our Arm? JoAnn Kluzik,1,3,4 Jo¨rn Diedrichsen,5 Reza Shadmehr,2 and Amy J. Bastian1,3 1 Department learn an internal model of the tool or adapt the model of our arm? J Neurophysiol 100: 14551464, 2008 or of our own arm? We considered a well-studied adaptation task in which people made goal-directed reaching
Pesaran, A.; Kim, G. H.; Smith, K.; Santhanagopalan, S.; Lee, K. J.
2012-05-01
This 2012 Annual Merit Review presentation gives an overview of the Computer-Aided Engineering of Batteries (CAEBAT) project and introduces the Multi-Scale, Multi-Dimensional model for modeling lithium-ion batteries for electric vehicles.
Data-Driven Optimization for Modeling in Computer Graphics and Vision
Yu, Lap Fai
2013-01-01
the recent trend of 3D printing in computer-aided design.3D display, and 3D printing have made these technologies
Data-Driven Optimization for Modeling in Computer Graphics and Vision
Yu, Lap Fai
2013-01-01
in 3D scanning, 3D display, and 3D printing have made thesethe recent trend of 3D printing in computer-aided design.
M.F. Simpson; K.-R. Kim
2010-12-01
In support of closing the nuclear fuel cycle using non-aqueous separations technology, this project aims to develop computational models of electrorefiners based on fundamental chemical and physical processes. Spent driver fuel from Experimental Breeder Reactor-II (EBR-II) is currently being electrorefined in the Fuel Conditioning Facility (FCF) at Idaho National Laboratory (INL). And Korea Atomic Energy Research Institute (KAERI) is developing electrorefining technology for future application to spent fuel treatment and management in the Republic of Korea (ROK). Electrorefining is a critical component of pyroprocessing, a non-aqueous chemical process which separates spent fuel into four streams: (1) uranium metal, (2) U/TRU metal, (3) metallic high-level waste containing cladding hulls and noble metal fission products, and (4) ceramic high-level waste containing sodium and active metal fission products. Having rigorous yet flexible electrorefiner models will facilitate process optimization and assist in trouble-shooting as necessary. To attain such models, INL/UI has focused on approaches to develop a computationally-light and portable two-dimensional (2D) model, while KAERI/SNU has investigated approaches to develop a computationally intensive three-dimensional (3D) model for detailed and fine-tuned simulation.
Determination of the proper operating range for the CAFCA IIB fuel cycle model
Warburton, Jamie (Jamie L.)
2007-01-01
The fuel cycle simulation tool, CAFCA II was previously modified to produce the most recent version, CAFCA IIB. The code tracks the mass distribution of transuranics in the fuel cycle in one model and also projects costs ...
Technosocial Modeling for Determining the Status and Nature of a State’s Nuclear Activities
Gastelum, Zoe N.; Harvey, Julia B.
2009-09-25
The International Atomic Energy Agency State Evaluation Process: The Role of Information Analysis in Reaching Safeguards Conclusions (Mathews et al. 2008), several examples of nonproliferation models using analytical software were developed that may assist the IAEA with collecting, visualizing, analyzing, and reporting information in support of the State Evaluation Process. This paper focuses on one of the examples a set of models developed in the Proactive Scenario Production, Evidence Collection, and Testing (ProSPECT) software that evaluates the status and nature of a state’s nuclear activities. The models use three distinct subject areas to perform this assessment: the presence of nuclear activities, the consistency of those nuclear activities with national nuclear energy goals, and the geopolitical context in which those nuclear activities are taking place. As a proof-of-concept for the models, a crude case study was performed. The study, which attempted to evaluate the nuclear activities taking place in Syria prior to September 2007, yielded illustrative, yet inconclusive, results. Due to the inconclusive nature of the case study results, changes that may improve the model’s efficiency and accuracy are proposed.
Illinois at Urbana-Champaign, University of
85721 Center for Reliable and High-Performance Computing Coordinated Science Laboratory UniversityFrom International Computer Performance and Dependability Symposium, Erlangen, Germany, April 1995, pp.285 294 MODELING RECYCLE: A CASE STUDY IN THE INDUSTRIAL USE OF MEASUREMENT AND MODELING Luai M
Eindhoven, Technische Universiteit
1 A COMPUTER SIMULATION MODEL FOR THE BLOW-BLOW FORMING PROCESS OF GLASS CONTAINERS C. G for industrial purposes. To achieve this both steps of the forming of glass containers, namely blow- blow needs conditions. In [1] the development of a computer model to be used for glass blowing was described
Coupling Multi-Component Models with MPH on Distributed Memory Computer Architectures
He, Yun; Ding, Chris
2005-01-01
Among these, NASA’s Earth System Models Framework (ESMF) [to facilitate coupling earth system model components and to
MODELING STRATEGIES TO COMPUTE NATURAL CIRCULATION USING CFD IN A VHTR AFTER A LOFA
Yu-Hsin Tung; Richard W. Johnson; Ching-Chang Chieng; Yuh-Ming Ferng
2012-11-01
A prismatic gas-cooled very high temperature reactor (VHTR) is being developed under the next generation nuclear plant program (NGNP) of the U.S. Department of Energy, Office of Nuclear Energy. In the design of the prismatic VHTR, hexagonal shaped graphite blocks are drilled to allow insertion of fuel pins, made of compacted TRISO fuel particles, and coolant channels for the helium coolant. One of the concerns for the reactor design is the effects of a loss of flow accident (LOFA) where the coolant circulators are lost for some reason, causing a loss of forced coolant flow through the core. In such an event, it is desired to know what happens to the (reduced) heat still being generated in the core and if it represents a problem for the fuel compacts, the graphite core or the reactor vessel (RV) walls. One of the mechanisms for the transport of heat out of the core is by the natural circulation of the coolant, which is still present. That is, how much heat may be transported by natural circulation through the core and upwards to the top of the upper plenum? It is beyond current capability for a computational fluid dynamic (CFD) analysis to perform a calculation on the whole RV with a sufficiently refined mesh to examine the full potential of natural circulation in the vessel. The present paper reports the investigation of several strategies to model the flow and heat transfer in the RV. It is found that it is necessary to employ representative geometries of the core to estimate the heat transfer. However, by taking advantage of global and local symmetries, a detailed estimate of the strength of the resulting natural circulation and the level of heat transfer to the top of the upper plenum is obtained.
Wang, S L; Singer, M A
2009-07-13
The purpose of this report is to evaluate the hemodynamic effects of renal vein inflow and filter position on unoccluded and partially occluded IVC filters using three-dimensional computational fluid dynamics. Three-dimensional models of the TrapEase and Gunther Celect IVC filters, spherical thrombi, and an IVC with renal veins were constructed. Hemodynamics of steady-state flow was examined for unoccluded and partially occluded TrapEase and Gunther Celect IVC filters in varying proximity to the renal veins. Flow past the unoccluded filters demonstrated minimal disruption. Natural regions of stagnant/recirculating flow in the IVC are observed superior to the bilateral renal vein inflows, and high flow velocities and elevated shear stresses are observed in the vicinity of renal inflow. Spherical thrombi induce stagnant and/or recirculating flow downstream of the thrombus. Placement of the TrapEase filter in the suprarenal vein position resulted in a large area of low shear stress/stagnant flow within the filter just downstream of thrombus trapped in the upstream trapping position. Filter position with respect to renal vein inflow influences the hemodynamics of filter trapping. Placement of the TrapEase filter in a suprarenal location may be thrombogenic with redundant areas of stagnant/recirculating flow and low shear stress along the caval wall due to the upstream trapping position and the naturally occurring region of stagnant flow from the renal veins. Infrarenal vein placement of IVC filters in a near juxtarenal position with the downstream cone near the renal vein inflow likely confers increased levels of mechanical lysis of trapped thrombi due to increased shear stress from renal vein inflow.
DEVELOPING A NEW APPROACH OF COMPUTER USE `KISS MODELING' FOR DESIGN-IDEAS ALTERNATIVES OF FORM
, YOSHIHIRO KOBAYASHI South Valley University, Faculty of Fine Arts at Luxor, Egypt. wael, and form generation through computational power is more prominent in the two dimensions than the three
Computational Model of Forward and Opposed Smoldering Combustion with Improved Chemical Kinetics
Rein, Guillermo
A computational study has been carried out to investigate smoldering ignition and propagation in polyurethane foam. The onedimensional, transient, governing equations for smoldering combustion in a porous fuel are solved ...
Garba, M.T.; Gonzales-Velez, H.; Roach, D.L.
2010-11-26
This paper examines the initial parallel implementation of SCATTER, a computationally intensive inelastic neutron scattering routine with polycrystalline averaging capability, for the General Utility Lattice Program (GULP). Of particular importance...
100 IEEE TRANSACTIONS ON COMPUTERS, VOL. 45, NO. 10, OCTOBER 1996 An Analytical Model
Mudge, Trevor
to spend money is the cheapest (rather than the fastest) cache level, particularly with small system are with the Advanced Com puter Architecture Laboratory, Department of Electrical Engineer ing and Computer Science
100 IEEE TRANSACTIONS ON COMPUTERS, VOL. 45, NO. 10, OCTOBER 1996 An Analytical Model
Mudge, Trevor
is the cheapest rather than the fastest cache level, particularly with small system budgets. Another is that money, Department of Electrical Engineer- ing and Computer Science, University of Michigan, Ann Arbor, MI 48109
Using the FLUENT computational fluid dynamics code to model the NACOK corrosion test
Parks, Benjamin T
2004-01-01
As a part of advancing nuclear technology, computational fluid dynamics (CFD) analysis offers safer and lower-cost results relative to experimental work. Its use as a safety analysis tool is gaining much broader acceptance ...
Christopher Beetle; Benjamin Bromley; Richard H. Price
2006-02-08
The periodic standing wave approach to binary inspiral assumes rigid rotation of gravitational fields and hence helically symmetric solutions. To exploit the symmetry, numerical computations must solve for ``helical scalars,'' fields that are functions only of corotating coordinates, the labels on the helical Killing trajectories. Here we present the formalism for describing linearized general relativity in terms of helical scalars and we present solutions to the mixed partial differential equations of the linearized gravity problem (and to a toy nonlinear problem) using the adapted coordinates and numerical techniques previously developed for scalar periodic standing wave computations. We argue that the formalism developed may suffice for periodic standing wave computations for post-Minkowskian computations and for full general relativity.
Computer Energy Modeling Techniques for Simulation Large Scale Correctional Institutes in Texas
Heneghan, T.; Haberl, J. S.; Saman, N.; Bou-Saada, T. E.
1996-01-01
Building energy simulation programs have undergone an increase in use for evaluating energy consumption and energy conservation retrofits in buildings. Utilization of computer simulation programs for large facilities with multiple buildings, however...
An experimental validation of the PRO model for parallel and distributed computation
Boyer, Edmond
model was introduced by Gebremedhin et al. [2002] as a frame- work for the design and analysis. The Parallel Resource-Optimal (PRO) model, introduced in Gebremedhin et al. [2002], is similar to the BSP model discussion of the PRO model, see Gebremedhin et al. [2002]. The PRO model The key features of PRO
Koniges, A; Eder, E; Liu, W; Barnard, J; Friedman, A; Logan, G; Fisher, A; Masers, N; Bertozzi, A
2011-11-04
The Neutralized Drift Compression Experiment II (NDCX II) is an induction accelerator planned for initial commissioning in 2012. The final design calls for a 3 MeV, Li+ ion beam, delivered in a bunch with characteristic pulse duration of 1 ns, and transverse dimension of order 1 mm. The NDCX II will be used in studies of material in the warm dense matter (WDM) regime, and ion beam/hydrodynamic coupling experiments relevant to heavy ion based inertial fusion energy. We discuss recent efforts to adapt the 3D ALE-AMR code to model WDM experiments on NDCX II. The code, which combines Arbitrary Lagrangian Eulerian (ALE) hydrodynamics with Adaptive Mesh Refinement (AMR), has physics models that include ion deposition, radiation hydrodynamics, thermal diffusion, anisotropic material strength with material time history, and advanced models for fragmentation. Experiments at NDCX-II will explore the process of bubble and droplet formation (two-phase expansion) of superheated metal solids using ion beams. Experiments at higher temperatures will explore equation of state and heavy ion fusion beam-to-target energy coupling efficiency. Ion beams allow precise control of local beam energy deposition providing uniform volumetric heating on a timescale shorter than that of hydrodynamic expansion. The ALE-AMR code does not have any export control restrictions and is currently running at the National Energy Research Scientific Computing Center (NERSC) at LBNL and has been shown to scale well to thousands of CPUs. New surface tension models that are being implemented and applied to WDM experiments. Some of the approaches use a diffuse interface surface tension model that is based on the advective Cahn-Hilliard equations, which allows for droplet breakup in divergent velocity fields without the need for imposed perturbations. Other methods require seeding or other methods for droplet breakup. We also briefly discuss the effects of the move to exascale computing and related computational changes on general modeling codes in fusion energy.
Bradley, D.R.; Gardner, D.R.; Brockmann, J.E.; Griffith, R.O. [Sandia National Labs., Albuquerque, NM (United States)
1993-10-01
The CORCON-Mod3 computer code was developed to mechanistically model the important core-concrete interaction phenomena, including those phenomena relevant to the assessment of containment failure and radionuclide release. The code can be applied to a wide range of severe accident scenarios and reactor plants. The code represents the current state of the art for simulating core debris interactions with concrete. This document comprises the user`s manual and gives a brief description of the models and the assumptions and limitations in the code. Also discussed are the input parameters and the code output. Two sample problems are also given.
Eric Wachsman; Keith L. Duncan
2006-09-30
This research was focused on two distinct but related issues. The first issue concerned using defect modeling to understand the relationship between point defect concentration and the electrochemical, thermo-chemical and mechano-chemical properties of typical solid oxide fuel cell (SOFC) materials. The second concerned developing relationships between the microstructural features of SOFC materials and their electrochemical performance. To understand the role point defects play in ceramics, a coherent analytical framework was used to develop expressions for the dependence of thermal expansion and elastic modulus on point defect concentration in ceramics. These models, collectively termed the continuum-level electrochemical model (CLEM), were validated through fits to experimental data from electrical conductivity, I-V characteristics, elastic modulus and thermo-chemical expansion experiments for (nominally pure) ceria, gadolinia-doped ceria (GDC) and yttria-stabilized zirconia (YSZ) with consistently good fits. The same values for the material constants were used in all of the fits, further validating our approach. As predicted by the continuum-level electrochemical model, the results reveal that the concentration of defects has a significant effect on the physical properties of ceramic materials and related devices. Specifically, for pure ceria and GDC, the elastic modulus decreased while the chemical expansion increased considerably in low partial pressures of oxygen. Conversely, the physical properties of YSZ remained insensitive to changes in oxygen partial pressure within the studied range. Again, the findings concurred exactly with the predictions of our analytical model. Indeed, further analysis of the results suggests that an increase in the point defect content weakens the attractive forces between atoms in fluorite-structured oxides. The reduction treatment effects on the flexural strength and the fracture toughness of pure ceria were also evaluated at room temperature. The results reveal that the flexural strength decreases significantly after heat treatment in very low oxygen partial pressure environments; however, in contrast, fracture toughness is increased by 30-40% when the oxygen partial pressure was decreased to 10{sup -20} to 10{sup -22} atm range. Fractographic studies show that microcracks developed at 800 oC upon hydrogen reduction are responsible for the decreased strength. To understand the role of microstructure on electrochemical performance, electrical impedance spectra from symmetric LSM/YSZ/LSM cells was de-convoluted to obtain the key electrochemical components of electrode performance, namely charge transfer resistance, surface diffusion of reactive species and bulk gas diffusion through the electrode pores. These properties were then related to microstructural features, such as triple-phase boundary length and tortuosity. From these experiments we found that the impedance due to oxygen adsorption obeys a power law with pore surface area, while the impedance due to charge transfer is found to obey a power-law with respect to triple phase boundary length. A model based on kinetic theory explaining the power-law relationships observed was then developed. Finally, during our EIS work on the symmetric LSM/YSZ/LSM cells a technique was developed to improve the quality of high-frequency impedance data and their subsequent de-convolution.
Refining and Extending the Business Model with Information Technology: Dell Computer Corporation
Kraemer, Kenneth L; Dedrick, Jason; Yamashiro, Sandra
1999-01-01
of Dell’s Direct Business Model Fuels Fifteenth ConsecutiveAND EXTENDING THE REFINING AND EXTENDING THE BUSINESS MODELBUSINESS MODEL CENTER FOR RESEARCH ON INFORMATION TECHNOLOGY
A non-CFD modeling system for computing 3D wind and concentration fields in urban environments
Nelson, Matthew A; Brown, Michael J; Williams, Michael D; Gowardhan, Akshay; Pardyjak, Eric R
2010-01-01
The Quick Urban & Industrial Complex (QUIC) Dispersion Modeling System has been developed to rapidly compute the transport and dispersion of toxic agent releases in the vicinity of buildings. It is composed of an empirical-diagnostic wind solver, an 'urbanized' Lagrangian random-walk model, and a graphical user interface. The code has been used for homeland security and environmental air pollution applications. In this paper, we discuss the wind solver methodology and improvements made to the original Roeckle schemes in order to better capture flow fields in dense built-up areas. The mode1-computed wind and concentration fields are then compared to measurements from several field experiments. Improvements to the QUIC Dispersion Modeling System have been made to account for the inhomogeneous and complex building layouts found in large cities. The logic that has been introduced into the code is described and comparisons of model output to full-scale outdoor urban measurements in Oklahoma City and New York City are given. Although far from perfect, the model agreed fairly well with measurements and in many cases performed equally to CFD codes.
Experimental determination and thermodynamic modeling of the Ni-Re binary system
Yaqoob, Khurram [Chimie Metallurgique des Terres Rares (CMTR), Institut de Chimie et des Materiaux Paris-Est (ICMPE), 2-8 rue Henri Dunant, 94320 Thiais Cedex (France)] [Chimie Metallurgique des Terres Rares (CMTR), Institut de Chimie et des Materiaux Paris-Est (ICMPE), 2-8 rue Henri Dunant, 94320 Thiais Cedex (France); Joubert, Jean-Marc, E-mail: jean-marc.joubert@icmpe.cnrs.fr [Chimie Metallurgique des Terres Rares (CMTR), Institut de Chimie et des Materiaux Paris-Est (ICMPE), 2-8 rue Henri Dunant, 94320 Thiais Cedex (France)] [Chimie Metallurgique des Terres Rares (CMTR), Institut de Chimie et des Materiaux Paris-Est (ICMPE), 2-8 rue Henri Dunant, 94320 Thiais Cedex (France)
2012-12-15
The phase diagram of the Ni-Re binary system has been partially reinvestigated by chemical, structural and thermal characterization of the arc melted alloys. The experimental results obtained during the present investigation were combined with the literature data and a new phase diagram of the Ni-Re binary system is proposed. In comparison with the Ni-Re phase diagram proposed by Nash et al. in 1985 [1], significant differences in the homogeneity domains, freezing ranges and peritectic reaction temperature were evidenced. On the other hand, thermodynamic modeling of the studied system by using the new experimental information has also been carried out with the help of the CALPHAD method. The calculated Ni-Re phase diagram showed a good agreement with the selected experimental information. - Graphical abstract: Ni-Re phase diagram according to the present study. Highlights: Black-Right-Pointing-Pointer Re-investigation of the Ni-Re phase diagram. Black-Right-Pointing-Pointer Extended phase field of the hcp phase. Black-Right-Pointing-Pointer Different freezing ranges and peritectic reaction temperature. Black-Right-Pointing-Pointer Thermodynamic modeling of the studied system by using the CALPHAD method.
Whitney, Daniel
, and Industry at Leeds University Daniel E. Whitney, Liaison Scientist, Manufacturing Summary The Computer Aided Engineering Unit in the Mechanical Engineering Department at Leeds University has built its expertise mechanics. Before coming to Leeds, de Pennington was at the Philips Production Automation Department
Campo, V. L. Jr.; Capelle, K.; Quintanilla, J.; Hooley, C.
2007-12-14
We propose an experiment to obtain the phase diagram of the fermionic Hubbard model, for any dimensionality, using cold atoms in optical lattices. It is based on measuring the total energy for a sequence of trap profiles. It combines finite-size scaling with an additional 'finite-curvature scaling' necessary to reach the homogeneous limit. We illustrate its viability in the 1D case, simulating experimental data in the Bethe-ansatz local-density approximation. Including experimental errors, the filling corresponding to the Mott transition can be determined with better than 3% accuracy.
Perry, Richard Jay
1985-01-01
4. 5 percentage units less fat (21. 64% vs 26, 1% carcass fat) than non-implanted steers of the same carcass weight and rate of gain (350 kg and ADG of 1. 0 kg/d). Fat as a percentage of gain averaged 67. 85, 52. 05 and 39. 59 for non... model which can be used for projecting changes in yield grade and quality grade with performance of cattle in the feedlot over time. The primary objectives of this research were: 1) To determine the "shelf life" of a steer in the feedlot. 2...
Nalewajko, Krzysztof; Begelman, Mitchell C.; Sikora, Marek
2014-11-20
Estimates of magnetic field strength in relativistic jets of active galactic nuclei, obtained by measuring the frequency-dependent radio core location, imply that the total magnetic fluxes in those jets are consistent with the predictions of the magnetically arrested disk (MAD) scenario of jet formation. On the other hand, the magnetic field strength determines the luminosity of the synchrotron radiation, which forms the low-energy bump of the observed blazar spectral energy distribution (SED). The SEDs of the most powerful blazars are strongly dominated by the high-energy bump, which is most likely due to the external radiation Compton mechanism. This high Compton dominance may be difficult to reconcile with the MAD scenario, unless (1) the geometry of external radiation sources (broad-line region, hot-dust torus) is quasi-spherical rather than flat, or (2) most gamma-ray radiation is produced in jet regions of low magnetization, e.g., in magnetic reconnection layers or in fast jet spines.
Computer Modeling of Transport of Oxidizing Species in Grain Boundaries during Zirconium Corrosion
Xian-Ming Bai; Yongfeng Zhang; Michael R. Tonks
2014-06-01
Zirconium (Zr) based alloys are widely used as the cladding materials in light-water reactors. The water-side corrosion of these alloys degrades their structural integrity and poses serious safety concerns. During the Zr corrosion process, a thin Zr oxide (ZrO2) layer forms on the alloy surface and serves as a barrier layer for further corrosion. The majority of the oxide has the monoclinic phase. At the transition region between the oxide and the metal, the oxide contains a thin layer of stabilized tetragonal phase. It is found that the texture of the tetragonal layer determines the protectiveness of the oxide for corrosion. The transport of oxidizing species, such as anion defects, cation defects, and electron through the tetragonal oxide layer could be the rate limiting step of the corrosion. The defect diffusion can be affected by the growing stresses and microstructures such as grain boundaries and dislocations. In this work molecular dynamics simulations are used to investigate the anion and cation diffusion in bulk and at grain boundaries in tetragonal ZrO2. The results show that defect diffusion at grain boundaries is complex and the behavior strongly depends on the grain boundary type. For most of the grain boundaries studied the defect diffusion are much slower than in the bulk, implying that grain boundaries may not be fast defect transport paths during corrosion. The connection between the modeling results and published experimental work will also be discussed. This work is funded by the Laboratory Directed Research and Development (LDRD) program at Idaho National Laboratory.
TECHNICAL REPORTS Today there are many well-established computer models that
Radcliffe, David
the monitoring data that is typically available. There is a need for more parsimonious models and monitoring data in the United States and Europe cover a spectrum from static risk-based index models such as the P-Index (Weld and Sharpley, 2007) and Source Apportionment (OSPAR, 2000) models to complex, process-based, dynamic models
Automated Analysis of Software Designs with Graphic Quality Models Department of Computer Science
Zhu, Hong
Automated Analysis of Software Designs with Graphic Quality Models QIAN ZHANG Department@brookes.ac.uk Abstract: Software quality models play a significant role in software quality assurance. Based on our previous work on graphic modelling of software quality, this paper extends the quality modelling language
Perez-Sanchez, Danyl [CIEMAT, Avenida Complutense 40, 28040, Madrid (Spain)] [CIEMAT, Avenida Complutense 40, 28040, Madrid (Spain)
2013-07-01
As a result of a pilot project developed at the old Spanish 'Junta de Energia Nuclear' to extract uranium from ores, tailings materials were generated. Most of these residual materials were sent back to different uranium mines, but a small amount of it was mixed with conventional building materials and deposited near the old plant until the surrounding ground was flattened. The affected land is included in an area under institutional control and used as recreational area. At the time of processing, uranium isotopes were separated but other radionuclides of the uranium decay series as Th-230, Ra-226 and daughters remain in the residue. Recently, the analyses of samples taken at different ground's depths confirmed their presence. This paper presents the methodology used to calculate the derived concentration level to ensure that the reference dose level of 0.1 mSv y-1 used as radiological criteria. In this study, a radiological impact assessment was performed modeling the area as recreational scenario. The modelization study was carried out with the code RESRAD considering as exposure pathways, external irradiation, inadvertent ingestion of soil, inhalation of resuspended particles, and inhalation of radon (Rn-222). As result was concluded that, if the concentration of Ra-226 in the first 15 cm of soil is lower than, 0.34 Bq g{sup -1}, the dose would not exceed the reference dose. Applying this value as a derived concentration level and comparing with the results of measurements on the ground, some areas with a concentration of activity slightly higher than latter were found. In these zones the remediation proposal has been to cover with a layer of 15 cm of clean material. This action represents a reduction of 85% of the dose and ensures compliance with the reference dose. (authors)
Rybkowski, Zofia K.; Wong, John-Michael; Ballard, Glenn; Tommelein, Iris D.
2008-01-01
1 Computer Push 20 46 55 0 3 0 58 Live Push 20 43 51 1 5 0 57* Computer Pull 12 46 1 0 1 0 2 Live Pull 12 39 1 1 0 0 2 *WS1 ran out of pieces at 5'20" DISCUSSION A purpose of many lean production simulations is to educate players... of Construction Engineering and Management, ASCE, 125(4), 265-276. Myers, J.L. and Well, A.D. (2003). Research Design and Statistical Analysis. 2nd Ed., Lawrence Erlbaum Assoc. Pubs., Mahwah, NJ. Pasqualini, F. and Zawislak, P.A. (2005). "Value Stream Mapping...
Stewart, James
, \\Shadows for bump-mapped surfaces", in Advanced Computer Graphics (Proceedings of Computer Graphics Tokyo. Hershberger, \\Finding the upper envelope of n line segments in O(nlogn) time", Inform. Process. Lett., vol. 33. The horizon computation can also be used to build horizon maps that are used in conjunction with bump maps
Stewart, James
. Max, ``Shadows for bumpmapped surfaces'', in Advanced Computer Graphics (Proceedings of Computer, pp. 92--99. [10] J. Hershberger, ``Finding the upper envelope of n line segments in O(n log n) time input to a progressive radiosity algorithm. The horizon computation can also be used to build horizon
Gribaudo, Marco
, Analysis, and Simulation of Computer and Telecommunications Systems (MASCOTS'04) 1526-7539/04 $20.00 © 2004, Analysis, and Simulation of Computer and Telecommunications Systems (MASCOTS'04) 1526-7539/04 $20.00 © 2004, Analysis, and Simulation of Computer and Telecommunications Systems (MASCOTS'04) 1526-7539/04 $20.00 © 2004
Gray, Alan; Harlen, Oliver G.; Harris, Sarah A.; Khalid, Syma; Leung, Yuk Ming; Lonsdale, Richard; Mulholland, Adrian J.; Pearson, Arwen R.; Read, Daniel J.; Richardson, Robin A.
2015-01-01
The current computational techniques available for biomolecular simulation are described, and the successes and limitations of each with reference to the experimental biophysical methods that they complement are presented. Despite huge advances in the computational techniques available for simulating biomolecules at the quantum-mechanical, atomistic and coarse-grained levels, there is still a widespread perception amongst the experimental community that these calculations are highly specialist and are not generally applicable by researchers outside the theoretical community. In this article, the successes and limitations of biomolecular simulation and the further developments that are likely in the near future are discussed. A brief overview is also provided of the experimental biophysical methods that are commonly used to probe biomolecular structure and dynamics, and the accuracy of the information that can be obtained from each is compared with that from modelling. It is concluded that progress towards an accurate spatial and temporal model of biomacromolecules requires a combination of all of these biophysical techniques, both experimental and computational.
Computing dynamics of copper-based SMA via centre manifold reduction of 3D models q
Melnik, Roderick
of applications in such areas as aerospace engineering, medicine, manufacturing (in- cluding automotive industry is based on the authors' contribution to the European Conference on Computational Mechanics, Munich, 1999 mechanics, and applied mathematics communities [4±13]. This interest has two major
Norman, Gethin
probabilistically quantified properties pertaining to buffer sizes, delays, energy usage etc., for each derived and Rajesh Gupta3 1School of Computer Science, University of Birmingham, Birmingham, UK 2Bradley Department). In this paper, we focus on the latter. Dynamic power management (DPM) is a way to save energy in devices which
Creativity as Pastiche: A Computational Model of Dynamic Blending and Textual Collage,
Veale, Tony
Reference to the Use of Blending in Cinematic Narratives Tony Veale, School of Computer Applications, Dublin combination within a framework of interconnected mental spaces. Through an analysis of cinematic pastiche, as exemplified by its use in cinematic pastiche, a process in which two or more narratives are blended together
Ewing, Richard E.
and Information Volume 1, Number 1, Pages 116 HIGH PERFORMANCE COMPUTING IN PETROLEUM APPLICATIONS RICHARD E in petroleum applications is to try to optimize the recovery of hydrocarbon from permeable underground the complex geomechanical, physical, and multiphase fluid flow processes that accompany the various recov- ery
Liquid Computing in a Simplified Model of Cortical Layer IV: Learning to Balance a Ball
Maass, Wolfgang
Probst1,3 , Wolfgang Maass2 , Henry Markram1 , and Marc-Oliver Gewaltig1 1 Blue Brain Project, ´Ecole that the cortical network can robustly learn this task using a supervised learning rule that penalizes the error function is then constructed by readout neurons or action cells which compute a weighted sum of the high
MapReduce Programming Model for .NET-based Distributed Computing Chao Jin and Rajkumar Buyya
Buyya, Rajkumar
,000 to several million commodity machines dispersed across about 25 data centers [10]. Microsoft also presents of Virtual Machines through the web service interface. Although the popularity of data centers is increasing the basis of `data center computer' [5]. #12;2 .NET is the standard platform of Windows applications
Reyes, Dasia Ann
2009-05-15
-Stokes (RANS) method. Preliminary investigations of PANS show promising re-sults but there still exist computational and physical issues that must be addressed.This study investigates the performance of the PANS method for ow past acylinder at a Reynolds number...
Brown, Jed
are needed to cope Computer architecture is increasingly hierarchical algorithms should conform the task of systematic truncation error testing and accuracy estimation. -- 1986 #12;Diffusive cooling(nlogk) Project vector onto eigenbasis in O(nlogk) Optimal-order graph partitioning and clustering Continuation
Modeling a Reconfigurable System for Computing the FFT in Place via Rewriting-Logic
Ayala-Rincón, Mauricio
implementation of the Fast Fourier Transform FFT using rewriting-logic. It is shown that rewriting-logic can purpose processors, reconfigurable computing delivers more processing power due to the implementation design makes it a good solution to the rapid changing electronic market. There are several taxonomies
College of Science Bachelor of Science in Computational Modeling and Data Analytics
Zallen, Richard
(Pre: MATH 1225) (4)( ) MATH 2114 Introduction to Linear Algebra (3)( ) MATH 2204 Intro Multivariable)( ) MATH 4144* Linear Algebra (Pre: MATH 3144) (3)( ) MATH 4425* Fourier Series PDE (Pre: MATH 2214, 2204 for computing "in-major GPA." Also included: additional STAT, MATH, CS courses at 3000 or 4000-level
Stat 586 Interpretation of Data I (Modern Applied Statistical Modeling and Computing)
Xie, Minge
, B.D., Modern Applied Statistics with S-Plus (Springer) Cody, R.P. and Simth, J.K. Applied Statistics and the SAS Programming Language Elliott, R. J. Learning SAS in the Computer Lab (Duxbury Press). Based allows) Students should expect to use either R or SAS or Splus statistical software packages in data
Development and Validation of a Computer Model for Energy-Efficient Shaded Fenestration Design
Oh, Kie Whan
2000-01-01
transfer through an unshaded window at any orientation. A computerized simulation model, called the Shaded Fenestration Design (SFD) model, was developed that is capable of calculating the amount of solar radiation for a clear day as well as calculating...
A critical review on modelling formalisms and simulation tools in computational biosystems
Machado, Daniel
Integration of different kinds of biological processes is an ultimate goal for whole-cell modelling. We briefly review modelling formalisms that have been used in Systems Biology and identify the criteria that must be ...
A Novel Computational Model for Tilting Pad Journal Bearings with Soft Pivot Stiffnesses
Tao, Yujiao 1988-
2012-12-10
A novel tilting pad journal bearing model including pivot flexibility as well as temporal fluid inertia effects on the thin film fluid flow aims to accurately predict the bearing forced performance. The predictive model also accounts for the thermal...
Flow field computation of the NREL S809 airfoil using various turbulence models
Chang, Y.L.; Yang, S.L.; Arici, O. [Michigan Technological Univ., Houghton, MI (United States). Mechanical Engineering-Engineering Mechanics Dept.
1996-10-01
Performance comparison of three popular turbulence models, namely Baldwin-Lomas algebraic model, Chien`s Low-Reynolds-Number {kappa}-{epsilon} model, and Wilcox`s Low-Reynolds-Number {kappa}-{omega} model, is given. These models were applied to calculate the flow field around the National Renewable Energy Laboratory S809 airfoil using Total Variational Diminishing scheme. Numerical results of C{sub P}, C{sub L}, and C{sub D} are presented along with the Delft experimental data. It is shown that all three models perform well for attached flow, i.e., no flow separation at low angles of attack. However, at high angles of attack with flow separation, convergence characteristics show Wilcox`s model outperforms the other models. Results of this study will be used to guide the authors in their dynamic stall research.
Chen, Shu-Ching
Florida Public Hurricane Loss Model Purpose: To develop and maintain a public computer model to assess hurricane wind, surge and flood related risk and to project annual expected insured residential for user defined scenarios. This public model can also be used to quantify the cost benefits of hurricane
Mathematical and Computer Modelling 43 (2006) 718726 www.elsevier.com/locate/mcm
Bohner, Martin
2006-01-01
maximization problem is much more flexible and realistic. For example, a consumer receives income at one point unify the standard discrete and continuous models in economics, but also, for example, allows models are set up in either continuous or discrete time. For example, in a discrete model, a consumer
A Computer Model of Field-directed Morphogenesis Part I -Julia Sets
Levin, Michael
are viewed as a map from ¢¡ to £¡ , and are modeled by the complex number algebra. Functions in complex that gene interactions can be modeled as a multi-dimensional algebra, and that only two interacting genes with actual pictures of any particular real organism. A brief introduction to top-down models and positional
Computer resources Computer resources
Yang, Zong-Liang
Computer resources 1 Computer resources available to the LEAD group Cédric David 30 September 2009 #12;Ouline · UT computer resources and services · JSG computer resources and services · LEAD computers· LEAD computers 2 #12;UT Austin services UT EID and Password 3 https://utdirect.utexas.edu #12;UT Austin
Westbrook, C K; Mizobuchi, Y; Poinsot, T J; Smith, P J; Warnatz, J
2004-08-26
Progress in the field of computational combustion over the past 50 years is reviewed. Particular attention is given to those classes of models that are common to most system modeling efforts, including fluid dynamics, chemical kinetics, liquid sprays, and turbulent flame models. The developments in combustion modeling are placed into the time-dependent context of the accompanying exponential growth in computer capabilities and Moore's Law. Superimposed on this steady growth, the occasional sudden advances in modeling capabilities are identified and their impacts are discussed. Integration of submodels into system models for spark ignition, diesel and homogeneous charge, compression ignition engines, surface and catalytic combustion, pulse combustion, and detonations are described. Finally, the current state of combustion modeling is illustrated by descriptions of a very large jet lifted 3D turbulent hydrogen flame with direct numerical simulation and 3D large eddy simulations of practical gas burner combustion devices.
PAD, a computer-aided molecular model building system utilizing two-dimensional graphical input
White, William Gerald
1980-01-01
and an artist. CONTENTS CHAPTER I. BACKGROUND: INTERFACING COMPUTER AND CHEMIcT Graphical Input Hardware Evolution. . . . . Software Systems: Using the Hardware, , Chemical Applications. Program PAD xn Perspective. CHAPTER II. YATERIA' S AND METHODS... are used. The 2-D coordinates of each atom input by the user are stored permanently in XS and YS, after temporary storage in IX and IY. Arrays XS and YS are each 99 elements long, as are all other arrays storing information about input atoms; therefore...
National Ignition Facility computational fluid dynamics modeling and light fixture case studies
Martin, R.; Bernardin, J.; Parietti, L.; Dennison, B.
1998-02-01
This report serves as a guide to the use of computational fluid dynamics (CFD) as a design tool for the National Ignition Facility (NIF) program Title I and Title II design phases at Lawrence Livermore National Laboratory. In particular, this report provides general guidelines on the technical approach to performing and interpreting any and all CFD calculations. In addition, a complete CFD analysis is presented to illustrate these guidelines on a NIF-related thermal problem.
Computational Tools for Accelerating Carbon Capture Process Development
Miller, David; Sahinidis, N.V,; Cozad, A; Lee, A; Kim, H; Morinelly, J.; Eslick, J.; Yuan, Z.
2013-06-04
This presentation reports development of advanced computational tools to accelerate next generation technology development. These tools are to develop an optimized process using rigorous models. They include: Process Models; Simulation-Based Optimization; Optimized Process; Uncertainty Quantification; Algebraic Surrogate Models; and Superstructure Optimization (Determine Configuration).