Error estimation of bathymetric grid models derived from historic and contemporary datasets
New Hampshire, University of
1 Error estimation of bathymetric grid models derived from historic and contemporary datasets and rapidly collecting dense bathymetric datasets. Sextants were replaced by radio navigation, then transit, to digitized contours; the test dataset shows examples of all of these types. From this database, we assign
Ma, Jia, S.M. Massachusetts Institute of Technology
2009-01-01T23:59:59.000Z
This thesis examines how the basis risk affects property derivative hedging in the UK market, based on the tracking error (basis risk) report from the Investment Property Forum study in 2007 (the IPF Study). The thesis ...
Deriving Human-Error Tolerance Requirements from Tasks Peter Wright, Bob Fields and Michael Harrison
Fields, Bob
- logy (SHARP) by employing a software engineering nota- tion (CSP) that provides a bridge between to human error, describe a task notationbased on CSP which helps us to elicit requirements on human settings (e.g., power and chemical plant control rooms, aircraft cockpits, etc.), users are not occasional
Remarks on statistical errors in equivalent widths
Klaus Vollmann; Thomas Eversberg
2006-07-03T23:59:59.000Z
Equivalent width measurements for rapid line variability in atomic spectral lines are degraded by increasing error bars with shorter exposure times. We derive an expression for the error of the line equivalent width $\\sigma(W_\\lambda)$ with respect to pure photon noise statistics and provide a correction value for previous calculations.
Olson, Eric J.
2013-06-11T23:59:59.000Z
An apparatus, program product, and method that run an algorithm on a hardware based processor, generate a hardware error as a result of running the algorithm, generate an algorithm output for the algorithm, compare the algorithm output to another output for the algorithm, and detect the hardware error from the comparison. The algorithm is designed to cause the hardware based processor to heat to a degree that increases the likelihood of hardware errors to manifest, and the hardware error is observable in the algorithm output. As such, electronic components may be sufficiently heated and/or sufficiently stressed to create better conditions for generating hardware errors, and the output of the algorithm may be compared at the end of the run to detect a hardware error that occurred anywhere during the run that may otherwise not be detected by traditional methodologies (e.g., due to cooling, insufficient heat and/or stress, etc.).
Abdelhamid Awad Aly Ahmed, Sala
2008-10-10T23:59:59.000Z
QUANTUM ERROR CONTROL CODES A Dissertation by SALAH ABDELHAMID AWAD ALY AHMED Submitted to the Office of Graduate Studies of Texas A&M University in partial fulfillment of the requirements for the degree of DOCTOR OF PHILOSOPHY May 2008 Major... Subject: Computer Science QUANTUM ERROR CONTROL CODES A Dissertation by SALAH ABDELHAMID AWAD ALY AHMED Submitted to the Office of Graduate Studies of Texas A&M University in partial fulfillment of the requirements for the degree of DOCTOR OF PHILOSOPHY...
Deterministic treatment of model error in geophysical data assimilation
Carrassi, Alberto
2015-01-01T23:59:59.000Z
This chapter describes a novel approach for the treatment of model error in geophysical data assimilation. In this method, model error is treated as a deterministic process fully correlated in time. This allows for the derivation of the evolution equations for the relevant moments of the model error statistics required in data assimilation procedures, along with an approximation suitable for application to large numerical models typical of environmental science. In this contribution we first derive the equations for the model error dynamics in the general case, and then for the particular situation of parametric error. We show how this deterministic description of the model error can be incorporated in sequential and variational data assimilation procedures. A numerical comparison with standard methods is given using low-order dynamical systems, prototypes of atmospheric circulation, and a realistic soil model. The deterministic approach proves to be very competitive with only minor additional computational c...
Dynamic Prediction of Concurrency Errors
Sadowski, Caitlin
2012-01-01T23:59:59.000Z
Relation 15 Must-Before Race Prediction 16 Implementation 17viii Abstract Dynamic Prediction of Concurrency Errors bySANTA CRUZ DYNAMIC PREDICTION OF CONCURRENCY ERRORS A
A Priori Error Estimates for Some Discontinuous Galerkin Immersed ...
2015-01-12T23:59:59.000Z
estimate in a mesh-dependant energy norm is derived, and this error ... 0 (Th), integrate both sides on each element K ? Th, and apply the Green's formula to.
Statistical Error in Particle Simulations of Low Mach Number Flows
Hadjiconstantinou, Nicolas G.
We present predictions for the statistical error due to finite sampling in the presence of thermal fluctuations in molecular simulation algorithms. Expressions for the fluid velocity, density and temperature are derived ...
Annual Energy Outlook 2013 [U.S. Energy Information Administration (EIA)]
AFDC Printable Version Share this resource Send a link to EERE: Alternative Fuels Data Center Home Page to someone by E-mail Share EERE: Alternative Fuels Data Center Home Page on Facebook Tweet about EERE: Alternative Fuels Data Center Home Page on Twitter Bookmark EERE:1 First Use of Energy for All Purposes (Fuel and Nonfuel),Feet) Year Jan Feb Mar Apr May Jun Jul(Summary) " ,"ClickPipelines AboutDecember 2005 (Thousand BarrelsNatural GasCold Fusion Error Unexpected Error Sorry!
Sandford, II, Maxwell T. (Los Alamos, NM); Handel, Theodore G. (Los Alamos, NM); Ettinger, J. Mark (Los Alamos, NM)
1999-01-01T23:59:59.000Z
A method of embedding auxiliary information into the digital representation of host data containing noise in the low-order bits. The method applies to digital data representing analog signals, for example digital images. The method reduces the error introduced by other methods that replace the low-order bits with auxiliary information. By a substantially reverse process, the embedded auxiliary data can be retrieved easily by an authorized user through use of a digital key. The modular error embedding method includes a process to permute the order in which the host data values are processed. The method doubles the amount of auxiliary information that can be added to host data values, in comparison with bit-replacement methods for high bit-rate coding. The invention preserves human perception of the meaning and content of the host data, permitting the addition of auxiliary data in the amount of 50% or greater of the original host data.
Approaches to Quantum Error Correction
Julia Kempe
2006-12-21T23:59:59.000Z
The purpose of this little survey is to give a simple description of the main approaches to quantum error correction and quantum fault-tolerance. Our goal is to convey the necessary intuitions both for the problems and their solutions in this area. After characterising quantum errors we present several error-correction schemes and outline the elements of a full fledged fault-tolerant computation, which works error-free even though all of its components can be faulty. We also mention alternative approaches to error-correction, so called error-avoiding or decoherence-free schemes. Technical details and generalisations are kept to a minimum.
Unequal Error Protection Turbo Codes
Henkel, Werner
Unequal Error Protection Turbo Codes Diploma Thesis Neele von Deetzen Arbeitsbereich Nachrichtentechnik School of Engineering and Science Bremen, February 28th, 2005 #12;Unequal Error Protection Turbo Convolutional Codes / Turbo Codes 18 3.1 Structure
Gasoline and Diesel Fuel Update (EIA)
AFDC Printable Version Share this resource Send a link to EERE: Alternative Fuels Data Center Home Page to someone by E-mail Share EERE: Alternative Fuels Data Center Home Page on Facebook Tweet about EERE: Alternative Fuels Data Center Home Page on Twitter Bookmark EERE:1 First Use of Energy for All Purposes (Fuel and Nonfuel),Feet) Year Jan Feb Mar Apr May Jun Jul(Summary) " ,"ClickPipelines About U.S.30Natural Gas Glossary529 633 6221,2372003of Energy for39 TableErrors of
Nested Quantum Error Correction Codes
Zhuo Wang; Kai Sun; Hen Fan; Vlatko Vedral
2009-09-28T23:59:59.000Z
The theory of quantum error correction was established more than a decade ago as the primary tool for fighting decoherence in quantum information processing. Although great progress has already been made in this field, limited methods are available in constructing new quantum error correction codes from old codes. Here we exhibit a simple and general method to construct new quantum error correction codes by nesting certain quantum codes together. The problem of finding long quantum error correction codes is reduced to that of searching several short length quantum codes with certain properties. Our method works for all length and all distance codes, and is quite efficient to construct optimal or near optimal codes. Two main known methods in constructing new codes from old codes in quantum error-correction theory, the concatenating and pasting, can be understood in the framework of nested quantum error correction codes.
Finding beam focus errors automatically
Lee, M.J.; Clearwater, S.H.; Kleban, S.D.
1987-01-01T23:59:59.000Z
An automated method for finding beam focus errors using an optimization program called COMFORT-PLUS. The steps involved in finding the correction factors using COMFORT-PLUS has been used to find the beam focus errors for two damping rings at the SLAC Linear Collider. The program is to be used as an off-line program to analyze actual measured data for any SLC system. A limitation on the application of this procedure is found to be that it depends on the magnitude of the machine errors. Another is that the program is not totally automated since the user must decide a priori where to look for errors. (LEW)
Data& Error Analysis 1 DATA and ERROR ANALYSIS
Mukasyan, Alexander
Data& Error Analysis 1 DATA and ERROR ANALYSIS Performing the experiment and collecting data learned, you might get a better grade.) Data analysis should NOT be delayed until all of the data. This will help one avoid the problem of spending an entire class collecting bad data because of a mistake
Experimental Uncertainties (Errors) Sources of Experimental Uncertainties (Experimental Errors)
Mukasyan, Alexander
the preparation of the lab report. A calculator should 1. Bevington, P. R., Data Reduction and Error Analysis for the Physical Sciences, New York: McGraw-Hill, 1969. 2. Taylor, J. R., An introduction to uncertainty analysis in the lab. In this laboratory, we keep to a very simple form of error analysis, our purpose being more
Static Detection of Disassembly Errors
Krishnamoorthy, Nithya; Debray, Saumya; Fligg, Alan K.
2009-10-13T23:59:59.000Z
Static disassembly is a crucial ?rst step in reverse engineering executable ?les, and there is a consider- able body of work in reverse-engineering of binaries, as well as areas such as semantics-based security anal- ysis, that assumes that the input executable has been correctly disassembled. However, disassembly errors, e.g., arising from binary obfuscations, can render this assumption invalid. This work describes a machine- learning-based approach, using decision trees, for stat- ically identifying possible errors in a static disassem- bly; such potential errors may then be examined more closely, e.g., using dynamic analyses. Experimental re- sults using a variety of input executables indicate that our approach performs well, correctly identifying most disassembly errors with relatively few false positives.
Unequal error protection of subband coded bits
Devalla, Badarinath
2012-06-07T23:59:59.000Z
Source coded data can be separated into different classes based on their susceptibility to channel errors. Errors in the Important bits cause greater distortion in the reconstructed signal. This thesis presents an Unequal Error Protection scheme...
Two-Layer Error Control Codes Combining Rectangular and Hamming Product Codes for Cache Error
Zhang, Meilin
We propose a novel two-layer error control code, combining error detection capability of rectangular codes and error correction capability of Hamming product codes in an efficient way, in order to increase cache error ...
Error Rate Performance of Coded Free-Space Optical Links over Gamma-Gamma Turbulence Channels
Li, Tiffany Jing
Error Rate Performance of Coded Free-Space Optical Links over Gamma-Gamma Turbulence Channels Murat be used over free-space optical (FSO) links to mitigate turbulence-induced fading. In this paper, we channels, considering the recently introduced gamma-gamma turbulence model. We derive a pairwise error
Error 404 - Document not found
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
AFDC Printable Version Share this resource Send a link to EERE: Alternative Fuels Data Center Home Page to someone by E-mail Share EERE: Alternative Fuels Data Center Home Page on Facebook Tweet about EERE: Alternative Fuels Data Center Home Page on Twitter Bookmark EERE:1 First Use of Energy for All Purposes (Fuel and Nonfuel),Feet) Year Jan Feb Mar Apr MayAtmospheric Optical Depth7-1D: Vegetation ProposedUsing ZirconiaPolicy and Assistance100 ton Stanat rollinggovErrors ERROR 404 - URL Not
Full protection of superconducting qubit systems from coupling errors
M. J. Storcz; J. Vala; K. R. Brown; J. Kempe; F. K. Wilhelm; K. B. Whaley
2005-08-09T23:59:59.000Z
Solid state qubits realized in superconducting circuits are potentially extremely scalable. However, strong decoherence may be transferred to the qubits by various elements of the circuits that couple individual qubits, particularly when coupling is implemented over long distances. We propose here an encoding that provides full protection against errors originating from these coupling elements, for a chain of superconducting qubits with a nearest neighbor anisotropic XY-interaction. The encoding is also seen to provide partial protection against errors deriving from general electronic noise.
Fitting Pulsar Wind Tori. II. Error Analysis and Applications
C. -Y. Ng; Roger W. Romani
2007-10-23T23:59:59.000Z
We have applied the torus fitting procedure described in Ng & Romani (2004) to PWNe observations in the Chandra data archive. This study provides quantitative measurement of the PWN geometry and we characterize the uncertainties in the fits, with statistical errors coming from the fit uncertainties and systematic errors estimated by varying the assumed fitting model. The symmetry axis $\\Psi$ of the PWN are generally well determined, and highly model-independent. We often derive a robust value for the spin inclination $\\zeta$. We briefly discuss the utility of these results in comparison with new radio and high energy pulse measurements
Fitting Pulsar Wind Tori. II. Error Analysis and Applications
Ng, C -Y
2007-01-01T23:59:59.000Z
We have applied the torus fitting procedure described in Ng & Romani (2004) to PWNe observations in the Chandra data archive. This study provides quantitative measurement of the PWN geometry and we characterize the uncertainties in the fits, with statistical errors coming from the fit uncertainties and systematic errors estimated by varying the assumed fitting model. The symmetry axis $\\Psi$ of the PWN are generally well determined, and highly model-independent. We often derive a robust value for the spin inclination $\\zeta$. We briefly discuss the utility of these results in comparison with new radio and high energy pulse measurements
Quantum Error Correction with magnetic molecules
José J. Baldoví; Salvador Cardona-Serra; Juan M. Clemente-Juan; Luis Escalera-Moreno; Alejandro Gaita-Ariño; Guillermo Mínguez Espallargas
2014-08-22T23:59:59.000Z
Quantum algorithms often assume independent spin qubits to produce trivial $|\\uparrow\\rangle=|0\\rangle$, $|\\downarrow\\rangle=|1\\rangle$ mappings. This can be unrealistic in many solid-state implementations with sizeable magnetic interactions. Here we show that the lower part of the spectrum of a molecule containing three exchange-coupled metal ions with $S=1/2$ and $I=1/2$ is equivalent to nine electron-nuclear qubits. We derive the relation between spin states and qubit states in reasonable parameter ranges for the rare earth $^{159}$Tb$^{3+}$ and for the transition metal Cu$^{2+}$, and study the possibility to implement Shor's Quantum Error Correction code on such a molecule. We also discuss recently developed molecular systems that could be adequate from an experimental point of view.
Distributed Error Confinement Extended Abstract
Patt-Shamir, Boaz
. These algorithms can serve as building blocks in more general reactive systems. Previous results in exploring locality in reactive systems were not error confined, and relied on the assump- tion (not used in current, that seems inherent for voting in reactive networks; its analysis leads to an interesting combinatorial
SHEAN (Simplified Human Error Analysis code) and automated THERP
Wilson, J.R.
1993-06-01T23:59:59.000Z
One of the most widely used human error analysis tools is THERP (Technique for Human Error Rate Prediction). Unfortunately, this tool has disadvantages. The Nuclear Regulatory Commission, realizing these drawbacks, commissioned Dr. Swain, the author of THERP, to create a simpler, more consistent tool for deriving human error rates. That effort produced the Accident Sequence Evaluation Program Human Reliability Analysis Procedure (ASEP), which is more conservative than THERP, but a valuable screening tool. ASEP involves answering simple questions about the scenario in question, and then looking up the appropriate human error rate in the indicated table (THERP also uses look-up tables, but four times as many). The advantages of ASEP are that human factors expertise is not required, and the training to use the method is minimal. Although not originally envisioned by Dr. Swain, the ASEP approach actually begs to be computerized. That WINCO did, calling the code SHEAN, for Simplified Human Error ANalysis. The code was done in TURBO Basic for IBM or IBM-compatible MS-DOS, for fast execution. WINCO is now in the process of comparing this code against THERP for various scenarios. This report provides a discussion of SHEAN.
Phase Errors and the Capture Effect
Blair, J., and Machorro, E.
2011-11-01T23:59:59.000Z
This slide-show presents analysis of spectrograms and the phase error of filtered noise in a signal. When the filtered noise is smaller than the signal amplitude, the phase error can never exceed 90{deg}, so the average phase error over many cycles is zero: this is called the capture effect because the largest signal captures the phase and frequency determination.
Approximate error conjugation gradient minimization methods
Kallman, Jeffrey S
2013-05-21T23:59:59.000Z
In one embodiment, a method includes selecting a subset of rays from a set of all rays to use in an error calculation for a constrained conjugate gradient minimization problem, calculating an approximate error using the subset of rays, and calculating a minimum in a conjugate gradient direction based on the approximate error. In another embodiment, a system includes a processor for executing logic, logic for selecting a subset of rays from a set of all rays to use in an error calculation for a constrained conjugate gradient minimization problem, logic for calculating an approximate error using the subset of rays, and logic for calculating a minimum in a conjugate gradient direction based on the approximate error. In other embodiments, computer program products, methods, and systems are described capable of using approximate error in constrained conjugate gradient minimization problems.
Error suppression in Hamiltonian based quantum computation using energy penalties
Adam D. Bookatz; Edward Farhi; Leo Zhou
2014-07-06T23:59:59.000Z
We consider the use of quantum error detecting codes, together with energy penalties against leaving the codespace, as a method for suppressing environmentally induced errors in Hamiltonian based quantum computation. This method was introduced in [1] in the context of quantum adiabatic computation, but we consider it more generally. Specifically, we consider a computational Hamiltonian, which has been encoded using the logical qubits of a single-qubit error detecting code, coupled to an environment of qubits by interaction terms that act one-locally on the system. Energy penalty terms are added that penalize states outside of the codespace. We prove that in the limit of infinitely large penalties, one-local errors are completely suppressed, and we derive some bounds for the finite penalty case. Our proof technique involves exact integration of the Schrodinger equation, making no use of master equations or their assumptions. We perform long time numerical simulations on a small (one logical qubit) computational system coupled to an environment and the results suggest that the energy penalty method achieves even greater protection than our bounds indicate.
Threshold error rates for the toric and surface codes
D. S. Wang; A. G. Fowler; A. M. Stephens; L. C. L. Hollenberg
2009-05-05T23:59:59.000Z
The surface code scheme for quantum computation features a 2d array of nearest-neighbor coupled qubits yet claims a threshold error rate approaching 1% (NJoP 9:199, 2007). This result was obtained for the toric code, from which the surface code is derived, and surpasses all other known codes restricted to 2d nearest-neighbor architectures by several orders of magnitude. We describe in detail an error correction procedure for the toric and surface codes, which is based on polynomial-time graph matching techniques and is efficiently implementable as the classical feed-forward processing step in a real quantum computer. By direct simulation of this error correction scheme, we determine the threshold error rates for the two codes (differing only in their boundary conditions) for both ideal and non-ideal syndrome extraction scenarios. We verify that the toric code has an asymptotic threshold of p = 15.5% under ideal syndrome extraction, and p = 7.8 10^-3 for the non-ideal case, in agreement with prior work. Simulations of the surface code indicate that the threshold is close to that of the toric code.
The Human Bathtub: Safety and Risk Predictions Including the Dynamic Probability of Operator Errors
Duffey, Romney B. [Atomic Energy of Canada, Ltd., 2251 Speakman Drive, Mississauga, ON, L5K 1B2 (Canada); Saull, John W. [International Federation of Airwothiness, 14 Railway Approach, East Grinstead, West Sussex, RH19 1BP (United Kingdom)
2006-07-01T23:59:59.000Z
Reactor safety and risk are dominated by the potential and major contribution for human error in the design, operation, control, management, regulation and maintenance of the plant, and hence to all accidents. Given the possibility of accidents and errors, now we need to determine the outcome (error) probability, or the chance of failure. Conventionally, reliability engineering is associated with the failure rate of components, or systems, or mechanisms, not of human beings in and interacting with a technological system. The probability of failure requires a prior knowledge of the total number of outcomes, which for any predictive purposes we do not know or have. Analysis of failure rates due to human error and the rate of learning allow a new determination of the dynamic human error rate in technological systems, consistent with and derived from the available world data. The basis for the analysis is the 'learning hypothesis' that humans learn from experience, and consequently the accumulated experience defines the failure rate. A new 'best' equation has been derived for the human error, outcome or failure rate, which allows for calculation and prediction of the probability of human error. We also provide comparisons to the empirical Weibull parameter fitting used in and by conventional reliability engineering and probabilistic safety analysis methods. These new analyses show that arbitrary Weibull fitting parameters and typical empirical hazard function techniques cannot be used to predict the dynamics of human errors and outcomes in the presence of learning. Comparisons of these new insights show agreement with human error data from the world's commercial airlines, the two shuttle failures, and from nuclear plant operator actions and transient control behavior observed in transients in both plants and simulators. The results demonstrate that the human error probability (HEP) is dynamic, and that it may be predicted using the learning hypothesis and the minimum failure rate, and can be utilized for probabilistic risk analysis purposes. (authors)
Estimating IMU heading error from SAR images.
Doerry, Armin Walter
2009-03-01T23:59:59.000Z
Angular orientation errors of the real antenna for Synthetic Aperture Radar (SAR) will manifest as undesired illumination gradients in SAR images. These gradients can be measured, and the pointing error can be calculated. This can be done for single images, but done more robustly using multi-image methods. Several methods are provided in this report. The pointing error can then be fed back to the navigation Kalman filter to correct for problematic heading (yaw) error drift. This can mitigate the need for uncomfortable and undesired IMU alignment maneuvers such as S-turns.
Flux recovery and a posteriori error estimators
2010-05-20T23:59:59.000Z
bility and the local efficiency bounds for this estimator are established provided that the ... For simple model problems, the energy norm of the true error is equal.
Wind Power Forecasting Error Distributions over Multiple Timescales (Presentation)
Hodge, B. M.; Milligan, M.
2011-07-01T23:59:59.000Z
This presentation presents some statistical analysis of wind power forecast errors and error distributions, with examples using ERCOT data.
Design error diagnosis and correction in digital circuits
Nayak, Debashis
1998-01-01T23:59:59.000Z
, each primary output would impose a con- straint on the on-set and off-set. These constraints should be combined together to derive the final on-set and off-set of the new function. Proposition 2: [9, 18, 17] Let i be the index of the primary outputs... to this equation are deleted. The work in [17] is also based on Boolean comparisons and applies to multiple errors. Overall, their method does not guarantee a solution. Test-vector simulation methods proposed for the DEDC problem include [20, 22, 26]. In [20...
Quantum Error Correction Beyond Completely Positive Maps
A. Shabani; D. A. Lidar
2009-10-21T23:59:59.000Z
By introducing an operator sum representation for arbitrary linear maps, we develop a generalized theory of quantum error correction (QEC) that applies to any linear map, in particular maps that are not completely positive (CP). This theory of "linear quantum error correction" is applicable in cases where the standard and restrictive assumption of a factorized initial system-bath state does not apply.
Meeting 12 February 25, 1999 Error Measure
California at Berkeley, University of
. The value of â?? is the corresponding eigenvalue. The eigenÂ values are the roots of the characteristic distances is nonÂnegative, so Q is posÂ itive semiÂdefinite. The error of an edge contraction is obtained paraboloid as illustrated in Figure 3. In other words, the preimage of a constant error value ffl, E \\Gamma1
Prediction Error and Event Boundaries 1 Running Head: PREDICTION ERROR AND EVENT BOUNDARIES
Zacks, Jeffrey M.
Prediction Error and Event Boundaries 1 Running Head: PREDICTION ERROR AND EVENT BOUNDARIES A computational model of event segmentation from perceptual prediction. Jeremy R. Reynolds, Jeffrey M. Zacks, and Todd S. Braver Washington University Manuscript #12;Prediction Error and Event Boundaries 2 People tend
Scalable extraction of error models from the output of error detection circuits
Austin G. Fowler; D. Sank; J. Kelly; R. Barends; John M. Martinis
2014-05-06T23:59:59.000Z
Accurate methods of assessing the performance of quantum gates are extremely important. Quantum process tomography and randomized benchmarking are the current favored methods. Quantum process tomography gives detailed information, but significant approximations must be made to reduce this information to a form quantum error correction simulations can use. Randomized benchmarking typically outputs just a single number, the fidelity, giving no information on the structure of errors during the gate. Neither method is optimized to assess gate performance within an error detection circuit, where gates will be actually used in a large-scale quantum computer. Specifically, the important issues of error composition and error propagation lie outside the scope of both methods. We present a fast, simple, and scalable method of obtaining exactly the information required to perform effective quantum error correction from the output of continuously running error detection circuits, enabling accurate prediction of large-scale behavior.
Error analysis of the chirp-z transform when implemented using waveform synthesizers and FFTs
Bielek, T.P.
1990-11-01T23:59:59.000Z
This report analyzes the effects of finite-precision arithmetic on discrete Fourier transforms (DFTs) calculated using the chirp-z transform algorithm. An introduction to the chirp-z transform is given together with a description of how the chirp-z transform is implemented in hardware. Equations for the effects of chirp rate errors, starting frequency errors, and starting phase errors on the frequency spectrum of the chirp-z transform are derived. Finally, the maximum possible errors in the chirp rate, the starting frequencies, and starting phases are calculated and used to compute the worst case effects on the amplitude and phase spectrums of the chirp-z transform. 1 ref., 6 figs.
Refined Error Estimates for the Riccati Equation with Applications to the Angular Teukolsky Equation
Felix Finster; Joel Smoller
2015-02-17T23:59:59.000Z
We derive refined rigorous error estimates for approximate solutions of Sturm-Liouville and Riccati equations with real or complex potentials. The approximate solutions include WKB approximations, Airy and parabolic cylinder functions, and certain Bessel functions. Our estimates are applied to solutions of the angular Teukolsky equation with a complex aspherical parameter in a rotating black hole Kerr geometry.
Quantum Error Correction for Quantum Memories
Barbara M. Terhal
2015-01-20T23:59:59.000Z
Active quantum error correction using qubit stabilizer codes has emerged as a promising, but experimentally challenging, engineering program for building a universal quantum computer. In this review we consider the formalism of qubit stabilizer and subsystem stabilizer codes and their possible use in protecting quantum information in a quantum memory. We review the theory of fault-tolerance and quantum error-correction, discuss examples of various codes and code constructions, the general quantum error correction conditions, the noise threshold, the special role played by Clifford gates and the route towards fault-tolerant universal quantum computation. The second part of the review is focused on providing an overview of quantum error correction using two-dimensional (topological) codes, in particular the surface code architecture. We discuss the complexity of decoding and the notion of passive or self-correcting quantum memories. The review does not focus on a particular technology but discusses topics that will be relevant for various quantum technologies.
Facemail : preventing common errors when composing email
Lieberman, Eric (Eric W.)
2006-01-01T23:59:59.000Z
Facemail is a system designed to investigate and prevent common errors that users make while composing emails. Users often accidentally send email to incorrect recipients by mistyping an email address, accidentally clicking ...
Organizational Errors: Directions for Future Research
Carroll, John Stephen
The goal of this chapter is to promote research about organizational errors—i.e., the actions of multiple organizational participants that deviate from organizationally specified rules and can potentially result in adverse ...
Quantum error-correcting codes and devices
Gottesman, Daniel (Los Alamos, NM)
2000-10-03T23:59:59.000Z
A method of forming quantum error-correcting codes by first forming a stabilizer for a Hilbert space. A quantum information processing device can be formed to implement such quantum codes.
Evaluating operating system vulnerability to memory errors.
Ferreira, Kurt Brian; Bridges, Patrick G. (University of New Mexico); Pedretti, Kevin Thomas Tauke; Mueller, Frank (North Carolina State University); Fiala, David (North Carolina State University); Brightwell, Ronald Brian
2012-05-01T23:59:59.000Z
Reliability is of great concern to the scalability of extreme-scale systems. Of particular concern are soft errors in main memory, which are a leading cause of failures on current systems and are predicted to be the leading cause on future systems. While great effort has gone into designing algorithms and applications that can continue to make progress in the presence of these errors without restarting, the most critical software running on a node, the operating system (OS), is currently left relatively unprotected. OS resiliency is of particular importance because, though this software typically represents a small footprint of a compute node's physical memory, recent studies show more memory errors in this region of memory than the remainder of the system. In this paper, we investigate the soft error vulnerability of two operating systems used in current and future high-performance computing systems: Kitten, the lightweight kernel developed at Sandia National Laboratories, and CLE, a high-performance Linux-based operating system developed by Cray. For each of these platforms, we outline major structures and subsystems that are vulnerable to soft errors and describe methods that could be used to reconstruct damaged state. Our results show the Kitten lightweight operating system may be an easier target to harden against memory errors due to its smaller memory footprint, largely deterministic state, and simpler system structure.
The Error-Pattern-Correcting Turbo Equalizer
Alhussien, Hakim
2010-01-01T23:59:59.000Z
The error-pattern correcting code (EPCC) is incorporated in the design of a turbo equalizer (TE) with aim to correct dominant error events of the inter-symbol interference (ISI) channel at the output of its matching Viterbi detector. By targeting the low Hamming-weight interleaved errors of the outer convolutional code, which are responsible for low Euclidean-weight errors in the Viterbi trellis, the turbo equalizer with an error-pattern correcting code (TE-EPCC) exhibits a much lower bit-error rate (BER) floor compared to the conventional non-precoded TE, especially for high rate applications. A maximum-likelihood upper bound is developed on the BER floor of the TE-EPCC for a generalized two-tap ISI channel, in order to study TE-EPCC's signal-to-noise ratio (SNR) gain for various channel conditions and design parameters. In addition, the SNR gain of the TE-EPCC relative to an existing precoded TE is compared to demonstrate the present TE's superiority for short interleaver lengths and high coding rates.
A systems approach to reducing utility billing errors
Ogura, Nori
2013-01-01T23:59:59.000Z
Many methods for analyzing the possibility of errors are practiced by organizations who are concerned about safety and error prevention. However, in situations where the error occurrence is random and difficult to track, ...
Error Detection and Recovery for Robot Motion Planning with Uncertainty
Donald, Bruce Randall
1987-07-01T23:59:59.000Z
Robots must plan and execute tasks in the presence of uncertainty. Uncertainty arises from sensing errors, control errors, and uncertainty in the geometry of the environment. The last, which is called model error, has ...
Running jobs error: "inet_arp_address_lookup"
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
jobs error: "inetarpaddresslookup" Resolved: Running jobs error: "inetarpaddresslookup" September 22, 2013 by Helen He (0 Comments) Symptom: After the Hopper August 14...
T-598: Apache Tomcat HTTP BIO Connector Error Discloses Information...
Broader source: Energy.gov (indexed) [DOE]
598: Apache Tomcat HTTP BIO Connector Error Discloses Information From Different Requests to Remote Users T-598: Apache Tomcat HTTP BIO Connector Error Discloses Information From...
V-235: Cisco Mobility Services Engine Configuration Error Lets...
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
5: Cisco Mobility Services Engine Configuration Error Lets Remote Users Login Anonymously V-235: Cisco Mobility Services Engine Configuration Error Lets Remote Users Login...
Model error in weather forecasting D. Orrell 1,2 , L. Smith 1,3 , J. Barkmeijer 4 , and T. Palmer 4
Smith, Leonard A
numerical weather prediction modÂ els. A simple law is derived to relate model error to likely shadowingModel error in weather forecasting D. Orrell 1,2 , L. Smith 1,3 , J. Barkmeijer 4 , and T. Palmer 4 in the model, and inacÂ curate initial conditions (Bjerknes, 1911). Because weather models are thought
Uncertainty and error in computational simulations
Oberkampf, W.L.; Diegert, K.V.; Alvin, K.F.; Rutherford, B.M.
1997-10-01T23:59:59.000Z
The present paper addresses the question: ``What are the general classes of uncertainty and error sources in complex, computational simulations?`` This is the first step of a two step process to develop a general methodology for quantitatively estimating the global modeling and simulation uncertainty in computational modeling and simulation. The second step is to develop a general mathematical procedure for representing, combining and propagating all of the individual sources through the simulation. The authors develop a comprehensive view of the general phases of modeling and simulation. The phases proposed are: conceptual modeling of the physical system, mathematical modeling of the system, discretization of the mathematical model, computer programming of the discrete model, numerical solution of the model, and interpretation of the results. This new view is built upon combining phases recognized in the disciplines of operations research and numerical solution methods for partial differential equations. The characteristics and activities of each of these phases is discussed in general, but examples are given for the fields of computational fluid dynamics and heat transfer. They argue that a clear distinction should be made between uncertainty and error that can arise in each of these phases. The present definitions for uncertainty and error are inadequate and. therefore, they propose comprehensive definitions for these terms. Specific classes of uncertainty and error sources are then defined that can occur in each phase of modeling and simulation. The numerical sources of error considered apply regardless of whether the discretization procedure is based on finite elements, finite volumes, or finite differences. To better explain the broad types of sources of uncertainty and error, and the utility of their categorization, they discuss a coupled-physics example simulation.
Bounds for Small-Error and Zero-Error Quantum Algorithms Harry Buhrman
de Wolf, Ronald
Bounds for Small-Error and Zero-Error Quantum Algorithms Harry Buhrman CWI Richard Cleve University algorithm with an auxiliary input r, which is uniformly distributed over some underlying sample space. In this case, for any x 2 f0;1gn, f(x) = 1 iff (9r 2 S)(A(x;r) = 1). Grover's quantum search algorithm [15
Retiming for Soft Error Minimization Under Error-Latching Window Constraints
Zhou, Hai
sensitivity to naturally- occurring radiation and the consequent soft error rates of CMOS circuits. Moreover Soft error, also known as single-event upsets (SEU), caused by radiation-induced charged particles circuits [3]: electrical masking occurs when SEUs are attenuated before being latched because
Laser Phase Errors in Seeded FELs
Ratner, D.; Fry, A.; Stupakov, G.; White, W.; /SLAC
2012-03-28T23:59:59.000Z
Harmonic seeding of free electron lasers has attracted significant attention from the promise of transform-limited pulses in the soft X-ray region. Harmonic multiplication schemes extend seeding to shorter wavelengths, but also amplify the spectral phase errors of the initial seed laser, and may degrade the pulse quality. In this paper we consider the effect of seed laser phase errors in high gain harmonic generation and echo-enabled harmonic generation. We use simulations to confirm analytical results for the case of linearly chirped seed lasers, and extend the results for arbitrary seed laser envelope and phase.
MEASUREMENT AND CORRECTION OF ULTRASONIC ANEMOMETER ERRORS
Heinemann, Detlev
commonly show systematic errors depending on wind speed due to inaccurate ultrasonic transducer mounting three- dimensional wind speed time series. Results for the variance and power spectra are shown. 1 wind speeds with ultrasonic anemometers: The measu- red flow is distorted by the probe head
Hierarchical Classification of Documents with Error Control
King, Kuo Chin Irwin
Hierarchical Classification of Documents with Error Control Chun-hung Cheng1 , Jian Tang2 , Ada Wai is a function that matches a new object with one of the predefined classes. Document classification is characterized by the large number of attributes involved in the objects (documents). The traditional method
Hierarchical Classification of Documents with Error Control
Fu, Ada Waichee
Hierarchical Classification of Documents with Error Control Chunhung Cheng 1 , Jian Tang 2 , Ada. Classification is a function that matches a new object with one of the predefined classes. Document classification is characterized by the large number of attributes involved in the objects (documents
Adjoint Error Correction for Integral Outputs
Pierce, Niles A.
a combustor; the total heat ux into a high pressure turbine blade from the surrounding ow; average noise. As an example, consider the wake behind a wing. To adequately resolve the wake requires a #12;ne grid locally in which the grid resolution is rather coarse. Grid adaptation based on error estimates that look
Verifying Volume Rendering Using Discretization Error Analysis
Kirby, Mike
Verifying Volume Rendering Using Discretization Error Analysis Tiago Etiene, Daniel Jo¨nsson, Timo--We propose an approach for verification of volume rendering correctness based on an analysis of the volume rendering integral, the basis of most DVR algorithms. With respect to the most common discretization
Distribution of Wind Power Forecasting Errors from Operational Systems (Presentation)
Hodge, B. M.; Ela, E.; Milligan, M.
2011-10-01T23:59:59.000Z
This presentation offers new data and statistical analysis of wind power forecasting errors in operational systems.
Statistical Analysis of CCD Data: Error Analysis/Noise Theorem
Peletier, Reynier
Statistical Analysis of CCD Data: Error Analysis/Noise Theorem Why Statistical Approach? Systematic Errors Random Errors (= Statistical Errors) Accuracy and Precision Best Estimator: Mean, Median Distribution Statistical CCD Data Analysis #12;Why do we need statistical analysis? (= Why do we need to worry
Error and jitter effect studies on the SLED for BEPCII-linac
Shi-Lun, Pei; Ou-Zheng, Xiao
2011-01-01T23:59:59.000Z
RF pulse compressor is a device to convert a long RF pulse to a short one with much higher peak RF magnitude. SLED can be regarded as the earliest RF pulse compressor used in large scale linear accelerators. It is widely studied around the world and applied in the BEPC and BEPCII linac for many years. During the routine operation, the error and jitter effects will deteriorate the SLED performance either on the output electromagnetic wave amplitude or phase. The error effects mainly include the frequency drift induced by cooling water temperature variation and the frequency/Q0/{\\beta} unbalances between the two energy storage cavities caused by mechanical fabrication or microwave tuning. The jitter effects refer to the PSK switching phase and time jitters. In this paper, we re-derived the generalized formulae for the conventional SLED used in the BEPCII linac. At last, the error and jitter effects on the SLED performance are investigated.
Influence of sky radiance measurement errors on inversion-retrieved aerosol properties
Torres, B.; Toledano, C.; Cachorro, V. E.; Bennouna, Y. S.; Fuertes, D.; Gonzalez, R.; Frutos, A. M. de [Atmospheric Optics Group (GOA), University of Valladolid, Valladolid (Spain); Berjon, A. J. [Izana Atmospheric Research Center, Meteorological State Agency of Spain (AEMET), Sta. Cruz de Tenerife (Spain); Dubovik, O.; Goloub, P.; Podvin, T.; Blarel, L. [Laboratory of Atmospheric Optics, Universite Lille 1, Villeneuve d'Ascq (France)
2013-05-10T23:59:59.000Z
Remote sensing of the atmospheric aerosol is a well-established technique that is currently used for routine monitoring of this atmospheric component, both from ground-based and satellite. The AERONET program, initiated in the 90's, is the most extended network and the data provided are currently used by a wide community of users for aerosol characterization, satellite and model validation and synergetic use with other instrumentation (lidar, in-situ, etc.). Aerosol properties are derived within the network from measurements made by ground-based Sun-sky scanning radiometers. Sky radiances are acquired in two geometries: almucantar and principal plane. Discrepancies in the products obtained following both geometries have been observed and the main aim of this work is to determine if they could be justified by measurement errors. Three systematic errors have been analyzed in order to quantify the effects on the inversion-derived aerosol properties: calibration, pointing accuracy and finite field of view. Simulations have shown that typical uncertainty in the analyzed quantities (5% in calibration, 0.2 Degree-Sign in pointing and 1.2 Degree-Sign field of view) yields to errors in the retrieved parameters that vary depending on the aerosol type and geometry. While calibration and pointing errors have relevant impact on the products, the finite field of view does not produce notable differences.
Systematic Errors in measurement of b1
Wood, S A
2014-10-27T23:59:59.000Z
A class of spin observables can be obtained from the relative difference of or asymmetry between cross sections of different spin states of beam or target particles. Such observables have the advantage that the normalization factors needed to calculate absolute cross sections from yields often divide out or cancel to a large degree in constructing asymmetries. However, normalization factors can change with time, giving different normalization factors for different target or beam spin states, leading to systematic errors in asymmetries in addition to those determined from statistics. Rapidly flipping spin orientation, such as what is routinely done with polarized beams, can significantly reduce the impact of these normalization fluctuations and drifts. Target spin orientations typically require minutes to hours to change, versus fractions of a second for beams, making systematic errors for observables based on target spin flips more difficult to control. Such systematic errors from normalization drifts are discussed in the context of the proposed measurement of the deuteron b(1) structure function at Jefferson Lab.
Tradeoff between energy and error in the discrimination of quantum-optical devices
Alessandro Bisio; Michele Dall'Arno; Giacomo Mauro D'Ariano
2011-07-11T23:59:59.000Z
We address the problem of energy-error tradeoff in the discrimination between two linear passive quantum optical devices with a single use. We provide an analytical derivation of the optimal strategy for beamsplitters and an iterative algorithm converging to the optimum in the general case. We then compare the optimal strategy with a simpler strategy using coherent input states and homodyne detection. It turns out that the former requires much less energy in order to achieve the same performances.
Correlated errors can lead to better performance of quantum codes
A. Shabani
2008-03-06T23:59:59.000Z
A formulation for evaluating the performance of quantum error correcting codes for a general error model is presented. In this formulation, the correlation between errors is quantified by a Hamiltonian description of the noise process. We classify correlated errors using the system-bath interaction: local versus nonlocal and two-body versus many-body interactions. In particular, we consider Calderbank-Shor-Steane codes and observe a better performance in the presence of correlated errors depending on the timing of the error recovery. We also find this timing to be an important factor in the design of a coding system for achieving higher fidelities.
Efficient Error Calculation for Multiresolution Texture-Based Volume Visualization
LaMar, E; Hamann, B; Joy, K I
2001-10-16T23:59:59.000Z
Multiresolution texture-based volume visualization is an excellent technique to enable interactive rendering of massive data sets. Interactive manipulation of a transfer function is necessary for proper exploration of a data set. However, multiresolution techniques require assessing the accuracy of the resulting images, and re-computing the error after each change in a transfer function is very expensive. They extend their existing multiresolution volume visualization method by introducing a method for accelerating error calculations for multiresolution volume approximations. Computing the error for an approximation requires adding individual error terms. One error value must be computed once for each original voxel and its corresponding approximating voxel. For byte data, i.e., data sets where integer function values between 0 and 255 are given, they observe that the set of error pairs can be quite large, yet the set of unique error pairs is small. instead of evaluating the error function for each original voxel, they construct a table of the unique combinations and the number of their occurrences. To evaluate the error, they add the products of the error function for each unique error pair and the frequency of each error pair. This approach dramatically reduces the amount of computation time involved and allows them to re-compute the error associated with a new transfer function quickly.
Human error contribution to nuclear materials-handling events
Sutton, Bradley (Bradley Jordan)
2007-01-01T23:59:59.000Z
This thesis analyzes a sample of 15 fuel-handling events from the past ten years at commercial nuclear reactors with significant human error contributions in order to detail the contribution of human error to fuel-handling ...
Quantum Error Correcting Subsystem Codes From Two Classical Linear Codes
Dave Bacon; Andrea Casaccino
2006-10-17T23:59:59.000Z
The essential insight of quantum error correction was that quantum information can be protected by suitably encoding this quantum information across multiple independently erred quantum systems. Recently it was realized that, since the most general method for encoding quantum information is to encode it into a subsystem, there exists a novel form of quantum error correction beyond the traditional quantum error correcting subspace codes. These new quantum error correcting subsystem codes differ from subspace codes in that their quantum correcting routines can be considerably simpler than related subspace codes. Here we present a class of quantum error correcting subsystem codes constructed from two classical linear codes. These codes are the subsystem versions of the quantum error correcting subspace codes which are generalizations of Shor's original quantum error correcting subspace codes. For every Shor-type code, the codes we present give a considerable savings in the number of stabilizer measurements needed in their error recovery routines.
Reply To "Comment on 'Quantum Convolutional Error-Correcting Codes' "
H. F. Chau
2005-06-02T23:59:59.000Z
In their comment, de Almedia and Palazzo \\cite{comment} discovered an error in my earlier paper concerning the construction of quantum convolutional codes (quant-ph/9712029). This error can be repaired by modifying the method of code construction.
Hardware-efficient autonomous quantum error correction
Zaki Leghtas; Gerhard Kirchmair; Brian Vlastakis; Robert Schoelkopf; Michel Devoret; Mazyar Mirrahimi
2013-01-16T23:59:59.000Z
We propose a new method to autonomously correct for errors of a logical qubit induced by energy relaxation. This scheme encodes the logical qubit as a multi-component superposition of coherent states in a harmonic oscillator, more specifically a cavity mode. The sequences of encoding, decoding and correction operations employ the non-linearity provided by a single physical qubit coupled to the cavity. We layout in detail how to implement these operations in a practical system. This proposal directly addresses the task of building a hardware-efficient and technically realizable quantum memory.
Error Analysis in Nuclear Density Functional Theory
Nicolas Schunck; Jordan D. McDonnell; Jason Sarich; Stefan M. Wild; Dave Higdon
2014-07-11T23:59:59.000Z
Nuclear density functional theory (DFT) is the only microscopic, global approach to the structure of atomic nuclei. It is used in numerous applications, from determining the limits of stability to gaining a deep understanding of the formation of elements in the universe or the mechanisms that power stars and reactors. The predictive power of the theory depends on the amount of physics embedded in the energy density functional as well as on efficient ways to determine a small number of free parameters and solve the DFT equations. In this article, we discuss the various sources of uncertainties and errors encountered in DFT and possible methods to quantify these uncertainties in a rigorous manner.
Edison Trouble Shooting and Error Messages
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
AFDC Printable Version Share this resource Send a link to EERE: Alternative Fuels Data Center Home Page to someone by E-mail Share EERE: Alternative Fuels Data Center Home Page on Facebook Tweet about EERE: Alternative Fuels Data Center Home Page on Twitter Bookmark EERE:1 First Use of Energy for All Purposes (Fuel and Nonfuel),Feet) Year Jan Feb Mar Apr May JunDatastreamsmmcrcalgovInstrumentsruc DocumentationP-Series toESnet4: Networking for the Future ofTrouble Shooting and Error Messages
Clustered Error Correction of Codeword-Stabilized Quantum Codes
Yunfan Li; Ilya Dumer; Leonid P. Pryadko
2010-03-08T23:59:59.000Z
Codeword stabilized (CWS) codes are a general class of quantum codes that includes stabilizer codes and many families of non-additive codes with good parameters. For such a non-additive code correcting all t-qubit errors, we propose an algorithm that employs a single measurement to test all errors located on a given set of t qubits. Compared with exhaustive error screening, this reduces the total number of measurements required for error recovery by a factor of about 3^t.
Standard errors of parameter estimates in the ETAS model
Schoenberg, Frederic Paik (Rick)
1 Standard errors of parameter estimates in the ETAS model Abstract Point process models of seismic catalogs and in short- term earthquake forecasting. The standard errors of parameter estimates of conventional standard error estimates based on the Hessian matrix of the log- likelihood function of the ETAS
Instruction Guide Paying Vendors: Checking Vouchers for Errors
Watson, Craig A.
Instruction Guide Paying Vendors: Checking Vouchers for Errors and Match Exceptions Updated the Pay Terms on the voucher Â· Changing the Vendor or Vendor address Â· Vouchering a PO that have receipts link or the Error Summary tab. #12;Instruction Guide Paying Vendors: Checking Vouchers for Errors
Lin, Shaowei
2014-07-02T23:59:59.000Z
The enactment of derivative action was expected to be actively used by shareholders to protect their interests. In fact, it turned out that this reform effort seemed futile as the right to engage in such actions was ...
Proulx, Paul
1985-01-01T23:59:59.000Z
of root. Different finals distinguish transitive from intransitive verbs. Some are abstract, some concrete. Elements are generally joined by one of two link vowels, e and o. Yurok nouns and particles are simpler in structure, but a few derivational finals...
Liu, Jun
2003-09-25T23:59:59.000Z
This paper studies the optimal investment strategy of an investor who can access not only the bond and the stock markets, but also the derivatives market. We consider the investment situation where, in addition to the usual ...
Karpovsky, Mark
On-line Self Error Detection with Equal Protection Against All Errors Mark G. Karpovsky, Konrad J and storage. We also present several design techniques for memories with self-error-detection based on the pro. The proposed robust codes require slightly larger overhead than standard and widely-used linear codes
E791 DATA ACQUISITION SYSTEM Error reports received ; no new errors reported
Fermi National Accelerator Laboratory
of events written to tape. 18 #12; Error and Status Displays Mailbox For Histogram Requests VaxÂonline Event Display VAX 11 / 780 Event Reconstruction Event Display Detector Monitoring 3 VAX Workstations 42 EXABYTE of the entire E791 DA system. The VAX 11/780 was the user interface to the VME part of the system, via the DA
Analysis of Errors in a Special Perturbations Satellite Orbit Propagator
Beckerman, M.; Jones, J.P.
1999-02-01T23:59:59.000Z
We performed an analysis of error densities for the Special Perturbations orbit propagator using data for 29 satellites in orbits of interest to Space Shuttle and International Space Station collision avoidance. We find that the along-track errors predominate. These errors increase monotonically over each 36-hour prediction interval. The predicted positions in the along-track direction progressively either leap ahead of or lag behind the actual positions. Unlike the along-track errors the radial and cross-track errors oscillate about their nearly zero mean values. As the number of observations per fit interval decline the along-track prediction errors, and amplitudes of the radial and cross-track errors, increase.
Evaluating and Minimizing Distributed Cavity Phase Errors in Atomic Clocks
Li, Ruoxin
2010-01-01T23:59:59.000Z
We perform 3D finite element calculations of the fields in microwave cavities and analyze the distributed cavity phase errors of atomic clocks that they produce. The fields of cylindrical cavities are treated as an azimuthal Fourier series. Each of the lowest components produces clock errors with unique characteristics that must be assessed to establish a clock's accuracy. We describe the errors and how to evaluate them. We prove that sharp structures in the cavity do not produce large frequency errors, even at moderately high powers, provided the atomic density varies slowly. We model the amplitude and phase imbalances of the feeds. For larger couplings, these can lead to increased phase errors. We show that phase imbalances produce a novel distributed cavity phase error that depends on the cavity detuning. We also design improved cavities by optimizing the geometry and tuning the mode spectrum so that there are negligible phase variations, allowing this source of systematic error to be dramatically reduced.
Evaluating and Minimizing Distributed Cavity Phase Errors in Atomic Clocks
Ruoxin Li; Kurt Gibble
2010-08-09T23:59:59.000Z
We perform 3D finite element calculations of the fields in microwave cavities and analyze the distributed cavity phase errors of atomic clocks that they produce. The fields of cylindrical cavities are treated as an azimuthal Fourier series. Each of the lowest components produces clock errors with unique characteristics that must be assessed to establish a clock's accuracy. We describe the errors and how to evaluate them. We prove that sharp structures in the cavity do not produce large frequency errors, even at moderately high powers, provided the atomic density varies slowly. We model the amplitude and phase imbalances of the feeds. For larger couplings, these can lead to increased phase errors. We show that phase imbalances produce a novel distributed cavity phase error that depends on the cavity detuning. We also design improved cavities by optimizing the geometry and tuning the mode spectrum so that there are negligible phase variations, allowing this source of systematic error to be dramatically reduced.
Plasma dynamics and a significant error of macroscopic averaging
Marek A. Szalek
2005-05-22T23:59:59.000Z
The methods of macroscopic averaging used to derive the macroscopic Maxwell equations from electron theory are methodologically incorrect and lead in some cases to a substantial error. For instance, these methods do not take into account the existence of a macroscopic electromagnetic field EB, HB generated by carriers of electric charge moving in a thin layer adjacent to the boundary of the physical region containing these carriers. If this boundary is impenetrable for charged particles, then in its immediate vicinity all carriers are accelerated towards the inside of the region. The existence of the privileged direction of acceleration results in the generation of the macroscopic field EB, HB. The contributions to this field from individual accelerated particles are described with a sufficient accuracy by the Lienard-Wiechert formulas. In some cases the intensity of the field EB, HB is significant not only for deuteron plasma prepared for a controlled thermonuclear fusion reaction but also for electron plasma in conductors at room temperatures. The corrected procedures of macroscopic averaging will induce some changes in the present form of plasma dynamics equations. The modified equations will help to design improved systems of plasma confinement.
Error analysis of nuclear forces and effective interactions
R. Navarro Perez; J. E. Amaro; E. Ruiz Arriola
2014-09-04T23:59:59.000Z
The Nucleon-Nucleon interaction is the starting point for ab initio Nuclear Structure and Nuclear reactions calculations. Those are effectively carried out via effective interactions fitting scattering data up to a maximal center of mass momentum. However, NN interactions are subjected to statistical and systematic uncertainties which are expected to propagate and have some impact on the predictive power and accuracy of theoretical calculations, regardless on the numerical accuracy of the method used to solve the many body problem. We stress the necessary conditions required for a correct and self-consistent statistical interpretation of the discrepancies between theory and experiment which enable a subsequent statistical error propagation and correlation analysis. We comprehensively discuss an stringent and recently proposed tail-sensitive normality test and provide a simple recipe to implement it. As an application, we analyze the deduced uncertainties and correlations of effective interactions in terms of Moshinsky-Skyrme parameters and effective field theory counterterms as derived from the bare NN potential containing One-Pion-Exchange and Chiral Two-Pion-Exchange interactions inferred from scattering data.
Pressure Change Measurement Leak Testing Errors
Pryor, Jeff M [ORNL] [ORNL; Walker, William C [ORNL] [ORNL
2014-01-01T23:59:59.000Z
A pressure change test is a common leak testing method used in construction and Non-Destructive Examination (NDE). The test is known as being a fast, simple, and easy to apply evaluation method. While this method may be fairly quick to conduct and require simple instrumentation, the engineering behind this type of test is more complex than is apparent on the surface. This paper intends to discuss some of the more common errors made during the application of a pressure change test and give the test engineer insight into how to correctly compensate for these factors. The principals discussed here apply to ideal gases such as air or other monoatomic or diatomic gasses; however these same principals can be applied to polyatomic gasses or liquid flow rate with altered formula specific to those types of tests using the same methodology.
Graphical Quantum Error-Correcting Codes
Sixia Yu; Qing Chen; C. H. Oh
2007-09-12T23:59:59.000Z
We introduce a purely graph-theoretical object, namely the coding clique, to construct quantum errorcorrecting codes. Almost all quantum codes constructed so far are stabilizer (additive) codes and the construction of nonadditive codes, which are potentially more efficient, is not as well understood as that of stabilizer codes. Our graphical approach provides a unified and classical way to construct both stabilizer and nonadditive codes. In particular we have explicitly constructed the optimal ((10,24,3)) code and a family of 1-error detecting nonadditive codes with the highest encoding rate so far. In the case of stabilizer codes a thorough search becomes tangible and we have classified all the extremal stabilizer codes up to 8 qubits.
Huang, Weidong
2011-01-01T23:59:59.000Z
Surface slope error of concentrator is one of the main factors to influence the performance of the solar concentrated collectors which cause deviation of reflected ray and reduce the intercepted radiation. This paper presents the general equation to calculate the standard deviation of reflected ray error from slope error through geometry optics, applying the equation to calculate the standard deviation of reflected ray error for 5 kinds of solar concentrated reflector, provide typical results. The results indicate that the slope error is transferred to the reflected ray in more than 2 folds when the incidence angle is more than 0. The equation for reflected ray error is generally fit for all reflection surfaces, and can also be applied to control the error in designing an abaxial optical system.
Hess-Flores, M
2011-11-10T23:59:59.000Z
Scene reconstruction from video sequences has become a prominent computer vision research area in recent years, due to its large number of applications in fields such as security, robotics and virtual reality. Despite recent progress in this field, there are still a number of issues that manifest as incomplete, incorrect or computationally-expensive reconstructions. The engine behind achieving reconstruction is the matching of features between images, where common conditions such as occlusions, lighting changes and texture-less regions can all affect matching accuracy. Subsequent processes that rely on matching accuracy, such as camera parameter estimation, structure computation and non-linear parameter optimization, are also vulnerable to additional sources of error, such as degeneracies and mathematical instability. Detection and correction of errors, along with robustness in parameter solvers, are a must in order to achieve a very accurate final scene reconstruction. However, error detection is in general difficult due to the lack of ground-truth information about the given scene, such as the absolute position of scene points or GPS/IMU coordinates for the camera(s) viewing the scene. In this dissertation, methods are presented for the detection, factorization and correction of error sources present in all stages of a scene reconstruction pipeline from video, in the absence of ground-truth knowledge. Two main applications are discussed. The first set of algorithms derive total structural error measurements after an initial scene structure computation and factorize errors into those related to the underlying feature matching process and those related to camera parameter estimation. A brute-force local correction of inaccurate feature matches is presented, as well as an improved conditioning scheme for non-linear parameter optimization which applies weights on input parameters in proportion to estimated camera parameter errors. Another application is in reconstruction pre-processing, where an algorithm detects and discards frames that would lead to inaccurate feature matching, camera pose estimation degeneracies or mathematical instability in structure computation based on a residual error comparison between two different match motion models. The presented algorithms were designed for aerial video but have been proven to work across different scene types and camera motions, and for both real and synthetic scenes.
Quantum root-mean-square error and measurement uncertainty relations
Paul Busch; Pekka Lahti; Reinhard F Werner
2014-10-10T23:59:59.000Z
Recent years have witnessed a controversy over Heisenberg's famous error-disturbance relation. Here we resolve the conflict by way of an analysis of the possible conceptualizations of measurement error and disturbance in quantum mechanics. We discuss two approaches to adapting the classic notion of root-mean-square error to quantum measurements. One is based on the concept of noise operator; its natural operational content is that of a mean deviation of the values of two observables measured jointly, and thus its applicability is limited to cases where such joint measurements are available. The second error measure quantifies the differences between two probability distributions obtained in separate runs of measurements and is of unrestricted applicability. We show that there are no nontrivial unconditional joint-measurement bounds for {\\em state-dependent} errors in the conceptual framework discussed here, while Heisenberg-type measurement uncertainty relations for {\\em state-independent} errors have been proven.
P. A. Sturrock; J. D. Scargle
2006-06-20T23:59:59.000Z
The purpose of this article is to carry out a power-spectrum analysis (based on likelihood methods) of the Super-Kamiokande 5-day dataset that takes account of the asymmetry in the error estimates. Whereas the likelihood analysis involves a linear optimization procedure for symmetrical error estimates, it involves a nonlinear optimization procedure for asymmetrical error estimates. We find that for most frequencies there is little difference between the power spectra derived from analyses of symmetrized error estimates and from asymmetrical error estimates. However, this proves not to be the case for the principal peak in the power spectra, which is found at 9.43 yr-1. A likelihood analysis which allows for a "floating offset" and takes account of the start time and end time of each bin and of the flux estimate and the symmetrized error estimate leads to a power of 11.24 for this peak. A Monte Carlo analysis shows that there is a chance of only 1% of finding a peak this big or bigger in the frequency band 1 - 36 yr-1 (the widest band that avoids artificial peaks). On the other hand, an analysis that takes account of the error asymmetry leads to a peak with power 13.24 at that frequency. A Monte Carlo analysis shows that there is a chance of only 0.1% of finding a peak this big or bigger in that frequency band 1 - 36 yr-1. From this perspective, power spectrum analysis that takes account of asymmetry of the error estimates gives evidence for variability that is significant at the 99.9% level. We comment briefly on an apparent discrepancy between power spectrum analyses of the Super-Kamiokande and SNO solar neutrino experiments.
Parallel Worldline Numerics: Implementation and Error Analysis
Dan Mazur; Jeremy S. Heyl
2014-07-28T23:59:59.000Z
We give an overview of the worldline numerics technique, and discuss the parallel CUDA implementation of a worldline numerics algorithm. In the worldline numerics technique, we wish to generate an ensemble of representative closed-loop particle trajectories, and use these to compute an approximate average value for Wilson loops. We show how this can be done with a specific emphasis on cylindrically symmetric magnetic fields. The fine-grained, massive parallelism provided by the GPU architecture results in considerable speedup in computing Wilson loop averages. Furthermore, we give a brief overview of uncertainty analysis in the worldline numerics method. There are uncertainties from discretizing each loop, and from using a statistical ensemble of representative loops. The former can be minimized so that the latter dominates. However, determining the statistical uncertainties is complicated by two subtleties. Firstly, the distributions generated by the worldline ensembles are highly non-Gaussian, and so the standard error in the mean is not a good measure of the statistical uncertainty. Secondly, because the same ensemble of worldlines is used to compute the Wilson loops at different values of $T$ and $x_\\mathrm{ cm}$, the uncertainties associated with each computed value of the integrand are strongly correlated. We recommend a form of jackknife analysis which deals with both of these problems.
Homological Error Correction: Classical and Quantum Codes
H. Bombin; M. A. Martin-Delgado
2006-05-10T23:59:59.000Z
We prove several theorems characterizing the existence of homological error correction codes both classically and quantumly. Not every classical code is homological, but we find a family of classical homological codes saturating the Hamming bound. In the quantum case, we show that for non-orientable surfaces it is impossible to construct homological codes based on qudits of dimension $D>2$, while for orientable surfaces with boundaries it is possible to construct them for arbitrary dimension $D$. We give a method to obtain planar homological codes based on the construction of quantum codes on compact surfaces without boundaries. We show how the original Shor's 9-qubit code can be visualized as a homological quantum code. We study the problem of constructing quantum codes with optimal encoding rate. In the particular case of toric codes we construct an optimal family and give an explicit proof of its optimality. For homological quantum codes on surfaces of arbitrary genus we also construct a family of codes asymptotically attaining the maximum possible encoding rate. We provide the tools of homology group theory for graphs embedded on surfaces in a self-contained manner.
Temperature-dependent errors in nuclear lattice simulations
Dean Lee; Richard Thomson
2007-01-17T23:59:59.000Z
We study the temperature dependence of discretization errors in nuclear lattice simulations. We find that for systems with strong attractive interactions the predominant error arises from the breaking of Galilean invariance. We propose a local "well-tempered" lattice action which eliminates much of this error. The well-tempered action can be readily implemented in lattice simulations for nuclear systems as well as cold atomic Fermi systems.
Optimized Learning with Bounded Error for Feedforward Neural Networks
Maggiore, Manfredi
Optimized Learning with Bounded Error for Feedforward Neural Networks A. Alessandri, M. Sanguineti-based learnings. A. Alessandri is with the Naval Automatio
Nonlinear local error bounds via a change of metric
Dominique Azé
2014-10-23T23:59:59.000Z
Oct 23, 2014 ... Abstract: In this work, we improve the approach of Corvellec-Motreanu to nonlinear error bounds for lowersemicontinuous functions on ...
New Fractional Error Bounds for Polynomial Systems with ...
2014-07-27T23:59:59.000Z
Our major result extends the existing error bounds from the system involving only a ... linear complementarity systems with polynomial data as well as high-order ...
Pushing schedule derivation method
Henriquez, B. [Compania Siderurgica Huachipato S.A., Talcahuano (Chile)
1996-12-31T23:59:59.000Z
The development of a Pushing Schedule Derivation Method has allowed the company to sustain the maximum production rate at CSH`s Coke Oven Battery, in spite of having single set oven machinery with a high failure index as well as a heat top tendency. The stated method provides for scheduled downtime of up to two hours for machinery maintenance purposes, periods of empty ovens for decarbonization and production loss recovery capability, while observing lower limits and uniformity of coking time.
Joseph M. Renes; Volkher B. Scholz
2014-02-26T23:59:59.000Z
We derive new Heisenberg-type uncertainty relations for both joint measurability and the error-disturbance tradeoff for arbitrary observables of finite-dimensional systems. The relations are formulated in terms of a directly operational quantity, namely the probability of distinguishing the actual operation of a device from its hypothetical ideal, by any possible testing procedure whatsoever. Moreover, they may be directly applied in information processing settings, for example to infer that devices which can faithfully transmit information regarding one observable do not leak any information about conjugate observables to the environment. Though intuitively apparent from Heisenberg's original arguments, only more limited versions of this statement have previously been formalized.
Luckas, W.J. Jr.; Lettieri, V.; Hall, R.E.
1982-02-01T23:59:59.000Z
This report provides a methodology for the initial quantification of specific categories of human errors made in conjunction with several instrumentation and control (I and C) system components operated, maintained, and tested in licensed nuclear power plants. The resultant human error rates (HER) provide the first real systems bases of comparison for the existing derived and/or best judgement equivalent set of such rates or probabilities. These calculated error rates also provide the first real indication of human performance as it relates directly to specific tasks in nuclear plants. This work of developing specific HERs is both an extension of and an outgrowth of the generic HERs developed for safety system pumpc and valves as reported in NUREG/CR-1880.
Luckas, W.J. Jr.; Lettieri, V.; Hall, R.E.
1982-02-01T23:59:59.000Z
This report provides a methodology for the initial quantification of specific categories of human errors made in conjunction with several instrumentation and control (I and C) system components operated, maintained, and tested in licensed nuclear power plants. The resultant human error rates (HER) provide the first real systems bases of comparison for the existing derived and/or best judgement equivalent set of such rates or probabilities. These calculated error rates also provide the first real indication of human performance as it relates directly to specific tasks in nuclear plants. This work of developing specific HERs is both an extension of and an outgrowth of the generic HERs developed for safety system pumps and valves as reported in NUREG/CR-1880.
Audenaert, Koenraad M. R., E-mail: koenraad.audenaert@rhul.ac.uk [Department of Mathematics, Royal Holloway University of London, Egham TW20 0EX (United Kingdom); Department of Physics and Astronomy, University of Ghent, S9, Krijgslaan 281, B-9000 Ghent (Belgium); Mosonyi, Milán, E-mail: milan.mosonyi@gmail.com [Física Teòrica: Informació i Fenomens Quàntics, Universitat Autònoma de Barcelona, ES-08193 Bellaterra, Barcelona (Spain); Mathematical Institute, Budapest University of Technology and Economics, Egry József u 1., Budapest 1111 (Hungary)
2014-10-15T23:59:59.000Z
We consider the multiple hypothesis testing problem for symmetric quantum state discrimination between r given states ?{sub 1}, …, ?{sub r}. By splitting up the overall test into multiple binary tests in various ways we obtain a number of upper bounds on the optimal error probability in terms of the binary error probabilities. These upper bounds allow us to deduce various bounds on the asymptotic error rate, for which it has been hypothesized that it is given by the multi-hypothesis quantum Chernoff bound (or Chernoff divergence) C(?{sub 1}, …, ?{sub r}), as recently introduced by Nussbaum and Szko?a in analogy with Salikhov's classical multi-hypothesis Chernoff bound. This quantity is defined as the minimum of the pairwise binary Chernoff divergences min{sub j
Uniform and optimal error estimates of an exponential wave ...
2014-05-01T23:59:59.000Z
of the error propagation, cut-off of the nonlinearity, and the energy method. ...... gives Lemma 3.4 for the local truncation error, which is of spectral order in ... estimates, we adopt a strategy similar to the finite difference method [4] (cf. diagram.
PROPAGATION OF ERRORS IN SPATIAL ANALYSIS Peter P. Siska
Hung, I-Kuai
, the conversion of data from analog to digital form used to be an extremely time-consuming process. At present process then the resulting error is inflated up to 20 percent for each grid cell of the final map. The magnitude of errors naturally increases with an addition of every new layer entering the overlay process
Quasi-sparse eigenvector diagonalization and stochastic error correction
Dean Lee
2000-08-30T23:59:59.000Z
We briefly review the diagonalization of quantum Hamiltonians using the quasi-sparse eigenvector (QSE) method. We also introduce the technique of stochastic error correction, which systematically removes the truncation error of the QSE result by stochastically sampling the contribution of the remaining basis states.
ERROR VISUALIZATION FOR TANDEM ACOUSTIC MODELING ON THE AURORA TASK
Ellis, Dan
ERROR VISUALIZATION FOR TANDEM ACOUSTIC MODELING ON THE AURORA TASK Manuel J. Reyes. This structure reduces the error rate on the Aurora 2 noisy English digits task by more than 50% compared development of tandem systems showed an improvement in the performance on the Aurora task [2] of these systems
Entanglement and Quantum Error Correction with Superconducting Qubits
Entanglement and Quantum Error Correction with Superconducting Qubits A Dissertation Presented David Reed All rights reserved. #12;Entanglement and Quantum Error Correction with Superconducting is to use superconducting quantum bits in the circuit quantum electro- dynamics (cQED) architecture. There
Grid-scale Fluctuations and Forecast Error in Wind Power
Bel, G; Toots, M; Bandi, M M
2015-01-01T23:59:59.000Z
The fluctuations in wind power entering an electrical grid (Irish grid) were analyzed and found to exhibit correlated fluctuations with a self-similar structure, a signature of large-scale correlations in atmospheric turbulence. The statistical structure of temporal correlations for fluctuations in generated and forecast time series was used to quantify two types of forecast error: a timescale error ($e_{\\tau}$) that quantifies the deviations between the high frequency components of the forecast and the generated time series, and a scaling error ($e_{\\zeta}$) that quantifies the degree to which the models fail to predict temporal correlations in the fluctuations of the generated power. With no $a$ $priori$ knowledge of the forecast models, we suggest a simple memory kernel that reduces both the timescale error ($e_{\\tau}$) and the scaling error ($e_{\\zeta}$).
Grid-scale Fluctuations and Forecast Error in Wind Power
G. Bel; C. P. Connaughton; M. Toots; M. M. Bandi
2015-03-29T23:59:59.000Z
The fluctuations in wind power entering an electrical grid (Irish grid) were analyzed and found to exhibit correlated fluctuations with a self-similar structure, a signature of large-scale correlations in atmospheric turbulence. The statistical structure of temporal correlations for fluctuations in generated and forecast time series was used to quantify two types of forecast error: a timescale error ($e_{\\tau}$) that quantifies the deviations between the high frequency components of the forecast and the generated time series, and a scaling error ($e_{\\zeta}$) that quantifies the degree to which the models fail to predict temporal correlations in the fluctuations of the generated power. With no $a$ $priori$ knowledge of the forecast models, we suggest a simple memory kernel that reduces both the timescale error ($e_{\\tau}$) and the scaling error ($e_{\\zeta}$).
A Sensing Error Aware MAC Protocol for Cognitive Radio Networks
Hu, Donglin
2011-01-01T23:59:59.000Z
Cognitive radios (CR) are intelligent radio devices that can sense the radio environment and adapt to changes in the radio environment. Spectrum sensing and spectrum access are the two key CR functions. In this paper, we present a spectrum sensing error aware MAC protocol for a CR network collocated with multiple primary networks. We explicitly consider both types of sensing errors in the CR MAC design, since such errors are inevitable for practical spectrum sensors and more important, such errors could have significant impact on the performance of the CR MAC protocol. Two spectrum sensing polices are presented, with which secondary users collaboratively sense the licensed channels. The sensing policies are then incorporated into p-Persistent CSMA to coordinate opportunistic spectrum access for CR network users. We present an analysis of the interference and throughput performance of the proposed CR MAC, and find the analysis highly accurate in our simulation studies. The proposed sensing error aware CR MAC p...
Optics measurement algorithms and error analysis for the proton energy frontier
Langner, A
2015-01-01T23:59:59.000Z
Optics measurement algorithms have been improved in preparation for the commissioning of the LHC at higher energy, i.e., with an increased damage potential. Due to machine protection considerations the higher energy sets tighter limits in the maximum excitation amplitude and the total beam charge, reducing the signal to noise ratio of optics measurements. Furthermore the precision in 2012 (4 TeV) was insufficient to understand beam size measurements and determine interaction point (IP) ?-functions (?). A new, more sophisticated algorithm has been developed which takes into account both the statistical and systematic errors involved in this measurement. This makes it possible to combine more beam position monitor measurements for deriving the optical parameters and demonstrates to significantly improve the accuracy and precision. Measurements from the 2012 run have been reanalyzed which, due to the improved algorithms, result in a significantly higher precision of the derived optical parameters and decreased...
Logical Error Rate Scaling of the Toric Code
Fern H. E. Watson; Sean D. Barrett
2014-09-26T23:59:59.000Z
To date, a great deal of attention has focused on characterizing the performance of quantum error correcting codes via their thresholds, the maximum correctable physical error rate for a given noise model and decoding strategy. Practical quantum computers will necessarily operate below these thresholds meaning that other performance indicators become important. In this work we consider the scaling of the logical error rate of the toric code and demonstrate how, in turn, this may be used to calculate a key performance indicator. We use a perfect matching decoding algorithm to find the scaling of the logical error rate and find two distinct operating regimes. The first regime admits a universal scaling analysis due to a mapping to a statistical physics model. The second regime characterizes the behavior in the limit of small physical error rate and can be understood by counting the error configurations leading to the failure of the decoder. We present a conjecture for the ranges of validity of these two regimes and use them to quantify the overhead -- the total number of physical qubits required to perform error correction.
Slope Error Measurement Tool for Solar Parabolic Trough Collectors: Preprint
Stynes, J. K.; Ihas, B.
2012-04-01T23:59:59.000Z
The National Renewable Energy Laboratory (NREL) has developed an optical measurement tool for parabolic solar collectors that measures the combined errors due to absorber misalignment and reflector slope error. The combined absorber alignment and reflector slope errors are measured using a digital camera to photograph the reflected image of the absorber in the collector. Previous work using the image of the reflection of the absorber finds the reflector slope errors from the reflection of the absorber and an independent measurement of the absorber location. The accuracy of the reflector slope error measurement is thus dependent on the accuracy of the absorber location measurement. By measuring the combined reflector-absorber errors, the uncertainty in the absorber location measurement is eliminated. The related performance merit, the intercept factor, depends on the combined effects of the absorber alignment and reflector slope errors. Measuring the combined effect provides a simpler measurement and a more accurate input to the intercept factor estimate. The minimal equipment and setup required for this measurement technique make it ideal for field measurements.
Wind Power Forecasting Error Distributions: An International Comparison; Preprint
Hodge, B. M.; Lew, D.; Milligan, M.; Holttinen, H.; Sillanpaa, S.; Gomez-Lazaro, E.; Scharff, R.; Soder, L.; Larsen, X. G.; Giebel, G.; Flynn, D.; Dobschinski, J.
2012-09-01T23:59:59.000Z
Wind power forecasting is expected to be an important enabler for greater penetration of wind power into electricity systems. Because no wind forecasting system is perfect, a thorough understanding of the errors that do occur can be critical to system operation functions, such as the setting of operating reserve levels. This paper provides an international comparison of the distribution of wind power forecasting errors from operational systems, based on real forecast data. The paper concludes with an assessment of similarities and differences between the errors observed in different locations.
Universal Framework for Quantum Error-Correcting Codes
Zhuo Li; Li-Juan Xing
2009-01-04T23:59:59.000Z
We present a universal framework for quantum error-correcting codes, i.e., the one that applies for the most general quantum error-correcting codes. This framework is established on the group algebra, an algebraic notation for the nice error bases of quantum systems. The nicest thing about this framework is that we can characterize the properties of quantum codes by the properties of the group algebra. We show how it characterizes the properties of quantum codes as well as generates some new results about quantum codes.
Derivation Tree Based Genetic Programming
Fernandez, Thomas
Derivation Tree Based Genetic Programming Summary of the Ph.D. Thesis by RÂ´obert VÂ´anyi supervisor presented in the Ph.D. thesis enti- tled Derivation Tree Based Genetic Programming. The thesis describes of the generated candidates. [6] 1.2 Proposed method The method presented in the thesis uses derivation trees
Diagnosing multiplicative error by lensing magnification of type Ia supernovae
Zhang, Pengjie
2015-01-01T23:59:59.000Z
Weak lensing causes spatially coherent fluctuations in flux of type Ia supernovae (SNe Ia). This lensing magnification allows for weak lensing measurement independent of cosmic shear. It is free of shape measurement errors associated with cosmic shear and can therefore be used to diagnose and calibrate multiplicative error. Although this lensing magnification is difficult to measure accurately in auto correlation, its cross correlation with cosmic shear and galaxy distribution in overlapping area can be measured to significantly higher accuracy. Therefore these cross correlations can put useful constraint on multiplicative error, and the obtained constraint is free of cosmic variance in weak lensing field. We present two methods implementing this idea and estimate their performances. We find that, with $\\sim 1$ million SNe Ia that can be achieved by the proposed D2k survey with the LSST telescope (Zhan et al. 2008), multiplicative error of $\\sim 0.5\\%$ for source galaxies at $z_s\\sim 1$ can be detected and la...
YELLOW SEA ACOUSTIC UNCERTAINTY CAUSED BY HYDROGRAPHIC DATA ERROR
Chu, Peter C.
the littoral and blue waters. After a weapon platform has detected its targets, the sensors on torpedoes, bathymetry, bottom type, and sound speed profiles. Here, the effect of sound speed errors (i.e., hydrographic
Strontium-90 Error Discovered in Subcontract Laboratory Spreadsheet
D. D. Brown A. S. Nagel
1999-07-31T23:59:59.000Z
West Valley Demonstration Project health physicists and environment scientists discovered a series of errors in a subcontractor's spreadsheet being used to reduce data as part of their strontium-90 analytical process.
Kinetic energy error in the NIMROD spheromak simulations Carl Sovinec
Sovinec, Carl
Kinetic energy error in the NIMROD spheromak simulations Carl Sovinec 10/25/00 Dmitri Ryutov at the ends (as in the spheromak simulations), it may lead to compression in a boundary layer.] The maximum
Error estimation and adaptive mesh refinement for aerodynamic flows
Hartmann, Ralf
Error estimation and adaptive mesh refinement for aerodynamic flows Ralf Hartmann1 and Paul Houston2 1 Institute of Aerodynamics and Flow Technology DLR (German Aerospace Center) Lilienthalplatz 7
An Approximation Algorithm for Constructing Error Detecting Prefix ...
2006-09-02T23:59:59.000Z
Sep 2, 2006 ... 2-bit Hamming prefix code problem. Our algorithm spends O(n log3 n) time to calculate a 2-bit. Hamming prefix code with an additive error of at ...
Secured Pace Web Server with Collaboration and Error Logging Capabilities
Tao, Lixin
: Secure Sockets Layer (SSL) using the Java Secure Socket Extension (JSSE) API, error logging............................................................................................ 8 Chapter 3 Secure Pace Web Server with SSL........................................................... 29 3.1 Introduction to SSL
Sample covariance based estimation of Capon algorithm error probabilities
Richmond, Christ D.
The method of interval estimation (MIE) provides a strategy for mean squared error (MSE) prediction of algorithm performance at low signal-to-noise ratios (SNR) below estimation threshold where asymptotic predictions fail. ...
Wind Power Forecasting Error Distributions over Multiple Timescales: Preprint
Hodge, B. M.; Milligan, M.
2011-03-01T23:59:59.000Z
In this paper, we examine the shape of the persistence model error distribution for ten different wind plants in the ERCOT system over multiple timescales. Comparisons are made between the experimental distribution shape and that of the normal distribution.
Servo control booster system for minimizing following error
Wise, William L. (Mountain View, CA)
1985-01-01T23:59:59.000Z
A closed-loop feedback-controlled servo system is disclosed which reduces command-to-response error to the system's position feedback resolution least increment, .DELTA.S.sub.R, on a continuous real-time basis for all operating speeds. The servo system employs a second position feedback control loop on a by exception basis, when the command-to-response error .gtoreq..DELTA.S.sub.R, to produce precise position correction signals. When the command-to-response error is less than .DELTA.S.sub.R, control automatically reverts to conventional control means as the second position feedback control loop is disconnected, becoming transparent to conventional servo control means. By operating the second unique position feedback control loop used herein at the appropriate clocking rate, command-to-response error may be reduced to the position feedback resolution least increment. The present system may be utilized in combination with a tachometer loop for increased stability.
Model Error Correction for Linear Methods in PET Neuroreceptor Measurements
Renaut, Rosemary
Model Error Correction for Linear Methods in PET Neuroreceptor Measurements Hongbin Guo address: hguo1@asu.edu (Hongbin Guo) Preprint submitted to NeuroImage December 11, 2008 #12;reached. A new
A Posteriori Error Estimation for - Department of Mathematics ...
Shuhao Cao supervised under Professor Zhiqiang Cai
2013-10-31T23:59:59.000Z
Oct 19, 2013 ... the “correct” Hilbert space the true flux µ?1?×u lies in, to recover a ...... The error heat map shows that ZZ-patch recovery estimator leads.
Quantum error correcting codes based on privacy amplification
Zhicheng Luo
2008-08-10T23:59:59.000Z
Calderbank-Shor-Steane (CSS) quantum error-correcting codes are based on pairs of classical codes which are mutually dual containing. Explicit constructions of such codes for large blocklengths and with good error correcting properties are not easy to find. In this paper we propose a construction of CSS codes which combines a classical code with a two-universal hash function. We show, using the results of Renner and Koenig, that the communication rates of such codes approach the hashing bound on tensor powers of Pauli channels in the limit of large block-length. While the bit-flip errors can be decoded as efficiently as the classical code used, the problem of efficiently decoding the phase-flip errors remains open.
Rateless and rateless unequal error protection codes for Gaussian channels
Boyle, Kevin P. (Kevin Patrick)
2007-01-01T23:59:59.000Z
In this thesis we examine two different rateless codes and create a rateless unequal error protection code, all for the additive white Gaussian noise (AWGN) channel. The two rateless codes are examined through both analysis ...
Systematic errors in current quantum state tomography tools
Christian Schwemmer; Lukas Knips; Daniel Richart; Tobias Moroder; Matthias Kleinmann; Otfried Gühne; Harald Weinfurter
2014-07-22T23:59:59.000Z
Common tools for obtaining physical density matrices in experimental quantum state tomography are shown here to cause systematic errors. For example, using maximum likelihood or least squares optimization for state reconstruction, we observe a systematic underestimation of the fidelity and an overestimation of entanglement. A solution for this problem can be achieved by a linear evaluation of the data yielding reliable and computational simple bounds including error bars.
An error correcting procedure for imperfect supervised, nonparametric classification
Ferrell, Dennis Ray
2012-06-07T23:59:59.000Z
AN ERROR CORRECTING PROCEDJJRE FOR IMPERFECTI, Y SUPERVISED, NONPARAMETRIC CLASSIFICATION A Thesis by DENNIS RAY FERRELL Submitted to the Graduate College of Texas AAM University in partial fulfillment of the requirement for the degree...) (Head f Depart en ) (Member) (Member) PE y (Memb ei) (Member) August 1973 ABSTRACT An Error Correcting Procedure For Imperfectly Supervised, Nonparametric Classification (August 1973) Dennis Ray Ferrell, B. S. , I, omar University Directed by...
Using doppler radar images to estimate aircraft navigational heading error
Doerry, Armin W. (Albuquerque, NM); Jordan, Jay D. (Albuquerque, NM); Kim, Theodore J. (Albuquerque, NM)
2012-07-03T23:59:59.000Z
A yaw angle error of a motion measurement system carried on an aircraft for navigation is estimated from Doppler radar images captured using the aircraft. At least two radar pulses aimed at respectively different physical locations in a targeted area are transmitted from a radar antenna carried on the aircraft. At least two Doppler radar images that respectively correspond to the at least two transmitted radar pulses are produced. These images are used to produce an estimate of the yaw angle error.
QUALITY OF ECHOTOP PRODUCTS AND DERIVED PROBABILITY OF HAIL AS A FUNCTION OF RANGE
Stoffelen, Ad
measurements are affected by various sources of error which tend to increase with the distance from the radar of reflectivity data measured by two or more radars are based on PPI or CAPPI products which does not allow (Netherlands) and RMI (Belgium) are derived from the height of the freezing level and the 45-dBZ echotop values
Hard Data on Soft Errors: A Large-Scale Assessment of Real-World Error Rates in GPGPU
Pratt, Vaughan
-GPGPU hardware in a controlled environment found no errors. However, our survey on Folding@home finds that carried out on over 50,000 GPUs on the Folding@home distributed computing network. MemtestG80
Error Monitoring: A Learning Strategy for Improving Academic Performance of LD Adolescents
Schumaker, Jean B.; Deshler, Donald D.; Nolan, Susan; Clark, Frances L.; Alley, Gordon R.; Warner, Michael M.
1981-04-01T23:59:59.000Z
Error monitoring, a learning strategy for detecting and correcting errors in written products, was taught to nine learning disabled adolescents. Students could detect and correct more errors after they received training ...
Assessing the Impact of Differential Genotyping Errors on Rare Variant Tests of Association
Fast, Shannon Marie
Genotyping errors are well-known to impact the power and type I error rate in single marker tests of association. Genotyping errors that happen according to the same process in cases and controls are known as non-differential ...
Estimation of the error for small-sample optimal binary filter design using prior knowledge
Sabbagh, David L
1999-01-01T23:59:59.000Z
Optimal binary filters estimate an unobserved ideal quantity from observed quantities. Optimality is with respect to some error criterion, which is usually mean absolute error MAE (or equivalently mean square error) for the binary values. Both...
Wind derivatives: hedging wind risk:.
Hoyer, S.A.
2013-01-01T23:59:59.000Z
??Wind derivatives are financial contracts that can be used to hedge or mitigate wind risk. In this thesis, the focus was on pricing these wind… (more)
Lattice gas models derived from effective field theory
Matthew Hamilton; Iyam Lynch; Dean Lee
2004-12-03T23:59:59.000Z
We start from a low-energy effective field theory for interacting fermions on the lattice and expand in the hopping parameter to derive the nearest-neighbor interactions for a lattice gas model. In this model the renormalization of couplings for different lattice spacings is inherited from the effective field theory, systematic errors can be estimated a priori, and the breakdown of the lattice gas model description at low temperatures can be understood quantitatively. We apply the lattice gas method to neutron matter and compare with results from a recent quantum simulation.
T-719:Apache mod_proxy_ajp HTTP Processing Error Lets Remote...
Broader source: Energy.gov (indexed) [DOE]
719:Apache modproxyajp HTTP Processing Error Lets Remote Users Deny Service T-719:Apache modproxyajp HTTP Processing Error Lets Remote Users Deny Service September 16, 2011 -...
E-Print Network 3.0 - achieved classification error Sample Search...
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
classifier design, and error estimation, which together form a microarray classification pipeline... error rate, reflects how well the classification rule can approximate the...
Heralded quantum gates with integrated error detection in optical cavitites
J. Borregaard; P. Kómár; E. M. Kessler; A. S. Sørensen; M. D. Lukin
2015-01-05T23:59:59.000Z
We propose and analyze heralded quantum gates between qubits in optical cavities. They employ an auxiliary qubit to report if a successful gate occurred. In this manner, the errors, which would have corrupted a deterministic gate, are converted into a non-unity probability of success: once successful the gate has a much higher fidelity than a similar deterministic gate. Specifically, we describe that a heralded , near-deterministic controlled phase gate (CZ-gate) with the conditional error arbitrarily close to zero and the success probability that approaches unity as the cooperativity of the system, C, becomes large. Furthermore, we describe an extension to near-deterministic N- qubit Toffoli gate with a favorable error scaling. These gates can be directly employed in quantum repeater networks to facilitate near-ideal entanglement swapping, thus greatly speeding up the entanglement distribution.
Development of an integrated system for estimating human error probabilities
Auflick, J.L.; Hahn, H.A.; Morzinski, J.A.
1998-12-01T23:59:59.000Z
This is the final report of a three-year, Laboratory Directed Research and Development (LDRD) project at the Los Alamos National Laboratory (LANL). This project had as its main objective the development of a Human Reliability Analysis (HRA), knowledge-based expert system that would provide probabilistic estimates for potential human errors within various risk assessments, safety analysis reports, and hazard assessments. HRA identifies where human errors are most likely, estimates the error rate for individual tasks, and highlights the most beneficial areas for system improvements. This project accomplished three major tasks. First, several prominent HRA techniques and associated databases were collected and translated into an electronic format. Next, the project started a knowledge engineering phase where the expertise, i.e., the procedural rules and data, were extracted from those techniques and compiled into various modules. Finally, these modules, rules, and data were combined into a nearly complete HRA expert system.
Hard Data on Soft Errors: A Large-Scale Assessment of Real-World Error Rates in GPGPU
Haque, Imran S
2009-01-01T23:59:59.000Z
Graphics processing units (GPUs) are gaining widespread use in computational chemistry and other scientific simulation contexts because of their huge performance advantages relative to conventional CPUs. However, the reliability of GPUs in error-intolerant applications is largely unproven. In particular, a lack of error checking and correcting (ECC) capability in the memory subsystems of graphics cards has been cited as a hindrance to the acceptance of GPUs as high-performance coprocessors, but the impact of this design has not been previously quantified. In this article we present MemtestG80, our software for assessing memory error rates on NVIDIA G80 and GT200-architecture-based graphics cards. Furthermore, we present the results of a large-scale assessment of GPU error rate, conducted by running MemtestG80 on over 20,000 hosts on the Folding@home distributed computing network. Our control experiments on consumer-grade and dedicated-GPGPU hardware in a controlled environment found no errors. However, our su...
Error estimates and specification parameters for functional renormalization
Schnoerr, David [Institute for Theoretical Physics, University of Heidelberg, D-69120 Heidelberg (Germany)] [Institute for Theoretical Physics, University of Heidelberg, D-69120 Heidelberg (Germany); Boettcher, Igor, E-mail: I.Boettcher@thphys.uni-heidelberg.de [Institute for Theoretical Physics, University of Heidelberg, D-69120 Heidelberg (Germany)] [Institute for Theoretical Physics, University of Heidelberg, D-69120 Heidelberg (Germany); Pawlowski, Jan M. [Institute for Theoretical Physics, University of Heidelberg, D-69120 Heidelberg (Germany) [Institute for Theoretical Physics, University of Heidelberg, D-69120 Heidelberg (Germany); ExtreMe Matter Institute EMMI, GSI Helmholtzzentrum für Schwerionenforschung mbH, D-64291 Darmstadt (Germany); Wetterich, Christof [Institute for Theoretical Physics, University of Heidelberg, D-69120 Heidelberg (Germany)] [Institute for Theoretical Physics, University of Heidelberg, D-69120 Heidelberg (Germany)
2013-07-15T23:59:59.000Z
We present a strategy for estimating the error of truncated functional flow equations. While the basic functional renormalization group equation is exact, approximated solutions by means of truncations do not only depend on the choice of the retained information, but also on the precise definition of the truncation. Therefore, results depend on specification parameters that can be used to quantify the error of a given truncation. We demonstrate this for the BCS–BEC crossover in ultracold atoms. Within a simple truncation the precise definition of the frequency dependence of the truncated propagator affects the results, indicating a shortcoming of the choice of a frequency independent cutoff function.
JLab SRF Cavity Fabrication Errors, Consequences and Lessons Learned
Frank Marhauser
2011-09-01T23:59:59.000Z
Today, elliptical superconducting RF (SRF) cavities are preferably made from deep-drawn niobium sheets as pursued at Jefferson Laboratory (JLab). The fabrication of a cavity incorporates various cavity cell machining, trimming and electron beam welding (EBW) steps as well as surface chemistry that add to forming errors creating geometrical deviations of the cavity shape from its design. An analysis of in-house built cavities over the last years revealed significant errors in cavity production. Past fabrication flaws are described and lessons learned applied successfully to the most recent in-house series production of multi-cell cavities.
Laser Phase Errors in Seeded Free Electron Lasers
Ratner, D.; Fry, A.; Stupakov, G.; White, W.; /SLAC
2012-04-17T23:59:59.000Z
Harmonic seeding of free electron lasers has attracted significant attention as a method for producing transform-limited pulses in the soft x-ray region. Harmonic multiplication schemes extend seeding to shorter wavelengths, but also amplify the spectral phase errors of the initial seed laser, and may degrade the pulse quality and impede production of transform-limited pulses. In this paper we consider the effect of seed laser phase errors in high gain harmonic generation and echo-enabled harmonic generation. We use simulations to confirm analytical results for the case of linearly chirped seed lasers, and extend the results for arbitrary seed laser envelope and phase.
Quantum error correcting codes and 4-dimensional arithmetic hyperbolic manifolds
Guth, Larry, E-mail: lguth@math.mit.edu [Department of Mathematics, MIT, Cambridge, Massachusetts 02139 (United States); Lubotzky, Alexander, E-mail: alex.lubotzky@mail.huji.ac.il [Institute of Mathematics, Hebrew University, Jerusalem 91904 (Israel)
2014-08-15T23:59:59.000Z
Using 4-dimensional arithmetic hyperbolic manifolds, we construct some new homological quantum error correcting codes. They are low density parity check codes with linear rate and distance n{sup ?}. Their rate is evaluated via Euler characteristic arguments and their distance using Z{sub 2}-systolic geometry. This construction answers a question of Zémor [“On Cayley graphs, surface codes, and the limits of homological coding for quantum error correction,” in Proceedings of Second International Workshop on Coding and Cryptology (IWCC), Lecture Notes in Computer Science Vol. 5557 (2009), pp. 259–273], who asked whether homological codes with such parameters could exist at all.
Considering Workload Input Variations in Error Coverage Estimation
Karlsson, Johan
different parts of the workload code to be executed different number of times. By using the results from in the workload input when estimating error detection coverage using fault injection are investigated. Results sequence based on results from fault injection experiments with another input sequence is presented
Stateful Testing: Finding More Errors in Code and Contracts
Meyer, Bertrand
. The generated test cases are designed to violate the dynamically inferred contracts (invariants) characterizing the existing test suite. As a consequence, they are in a good position to detect new faults, and alsoStateful Testing: Finding More Errors in Code and Contracts Yi Wei Â· Hannes Roth Â· Carlo A. Furia
Error magnitude in the conservation of energy in the
van Thienen, Peter
Appendix A Error magnitude in the conservation of energy in the approximate melt segregation scheme A.1 Conservation of energy The approximate melt segregation used in the thermochemical convection models of chap- ters 6 and 7 has an impact on the conservation of energy, because although 'segregated
The contour method cutting assumption: error minimization and correction
Prime, Michael B [Los Alamos National Laboratory; Kastengren, Alan L [ANL
2010-01-01T23:59:59.000Z
The recently developed contour method can measure 2-D, cross-sectional residual-stress map. A part is cut in two using a precise and low-stress cutting technique such as electric discharge machining. The contours of the new surfaces created by the cut, which will not be flat if residual stresses are relaxed by the cutting, are then measured and used to calculate the original residual stresses. The precise nature of the assumption about the cut is presented theoretically and is evaluated experimentally. Simply assuming a flat cut is overly restrictive and misleading. The critical assumption is that the width of the cut, when measured in the original, undeformed configuration of the body is constant. Stresses at the cut tip during cutting cause the material to deform, which causes errors. The effect of such cutting errors on the measured stresses is presented. The important parameters are quantified. Experimental procedures for minimizing these errors are presented. An iterative finite element procedure to correct for the errors is also presented. The correction procedure is demonstrated on experimental data from a steel beam that was plastically bent to put in a known profile of residual stresses.
MULTITARGET ERROR ESTIMATION AND ADAPTIVITY IN AERODYNAMIC FLOW SIMULATIONS
Hartmann, Ralf
MULTIÂTARGET ERROR ESTIMATION AND ADAPTIVITY IN AERODYNAMIC FLOW SIMULATIONS RALF HARTMANN # Abstract. Important quantities in aerodynamic flow simulations are the aerodynamic force coe subject classifications. 65N12,65N15,65N30 1. Introduction. In aerodynamic computations like compressible
Error estimation and adaptive mesh refinement for aerodynamic flows
Hartmann, Ralf
Error estimation and adaptive mesh refinement for aerodynamic flows Ralf Hartmann, Joachim Held-oriented mesh refinement for single and multiple aerodynamic force coefficients as well as residual-based mesh refinement applied to various three-dimensional lam- inar and turbulent aerodynamic test cases defined
MULTITARGET ERROR ESTIMATION AND ADAPTIVITY IN AERODYNAMIC FLOW SIMULATIONS
Hartmann, Ralf
MULTITARGET ERROR ESTIMATION AND ADAPTIVITY IN AERODYNAMIC FLOW SIMULATIONS RALF HARTMANN Abstract. Important quantities in aerodynamic flow simulations are the aerodynamic force coefficients including Navier-Stokes equations AMS subject classifications. 65N12,65N15,65N30 1. Introduction. In aerodynamic
Analysis of possible systematic errors in the Oslo method
Larsen, A. C.; Guttormsen, M.; Buerger, A.; Goergen, A.; Nyhus, H. T.; Rekstad, J.; Siem, S.; Toft, H. K.; Tveten, G. M.; Wikan, K. [Department of Physics, University of Oslo, N-0316 Oslo (Norway); Krticka, M. [Institute of Particle and Nuclear Physics, Charles University, Prague (Czech Republic); Betak, E. [Institute of Physics SAS, 84511 Bratislava (Slovakia); Faculty of Philosophy and Science, Silesian University, 74601 Opava (Czech Republic); Schiller, A.; Voinov, A. V. [Department of Physics and Astronomy, Ohio University, Athens, Ohio 45701 (United States)
2011-03-15T23:59:59.000Z
In this work, we have reviewed the Oslo method, which enables the simultaneous extraction of the level density and {gamma}-ray transmission coefficient from a set of particle-{gamma} coincidence data. Possible errors and uncertainties have been investigated. Typical data sets from various mass regions as well as simulated data have been tested against the assumptions behind the data analysis.
Multilayer Perceptron Error Surfaces: Visualization, Structure and Modelling
Gallagher, Marcus
. This is commonly formulated as a multivariate nonÂlinear optimization problem over a very highÂdimensional space of analysis are not wellÂsuited to this problem. Visualizing and describÂ ing the error surface are also three related methods. Firstly, Principal Component Analysis (PCA) is proposed as a method
Multi-layer Perceptron Error Surfaces: Visualization, Structure and Modelling
Gallagher, Marcus
. This is commonly formulated as a multivariate non-linear optimization problem over a very high-dimensional space of analysis are not well-suited to this problem. Visualizing and describ- ing the error surface are also three related methods. Firstly, Principal Component Analysis (PCA) is proposed as a method
Analysis of possible systematic errors in the Oslo method
A. C. Larsen; M. Guttormsen; M. Krticka; E. Betak; A. Bürger; A. Görgen; H. T. Nyhus; J. Rekstad; A. Schiller; S. Siem; H. K. Toft; G. M. Tveten; A. V. Voinov; K. Wikan
2012-11-27T23:59:59.000Z
In this work, we have reviewed the Oslo method, which enables the simultaneous extraction of level density and gamma-ray transmission coefficient from a set of particle-gamma coincidence data. Possible errors and uncertainties have been investigated. Typical data sets from various mass regions as well as simulated data have been tested against the assumptions behind the data analysis.
Error Control Based Model Reduction for Parameter Optimization of Elliptic
of technical devices that rely on multiscale processes, such as fuel cells or batteries. As the solutionError Control Based Model Reduction for Parameter Optimization of Elliptic Homogenization Problems optimization of elliptic multiscale problems with macroscopic optimization functionals and microscopic material
Verification of unfold error estimates in the unfold operator code
Fehl, D.L.; Biggs, F. [Sandia National Laboratories, Albuquerque, New Mexico 87185 (United States)] [Sandia National Laboratories, Albuquerque, New Mexico 87185 (United States)
1997-01-01T23:59:59.000Z
Spectral unfolding is an inverse mathematical operation that attempts to obtain spectral source information from a set of response functions and data measurements. Several unfold algorithms have appeared over the past 30 years; among them is the unfold operator (UFO) code written at Sandia National Laboratories. In addition to an unfolded spectrum, the UFO code also estimates the unfold uncertainty (error) induced by estimated random uncertainties in the data. In UFO the unfold uncertainty is obtained from the error matrix. This built-in estimate has now been compared to error estimates obtained by running the code in a Monte Carlo fashion with prescribed data distributions (Gaussian deviates). In the test problem studied, data were simulated from an arbitrarily chosen blackbody spectrum (10 keV) and a set of overlapping response functions. The data were assumed to have an imprecision of 5{percent} (standard deviation). One hundred random data sets were generated. The built-in estimate of unfold uncertainty agreed with the Monte Carlo estimate to within the statistical resolution of this relatively small sample size (95{percent} confidence level). A possible 10{percent} bias between the two methods was unresolved. The Monte Carlo technique is also useful in underdetermined problems, for which the error matrix method does not apply. UFO has been applied to the diagnosis of low energy x rays emitted by Z-pinch and ion-beam driven hohlraums. {copyright} {ital 1997 American Institute of Physics.}
Achievable Error Exponents for the Private Fingerprinting Game
Merhav, Neri
Achievable Error Exponents for the Private Fingerprinting Game Anelia Somekh-Baruch and Neri Merhav a forgery of the data while aiming at erasing the fingerprints in order not to be detected. Their action have presented and analyzed a game-theoretic model of private2 fingerprinting systems in the presence
RESOLVE Upgrades for on Line Lattice Error Analysis
Lee, M.; Corbett, J.; White, G.; /SLAC; Zambre, Y.; /Unlisted
2011-08-25T23:59:59.000Z
We have increased the speed and versatility of the orbit analysis process by adding a command file, or 'script' language, to RESOLVE. This command file feature enables us to automate data analysis procedures to detect lattice errors. We describe the RESOLVE command file and present examples of practical applications.
Stereoscopic Light Stripe Scanning: Interference Rejection, Error Minimization and Calibration
This paper addresses the problem of rejecting interfer- ence due to secondary specular reflections, cross structure, acquisition delay, lack of error recovery, and incorrect modelling of measurement noise. We cause secondary reflections, edges and textures may have a stripe-like appearance, and cross-talk can
Effects of errors in the solar radius on helioseismic inferences
Sarbani Basu
1997-12-09T23:59:59.000Z
Frequencies of intermediate-degree f-modes of the Sun seem to indicate that the solar radius is smaller than what is normally used in constructing solar models. We investigate the possible consequences of an error in radius on results for solar structure obtained using helioseismic inversions. It is shown that solar sound speed will be overestimated if oscillation frequencies are inverted using reference models with a larger radius. Using solar models with radius of 695.78 Mm and new data sets, the base of the solar convection zone is estimated to be at radial distance of $0.7135\\pm 0.0005$ of the solar radius. The helium abundance in the convection zone as determined using models with OPAL equation of state is $0.248\\pm 0.001$, where the errors reflect the estimated systematic errors in the calculation, the statistical errors being much smaller. Assuming that the OPAL opacities used in the construction of the solar models are correct, the surface $Z/X$ is estimated to be $0.0245\\pm 0.0006$.
Error field and magnetic diagnostic modeling for W7-X
Lazerson, Sam A. [PPPL; Gates, David A. [PPPL; NEILSON, GEORGE H. [PPPL; OTTE, M.; Bozhenkov, S.; Pedersen, T. S.; GEIGER, J.; LORE, J.
2014-07-01T23:59:59.000Z
The prediction, detection, and compensation of error fields for the W7-X device will play a key role in achieving a high beta (? = 5%), steady state (30 minute pulse) operating regime utilizing the island divertor system [1]. Additionally, detection and control of the equilibrium magnetic structure in the scrape-off layer will be necessary in the long-pulse campaign as bootstrapcurrent evolution may result in poor edge magnetic structure [2]. An SVD analysis of the magnetic diagnostics set indicates an ability to measure the toroidal current and stored energy, while profile variations go undetected in the magnetic diagnostics. An additional set of magnetic diagnostics is proposed which improves the ability to constrain the equilibrium current and pressure profiles. However, even with the ability to accurately measure equilibrium parameters, the presence of error fields can modify both the plasma response and diverter magnetic field structures in unfavorable ways. Vacuum flux surface mapping experiments allow for direct measurement of these modifications to magnetic structure. The ability to conduct such an experiment is a unique feature of stellarators. The trim coils may then be used to forward model the effect of an applied n = 1 error field. This allows the determination of lower limits for the detection of error field amplitude and phase using flux surface mapping. *Research supported by the U.S. DOE under Contract No. DE-AC02-09CH11466 with Princeton University.
Designing Automation to Reduce Operator Errors Nancy G. Leveson
Leveson, Nancy
Designing Automation to Reduce Operator Errors Nancy G. Leveson Computer Science and Engineering University of Washington Everett Palmer NASA Ames Research Center Introduction Advanced automation has been of modeÂrelated problems [SW95]. After studying accidents and incidents in the new, highly automated
MODELS FOR DIAGNOSING ROBOT ERROR SOURCES Louis J. Everett
Everett, Louis J.
of industrial robots and on some machine tools, calibration methods have significantly improved positionMODELS FOR DIAGNOSING ROBOT ERROR SOURCES Louis J. Everett Mechanical Engineering Texas A that the somewhat ad-hoc modelling methods used for robot calibration, although satisfactory for improving accu
Removing Systematic Errors from Rotating Shadowband Pyranometer Data Frank Vignola
Oregon, University of
Removing Systematic Errors from Rotating Shadowband Pyranometer Data Frank Vignola Solar Radiation irradiance be- cause they do not require manual adjustment of trackers. However, a RSP requires the use of solar cell based pyranometers which underestimate diffuse irradiance by 20- 30% under clear sky
Two infinite families of nonadditive quantum error-correcting codes
Sixia Yu; Qing Chen; C. H. Oh
2009-01-14T23:59:59.000Z
We construct explicitly two infinite families of genuine nonadditive 1-error correcting quantum codes and prove that their coding subspaces are 50% larger than those of the optimal stabilizer codes of the same parameters via the linear programming bound. All these nonadditive codes can be characterized by a stabilizer-like structure and thus their encoding circuits can be designed in a straightforward manner.
Quantum Error Correction of Continuous Variable States against Gaussian Noise
T. C. Ralph
2011-05-22T23:59:59.000Z
We describe a continuous variable error correction protocol that can correct the Gaussian noise induced by linear loss on Gaussian states. The protocol can be implemented using linear optics and photon counting. We explore the theoretical bounds of the protocol as well as the expected performance given current knowledge and technology.
A Method for Treating Discretization Error in Nondeterministic Analysis
Alvin, K.F.
1999-01-27T23:59:59.000Z
A response surface methodology-based technique is presented for treating discretization error in non-deterministic analysis. The response surface, or metamodel, is estimated from computer experiments which vary both uncertain physical parameters and the fidelity of the computational mesh. The resultant metamodel is then used to propagate the variabilities in the continuous input parameters, while the mesh size is taken to zero, its asymptotic limit. With respect to mesh size, the metamodel is equivalent to Richardson extrapolation, in which solutions on coarser and finer meshes are used to estimate discretization error. The method is demonstrated on a one dimensional prismatic bar, in which uncertainty in the third vibration frequency is estimated by propagating variations in material modulus, density, and bar length. The results demonstrate the efficiency of the method for combining non-deterministic analysis with error estimation to obtain estimates of total simulation uncertainty. The results also show the relative sensitivity of failure estimates to solution bias errors in a reliability analysis, particularly when the physical variability of the system is low.
Error Control of Iterative Linear Solvers for Integrated Groundwater Models
Bai, Zhaojun
and presentation of GMRES performance benchmarking results. Introduction As the groundwater model infrastructureError Control of Iterative Linear Solvers for Integrated Groundwater Models by Matthew F. Dixon1 for integrated groundwater models, which are implicitly coupled to another model, such as surface water models
A-posteriori estimation and adaptive control of the error in the solution quantity of interest
Datta, Dibyendu Kumar, Dd 1973-
1997-01-01T23:59:59.000Z
-posteriori estimation and adaptive control of the error in the quantity of interest. The major tool for the estimation of the error in the desired quantity is the splitting of the error into two components: the near-field or local error, and the far-field or pollution...
Global tropospheric ozone modeling: Quantifying errors due to grid resolution
Wild, Oliver; Prather, Michael J
2006-01-01T23:59:59.000Z
model-derived monthly climatologies [Tegen et al. , 1997],extent that the MODIS climatology does. [ 11 ] The standarduse the optional aerosol climatology. These simulations are
Quantum computing with nearest neighbor interactions and error rates over 1%
David S. Wang; Austin G. Fowler; Lloyd C. L. Hollenberg
2010-09-20T23:59:59.000Z
Large-scale quantum computation will only be achieved if experimentally implementable quantum error correction procedures are devised that can tolerate experimentally achievable error rates. We describe a quantum error correction procedure that requires only a 2-D square lattice of qubits that can interact with their nearest neighbors, yet can tolerate quantum gate error rates over 1%. The precise maximum tolerable error rate depends on the error model, and we calculate values in the range 1.1--1.4% for various physically reasonable models. Even the lowest value represents the highest threshold error rate calculated to date in a geometrically constrained setting, and a 50% improvement over the previous record.
Alvarez-Ramirez, J.; Monroy-Loperena, R.; Cervantes, I.; Morales, A.
2000-02-01T23:59:59.000Z
The aim of this paper is to propose a novel proportional-integral-derivative (PID) control configuration based on an observer structure. Batch distillation is used as the base case study where the regulated output is the distillate composition. The proposed PID control law is derived in the framework of robust nonlinear control with modeling error compensation techniques. A reduced-order observer is proposed to estimate both the derivative of the regulated output and the underlying modeling error. These observations are subsequently used in a control loop to feedback variations of distillate composition (derivative feedback) and to counteract the effects of modeling errors. It is shown that, under certain conditions, the resulting control law is equivalent to a classical PID controller with an antireset windup scheme. Moreover, the tuning of the controller is performed very easily in terms of a prescribed closed-loop time constant and an estimation time constant. Numerical results are provided for binary and multicomponent separations. Sampled/delayed measurements and several sources of uncertainties are considered in order to provide a realistic test scenario for the proposed control design procedure.
Joachim Wuttke
2012-09-01T23:59:59.000Z
The C library \\texttt{libkww} provides functions to compute the Kohlrausch-Williams-Watts function, i.e.\\ the Laplace-Fourier transform of the stretched (or compressed) exponential function $\\exp(-t^\\beta)$ for exponents $\\beta$ between 0.1 and 1.9 with sixteen-digits accuracy. Analytic error bounds are derived for the low and high frequency series expansions. For intermediate frequencies the numeric integration is enormously accelerated by using the Ooura-Mori double exponential transformation. The source code is available from the project home page \\url{http://apps.jcns.fz-juelich.de/doku/sc/kww}.
National Nuclear Security Administration (NNSA)
AFDC Printable Version Share this resource Send a link to EERE: Alternative Fuels Data Center Home Page to someone by E-mail Share EERE: Alternative Fuels Data Center Home Page on Facebook Tweet about EERE: Alternative Fuels Data Center Home Page on Twitter Bookmark EERE:1 First Use of Energy for All Purposes (Fuel and Nonfuel),Feet) Year Jan Feb Mar Apr May Jun Jul(Summary) "ofEarlyEnergyDepartmentNationalRestart of the Review of theOFFICEACME |Supplement Analysis,/U U U10
Derivative-Free Optimization Proximal Point Methods Derivative-Free Proximal Point Conclusion Derivative-Free Optimization via Proximal Point Methods Yves Lucet & Warren Hare July 24, 2013 1 / 26 #12;Derivative-Free Optimization Proximal Point Methods Derivative-Free Proximal Point Conclusion Outline 1
Comparison of Wind Power and Load Forecasting Error Distributions: Preprint
Hodge, B. M.; Florita, A.; Orwig, K.; Lew, D.; Milligan, M.
2012-07-01T23:59:59.000Z
The introduction of large amounts of variable and uncertain power sources, such as wind power, into the electricity grid presents a number of challenges for system operations. One issue involves the uncertainty associated with scheduling power that wind will supply in future timeframes. However, this is not an entirely new challenge; load is also variable and uncertain, and is strongly influenced by weather patterns. In this work we make a comparison between the day-ahead forecasting errors encountered in wind power forecasting and load forecasting. The study examines the distribution of errors from operational forecasting systems in two different Independent System Operator (ISO) regions for both wind power and load forecasts at the day-ahead timeframe. The day-ahead timescale is critical in power system operations because it serves the unit commitment function for slow-starting conventional generators.
Efficient Semiparametric Estimators for Biological, Genetic, and Measurement Error Applications
Garcia, Tanya
2012-10-19T23:59:59.000Z
) as pW,Y,Z(w, y, z; ?, ?1, ?2, ?3) which equals ? pW |X,Z(w|x, z)?1(x, z)?2{y ? m(x, z; ?), x, z}?3(z)dx, (2.1) where ? is the finite p-dimensional parameter of interest, ?1(x, z) ? pX|Z(x|z), ?2(#15;, x, z) ? p#15;|X,Z(#15;|x, z), and ?3(z) ? pZ(z...) are infinite dimensional nuisance parameters. Doing so, we see that pW,Y,Z , the RMM with measurement error, is tightly linked to the RMM without measurement error with probability density expressed as pX,Y,Z ? ?1(x, z)?2{y ? m(x, z; ?), x, z}?3(z). The lack...
Method and system for reducing errors in vehicle weighing systems
Hively, Lee M. (Philadelphia, TN); Abercrombie, Robert K. (Knoxville, TN)
2010-08-24T23:59:59.000Z
A method and system (10, 23) for determining vehicle weight to a precision of <0.1%, uses a plurality of weight sensing elements (23), a computer (10) for reading in weighing data for a vehicle (25) and produces a dataset representing the total weight of a vehicle via programming (40-53) that is executable by the computer (10) for (a) providing a plurality of mode parameters that characterize each oscillatory mode in the data due to movement of the vehicle during weighing, (b) by determining the oscillatory mode at which there is a minimum error in the weighing data; (c) processing the weighing data to remove that dynamical oscillation from the weighing data; and (d) repeating steps (a)-(c) until the error in the set of weighing data is <0.1% in the vehicle weight.
Error message recording and reporting in the SLC control system
Spencer, N.; Bogart, J.; Phinney, N.; Thompson, K.
1985-10-01T23:59:59.000Z
Error or information messages that are signaled by control software either in the VAX host computer or the local microprocessor clusters are handled by a dedicated VAX process (PARANOIA). Messages are recorded on disk for further analysis and displayed at the appropriate console. Another VAX process (ERRLOG) can be used to sort, list and histogram various categories of messages. The functions performed by these processes and the algorithms used are discussed.
Error message recording and reporting in the SLC control system
Spencer, N.; Bogart, J.; Phinney, N.; Thompson, K.
1985-04-01T23:59:59.000Z
Error or information messages that are signaled by control software either in the VAX host computer or the local microprocessor clusters are handled by a dedicated VAX process (PARANOIA). Messages are recorded on disk for further analysis and displayed at the appropriate console. Another VAX process (ERRLOG) can be used to sort, list and histogram various categories of messages. The functions performed by these processes and the algorithms used are discussed.
Topics in measurement error and missing data problems
Liu, Lian
2009-05-15T23:59:59.000Z
reasons. In this research, the impact of missing genotypes is investigated for high resolution combined linkage and association mapping of quantitative trait loci (QTL). We assume that the genotype data are missing completely at random (MCAR). Two... and asymptotic properties. In the genetics study, a new method is proposed to account for the missing genotype in a combined linkage and association study. We have concluded that this method does not improve power but it will provide better type I error rates...
Magnetic error analysis of recycler pbar injection transfer line
Yang, M.J.; /Fermilab
2007-06-01T23:59:59.000Z
Detailed study of Fermilab Recycler Ring anti-proton injection line became feasible with its BPM system upgrade, though the beamline has been in existence and operational since year 2000. Previous attempts were not fruitful due to limitations in the BPM system. Among the objectives are the assessment of beamline optics and the presence of error fields. In particular the field region of the permanent Lambertson magnets at both ends of R22 transfer line will be scrutinized.
Error rate and power dissipation in nano-logic devices
Kim, Jong Un
2004-01-01T23:59:59.000Z
and content by: Laszlo B. Kish (Chair of Commitpte) Ed anc ez-Sinencio (Member) song g (Member) William Marlow (Member) nan Singh (Head of Department) May 2004 Major Subject: Electrical Engineering 111 ABSTRACT Error Rate and Power Dissipation... in Nano-Logic Devices. (May 2004) Jong Un Kim, B. S. ; M. S. ; Ph. D. , Seoul National University Chair of Advisory Committee: Dr. Laszlo B. Kish Current-controlled logic and single electron logic processors have been investigated with respect...
Runtime Detection of C-Style Errors in UPC Code
Pirkelbauer, P; Liao, C; Panas, T; Quinlan, D
2011-09-29T23:59:59.000Z
Unified Parallel C (UPC) extends the C programming language (ISO C 99) with explicit parallel programming support for the partitioned global address space (PGAS), which provides a global memory space with localized partitions to each thread. Like its ancestor C, UPC is a low-level language that emphasizes code efficiency over safety. The absence of dynamic (and static) safety checks allows programmer oversights and software flaws that can be hard to spot. In this paper, we present an extension of a dynamic analysis tool, ROSE-Code Instrumentation and Runtime Monitor (ROSECIRM), for UPC to help programmers find C-style errors involving the global address space. Built on top of the ROSE source-to-source compiler infrastructure, the tool instruments source files with code that monitors operations and keeps track of changes to the system state. The resulting code is linked to a runtime monitor that observes the program execution and finds software defects. We describe the extensions to ROSE-CIRM that were necessary to support UPC. We discuss complications that arise from parallel code and our solutions. We test ROSE-CIRM against a runtime error detection test suite, and present performance results obtained from running error-free codes. ROSE-CIRM is released as part of the ROSE compiler under a BSD-style open source license.
Estimating the error in simulation prediction over the design space
Shinn, R. (Rachel); Hemez, F. M. (François M.); Doebling, S. W. (Scott W.)
2003-01-01T23:59:59.000Z
This study addresses the assessrnent of accuracy of simulation predictions. A procedure is developed to validate a simple non-linear model defined to capture the hardening behavior of a foam material subjected to a short-duration transient impact. Validation means that the predictive accuracy of the model must be established, not just in the vicinity of a single testing condition, but for all settings or configurations of the system. The notion of validation domain is introduced to designate the design region where the model's predictive accuracy is appropriate for the application of interest. Techniques brought to bear to assess the model's predictive accuracy include test-analysis coi-relation, calibration, bootstrapping and sampling for uncertainty propagation and metamodeling. The model's predictive accuracy is established by training a metalnodel of prediction error. The prediction error is not assumed to be systcmatic. Instead, it depends on which configuration of the system is analyzed. Finally, the prediction error's confidence bounds are estimated by propagating the uncertainty associated with specific modeling assumptions.
On the efficiency of nondegenerate quantum error correction codes for Pauli channels
Gunnar Bjork; Jonas Almlof; Isabel Sainz
2009-05-19T23:59:59.000Z
We examine the efficiency of pure, nondegenerate quantum-error correction-codes for Pauli channels. Specifically, we investigate if correction of multiple errors in a block is more efficient than using a code that only corrects one error per block. Block coding with multiple-error correction cannot increase the efficiency when the qubit error-probability is below a certain value and the code size fixed. More surprisingly, existing multiple-error correction codes with a code length equal or less than 256 qubits have lower efficiency than the optimal single-error correcting codes for any value of the qubit error-probability. We also investigate how efficient various proposed nondegenerate single-error correcting codes are compared to the limit set by the code redundancy and by the necessary conditions for hypothetically existing nondegenerate codes. We find that existing codes are close to optimal.
Reliable random error estimation in the measurement of line-strength indices
N. Cardiel; J. Gorgas; J. Cenarro; J. J. Gonzalez
1997-06-12T23:59:59.000Z
We present a new set of accurate formulae for the computation of random errors in the measurement of atomic and molecular indices. The new expressions are in excellent agreement with numerical simulations. We have found that, in some cases, the use of approximated equations can give misleading line-strength index errors. It is important to note that accurate errors can only be achieved after a full control of the error propagation throughout the data reduction with a parallel processing of data and error frames. Finally, simple recipes for the estimation of the required signal-to-noise ratio to achieve a fixed index error are presented.
Quaternion Derivatives: The GHR Calculus
Dongpo Xu; Cyrus Jahanchahi; Clive C. Took; Danilo P. Mandic
2014-09-25T23:59:59.000Z
Quaternion derivatives in the mathematical literature are typically defined only for analytic (regular) functions. However, in engineering problems, functions of interest are often real-valued and thus not analytic, such as the standard cost function. The HR calculus is a convenient way to calculate formal derivatives of both analytic and non-analytic functions of quaternion variables, however, both the HR and other functional calculus in quaternion analysis have encountered an essential technical obstacle, that is, the traditional product rule is invalid due to the non- commutativity of the quaternion algebra. To address this issue, a generalized form of the HR derivative is proposed based on a general orthogonal system. The so introduced generalization, called the generalized HR (GHR) calculus, encompasses not just the left- and right-hand versions of quaternion derivative, but also enables solutions to some long standing problems, such as the novel product rule, the chain rule, the mean-valued theorem and Taylor's theorem. At the core of the proposed approach is the quaternion rotation, which can naturally be applied to other functional calculi in non-commutative settings. Examples on using the GHR calculus in adaptive signal processing support the analysis.
Sensitivity of OFDM Systems to Synchronization Errors and Spatial Diversity
Zhou, Yi
2012-02-14T23:59:59.000Z
of observations. Computer simulations show that the asymptotic bound is tight with more than 64 observations. Therefore, a good bench- mark for comparing the performance of different timing synchronizers is obtained. The selection of several system design.... 4. The curve marked with circles represents the actual ICI power from computer simulations while the curve with stars denotes the upper bound derived in (A.11). Fig. 4 illustrates that the derived upper bound predicts very accurately the ICI power...
T-609: Adobe Acrobat/Reader Memory Corruption Error in CoolType...
Broader source: Energy.gov (indexed) [DOE]
09: Adobe AcrobatReader Memory Corruption Error in CoolType Library Lets Remote Users Execute Arbitrary Code T-609: Adobe AcrobatReader Memory Corruption Error in CoolType...
Cognitive analysis of students' errors and misconceptions in variables, equations, and functions
Li, Xiaobao
2009-05-15T23:59:59.000Z
such issues, three basic algebra concepts - variable, equation, and function – are used to analyze students’ errors, possible buggy algorithms, and the conceptual basis of these errors: misconceptions. Through the research on these three basic concepts...
Recompile if your codes run into MPICH error after the maintenance...
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
Recompile if your codes run into MPICH errors after the maintenance on 6252014 Recompile if your codes run into MPICH error after the maintenance on 6252014 June 27, 2014 (0...
Reducing Quantum Errors and Improving Large Scale Quantum Cryptography
T. Mor
1996-08-15T23:59:59.000Z
Noise causes severe difficulties in implementing quantum computing and quantum cryptography. Several schemes have been suggested to reduce this problem, mainly focusing on quantum computation. Motivated by quantum cryptography, we suggest a coding which uses $N$ quantum bits ($N=n^2$) to encode one quantum bit, and reduces the error exponentially with $n$. Our result suggests the possibility of distributing a secure key over very long distances, and maintaining quantum states for very long times. It also provides a new quantum privacy amplification against a strong adversary.
Topological Quantum Computation and Error Correction by Biological Cells
J T Lofthouse
2005-02-02T23:59:59.000Z
A Topological examination of phospholipid dynamics in the Far from Equilibrium state has demonstrated that metabolically active cells use waste heat to generate spatially patterned membrane flows by forced convection and shear. This paper explains the resemblance between this nonlinear membrane model and Witten Kitaev type Topological Quantum Computation systems, and demonstrates how this self-organising membrane enables biological cells to circumvent the decoherence problem, perform error correction procedures, and produce classical level output as shielded current flow through cytoskeletal protein conduit. Cellular outputs are shown to be Turing compatible as they are determined by computable in principle hydromagnetic fluid flows, and importantly, are Adaptive from an Evolutionary perspective.
Error 401 on upload? | OpenEI Community
AFDC Printable Version Share this resource Send a link to EERE: Alternative Fuels Data Center Home Page to someone by E-mail Share EERE: Alternative Fuels Data Center Home Page on Facebook Tweet about EERE: Alternative Fuels Data Center Home Page on Twitter Bookmark EERE: Alternative Fuels Data Center Home Page on Google Bookmark EERE: Alternative Fuels Data Center Home5b9fcbce19 No revision has beenFfe2fb55-352f-473b-a2dd-50ae8b27f0a6 NoSan Leandro,Law and Policy CenterTODO:Erie CountyError 401
Deriving confinement via RG decimations
E. T. Tomboulis
2007-10-10T23:59:59.000Z
We present the general framework and building blocks of a recent derivation of the fact that the SU(2) LGT is in a confining phase for all values of the coupling $0 < \\beta < \\infty$, for space-time dimension $d \\leq 4$. The method employs approximate but explicitly computable RG decimations that are shown to constrain the exact partition function and order parameters from above and below, and flow from the weak to the strong coupling regime without encountering a fixed point.
Error Estimation for High Speed Flows Using Continuous and Discrete Adjoints
Alonso, Juan J.
the fullest extent possible) strategy to control the error in multi-physics simulations of Scramjet propulsion
Local Estimation of Modeling Error in Multi-Scale Modeling of Heterogeneous Elastic Solids
Moody, Tristan
2008-03-19T23:59:59.000Z
Global Enhancement . . . . . . . . . . . . . . . . . . . . 24 3.3.3 The Adaptive Process . . . . . . . . . . . . . . . . . . . 24 Chapter 4. Modeling Error Estimation 30 4.1 Residual-Based Error Estimation . . . . . . . . . . . . . . . . . 30 4.2 Error... with effectivity indices . 80 ix 5.18 Average estimates of enhancement errors in quantity of interest Qx(u) using three enhancements with effectivity indices . . . . 80 x List of Figures 1.1 Examples of composite failure modes . . . . . . . . . . . . . . 4 2...
Using Graphs for Fast Error Term Approximation of Time-varying Datasets
Nuber, C; LaMar, E C; Pascucci, V; Hamann, B; Joy, K I
2003-02-27T23:59:59.000Z
We present a method for the efficient computation and storage of approximations of error tables used for error estimation of a region between different time steps in time-varying datasets. The error between two time steps is defined as the distance between the data of these time steps. Error tables are used to look up the error between different time steps of a time-varying dataset, especially when run time error computation is expensive. However, even the generation of error tables itself can be expensive. For n time steps, the exact error look-up table (which stores the error values for all pairs of time steps in a matrix) has a memory complexity and pre-processing time complexity of O(n2), and O(1) for error retrieval. Our approximate error look-up table approach uses trees, where the leaf nodes represent original time steps, and interior nodes contain an average (or best-representative) of the children nodes. The error computed on an edge of a tree describes the distance between the two nodes on that edge. Evaluating the error between two different time steps requires traversing a path between the two leaf nodes, and accumulating the errors on the traversed edges. For n time steps, this scheme has a memory complexity and pre-processing time complexity of O(nlog(n)), a significant improvement over the exact scheme; the error retrieval complexity is O(log(n)). As we do not need to calculate all possible n2 error terms, our approach is a fast way to generate the approximation.
Bayesian Semiparametric Density Deconvolution and Regression in the Presence of Measurement Errors
Sarkar, Abhra
2014-06-24T23:59:59.000Z
BAYESIAN SEMIPARAMETRIC DENSITY DECONVOLUTION AND REGRESSION IN THE PRESENCE OF MEASUREMENT ERRORS A Dissertation by ABHRA SARKAR Submitted to the Office of Graduate and Professional Studies of Texas A&M University in partial fulfillment... Copyright 2014 Abhra Sarkar ABSTRACT Although the literature on measurement error problems is quite extensive, so- lutions to even the most fundamental measurement error problems like density de- convolution and regression with errors...
Derivation of Hamiltonians for accelerators
Symon, K.R.
1997-09-12T23:59:59.000Z
In this report various forms of the Hamiltonian for particle motion in an accelerator will be derived. Except where noted, the treatment will apply generally to linear and circular accelerators, storage rings, and beamlines. The generic term accelerator will be used to refer to any of these devices. The author will use the usual accelerator coordinate system, which will be introduced first, along with a list of handy formulas. He then starts from the general Hamiltonian for a particle in an electromagnetic field, using the accelerator coordinate system, with time t as independent variable. He switches to a form more convenient for most purposes using the distance s along the reference orbit as independent variable. In section 2, formulas will be derived for the vector potentials that describe the various lattice components. In sections 3, 4, and 5, special forms of the Hamiltonian will be derived for transverse horizontal and vertical motion, for longitudinal motion, and for synchrobetatron coupling of horizontal and longitudinal motions. Hamiltonians will be expanded to fourth order in the variables.
Maintaining Standards: Differences between the Standard Deviation and Standard Error, and
California at Santa Cruz, University of
Maintaining Standards: Differences between the Standard Deviation and Standard Error, and When to Use Each David L Streiner, PhD1 Many people confuse the standard deviation (SD) and the standard error of the findings. (Can J Psychiatry 1996;41:498Â502) Key Words: statistics, standard deviation, standard error
TECHNICAL ADVANCES Dye shift: a neglected source of genotyping error in molecular
Jamieson, Ian
TECHNICAL ADVANCES Dye shift: a neglected source of genotyping error in molecular ecology JOLENE T for genotyping error, yet potential errors stemming from dye-induced mobility shift (dye shift) may be frequently left uncorrected, dye shift can lead to mis-scoring alleles and even to fal- sely calling new alleles
Exposure Measurement Error in Time-Series Studies of Air Pollution: Concepts and Consequences
Dominici, Francesca
of air pollution and health. Because measurement error may have substantial implications for interpreting1 Exposure Measurement Error in Time-Series Studies of Air Pollution: Concepts and Consequences S in time-series studies 1 11/11/99 Keywords: measurement error, air pollution, time series, exposure
Database Error Trapping and Prediction Mike West & Robert L. Winkler \\Lambda
West, Mike
Database Error Trapping and Prediction By Mike West & Robert L. Winkler \\Lambda Duke University of errors in databases. In particular, we study two error detection methods. In the duplicate performance method, all items in a database are processed by two individuals (or machines), and the resulting records
Error Analysis of Ia Supernova and Query on Cosmic Dark Energy
Qiuhe Peng; Yiming Hu; Kun Wang; Yu Liang
2012-01-16T23:59:59.000Z
Some serious faults in error analysis of observations for SNIa have been found. Redoing the same error analysis of SNIa, by our idea, it is found that the average total observational error of SNIa is obviously greater than $0.55^m$, so we can't decide whether the universe is accelerating expansion or not.
Verification of unfold error estimates in the UFO code
Fehl, D.L.; Biggs, F.
1996-07-01T23:59:59.000Z
Spectral unfolding is an inverse mathematical operation which attempts to obtain spectral source information from a set of tabulated response functions and data measurements. Several unfold algorithms have appeared over the past 30 years; among them is the UFO (UnFold Operator) code. In addition to an unfolded spectrum, UFO also estimates the unfold uncertainty (error) induced by running the code in a Monte Carlo fashion with prescribed data distributions (Gaussian deviates). In the problem studied, data were simulated from an arbitrarily chosen blackbody spectrum (10 keV) and a set of overlapping response functions. The data were assumed to have an imprecision of 5% (standard deviation). 100 random data sets were generated. The built-in estimate of unfold uncertainty agreed with the Monte Carlo estimate to within the statistical resolution of this relatively small sample size (95% confidence level). A possible 10% bias between the two methods was unresolved. The Monte Carlo technique is also useful in underdetemined problems, for which the error matrix method does not apply. UFO has been applied to the diagnosis of low energy x rays emitted by Z-Pinch and ion-beam driven hohlraums.
Effects of imbalance and geometric error on precision grinding machines
Bibler, J.E.
1997-06-01T23:59:59.000Z
To study balancing in grinding, a simple mechanical system was examined. It was essential to study such a well-defined system, as opposed to a large, complex system such as a machining center. The use of a compact, well-defined system enabled easy quantification of the imbalance force input, its phase angle to any geometric decentering, and good understanding of the machine mode shapes. It is important to understand a simple system such as the one I examined given that imbalance is so intimately coupled to machine dynamics. It is possible to extend the results presented here to industrial machines, although that is not part of this work. In addition to the empirical testing, a simple mechanical system to look at how mode shapes, balance, and geometric error interplay to yield spindle error motion was modelled. The results of this model will be presented along with the results from a more global grinding model. The global model, presented at ASPE in November 1996, allows one to examine the effects of changing global machine parameters like stiffness and damping. This geometrically abstract, one-dimensional model will be presented to demonstrate the usefulness of an abstract approach for first-order understanding but it will not be the main focus of this thesis. 19 refs., 36 figs., 10 tables.
Deriving Ontologies from XML Schema
Bedini, Ivan; Nguyen, Benjamin
2010-01-01T23:59:59.000Z
In this paper, we present a method and a tool for deriving a skeleton of an ontology from XML schema files. We first recall what an is ontology and its relationships with XML schemas. Next, we focus on ontology building methodology and associated tool requirements. Then, we introduce Janus, a tool for building an ontology from various XML schemas in a given domain. We summarize the main features of Janus and illustrate its functionalities through a simple example. Finally, we compare our approach to other existing ontology building tools.
Organic derivatives of zirconium phosphate
Fine, Steven Beryl
1980-01-01T23:59:59.000Z
-zirconium phosphate derivatives. 9 Thermogravimetric analysis of y-zirconium 2-hydroxy- 1-ethyl phosphate. 25 10 Thermogravimetric analysis of y-zirconium 2-carboxy- 2-hydroxy-1-ethyl phosphate. 28 11 Thermo gravimetric 1-ethyl phosphate. analysis of n...-ZrP Figure 3. A comparison of the layered structures of c and y zirconium phosphate. according to Yamanaka [9], ethylene oxide is 18. 4 A [ 10] . When this phase is dehydrated at 150 for one hour, the d-spacing decreases to 17. 2 A, but when left in air...
Higher order treatment on temporal derivative of angular flux for time-dependent MOC
Tsujita, K.; Endo, T.; Yamamoto, A. [Nagoya University, Department of Material, Physics and Energy Engineering, Furo-cho, Chikusa-ku, Nagoya, 464-8603 (Japan); Kamiyama, Y.; Kirimura, K. [Mitsubishi Heavy Industries, Ltd., Wadasakicho1-1-1, Hyogo-ku, Kobe, 652-8585 (Japan)
2013-07-01T23:59:59.000Z
A new kinetic analysis method, whose angular dependence of temporal derivative for angular flux is accurately treated within practical memory requirement, is proposed. The method of characteristics (MOC) is being widely used for reactor analysis thanks to the advances of numerical algorithms and computer hardware. However, the computational resources, i.e., the memory capacity, can be still a crucial problem for rigorous kinetic calculations using MOC. In the straightforward approach for kinetic calculation using MOC, the segment-averaged angular fluxes should be stored on the memory in order to explicitly calculate the temporal derivative of the angular flux, which would require huge memory. Thus, in the conventional kinetic calculation code using MOC, the temporal derivative of the angular flux has been approximated as angularly isotropic in order to reduce the memory requirement (isotropic assumption). However, the approximation error caused by the conventional isotropic assumption has not been thoroughly and quantitatively investigated so far and an accurate kinetic calculation method, which can quantitatively estimate the above approximation error within practical memory storage, has not been developed. The present study tries to address this issue with a newly developed approach. Effect of the approximate treatment for the temporal derivative of angular flux is evaluated through benchmark calculations. (authors)
Coordinated joint motion control system with position error correction
Danko, George (Reno, NV)
2011-11-22T23:59:59.000Z
Disclosed are an articulated hydraulic machine supporting, control system and control method for same. The articulated hydraulic machine has an end effector for performing useful work. The control system is capable of controlling the end effector for automated movement along a preselected trajectory. The control system has a position error correction system to correct discrepancies between an actual end effector trajectory and a desired end effector trajectory. The correction system can employ one or more absolute position signals provided by one or more acceleration sensors supported by one or more movable machine elements. Good trajectory positioning and repeatability can be obtained. A two-joystick controller system is enabled, which can in some cases facilitate the operator's task and enhance their work quality and productivity.
Significance of gauge line error in orifice measurement
Bowen, J.W. [ANR Pipeline Co., Detroit, MI (United States)
1995-12-01T23:59:59.000Z
Pulsation induced gauge line amplification can cause errors in the recorded differential signal used to calculate flow. Its presence may be detected using dual transmitters (one connected at the orifice taps, the other at the end of the gauge lines) and comparing the relative peak to peak amplitudes. Its affect on recorded differential may be determined by averaging both signals with a PC based data acquisition and analysis system. Remedial action is recommended in all cases where amplification is detected. Use of close connect, full opening manifolds, is suggested to decouple the gauge lines` resonant frequency from that of the excitation`s, by positioning the recording device as close to the process signal`s origin as possible.
Statistical Error analysis of Nucleon-Nucleon phenomenological potentials
R. Navarro Perez; J. E. Amaro; E. Ruiz Arriola
2014-06-10T23:59:59.000Z
Nucleon-Nucleon potentials are commonplace in nuclear physics and are determined from a finite number of experimental data with limited precision sampling the scattering process. We study the statistical assumptions implicit in the standard least squares fitting procedure and apply, along with more conventional tests, a tail sensitive quantile-quantile test as a simple and confident tool to verify the normality of residuals. We show that the fulfilment of normality tests is linked to a judicious and consistent selection of a nucleon-nucleon database. These considerations prove crucial to a proper statistical error analysis and uncertainty propagation. We illustrate these issues by analyzing about 8000 proton-proton and neutron-proton scattering published data. This enables the construction of potentials meeting all statistical requirements necessary for statistical uncertainty estimates in nuclear structure calculations.
A study on real estate derivatives
Lim, Jong Yoon, S.M. Massachusetts Institute of Technology
2006-01-01T23:59:59.000Z
All major asset classes including stocks and bonds have a well developed derivative market. Derivatives enable counterparties to reflect a view on a particular market, without having to trade the underlying asset. This ...
Notes on Derived Functors and Grothendieck Duality.
2008-12-02T23:59:59.000Z
59. 2.7. Acyclic objects; finite-dimensional derived functors . . . . . 65. Chapter 3. Derived ...... isomorphism s: I ? Y such that sf = 0 [H, p.37]; and if s has a left.
Reduction of derived Hochschild functors over commutative ...
2012-01-09T23:59:59.000Z
Jan 9, 2012 ... LUCHEZAR L. AVRAMOV, SRIKANTH B. IYENGAR, JOSEPH LIPMAN,. AND SURESH NAYAK. Abstract. We study functors underlying derived ...
Postgraduate Scholarship Pricing temperature derivatives and modelling
Banaji,. Murad
the volumetric risk of the energy units sold, rather than the price risk of each unit. Weather derivativesPostgraduate Scholarship Pricing temperature derivatives and modelling the market price of risk: Pricing temperature derivatives and modelling the market price of risk. Main Supervisor: A. Alexandridis
The Fourth Partial Derivative In Transport Dynamics
Trinh Khanh Tuoc
2010-01-11T23:59:59.000Z
A new fourth partial derivative is introduced for the study of transport dynamics. It is a Lagrangian partial derivative following the path of diffusion, not the path of convection. Use of this derivative decouples the effect of diffusion and convection and simplifies the analysis of transport processes.
Remarkable Degenerate Quantum Stabilizer Codes Derived from Duadic Codes
Salah A. Aly; Andreas Klappenecker; Pradeep Kiran Sarvepalli
2006-01-18T23:59:59.000Z
Good quantum codes, such as quantum MDS codes, are typically nondegenerate, meaning that errors of small weight require active error-correction, which is--paradoxically--itself prone to errors. Decoherence free subspaces, on the other hand, do not require active error correction, but perform poorly in terms of minimum distance. In this paper, examples of degenerate quantum codes are constructed that have better minimum distance than decoherence free subspaces and allow some errors of small weight that do not require active error correction. In particular, two new families of [[n,1,>= sqrt(n)
Minimum error discrimination between similarity-transformed quantum states
Jafarizadeh, M. A. [Department of Theoretical Physics and Astrophysics, University of Tabriz, Tabriz 51664 (Iran, Islamic Republic of); Institute for Studies in Theoretical Physics and Mathematics, Tehran 19395-1795 (Iran, Islamic Republic of); Research Institute for Fundamental Sciences, Tabriz 51664 (Iran, Islamic Republic of); Sufiani, R. [Department of Theoretical Physics and Astrophysics, University of Tabriz, Tabriz 51664 (Iran, Islamic Republic of); Institute for Studies in Theoretical Physics and Mathematics, Tehran 19395-1795 (Iran, Islamic Republic of); Mazhari Khiavi, Y. [Department of Theoretical Physics and Astrophysics, University of Tabriz, Tabriz 51664 (Iran, Islamic Republic of)
2011-07-15T23:59:59.000Z
Using the well-known necessary and sufficient conditions for minimum error discrimination (MED), we extract an equivalent form for the MED conditions. In fact, by replacing the inequalities corresponding to the MED conditions with an equivalent but more suitable and convenient identity, the problem of mixed state discrimination with optimal success probability is solved. Moreover, we show that the mentioned optimality conditions can be viewed as a Helstrom family of ensembles under some circumstances. Using the given identity, MED between N similarity transformed equiprobable quantum states is investigated. In the case that the unitary operators are generating a set of irreducible representation, the optimal set of measurements and corresponding maximum success probability of discrimination can be determined precisely. In particular, it is shown that for equiprobable pure states, the optimal measurement strategy is the square-root measurement (SRM), whereas for the mixed states, SRM is not optimal. In the case that the unitary operators are reducible, there is no closed-form formula in the general case, but the procedure can be applied in each case in accordance to that case. Finally, we give the maximum success probability of optimal discrimination for some important examples of mixed quantum states, such as generalized Bloch sphere m-qubit states, spin-j states, particular nonsymmetric qudit states, etc.
A surrogate-based uncertainty quantification with quantifiable errors
Bang, Y.; Abdel-Khalik, H. S. [North Carolina State Univ., Raleigh, NC 27695 (United States)
2012-07-01T23:59:59.000Z
Surrogate models are often employed to reduce the computational cost required to complete uncertainty quantification, where one is interested in propagating input parameters uncertainties throughout a complex engineering model to estimate responses uncertainties. An improved surrogate construction approach is introduced here which places a premium on reducing the associated computational cost. Unlike existing methods where the surrogate is constructed first, then employed to propagate uncertainties, the new approach combines both sensitivity and uncertainty information to render further reduction in the computational cost. Mathematically, the reduction is described by a range finding algorithm that identifies a subspace in the parameters space, whereby parameters uncertainties orthogonal to the subspace contribute negligible amount to the propagated uncertainties. Moreover, the error resulting from the reduction can be upper-bounded. The new approach is demonstrated using a realistic nuclear assembly model and compared to existing methods in terms of computational cost and accuracy of uncertainties. Although we believe the algorithm is general, it will be applied here for linear-based surrogates and Gaussian parameters uncertainties. The generalization to nonlinear models will be detailed in a separate article. (authors)
Implications of Monte Carlo Statistical Errors in Criticality Safety Assessments
Pevey, Ronald E.
2005-09-15T23:59:59.000Z
Most criticality safety calculations are performed using Monte Carlo techniques because of Monte Carlo's ability to handle complex three-dimensional geometries. For Monte Carlo calculations, the more histories sampled, the lower the standard deviation of the resulting estimates. The common intuition is, therefore, that the more histories, the better; as a result, analysts tend to run Monte Carlo analyses as long as possible (or at least to a minimum acceptable uncertainty). For Monte Carlo criticality safety analyses, however, the optimization situation is complicated by the fact that procedures usually require that an extra margin of safety be added because of the statistical uncertainty of the Monte Carlo calculations. This additional safety margin affects the impact of the choice of the calculational standard deviation, both on production and on safety. This paper shows that, under the assumptions of normally distributed benchmarking calculational errors and exact compliance with the upper subcritical limit (USL), the standard deviation that optimizes production is zero, but there is a non-zero value of the calculational standard deviation that minimizes the risk of inadvertently labeling a supercritical configuration as subcritical. Furthermore, this value is shown to be a simple function of the typical benchmarking step outcomes--the bias, the standard deviation of the bias, the upper subcritical limit, and the number of standard deviations added to calculated k-effectives before comparison to the USL.
Subband coding of monochrome images over binary symmetric channels with error correction
Sheppard, Denise M
2012-06-07T23:59:59.000Z
for the degree of MASTER OF SCIENCE December 1992 Major Subject: Electrical Engineering SUBBAND CODING OF MONOCHROME IMAGES OVER BINARY SYMMETRIC CHANNELS WITH ERROR CORRECTION A Thesis by DENISE M. SHEPPARD Approved as to style and content by: V.... State diagram analysis of error recovery . B. Codebook design C. Performance results . 16 18 23 29 42 48 V VI ERROR CORRECTION A. Algorithm B. Performance Results CONCLUSION REFERENCES APPENDIX A . APPENDIX B . 52 54 61 67 70...
Method and apparatus for detecting timing errors in a system oscillator
Gliebe, Ronald J. (Library, PA); Kramer, William R. (Bethel Park, PA)
1993-01-01T23:59:59.000Z
A method of detecting timing errors in a system oscillator for an electronic device, such as a power supply, includes the step of comparing a system oscillator signal with a delayed generated signal and generating a signal representative of the timing error when the system oscillator signal is not identical to the delayed signal. An LED indicates to an operator that a timing error has occurred. A hardware circuit implements the above-identified method.
Unconventional fuel: Tire derived fuel
Hope, M.W. [Waste Recovery, Inc., Portland, OR (United States)
1995-09-01T23:59:59.000Z
Material recovery of scrap tires for their fuel value has moved from a pioneering concept in the early 1980`s to a proven and continuous use in the United States` pulp and paper, utility, industrial, and cement industry. Pulp and paper`s use of tire derived fuel (TDF) is currently consuming tires at the rate of 35 million passenger tire equivalents (PTEs) per year. Twenty mills are known to be burning TDF on a continuous basis. The utility industry is currently consuming tires at the rate of 48 million PTEs per year. Thirteen utilities are known to be burning TDF on a continuous basis. The cement industry is currently consuming tires at the rate of 28 million PTEs per year. Twenty two cement plants are known to be burning TDF on a continuous basis. Other industrial boilers are currently consuming tires at the rate of 6.5 million PTEs per year. Four industrial boilers are known to be burning TDF on a continuous basis. In total, 59 facilities are currently burning over 117 million PTEs per year. Although 93% of these facilities were not engineered to burn TDF, it has become clear that TDF has found acceptance as a supplemental fuel when blending with conventional fuels in existing combustion devices designed for normal operating conditions. The issues of TDF as a supplemental fuel and its proper specifications are critical to the successful development of this fuel alternative. This paper will focus primarily on TDF`s use in a boiler type unit.
St Andrews, University of
Eccentricity Error Correction for Automated Estimation of Polyethylene Wear after Total Hip. Wire markers are typically attached to the polyethylene acetabular component of the prosthesis so
E-Print Network 3.0 - analytic error estimates Sample Search...
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
46 To: Supersite Principal Investigators, Data Managers and Research From: Dr. Paul Solomon, ORD, Dennis Mikel, OAQPS; Mike Jones, OAQPS Summary: of random error (precision)...
Efficient Small Area Estimation in the Presence of Measurement Error in Covariates
Singh, Trijya
2012-10-19T23:59:59.000Z
for the four estimators, yi, eYiS, bYiME, bYiSIMEX when the number of small areas is 100, measure- ment error variance Ci = 3 and 2v = 4. k is the percentage of areas having auxiliary information measured with error. : : : : : : : 52 2 Absolute value... 3 Jackknife estimates of the mean squared error of the Lohr-Ybarra estimator bYiME and the SIMEX estimator bYiSIMEX when the num- ber of small areas is 100, measurement error variance Ci = 2 and 2v = 4. k is the percentage of areas having...
SPPI ORIGINAL PAPER October 11, 2011 GROSS ERRORS IN THE IPCC-AR4
Gray, William
SPPI ORIGINAL PAPER October 11, 2011 GROSS ERRORS IN THE IPCC-AR4 REPORT REGARDING PAST & FUTURE FIGURE AND GEORGE WILL QUOTE.....................
Lipnikov, Konstantin [Los Alamos National Laboratory; Agouzal, Abdellatif [UNIV DE LYON; Vassilevski, Yuri [Los Alamos National Laboratory
2009-01-01T23:59:59.000Z
We present a new technology for generating meshes minimizing the interpolation and discretization errors or their gradients. The key element of this methodology is construction of a space metric from edge-based error estimates. For a mesh with N{sub h} triangles, the error is proportional to N{sub h}{sup -1} and the gradient of error is proportional to N{sub h}{sup -1/2} which are optimal asymptotics. The methodology is verified with numerical experiments.
Correction of motion measurement errors beyond the range resolution of a synthetic aperture radar
Doerry, Armin W. (Albuquerque, NM); Heard, Freddie E. (Albuquerque, NM); Cordaro, J. Thomas (Albuquerque, NM)
2008-06-24T23:59:59.000Z
Motion measurement errors that extend beyond the range resolution of a synthetic aperture radar (SAR) can be corrected by effectively decreasing the range resolution of the SAR in order to permit measurement of the error. Range profiles can be compared across the slow-time dimension of the input data in order to estimate the error. Once the error has been determined, appropriate frequency and phase correction can be applied to the uncompressed input data, after which range and azimuth compression can be performed to produce a desired SAR image.
Neutron Soft Errors in Xilinx FPGAs at Lawrence Berkeley National Laboratory
George, Jeffrey S.
2008-01-01T23:59:59.000Z
Quasi-Monoenergetic Neutron Beam from Deuteron Breakup”, inexperiments of atmospheric neutron effects on deep sub-Neutron Soft Errors in Xilinx FPGAs at Lawrence Berkeley
Abstract Error Groups Via Jones Unitary Braid Group Representations at q=i
Yong Zhang
2009-02-02T23:59:59.000Z
In this paper, we classify a type of abstract groups by the central products of dihedral groups and quaternion groups. We recognize them as abstract error groups which are often not isomorphic to the Pauli groups in the literature. We show the corresponding nice error bases equivalent to the Pauli error bases modulo phase factors. The extension of these abstract groups by the symmetric group are finite images of the Jones unitary representations (or modulo a phase factor) of the braid group at q=i or r=4. We hope this work can finally lead to new families of quantum error correction codes via the representation theory of the braid group.
Van Peursem, David J.
1991-01-01T23:59:59.000Z
SIMULATION STUDY ON ERROR PROPAGATION EFFECTS WHEN DETERMINING SECOND VIRIAL COEFFICIENTS FROM THE SPEED-OF-SOUND OR THE JOULE-THOMSON EXPERIMENT A Thesis by DAVID J. VAN PEURSEM Submitted to the Office of Graduate Studies of Texas A&M...-hexane 0 a a Al-Bizreh a Wormald [10] ~ Dymond & Smith [9l 280 300 320 340 360 380 400 420 440 460 T (K) Thomson coefficient data are propagated into the second virial coefficients derived from these experiments. This will be done by performing a...
DysList: An Annotated Resource of Dyslexic Errors Luz Rello,1
of texts written by people with dyslexia. Each of the errors was annotated with a set of characteristics of this kind, especially given the difficulty of finding texts written by people with dyslexia. Keywords: Errors, Dyslexia, Visual, Phonetics, Resource 1. Introduction Dyslexia is a reading and spelling disorder
Simple Loran Cycle Error Detection Algorithms for Maritime Harbor Entrance Approach
Stanford University
Simple Loran Cycle Error Detection Algorithms for Maritime Harbor Entrance Approach Operations cycle. This paper details and examines some of the algorithms being developed and analyzed by SC127. SC 127 is developing simplified eLoran cycle error detection algorithms for the eLoran HEA MPS. Correct
State preservation by repetitive error detection in a superconducting quantum circuit
J. Kelly; R. Barends; A. G. Fowler; A. Megrant; E. Jeffrey; T. C. White; D. Sank; J. Y. Mutus; B. Campbell; Yu Chen; Z. Chen; B. Chiaro; A. Dunsworth; I. -C. Hoi; C. Neill; P. J. J. O'Malley; C. Quintana; P. Roushan; A. Vainsencher; J. Wenner; A. N. Cleland; John M. Martinis
2014-11-26T23:59:59.000Z
Quantum computing becomes viable when a quantum state can be preserved from environmentally-induced error. If quantum bits (qubits) are sufficiently reliable, errors are sparse and quantum error correction (QEC) is capable of identifying and correcting them. Adding more qubits improves the preservation by guaranteeing increasingly larger clusters of errors will not cause logical failure - a key requirement for large-scale systems. Using QEC to extend the qubit lifetime remains one of the outstanding experimental challenges in quantum computing. Here, we report the protection of classical states from environmental bit-flip errors and demonstrate the suppression of these errors with increasing system size. We use a linear array of nine qubits, which is a natural precursor of the two-dimensional surface code QEC scheme, and track errors as they occur by repeatedly performing projective quantum non-demolition (QND) parity measurements. Relative to a single physical qubit, we reduce the failure rate in retrieving an input state by a factor of 2.7 for five qubits and a factor of 8.5 for nine qubits after eight cycles. Additionally, we tomographically verify preservation of the non-classical Greenberger-Horne-Zeilinger (GHZ) state. The successful suppression of environmentally-induced errors strongly motivates further research into the many exciting challenges associated with building a large-scale superconducting quantum computer.
Distributed Forcing of Forecast and Assimilation Error Systems BRIAN F. FARRELL
Farrell, Brian F.
Distributed Forcing of Forecast and Assimilation Error Systems BRIAN F. FARRELL Division forecast system gov- erning forecast error growth and the tangent linear observer system governing deterministic and stochastic forcings of the forecast and observer systems over a chosen time interval
Evaluation of servo, geometric and dynamic error sources on five axis high-speed machine tool
Andolfatto, Loïc; Mayer, René
2011-01-01T23:59:59.000Z
Many sources of errors exist in the manufacturing process of complex shapes. Some approximations occur at each step from the design geometry to the machined part. The aim of the paper is to present a method to evaluate the effect of high speed and high dynamic load on volumetric errors at the tool center point. The interpolator output signals and the machine encoder signals are recorded and compared to evaluate the contouring errors resulting from each axis follow-up error. The machine encoder signals are also compared to the actual tool center point position as recorded with a non-contact measuring instrument called CapBall to evaluate the total geometric errors. The novelty of the work lies in the method that is proposed to decompose the geometric errors in two categories: the quasi-static geometric errors independent from the speed of the trajectory and the dynamic geometric errors, dependent on the programmed feed rate and resulting from the machine structure deflection during the acceleration of its axes...
The Impact of Background Error on Incomplete Observations for 4D-Var Data Assimilation with
Navon, Michael
The Impact of Background Error on Incomplete Observations for 4D-Var Data Assimilation with the FSU on the 4D- Var data assimilation, twin experiments were carried out with the dy- namical core of the new is also investigated. Keywords: Data assimilation, incomplete observations, background error. 1
Adjoint goal-based error norms for adaptive mesh ocean modelling
Navon, Michael
Adjoint goal-based error norms for adaptive mesh ocean modelling P.W. Power a , M.D. Piggott a,*, F. The use of dynamically-adaptive meshes has many potential advantages but needs to be guided by an error where resolution should be changed. A barotropic wind driven gyre problem is used to demonstrate
Paragios, Nikos
Fast Illumination-invariant Background Subtraction using Two Views: Error Analysis, SensorÂ£ Abstract Background modeling and subtraction to detect new or moving objects in a scene is an important a detailed analysis of such errors. Then, we propose a sensor configuration that eliminates false de
A Posteriori Error Estimates with Post-Processing for Nonconforming Finite Elements
Schieweck, Friedhelm
that it has the same asymptotic behavior as the energy norm of the real discretization error itself. We show, we propose an a posteriori error estimate in the energy norm which uses as an additive term the \\post in the global energy norm, we demonstrate that the concept of using a conforming approximation
A Case for Soft Error Detection and Correction in Computational Chemistry
van Dam, Hubertus JJ; Vishnu, Abhinav; De Jong, Wibe A.
2013-09-10T23:59:59.000Z
High performance computing platforms are expected to deliver 10(18) floating operations per second by the year 2022 through the deployment of millions of cores. Even if every core is highly reliable the sheer number of the them will mean that the mean time between failures will become so short that most applications runs will suffer at least one fault. In particular soft errors caused by intermittent incorrect behavior of the hardware are a concern as they lead to silent data corruption. In this paper we investigate the impact of soft errors on optimization algorithms using Hartree-Fock as a particular example. Optimization algorithms iteratively reduce the error in the initial guess to reach the intended solution. Therefore they may intuitively appear to be resilient to soft errors. Our results show that this is true for soft errors of small magnitudes but not for large errors. We suggest error detection and correction mechanisms for different classes of data structures. The results obtained with these mechanisms indicate that we can correct more than 95% of the soft errors at moderate increases in the computational cost.
Goal-oriented error estimation for reduced basis method, with application to certified sensitivity
Boyer, Edmond
of interest computed using the reduced model is tainted by a reduction error. We present a new, efficiently- tions (PDEs). These models require input data (e.g., the physical features of the considered system papers showed that using an adapted basis could lead to a great improvement of reduction error
Systematic Errors in Future Weak Lensing Surveys: Requirements and Prospects for Self-Calibration
Dragan Huterer; Masahiro Takada; Gary Bernstein; Bhuvnesh Jain
2005-06-02T23:59:59.000Z
We study the impact of systematic errors on planned weak lensing surveys and compute the requirements on their contributions so that they are not a dominant source of the cosmological parameter error budget. The generic types of error we consider are multiplicative and additive errors in measurements of shear, as well as photometric redshift errors. In general, more powerful surveys have stronger systematic requirements. For example, for a SNAP-type survey the multiplicative error in shear needs to be smaller than 1%(fsky/0.025)^{-1/2} of the mean shear in any given redshift bin, while the centroids of photometric redshift bins need to be known to better than 0.003(fsky/0.025)^{-1/2}. With about a factor of two degradation in cosmological parameter errors, future surveys can enter a self-calibration regime, where the mean systematic biases are self-consistently determined from the survey and only higher-order moments of the systematics contribute. Interestingly, once the power spectrum measurements are combined with the bispectrum, the self-calibration regime in the variation of the equation of state of dark energy w_a is attained with only a 20-30% error degradation.
Dimension Augmentation and Combinatorial Criteria for Efficient Error-resistant DNA Self-assembly
Goel, Ashish
Dimension Augmentation and Combinatorial Criteria for Efficient Error-resistant DNA Self-assembly Abstract DNA self-assembly has emerged as a rich and promising primitive for nano-technology. Experimental-correction mech- anisms have been proposed for the tile model of self- assembly. These error-correction mechanisms
Priority-Based Broadcasting of Sensitive Data in Error-Prone Wireless Networks
Wu, Jie
in its Binary-Coded Decimal (BCD) representation. In BCD, each decimal digit is represented as a 4 bits that benefits from network coding. We also consider the case of burst errors and discuss how can we make our--Symbol-level coding, broadcasting, reliability, burst error, random linear network coding, priority, wireless networks
Integrated Control-Path Design and Error Recovery in the Synthesis of Digital
Chakrabarty, Krishnendu
11 Integrated Control-Path Design and Error Recovery in the Synthesis of Digital Microfluidic Lab-on-Chip YANG ZHAO, TAO XU, and KRISHNENDU CHAKRABARTY Duke University Recent advances in digital microfluidics that incorporates control paths and an error- recovery mechanism in the design of a digital microfluidic lab
Error Modeling in the ACT-R Production System Christian Lebire
Reder, Lynne
Error Modeling in the ACT-R Production System Christian LebiÃ¨re Department of Psychology Carnegie to extend the ACT-R production system to model human errors in the performance of a high-level cognitive be successfully duplicated in production system models. Introduction ACT-R (Anderson, 1993) is a model of human
Threshold analysis with fault-tolerant operations for nonbinary quantum error correcting codes
Kanungo, Aparna
2005-11-01T23:59:59.000Z
Quantum error correcting codes have been introduced to encode the data bits in extra redundant bits in order to accommodate errors and correct them. However, due to the delicate nature of the quantum states or faulty gate operations, there is a...
GLOBAL CONVERGENCE OF GENERAL DERIVATIVE-FREE ...
2006-10-26T23:59:59.000Z
a relatively recent topic [1, 3, 7, 10]. In this paper we address trust-region methods for unconstrained derivative-free optimization. These methods ... Page 2
Zhu, Lingxue
2012-01-01T23:59:59.000Z
In this paper, which is part II in a series of two, the pre-asymptotic error analysis of the continuous interior penalty finite element method (CIP-FEM) and the FEM for the Helmholtz equation in two and three dimensions is continued. While part I contained results on the linear CIP-FEM and FEM, the present part deals with approximation spaces of order $p \\ge 1$. By using a modified duality argument, pre-asymptotic error estimates are derived for both methods under the condition of $\\frac{kh}{p}\\le C_0\\big(\\frac{p}{k}\\big)^{\\frac{1}{p+1}}$, where $k$ is the wave number, $h$ is the mesh size, and $C_0$ is a constant independent of $k, h, p$, and the penalty parameters. It is shown that the pollution errors of both methods in $H^1$-norm are $O(k^{2p+1}h^{2p})$ if $p=O(1)$ and are $O\\Big(\\frac{k}{p^2}\\big(\\frac{kh}{\\sigma p}\\big)^{2p}\\Big)$ if the exact solution $u\\in H^2(\\Om)$ which coincide with existent dispersion analyses for the FEM on Cartesian grids. Here $\\si$ is a constant independent of $k, h, p$, and t...
A BAYESIAN APPROACH TO DERIVING AGES OF INDIVIDUAL FIELD WHITE DWARFS
O'Malley, Erin M. [Department of Physics and Astronomy, Siena College, Loudonville, NY 12211 (United States); Von Hippel, Ted [Department of Physical Sciences, Embry-Riddle Aeronautical University, Daytona Beach, FL 32114 (United States); Van Dyk, David A., E-mail: ted.vonhippel@erau.edu, E-mail: dvandyke@imperial.ac.uk [Statistics Section, Department of Mathematics, Imperial College London, SW7 2AZ (United Kingdom)
2013-09-20T23:59:59.000Z
We apply a self-consistent and robust Bayesian statistical approach to determine the ages, distances, and zero-age main sequence (ZAMS) masses of 28 field DA white dwarfs (WDs) with ages of approximately 4-8 Gyr. Our technique requires only quality optical and near-infrared photometry to derive ages with <15% uncertainties, generally with little sensitivity to our choice of modern initial-final mass relation. We find that age, distance, and ZAMS mass are correlated in a manner that is too complex to be captured by traditional error propagation techniques. We further find that the posterior distributions of age are often asymmetric, indicating that the standard approach to deriving WD ages can yield misleading results.
E-Print Network 3.0 - anthraquinone derivatives derived Sample...
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
derived Page: << < 1 2 3 4 5 > >> 1 Langmuir Films of Anthracene Derivatives on Liquid Mercury I: Symmetric Molecules Department of Physics, Bar-Ilan UniVersity, Ramat-Gan 52900,...
A self-checking fiber optic dosimeter for monitoring common errors in brachytherapy applications
Yin, Y.; Lambert, J.; Yang, S.; McKenzie, D. R.; Jackson, M.; Suchowerska, N. [Physics School, University of Sydney, New South Wales 2006 (Australia); Physics School, University of Sydney, New South Wales 2006 (Australia) and Department of Radiation Oncology, Royal Prince Alfred Hospital, New South Wales 2050 (Australia); Physics School, University of Sydney, New South Wales 2006 (Australia); Department of Radiation Oncology, Royal Prince Alfred Hospital, New South Wales 2050 (Australia); Physics School, University of Sydney, New South Wales 2006 (Australia) and Department of Radiation Oncology, Royal Prince Alfred Hospital, New South Wales 2050 (Australia)
2009-07-15T23:59:59.000Z
Scintillation dosimetry with optical fiber readout [fiber optic dosimetry (FOD)] requires accurate measurement of light intensity. It is therefore vulnerable to loss of calibration if any changes occur in the efficiency of the optical pathway between the scintillator and the light detector. The authors show in this article that common types of errors that arise during clinical use for brachytherapy applications can be quantified using a light emitting diode to stimulate the scintillator, the so-called LED-FOD method, in an integrated and easy-to-use control unit that incorporates a compact peripheral component interconnect extension for instrumentation. Common sources of error include bending and mechanical compression of the fiber optic components and changes in the temperature of the scintillator. The authors show that the method can detect all the common errors studied in this work and that different types of errors can result in different correlations between the LED stimulated signal and the brachytherapy source signal. For a single-type error the LED-FOD can be used easily for system diagnosis and validation with the possibility to correct the dosimeter reading if the correlation between the LED stimulated signal and the brachytherapy source signal can be defined. For more complex errors, resulting from two or more errors occurring simultaneously, the LED-FOD method can also allow the clinician to make a judgment on the reliability of the dosimeter reading. This self-checking method can enhance the clinical robustness of the FOD for achieving accurate dose control.
Modern Palliative Radiation Treatment: Do Complexity and Workload Contribute to Medical Errors?
D'Souza, Neil, E-mail: neil.dsouza@sunnybrook.ca [Department of Radiation Oncology, University of Toronto, Toronto, Ontario (Canada) [Department of Radiation Oncology, University of Toronto, Toronto, Ontario (Canada); Odette Cancer Centre, Sunnybrook Health Sciences Centre, Toronto, Ontario (Canada); Holden, Lori [Department of Radiation Oncology, University of Toronto, Toronto, Ontario (Canada) [Department of Radiation Oncology, University of Toronto, Toronto, Ontario (Canada); Odette Cancer Centre, Sunnybrook Health Sciences Centre, Toronto, Ontario (Canada); Robson, Sheila [Odette Cancer Centre, Sunnybrook Health Sciences Centre, Toronto, Ontario (Canada)] [Odette Cancer Centre, Sunnybrook Health Sciences Centre, Toronto, Ontario (Canada); Mah, Kathy; Di Prospero, Lisa; Wong, C. Shun; Chow, Edward; Spayne, Jacqueline [Department of Radiation Oncology, University of Toronto, Toronto, Ontario (Canada) [Department of Radiation Oncology, University of Toronto, Toronto, Ontario (Canada); Odette Cancer Centre, Sunnybrook Health Sciences Centre, Toronto, Ontario (Canada)
2012-09-01T23:59:59.000Z
Purpose: To examine whether treatment workload and complexity associated with palliative radiation therapy contribute to medical errors. Methods and Materials: In the setting of a large academic health sciences center, patient scheduling and record and verification systems were used to identify patients starting radiation therapy. All records of radiation treatment courses delivered during a 3-month period were retrieved and divided into radical and palliative intent. 'Same day consultation, planning and treatment' was used as a proxy for workload and 'previous treatment' and 'multiple sites' as surrogates for complexity. In addition, all planning and treatment discrepancies (errors and 'near-misses') recorded during the same time frame were reviewed and analyzed. Results: There were 365 new patients treated with 485 courses of palliative radiation therapy. Of those patients, 128 (35%) were same-day consultation, simulation, and treatment patients; 166 (45%) patients had previous treatment; and 94 (26%) patients had treatment to multiple sites. Four near-misses and 4 errors occurred during the audit period, giving an error per course rate of 0.82%. In comparison, there were 10 near-misses and 5 errors associated with 1100 courses of radical treatment during the audit period. This translated into an error rate of 0.45% per course. An association was found between workload and complexity and increased palliative therapy error rates. Conclusions: Increased complexity and workload may have an impact on palliative radiation treatment discrepancies. This information may help guide the necessary recommendations for process improvement for patients who require palliative radiation therapy.
The Theory of Quaternion Matrix Derivatives
Dongpo Xu; Danilo P. Mandic
2014-10-01T23:59:59.000Z
A systematic theory is introduced for calculating the derivatives of quaternion matrix function with respect to quaternion matrix variables. The proposed methodology is equipped with the matrix product rule and chain rule and it is able to handle both analytic and nonanalytic functions. This corrects a flaw in the existing methods, that is, the incorrect use of the traditional product rule. In the framework introduced, the derivatives of quaternion matrix functions can be calculated directly without the differential of this function. Key results are summarized in tables. Several examples show how the quaternion matrix derivatives can be used as an important tool for solving problems related to signal processing.
Scher, Aaron David
2005-08-29T23:59:59.000Z
In this thesis, two separate research topics are undertaken both in the general area of compact RF/microwave circuit design. The first topic involves characterizing the parasitic effects and error due to unused post-production tuning bars...
Gross Error Detection in Chemical Plants and Refineries for On-Line Optimization
Pike, Ralph W.
Gross Error Detection in Chemical Plants and Refineries for On-Line Optimization Xueyu Chen, Derya) British Petroleum Applications mainly crude units in refineries and ethylene plants #12;Companies
Efficient error correction for speech systems using constrained re-recognition
Yu, Gregory T
2008-01-01T23:59:59.000Z
Efficient error correction of recognition output is a major barrier in the adoption of speech interfaces. This thesis addresses this problem through a novel correction framework and user interface. The system uses constraints ...
Error analysis of motion transmission mechanisms : design of a parabolic solar trough
Koniski, Cyril (Cyril A.)
2009-01-01T23:59:59.000Z
This thesis presents the error analysis pertaining to the design of an innovative solar trough for use in solar thermal energy generation fields. The research was a collaborative effort between Stacy Figueredo from Prof. ...
Benestad, R E
2013-01-01T23:59:59.000Z
Comment on Scafetta, Nicola. 'Discussion on Common Errors in Analyzing Sea Level Accelerations, Solar Trends and Global Warming.' arXiv:1305.2812 (May 13, 2013a). doi:10.5194/prp-1-37-2013.
Reward Prediction Error Signaling in Posterior Dorsomedial Striatum Is Action Specific
Ogawa, Masaaki
Neural correlates of reward prediction errors (RPEs) have been found in dorsal striatum. Such signals may be important for updating associative action representations within striatum. In order that the appropriate ...
Minimizing Actuator-Induced Residual Error in Active Space Telescope Primary Mirrors
. Smith, David W. Miller September 2010 SSL #12-10 #12;#12;Minimizing Actuator-Induced Residual Error in Active Space Telescope Primary Mirrors Matthew W. Smith, David W. Miller September 2010 SSL #12
Error Field Correction in DIII-D Ohmic Plasmas With Either Handedness
Jong-Kyu Park, Michael J. Schaffer, Robert J. La Haye,Timothy J. Scoville and Jonathan E. Menard
2011-05-16T23:59:59.000Z
Error field correction results in DIII-D plasmas are presented in various configurations. In both left-handed and right-handed plasma configurations, where the intrinsic error fields become different due to the opposite helical twist (handedness) of the magnetic field, the optimal error correction currents and the toroidal phases of internal(I)-coils are empirically established. Applications of the Ideal Perturbed Equilibrium Code to these results demonstrate that the field component to be minimized is not the resonant component of the external field, but the total field including ideal plasma responses. Consistency between experiment and theory has been greatly improved along with the understanding of ideal plasma responses, but non-ideal plasma responses still need to be understood to achieve the reliable predictability in tokamak error field correction.
Evans, Jonah Wy
2007-09-17T23:59:59.000Z
procedure, participants misidentified the tracks of 12 species as otter. Inaccurate identification of indirect signs is a likely source of error in wildlife studies. I recommend that observer skill in identification of indirect signs be measured in order...
Errors in Estimating River Discharge from Remote Sensing based on Manning's Equation
Washington at Seattle, University of
Â° inclination; all rivers, lakes, reservoirs observed at least twice every 22 days. Will measure reach (2008) estimated errors in width resulting from water coherence time effects (due to wind and turbulence
Locatelli, R.
A modelling experiment has been conceived to assess the impact of transport model errors on methane emissions estimated in an atmospheric inversion system. Synthetic methane observations, obtained from 10 different model ...
Combined wavelet video coding and error control for internet streaming and multicast
Chu, Tianli
2002-01-01T23:59:59.000Z
In the past several years, advances in Internet video streaming have been tremendous. Originally designed without error protection, Receiver-driven layered multicast (RLM) has proved to be a very effective scheme for scalable video multicast. Though...
Demonstration Integrated Knowledge-Based System for Estimating Human Error Probabilities
Auflick, Jack L.
1999-04-21T23:59:59.000Z
Human Reliability Analysis (HRA) is currently comprised of at least 40 different methods that are used to analyze, predict, and evaluate human performance in probabilistic terms. Systematic HRAs allow analysts to examine human-machine relationships, identify error-likely situations, and provide estimates of relative frequencies for human errors on critical tasks, highlighting the most beneficial areas for system improvements. Unfortunately, each of HRA's methods has a different philosophical approach, thereby producing estimates of human error probabilities (HEPs) that area better or worse match to the error likely situation of interest. Poor selection of methodology, or the improper application of techniques can produce invalid HEP estimates, where that erroneous estimation of potential human failure could have potentially severe consequences in terms of the estimated occurrence of injury, death, and/or property damage.
Error and uncertainty in estimates of Reynolds stress using ADCP in an energetic ocean state
Rapo, Mark Andrew.
2006-01-01T23:59:59.000Z
(cont.) To that end, the space-time correlations of the error, turbulence, and wave processes are developed and then utilized to find the extent to which the environmental and internal processing parameters contribute to ...
On the evaluation of human error probabilities for post-initiating events
Presley, Mary R
2006-01-01T23:59:59.000Z
Quantification of human error probabilities (HEPs) for the purpose of human reliability assessment (HRA) is very complex. Because of this complexity, the state of the art includes a variety of HRA models, each with its own ...
SCM Forcing Data Derived from NWP Analyses
DOE Data Explorer [Office of Scientific and Technical Information (OSTI)]
Jakob, Christian
Forcing data, suitable for use with single column models (SCMs) and cloud resolving models (CRMs), have been derived from NWP analyses for the ARM (Atmospheric Radiation Measurement) Tropical Western Pacific (TWP) sites of Manus Island and Nauru.
Background and Motivation Biomass derived syngas contains
Das, Suman
Background and Motivation · Biomass derived syngas contains: CO, H2, small hydrocarbons, H2S prepared by SEA contain smaller metal particle sizes and have higher dispersion · The quantity of reducible
Modelling and Simulating of Rain Derivatives
Modelling and Simulating of Rain Derivatives Master thesis Cathrin van Emmerich Supervisor, Februar 2005 Cathrin van Emmerich i #12;Table of Contents Table of Contents ii 1 Introduction 1 2
Entanglement entropy in higher derivative holography
Arpan Bhattacharyya; Apratim Kaviraj; Aninda Sinha
2013-08-16T23:59:59.000Z
We consider holographic entanglement entropy in higher derivative gravity theories. Recently Lewkowycz and Maldacena arXiv:1304.4926 have provided a method to derive the equations for the entangling surface from first principles. We use this method to compute the entangling surface in four derivative gravity. Certain interesting differences compared to the two derivative case are pointed out. For Gauss-Bonnet gravity, we show that in the regime where this method is applicable, the resulting equations coincide with proposals in the literature as well as with what follows from considerations of the stress tensor on the entangling surface. Finally we demonstrate that the area functional in Gauss-Bonnet holography arises as a counterterm needed to make the Euclidean action free of power law divergences.
Comprehension by Derivation Douglas R. Smith
Comprehension by Derivation Douglas R. Smith Kestrel Institute 3260 Hillview Avenue Palo Alto, California 94304 USA smith@kestrel.edu Abstract We argue that to comprehend a software system is to have
Direct synthesis of pyridine and pyrimidine derivatives
Hill, Matthew D. (Matthew Dennis)
2008-01-01T23:59:59.000Z
I. Synthesis of Substituted Pyridine Derivatives via the Ruthenium-Catalyzed Cycloisomerization of 3-Azadienynes. The two-step conversion of various N-vinyl and N-aryl amides to the corresponding substituted pyridines and ...
Fraden, Seth
is at the following: U.S. Customs and Border Protection Boston, Logan International Airport Terminal E, Lower Level the ISSO at isso@brandeis.edu or 781 736 3480. If you are still at the airport while you notice the error the airport, you do not necessarily have to return to the same airport you flew through but you do need to go
Validation of error estimators and superconvergence by a computer-based approach
Upadhyay, Chandra Shekhar
2012-06-07T23:59:59.000Z
ESTIMATORS FOR PATCHWISE UNIFORM MESHES 5. 1 The methodology for checking the estimators. 5. 2 Numerical study of the estimators. . . . . 5. 3 Major results. 58 . . 58 65 160 CHAPTER VI STUDY OF THE ERROR ESTIMATORS FOR GENERAL MESHES . . 6. 1... Definition of the robustness index 6. 2 The computational methodology for general grids . . 6. 3 Numerical studies of robustness of various error estimators . . 6. 4 Major results. vn Page 164 164 166 167 , 198 CHAPTER VII STUDY OF SUPERCONVERGENCE...
Eldred, Michael Scott; Subia, Samuel Ramirez; Neckels, David; Hopkins, Matthew Morgan; Notz, Patrick K.; Adams, Brian M.; Carnes, Brian; Wittwer, Jonathan W.; Bichon, Barron J.; Copps, Kevin D.
2006-10-01T23:59:59.000Z
This report documents the results for an FY06 ASC Algorithms Level 2 milestone combining error estimation and adaptivity, uncertainty quantification, and probabilistic design capabilities applied to the analysis and design of bistable MEMS. Through the use of error estimation and adaptive mesh refinement, solution verification can be performed in an automated and parameter-adaptive manner. The resulting uncertainty analysis and probabilistic design studies are shown to be more accurate, efficient, reliable, and convenient.
Design consistency and driver error as reflected by driver workload and accident rates
Wooldridge, Mark Douglas
1992-01-01T23:59:59.000Z
DESIGN CONSISTENCY AND DRIVER ERROR AS REFLECTED BY DRIVER WORKLOAD AND ACCIDENT RATES A Thesis by MARK DOUGLAS WOOLDRIDGE Approved as to style and content by: Daniel B. Fambro (Chair of Committee) Raymond A. Krammes (Member) Olga J.... Pendleton (Member) James T. P. Yao (Head of Department) May 1992 ABSTRACT Design Consistency and Driver Error as Reflected by Driver Workload and Accident Rates (May 1992) Mark Douglas Wooldridge, B. S. , Texas A&M University Chair of Advisory...
Effects and Correction of Closed Orbit Magnet Errors in the SNS Ring
Bunch, S.C.; Holmes, J.
2004-01-01T23:59:59.000Z
We consider the effect and correction of three types of orbit errors in SNS: quadrupole displacement errors, dipole displacement errors, and dipole field errors. Using the ORBIT beam dynamics code, we focus on orbit deflection of a standard pencil beam and on beam losses in a high intensity injection simulation. We study the correction of these orbit errors using the proposed system of 88 (44 horizontal and 44 vertical) ring beam position monitors (BPMs) and 52 (24 horizontal and 28 vertical) dipole corrector magnets. Correction is carried out numerically by adjusting the kick strengths of the dipole corrector magnets to minimize the sum of the squares of the BPM signals for the pencil beam. In addition to using the exact BPM signals as input to the correction algorithm, we also consider the effect of random BPM signal errors. For all three types of error and for perturbations of individual magnets, the correction algorithm always chooses the three-bump method to localize the orbit displacement to the region between the magnet and its adjacent correctors. The values of the BPM signals resulting from specified settings of the dipole corrector kick strengths can be used to set up the orbit response matrix, which can then be applied to the correction in the limit that the signals from the separate errors add linearly. When high intensity calculations are carried out to study beam losses, it is seen that the SNS orbit correction system, even with BPM uncertainties, is sufficient to correct losses to less than 10-4 in nearly all cases, even those for which uncorrected losses constitute a large portion of the beam.
Quantification of model mismatch errors of the dynamic energy distribution in a stirred-tank reactor
Kimmich, Mark Raymond
1987-01-01T23:59:59.000Z
QUANTIFICATION OF MODEL MISMATCH ERRORS OF THE DYNAMIC ENERGY DISTRIBUTION IN A STIRRED- TANK REACTOR A Thesis by MARK RAYMOND KIMMICH Submitted to the Graduate College of Texas AkM University in partial fulfillment of the requirement... for the degree of MASTER OF SCIENCE August 198i Major Subject: Chemical Engineering QUANTIFICATION OF MODEL MISMATCH ERRORS OF THE DYNAMIC ENERGY DISTRIBUTION IN A STIRRED-TANK REACTOR A Thesis by MARK RAYMOND KIMMICH Approved as to style and content by...
Feldman, Joel
() M, f( i )2 - LM + 1 4 L2 2 f( i ) Â± 1 2 L 2 f( i )2 + LM + 1 4 L2 2 Hence 1 2 f( i )2 - 1 2 LM2 + 1 8 L2 3 area of sector #i 1 2 f( i )2 + 1 2 LM2 + 1 8 L2 3 and the total area A obeys n i=1 1 2 f( i )2 - 1 2 LM2 + 1 8 L2 3 A n i=1 1 2 f( i )2 + 1 2 LM2 + 1 8 L2 3 1 2 n i=1 f( i )2 - 1 2 nLM
Applying lessons learned to enhance human performance and reduce human error for ISS operations
Nelson, W.R.
1998-09-01T23:59:59.000Z
A major component of reliability, safety, and mission success for space missions is ensuring that the humans involved (flight crew, ground crew, mission control, etc.) perform their tasks and functions as required. This includes compliance with training and procedures during normal conditions, and successful compensation when malfunctions or unexpected conditions occur. A very significant issue that affects human performance in space flight is human error. Human errors can invalidate carefully designed equipment and procedures. If certain errors combine with equipment failures or design flaws, mission failure or loss of life can occur. The control of human error during operation of the International Space Station (ISS) will be critical to the overall success of the program. As experience from Mir operations has shown, human performance plays a vital role in the success or failure of long duration space missions. The Department of Energy`s Idaho National Engineering and Environmental Laboratory (INEEL) is developed a systematic approach to enhance human performance and reduce human errors for ISS operations. This approach is based on the systematic identification and evaluation of lessons learned from past space missions such as Mir to enhance the design and operation of ISS. This paper describes previous INEEL research on human error sponsored by NASA and how it can be applied to enhance human reliability for ISS.
Orifice-meter measurement errors caused by gas-system pulsations can be controlled
Gegg, D. (Arkla Energy Resources, Shreveport, LA (US))
1989-10-16T23:59:59.000Z
Pulsation-induced inaccuracies in orifice meter gas-flow measurement can be eliminated or at least better controlled. In today's increasingly competitive gas-supply marketplace, such errors can no longer be ignored. In some instances, pulsations have caused errors of 30-50% in volumes. Policies and procedures of Arkla Energy Resources, Shreveport, La., reflect current efforts to combat the problem. Orifice-meter pulsation error can be divided into three categories: those associated with the primary element itself (across the flange taps); those in the gauge line/manifold/transducer system;; and those within the recording and analysis system. Each category is significant in that total meter error contains all three types (box). While it would be desirable to segregate the problems and solve them independently, it is very difficult to do. It becomes apparent quickly that all three occur in many cases. The focus of this discussion is on primary-element errors; if these are reduced, other errors tend to be minimal.
Entanglement-Assisted Quantum Error-Correcting Codes with Imperfect Ebits
Ching-Yi Lai; Todd A. Brun
2012-04-04T23:59:59.000Z
The scheme of entanglement-assisted quantum error-correcting (EAQEC) codes assumes that the ebits of the receiver are error-free. In practical situations, errors on these ebits are unavoidable, which diminishes the error-correcting ability of these codes. We consider two different versions of this problem. We first show that any (nondegenerate) standard stabilizer code can be transformed into an EAQEC code that can correct errors on the qubits of both sender and receiver. These EAQEC codes are equivalent to standard stabilizer codes, and hence the decoding techniques of standard stabilizer codes can be applied. Several EAQEC codes of this type are found to be optimal. In a second scheme, the receiver uses a standard stabilizer code to protect the ebits, which we call a "combination code." The performances of different quantum codes are compared in terms of the channel fidelity over the depolarizing channel. We give a formula for the channel fidelity over the depolarizing channel (or any Pauli error channel), and show that it can be efficiently approximated by a Monte Carlo calculation. Finally, we discuss the tradeoff between performing extra entanglement distillation and applying an EAQEC code with imperfect ebits.
Leahy, Frank Byron
1999-01-01T23:59:59.000Z
for the area 1000 km within the IMPACT detector location. A site error correction curve was used to correct for azimuthal errors in bearings to CG stroke locations. Errors were reduced to less than 4° in most cases. The IMPACT detector uses a crossed...
Detecting bit-flip errors in a logical qubit using stabilizer measurements
D. Ristè; S. Poletto; M. -Z. Huang; A. Bruno; V. Vesterinen; O. -P. Saira; L. DiCarlo
2014-11-20T23:59:59.000Z
Quantum data is susceptible to decoherence induced by the environment and to errors in the hardware processing it. A future fault-tolerant quantum computer will use quantum error correction (QEC) to actively protect against both. In the smallest QEC codes, the information in one logical qubit is encoded in a two-dimensional subspace of a larger Hilbert space of multiple physical qubits. For each code, a set of non-demolition multi-qubit measurements, termed stabilizers, can discretize and signal physical qubit errors without collapsing the encoded information. Experimental demonstrations of QEC to date, using nuclear magnetic resonance, trapped ions, photons, superconducting qubits, and NV centers in diamond, have circumvented stabilizers at the cost of decoding at the end of a QEC cycle. This decoding leaves the quantum information vulnerable to physical qubit errors until re-encoding, violating a basic requirement for fault tolerance. Using a five-qubit superconducting processor, we realize the two parity measurements comprising the stabilizers of the three-qubit repetition code protecting one logical qubit from physical bit-flip errors. We construct these stabilizers as parallelized indirect measurements using ancillary qubits, and evidence their non-demolition character by generating three-qubit entanglement from superposition states. We demonstrate stabilizer-based quantum error detection (QED) by subjecting a logical qubit to coherent and incoherent bit-flip errors on its constituent physical qubits. While increased physical qubit coherence times and shorter QED blocks are required to actively safeguard quantum information, this demonstration is a critical step toward larger codes based on multiple parity measurements.
Huang, Weidong
2011-01-01T23:59:59.000Z
This paper presents the general equation to calculate the standard deviation of reflected ray error from optical error through geometry optics, applying the equation to calculate the standard deviation of reflected ray error for 8 kinds of solar concentrated reflector, provide typical results. The results indicate that the slope errors in two direction is transferred to any one direction of the focus ray when the incidence angle is more than 0 for solar trough and heliostats reflector; for point focus Fresnel lens, point focus parabolic glass mirror, line focus parabolic galss mirror, the error transferring coefficient from optical to focus ray will increase when the rim angle increase; for TIR-R concentrator, it will decrease; for glass heliostat, it relates to the incidence angle and azimuth of the reflecting point. Keywords: optic error, standard deviation, refractive ray error, concentrated solar collector
Moment Methods for Exotic Volatility Derivatives
Albanese, Claudio
2007-01-01T23:59:59.000Z
The latest generation of volatility derivatives goes beyond variance and volatility swaps and probes our ability to price realized variance and sojourn times along bridges for the underlying stock price process. In this paper, we give an operator algebraic treatment of this problem based on Dyson expansions and moment methods and discuss applications to exotic volatility derivatives. The methods are quite flexible and allow for a specification of the underlying process which is semi-parametric or even non-parametric, including state-dependent local volatility, jumps, stochastic volatility and regime switching. We find that volatility derivatives are particularly well suited to be treated with moment methods, whereby one extrapolates the distribution of the relevant path functionals on the basis of a few moments. We consider a number of exotics such as variance knockouts, conditional corridor variance swaps, gamma swaps and variance swaptions and give valuation formulas in detail.
Shi, Di-Fu; Qian, Bao-Liang; Wang, Hong-Gang; Li, Wei [College of Opto-electric Science and Engineering, National University of Defense Technology, Changsha 410073 (China)] [College of Opto-electric Science and Engineering, National University of Defense Technology, Changsha 410073 (China)
2013-12-15T23:59:59.000Z
Field analysis method is used to derive the dispersion relation of rising-sun magnetron with sectorial and rectangular cavities. This dispersion relation is then extended to the general case in which the rising-sun magnetron can be with multi-group cavities of different shapes and sizes, and from which the dispersion relations of conventional magnetron, rising-sun magnetron, and magnetron-like device can be obtained directly. The results show that the relative errors between the theoretical and simulation values of the dispersion relation are less than 3%, the relative errors between the theoretical and simulation values of the cutoff frequencies of ? mode are less than 2%. In addition, the influences of each structure parameter of the magnetron on the cutoff frequency of ? mode and on the mode separation are investigated qualitatively and quantitatively, which may be of great interest to designing a frequency tuning magnetron.
High ethanol producing derivatives of Thermoanaerobacter ethanolicus
Ljungdahl, L.G.; Carriera, L.H.
1983-05-24T23:59:59.000Z
Derivatives of the newly discovered microorganism Thermoanaerobacter ethanolicus which under anaerobic and thermophilic conditions continuously ferment substrates such as starch, cellobiose, glucose, xylose and other sugars to produce recoverable amounts of ethanol solving the problem of fermentations yielding low concentrations of ethanol using the parent strain of the microorganism Thermoanaerobacter ethanolicus are disclosed. These new derivatives are ethanol tolerant up to 10% (v/v) ethanol during fermentation. The process includes the use of an aqueous fermentation medium, containing the substrate at a substrate concentration greater than 1% (w/v).
High ethanol producing derivatives of Thermoanaerobacter ethanolicus
Ljungdahl, Lars G. (Athens, GA); Carriera, Laura H. (Athens, GA)
1983-01-01T23:59:59.000Z
Derivatives of the newly discovered microorganism Thermoanaerobacter ethanolicus which under anaerobic and thermophilic conditions continuously ferment substrates such as starch, cellobiose, glucose, xylose and other sugars to produce recoverable amounts of ethanol solving the problem of fermentations yielding low concentrations of ethanol using the parent strain of the microorganism Thermoanaerobacter ethanolicus are disclosed. These new derivatives are ethanol tolerant up to 10% (v/v) ethanol during fermentation. The process includes the use of an aqueous fermentation medium, containing the substrate at a substrate concentration greater than 1% (w/v).
Extreme density-driven delocalization error for a model solvated-electron system
Johnson, Erin R., E-mail: ejohnson29@ucmerced.edu; Otero-de-la-Roza, A., E-mail: aoterodelaroza@ucmerced.edu; Dale, Stephen G., E-mail: sdale@ucmerced.edu [Chemistry and Chemical Biology, School of Natural Sciences, University of California, Merced, 5200 North Lake Road, Merced, California 95343 (United States)
2013-11-14T23:59:59.000Z
Delocalization (or charge-transfer) error is one of the scarce but spectacular failures of density-functional theory. It is particularly apparent in extensively delocalized molecules, and manifests in the calculation of bandgaps, reaction barriers, and dissociation limits. Even though delocalization error is always present in the self-consistent electron density, the differences from reference densities are often quite subtle and the error tends to be driven by the exchange-correlation energy expression. In this article, we propose a model system (the Kevan model) where approximate density functionals predict dramatically different charge distributions because of delocalization error. The model system consists of an electron trapped in a water hexamer and is a finite representation of an experimentally observed class of solids: electrides. The Kevan model is of fundamental interest because it allows the estimation of charge transfer error without recourse to fractional charge calculations, but our results are also relevant in the context of the modeling of confined electrons in density-functional theory.
Error Compensation of Single-Qubit Gates in a Surface Electrode Ion Trap Using Composite Pulses
Emily Mount; Chingiz Kabytayev; Stephen Crain; Robin Harper; So-Young Baek; Geert Vrijsen; Steven Flammia; Kenneth R. Brown; Peter Maunz; Jungsang Kim
2015-04-06T23:59:59.000Z
The trapped atomic ion qubits feature desirable properties for use in a quantum computer such as long coherence times (Langer et al., 2005), high qubit measurement fidelity (Noek et al., 2013), and universal logic gates (Home et al., 2009). The quality of quantum logic gate operations on trapped ion qubits has been limited by the stability of the control fields at the ion location used to implement the gate operations. For this reason, the logic gates utilizing microwave fields (Brown et al., 2011; Shappert et al., 2013; Harty et al., 2014) have shown gate fidelities several orders of magnitude better than those using laser fields (Knill et al., 2008; Benhelm et al., 2008; Ballance et al., 2014). Here, we demonstrate low-error single-qubit gates performed using stimulated Raman transitions on an ion qubit trapped in a microfabricated chip trap. Gate errors are measured using a randomized benchmarking protocol (Knill et al., 2008; Wallman et al., 2014; Magesan et al., 2012), where amplitude error in the control beam is compensated using various pulse sequence techniques (Wimperis, 1994; Low et al., 2014). Using B2 compensation (Wimperis, 1994), we demonstrate single qubit gates with an average error per randomized Clifford group gate of $3.6(3)\\times10^{-4}$. We also show that compact palindromic pulse compensation sequences (PD$n$) (Low et al., 2014) compensate for amplitude errors as designed.
A Comparison of Forecast Error Generators for Modeling Wind and Load Uncertainty
Lu, Ning; Diao, Ruisheng; Hafen, Ryan P.; Samaan, Nader A.; Makarov, Yuri V.
2013-07-25T23:59:59.000Z
This paper presents four algorithms to generate random forecast error time series. The performance of four algorithms is compared. The error time series are used to create real-time (RT), hour-ahead (HA), and day-ahead (DA) wind and load forecast time series that statistically match historically observed forecasting data sets used in power grid operation to study the net load balancing need in variable generation integration studies. The four algorithms are truncated-normal distribution models, state-space based Markov models, seasonal autoregressive moving average (ARMA) models, and a stochastic-optimization based approach. The comparison is made using historical DA load forecast and actual load values to generate new sets of DA forecasts with similar stoical forecast error characteristics (i.e., mean, standard deviation, autocorrelation, and cross-correlation). The results show that all methods generate satisfactory results. One method may preserve one or two required statistical characteristics better the other methods, but may not preserve other statistical characteristics as well compared with the other methods. Because the wind and load forecast error generators are used in wind integration studies to produce wind and load forecasts time series for stochastic planning processes, it is sometimes critical to use multiple methods to generate the error time series to obtain a statistically robust result. Therefore, this paper discusses and compares the capabilities of each algorithm to preserve the characteristics of the historical forecast data sets.
S. D. Bloom; D. A. Dale; R. Cool; K. Dupczak; C. Miller; A. Haugsjaa; C. Peters; M. Tornikoski; P. Wallace; M. Pierce
2004-04-02T23:59:59.000Z
We present the most recent results of an optical survey of the position error contours ("error boxes") of unidentified high energy gamma-ray sources.
PRICING ENERGY DERIVATIVES BY LINEAR PROGRAMMING: TOLLING AGREEMENT CONTRACTS
Uryasev, Stanislav
PRICING ENERGY DERIVATIVES BY LINEAR PROGRAMMING: TOLLING AGREEMENT CONTRACTS Valeriy Ryabchenko for pricing energy derivatives known as tolling agreement contracts. The pricing problem is reduced energy derivatives. The problem of pricing such contracts falls into the class of multiple optimal
Carbide-derived carbons - From porous networks to nanotubes and...
Office of Scientific and Technical Information (OSTI)
Carbide-derived carbons - From porous networks to nanotubes and graphene Re-direct Destination: Carbide-derived carbons (CDCs) are a large family of carbon materials derived from...
Wave function derivation of the JIMWLK equation
Alexey V. Popov
2008-12-16T23:59:59.000Z
Using the stationary lightcone perturbation theory, we propose the complete and careful derivation the JIMWLK equation. We show that the rigorous treatment requires the knowledge of a boosted wave function with second order accuracy. Previous wave function approaches are incomplete and implicitly used the time ordered perturbation theory, which requires a usage of an external target field.
Benchmarking Derivative-Free Optimization Algorithms
2008-05-13T23:59:59.000Z
Science, U.S. Department of Energy, under Contract DE-AC02-06CH11357 ... deterministic simulation based on solving the equations (for example, ... most ambitious work in this direction [7] is a comparison of six derivative-free optimization ...... [5] Elizabeth D. Dolan and Jorge J. Moré, Benchmarking optimization software.
Derivative Free Optimization Methods for Optimizing Stirrer ...
2007-02-01T23:59:59.000Z
It also gives a comparison of the two optimization algorithms. Key words: Energy, Numerical Optimization, Derivative-Free Optimization,. Computational Fluid ... numerical simulations are needed for the computation of the fluid flow inside the stirrer for ..... Optimization Methods and Software, 20 (2005) 493-508. [6] A. R. Conn ...
High speed point derivative microseismic detector
Uhl, J.E.; Warpinski, N.R.; Whetten, E.B.
1998-06-30T23:59:59.000Z
A high speed microseismic event detector constructed in accordance with the present invention uses a point derivative comb to quickly and accurately detect microseismic events. Compressional and shear waves impinging upon microseismic receiver stations disposed to collect waves are converted into digital data and analyzed using a point derivative comb including assurance of quiet periods prior to declaration of microseismic events. If a sufficient number of quiet periods have passed, the square of a two point derivative of the incoming digital signal is compared to a trip level threshold exceeding the determined noise level to declare a valid trial event. The squaring of the derivative emphasizes the differences between noise and signal, and the valid event is preferably declared when the trip threshold has been exceeded over a temporal comb width to realize a comb over a given time period. Once a trial event has been declared, the event is verified through a spatial comb, which applies the temporal event comb to additional stations. The detector according to the present invention quickly and accurately detects initial compressional waves indicative of a microseismic event which typically exceed the ambient cultural noise level by a small amount, and distinguishes the waves from subsequent larger amplitude shear waves. 9 figs.
High speed point derivative microseismic detector
Uhl, James Eugene (Albuquerque, NM); Warpinski, Norman Raymond (Albuquerque, NM); Whetten, Ernest Blayne (Albuquerque, NM)
1998-01-01T23:59:59.000Z
A high speed microseismic event detector constructed in accordance with the present invention uses a point derivative comb to quickly and accurately detect microseismic events. Compressional and shear waves impinging upon microseismic receiver stations disposed to collect waves are converted into digital data and analyzed using a point derivative comb including assurance of quiet periods prior to declaration of microseismic events. If a sufficient number of quiet periods have passed, the square of a two point derivative of the incoming digital signal is compared to a trip level threshold exceeding the determined noise level to declare a valid trial event. The squaring of the derivative emphasizes the differences between noise and signal, and the valid event is preferably declared when the trip threshold has been exceeded over a temporal comb width to realize a comb over a given time period. Once a trial event has been declared, the event is verified through a spatial comb, which applies the temporal event comb to additional stations. The detector according to the present invention quickly and accurately detects initial compressional waves indicative of a microseismic event which typically exceed the ambient cultural noise level by a small amount, and distinguishes the waves from subsequent larger amplitude shear waves.
Biofuels and bio-products derived from
Ginzel, Matthew
NEED Biofuels and bio- products derived from lignocellulosic biomass (plant materials) are part improve the energy and carbon efficiencies of biofuels production from a barrel of biomass using chemical and thermal catalytic mechanisms. The Center for Direct Catalytic Conversion of Biomass to Biofuels IMPACT
Regular-expression derivatives reexamined SCOTT OWENS
Strickland, Stevie
. For regular sets of strings, i.e., sets defined by regular expressions (REs), the derivative is also a regular is elegant and easily supports extended regular expressions; i.e., REs extended with Boolean operations expressions, but since the extensions are conservative (i.e., regular languages are closed under Boolean
Higher Derivative D-brane Couplings
Guo, Guangyu
2012-10-19T23:59:59.000Z
supersymmetry. In the third part, we obtain the higher derivative D-brane action by using both linearized T-duality and string disc amplitude computation. We evaluate disc amplitude of one R-R field C^(p-3) and two NS-NS fields in the presence of a single Dp...
Derivation of a Stochastic Neutron Transport Equation
Edward J. Allen
2010-04-14T23:59:59.000Z
Stochastic difference equations and a stochastic partial differential equation (SPDE) are simultaneously derived for the time-dependent neutron angular density in a general three-dimensional medium where the neutron angular density is a function of position, direction, energy, and time. Special cases of the equations are given such as transport in one-dimensional plane geometry with isotropic scattering and transport in a homogeneous medium. The stochastic equations are derived from basic principles, i.e., from the changes that occur in a small time interval. Stochastic difference equations of the neutron angular density are constructed, taking into account the inherent randomness in scatters, absorptions, and source neutrons. As the time interval decreases, the stochastic difference equations lead to a system of Ito stochastic differential equations (SDEs). As the energy, direction, and position intervals decrease, an SPDE is derived for the neutron angular density. Comparisons between numerical solutions of the stochastic difference equations and independently formulated Monte Carlo calculations support the accuracy of the derivations.
Isatin Derivatives as Inhibitors of Microtubule Assembly
Beckman, Karen
2008-09-04T23:59:59.000Z
This thesis describes the rationale, design, and syntheses of derivatives of isatin (1-H-indole-2,3-dione). Isatin was identified, during a high throughput screen of 10,000 compounds, as a potential scaffold for microtubule-destabilizing agents...
HUMAN ERROR QUANTIFICATION USING PERFORMANCE SHAPING FACTORS IN THE SPAR-H METHOD
Harold S. Blackman; David I. Gertman; Ronald L. Boring
2008-09-01T23:59:59.000Z
This paper describes a cognitively based human reliability analysis (HRA) quantification technique for estimating the human error probabilities (HEPs) associated with operator and crew actions at nuclear power plants. The method described here, Standardized Plant Analysis Risk-Human Reliability Analysis (SPAR-H) method, was developed to aid in characterizing and quantifying human performance at nuclear power plants. The intent was to develop a defensible method that would consider all factors that may influence performance. In the SPAR-H approach, calculation of HEP rates is especially straightforward, starting with pre-defined nominal error rates for cognitive vs. action-oriented tasks, and incorporating performance shaping factor multipliers upon those nominal error rates.
Ricciardi, S; Natoli, P; Polenta, G; Baccigalupi, C; Salerno, E; Kayabol, K; Bedini, L; De Zotti, G; 10.1111/j.1365-2966.2010.16819.x
2010-01-01T23:59:59.000Z
We present a data analysis pipeline for CMB polarization experiments, running from multi-frequency maps to the power spectra. We focus mainly on component separation and, for the first time, we work out the covariance matrix accounting for errors associated to the separation itself. This allows us to propagate such errors and evaluate their contributions to the uncertainties on the final products.The pipeline is optimized for intermediate and small scales, but could be easily extended to lower multipoles. We exploit realistic simulations of the sky, tailored for the Planck mission. The component separation is achieved by exploiting the Correlated Component Analysis in the harmonic domain, that we demonstrate to be superior to the real-space application (Bonaldi et al. 2006). We present two techniques to estimate the uncertainties on the spectral parameters of the separated components. The component separation errors are then propagated by means of Monte Carlo simulations to obtain the corresponding contributi...
Error correcting code with chip kill capability and power saving enhancement
Gara, Alan G. (Mount Kisco, NY); Chen, Dong (Croton On Husdon, NY); Coteus, Paul W. (Yorktown Heights, NY); Flynn, William T. (Rochester, MN); Marcella, James A. (Rochester, MN); Takken, Todd (Brewster, NY); Trager, Barry M. (Yorktown Heights, NY); Winograd, Shmuel (Scarsdale, NY)
2011-08-30T23:59:59.000Z
A method and system are disclosed for detecting memory chip failure in a computer memory system. The method comprises the steps of accessing user data from a set of user data chips, and testing the user data for errors using data from a set of system data chips. This testing is done by generating a sequence of check symbols from the user data, grouping the user data into a sequence of data symbols, and computing a specified sequence of syndromes. If all the syndromes are zero, the user data has no errors. If one of the syndromes is non-zero, then a set of discriminator expressions are computed, and used to determine whether a single or double symbol error has occurred. In the preferred embodiment, less than two full system data chips are used for testing and correcting the user data.
Joint Seasonal ARMA Approach for Modeling of Load Forecast Errors in Planning Studies
Hafen, Ryan P.; Samaan, Nader A.; Makarov, Yuri V.; Diao, Ruisheng; Lu, Ning
2014-04-14T23:59:59.000Z
To make informed and robust decisions in the probabilistic power system operation and planning process, it is critical to conduct multiple simulations of the generated combinations of wind and load parameters and their forecast errors to handle the variability and uncertainty of these time series. In order for the simulation results to be trustworthy, the simulated series must preserve the salient statistical characteristics of the real series. In this paper, we analyze day-ahead load forecast error data from multiple balancing authority locations and characterize statistical properties such as mean, standard deviation, autocorrelation, correlation between series, time-of-day bias, and time-of-day autocorrelation. We then construct and validate a seasonal autoregressive moving average (ARMA) model to model these characteristics, and use the model to jointly simulate day-ahead load forecast error series for all BAs.
Fade-resistant forward error correction method for free-space optical communications systems
Johnson, Gary W. (Livermore, CA); Dowla, Farid U. (Castro Valley, CA); Ruggiero, Anthony J. (Livermore, CA)
2007-10-02T23:59:59.000Z
Free-space optical (FSO) laser communication systems offer exceptionally wide-bandwidth, secure connections between platforms that cannot other wise be connected via physical means such as optical fiber or cable. However, FSO links are subject to strong channel fading due to atmospheric turbulence and beam pointing errors, limiting practical performance and reliability. We have developed a fade-tolerant architecture based on forward error correcting codes (FECs) combined with delayed, redundant, sub-channels. This redundancy is made feasible though dense wavelength division multiplexing (WDM) and/or high-order M-ary modulation. Experiments and simulations show that error-free communications is feasible even when faced with fades that are tens of milliseconds long. We describe plans for practical implementation of a complete system operating at 2.5 Gbps.
Hodge, B. M.; Lew, D.; Milligan, M.
2013-01-01T23:59:59.000Z
Load forecasting in the day-ahead timescale is a critical aspect of power system operations that is used in the unit commitment process. It is also an important factor in renewable energy integration studies, where the combination of load and wind or solar forecasting techniques create the net load uncertainty that must be managed by the economic dispatch process or with suitable reserves. An understanding of that load forecasting errors that may be expected in this process can lead to better decisions about the amount of reserves necessary to compensate errors. In this work, we performed a statistical analysis of the day-ahead (and two-day-ahead) load forecasting errors observed in two independent system operators for a one-year period. Comparisons were made with the normal distribution commonly assumed in power system operation simulations used for renewable power integration studies. Further analysis identified time periods when the load is more likely to be under- or overforecast.
Maps for general open quantum systems and a theory of linear quantum error correction
A. Shabani; D. A. Lidar
2009-02-14T23:59:59.000Z
We show that quantum subdynamics of an open quantum system can always be described by a Hermitian map, irrespective of the form of the initial total system state. Since the theory of quantum error correction was developed based on the assumption of completely positive (CP) maps, we present a generalized theory of linear quantum error correction, which applies to any linear map describing the open system evolution. In the physically relevant setting of Hermitian maps, we show that the CP-map based version of quantum error correction theory applies without modifications. However, we show that a more general scenario is also possible, where the recovery map is Hermitian but not CP. Since non-CP maps have non-positive matrices in their range, we provide a geometric characterization of the positivity domain of general linear maps. In particular, we show that this domain is convex, and that this implies a simple algorithm for finding its boundary.
Effects of Magnet Errors in the ILC 14 mrad Extraction Line
Toprek, Dragan; /VINCA Inst. Nucl. Sci., Belgrade; Nosochkov, Yuri; /SLAC
2009-05-08T23:59:59.000Z
The ILC baseline extraction line is designed for 14 mrad horizontal crossing angle between e{sup +} and e{sup -} colliding beams at Interaction Point (IP). The extraction optics in the Interaction Region (IR) includes a detector integrated dipole field (anti-DID) to reduce orbit perturbation caused by the detector solenoid and minimize detector background. This paper presents a study of random field and alignment errors in the extraction magnets, compensation of the induced orbit perturbation, and effects of errors on extraction beam power loss. The results are obtained for the baseline ILC energy of 500 GeV center-of-mass and three options of beam parameters.
Lobach, Iryna
2009-05-15T23:59:59.000Z
) are binary and probability of disease is known. Environmental variable is measured with error with misclassi cation probabilities pr(W = 0jX = 1) = 0:20 and pr(W = 1jX = 0) = 0:10. The results are based on a simulation study with 500 replications for 1000... variant (G), and environmental covariate (X) are binary and probability of disease is unknown. Environmen- tal variable is measured with error with misclassi cation probabilities pr(W = 0jX = 1) = 0:20 and pr(W = 1jX = 0) = 0:10. The results are based on a...
Comer, K.; Gaddy, C.D.; Seaver, D.A.; Stillwell, W.G.
1985-01-01T23:59:59.000Z
The US Nuclear Regulatory Commission and Sandia National Laboratories sponsored a project to evaluate psychological scaling techniques for use in generating estimates of human error probabilities. The project evaluated two techniques: direct numerical estimation and paired comparisons. Expert estimates were found to be consistent across and within judges. Convergent validity was good, in comparison to estimates in a handbook of human reliability. Predictive validity could not be established because of the lack of actual relative frequencies of error (which will be a difficulty inherent in validation of any procedure used to estimate HEPs). Application of expert estimates in probabilistic risk assessment and in human factors is discussed.
Estimating rock properties in two phase petroleum reservoirs: an error analysis
Paul, Anthony Ian
1983-01-01T23:59:59.000Z
. 60x10 1. 16x10 4. 16x10 1. 01 9. 13x10 1. 51 2. 07%10 6. 51x10 5. 42x10 8. 18 3. 16x10 26 function approximated in the tt parameter space gives a mean prediction error which is essentially zero. However, using the permeability estimates obtained.... G ver (Mem er) . L. Curry . D. Ho land ( d of Department) December 1983 ABSTRACT Estimating Rock Properties In Two Phase Petroleum Reservoirs: An Error Analysis. (December 1983) Anthony Ian Paul B. Sc. , Imperial College, London University...
Errors Due to Forcing and Physics Can Be Distinguished by Intercomparing Simulations
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
AFDC Printable Version Share this resource Send a link to EERE: Alternative Fuels Data Center Home Page to someone by E-mail Share EERE: Alternative Fuels Data Center Home Page on Facebook Tweet about EERE: Alternative Fuels Data Center Home Page on Twitter Bookmark EERE:1 First Use of Energy for All Purposes (Fuel and Nonfuel),Feet) Year Jan Feb Mar Apr MayAtmospheric Optical Depth7-1D: Vegetation ProposedUsing ZirconiaPolicy and Assistance100 ton Stanat rollinggovErrors ERROR 404 - URL
Interaction of coal-derived synthesis gas impurities with solid...
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
coal-derived synthesis gas impurities with solid oxide fuel cell metallic components. Interaction of coal-derived synthesis gas impurities with solid oxide fuel cell metallic...
Low-Emissions Burner Technology using Biomass-Derived Liquid...
Broader source: Energy.gov (indexed) [DOE]
Emissions Burner Technology using Biomass-Derived Liquid Fuels Low-Emissions Burner Technology using Biomass-Derived Liquid Fuels This factsheet describes a project that developed...
Progress toward Biomass and Coal-Derived Syngas Warm Cleanup...
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
Progress toward Biomass and Coal-Derived Syngas Warm Cleanup: Proof-of-Concept Process Demonstration of Multicontaminant Removal Progress toward Biomass and Coal-Derived Syngas...
Exploring Hydrogen Generation from Biomass-Derived Sugar and...
Office of Energy Efficiency and Renewable Energy (EERE) Indexed Site
Exploring Hydrogen Generation from Biomass-Derived Sugar and Sugar Alcohols to Reduce Costs Exploring Hydrogen Generation from Biomass-Derived Sugar and Sugar Alcohols to Reduce...
Agenda for the Derived Liquids to Hydrogen Distributed Reforming...
Broader source: Energy.gov (indexed) [DOE]
Agenda for the Derived Liquids to Hydrogen Distributed Reforming Working Group (BILIWG) Hydrogen Production Technical Team Research Review Agenda for the Derived Liquids to...
Precise attention filters for Weber contrast derived from centroid estimations
Sperling, George
Precise attention filters for Weber contrast derived from centroid estimations Department attention filters for Weber contrast derived from centroid estimations. Journal of Vision, 10(10):20, 1
acetic acid derivatives: Topics by E-print Network
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
affects the values of the derivatives. Our constructions use randomness extractors and expander graphs. We study our derivatives in a more general setting than Arora et al. In...
amino acetal derivative: Topics by E-print Network
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
affects the values of the derivatives. Our constructions use randomness extractors and expander graphs. We study our derivatives in a more general setting than Arora et al. In...
Automation of Nested Matrix and Derivative Operations Robert Kalaba
Tesfatsion, Leigh
Automation of Nested Matrix and Derivative Operations Robert Kalaba Departments of Electrical of expressions involving nested matrix and derivative operations. The need to differentiate such expressions
Phosphine oxide derivatives as hosts for blue phosphors: A joint...
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
oxide derivatives as hosts for blue phosphors: A joint theoretical and experimental study of their electronic Phosphine oxide derivatives as hosts for blue phosphors: A joint...
Goutte, Stéphane; Russo, Francesco
2012-01-01T23:59:59.000Z
Given a process with independent increments $X$ (not necessarily a martingale) and a large class of square integrable r.v. $H=f(X_T)$, $f$ being the Fourier transform of a finite measure $\\mu$, we provide explicit Kunita-Watanabe and F\\"ollmer-Schweizer decompositions. The representation is expressed by means of two significant maps: the expectation and derivative operators related to the characteristics of $X$. We also provide an explicit expression for the variance optimal error when hedging the claim $H$ with underlying process $X$. Those questions are motivated by finding the solution of the celebrated problem of global and local quadratic risk minimization in mathematical finance.
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
AFDC Printable Version Share this resource Send a link to EERE: Alternative Fuels Data Center Home Page to someone by E-mail Share EERE: Alternative Fuels Data Center Home Page on Facebook Tweet about EERE: Alternative Fuels Data Center Home Page on Twitter Bookmark EERE:1 First Use of Energy for All Purposes (Fuel and Nonfuel),Feet) Year Jan Feb Mar Apr MayAtmosphericNuclear Security AdministrationcontrollerNanocrystallineForeign Object Damage 3Nationalmimic keyProcessing
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
AFDC Printable Version Share this resource Send a link to EERE: Alternative Fuels Data Center Home Page to someone by E-mail Share EERE: Alternative Fuels Data Center Home Page on Facebook Tweet about EERE: Alternative Fuels Data Center Home Page on Twitter Bookmark EERE:1 First Use of Energy for All Purposes (Fuel and Nonfuel),Feet) Year Jan Feb Mar Apr May JunDatastreamsmmcrcalgovInstrumentsruc DocumentationP-Series toESnet4:Epitaxial Thin Film XRD Epitaxial Thin FilmErinwrites out the
Kalapurakal, John A., E-mail: j-kalapurakal@northwestern.edu [Department of Radiation Oncology, Northwestern University Feinberg School of Medicine, Chicago, Illinois (United States); Zafirovski, Aleksandar; Smith, Jeffery; Fisher, Paul; Sathiaseelan, Vythialingam [Department of Radiation Oncology, Northwestern University Feinberg School of Medicine, Chicago, Illinois (United States)] [Department of Radiation Oncology, Northwestern University Feinberg School of Medicine, Chicago, Illinois (United States); Barnard, Cynthia [Department of Quality Strategies, Northwestern Memorial Hospital, Chicago, Illinois (United States)] [Department of Quality Strategies, Northwestern Memorial Hospital, Chicago, Illinois (United States); Rademaker, Alfred W. [Department of Preventive Medicine, Northwestern University Feinberg School of Medicine, Chicago, Illinois (United States)] [Department of Preventive Medicine, Northwestern University Feinberg School of Medicine, Chicago, Illinois (United States); Rave, Nick [Department of Physicians Services, Northwestern Memorial Hospital, Chicago, Illinois (United States)] [Department of Physicians Services, Northwestern Memorial Hospital, Chicago, Illinois (United States); Mittal, Bharat B. [Department of Radiation Oncology, Northwestern University Feinberg School of Medicine, Chicago, Illinois (United States)] [Department of Radiation Oncology, Northwestern University Feinberg School of Medicine, Chicago, Illinois (United States)
2013-06-01T23:59:59.000Z
Purpose: This report describes the value of a voluntary error reporting system and the impact of a series of quality assurance (QA) measures including checklists and timeouts on reported error rates in patients receiving radiation therapy. Methods and Materials: A voluntary error reporting system was instituted with the goal of recording errors, analyzing their clinical impact, and guiding the implementation of targeted QA measures. In response to errors committed in relation to treatment of the wrong patient, wrong treatment site, and wrong dose, a novel initiative involving the use of checklists and timeouts for all staff was implemented. The impact of these and other QA initiatives was analyzed. Results: From 2001 to 2011, a total of 256 errors in 139 patients after 284,810 external radiation treatments (0.09% per treatment) were recorded in our voluntary error database. The incidence of errors related to patient/tumor site, treatment planning/data transfer, and patient setup/treatment delivery was 9%, 40.2%, and 50.8%, respectively. The compliance rate for the checklists and timeouts initiative was 97% (P<.001). These and other QA measures resulted in a significant reduction in many categories of errors. The introduction of checklists and timeouts has been successful in eliminating errors related to wrong patient, wrong site, and wrong dose. Conclusions: A comprehensive QA program that regularly monitors staff compliance together with a robust voluntary error reporting system can reduce or eliminate errors that could result in serious patient injury. We recommend the adoption of these relatively simple QA initiatives including the use of checklists and timeouts for all staff to improve the safety of patients undergoing radiation therapy in the modern era.
Triamine chelants, their derivatives, complexes and conjugates
Troutner, D.E.; John, C.S.; Pillai, M.R.A.
1995-03-07T23:59:59.000Z
A group of functionalized triamine chelants and their derivatives that form complexes with radioactive metal ions are disclosed. The complexes can be covalently attached to a protein or an antibody or antibody fragment and used for therapeutic and/or diagnostic purposes. The chelants are of the formula, as shown in the accompanying diagrams, wherein n, m, R, R{sup 1}, R{sup 2} and L are defined in the specification.
A stochastic derivation of the geodesic rule
Nikos Kalogeropoulos
2006-02-09T23:59:59.000Z
We argue that the geodesic rule, for global defects, is a consequence of the randomness of the values of the Goldstone field $\\phi$ in each causally connected volume. As these volumes collide and coalescence, $\\phi$ evolves by performing a random walk on the vacuum manifold $\\mathcal{M}$. We derive a Fokker-Planck equation that describes the continuum limit of this process. Its fundamental solution is the heat kernel on $\\mathcal{M}$, whose leading asymptotic behavior establishes the geodesic rule.
Triamine chelants, their derivatives, complexes and conjugates
Troutner, David E. (Phoenixville, PA); John, Christy S. (Gaithersburg, MD); Pillai, Maroor R. A. (Vashi, IN)
1995-01-01T23:59:59.000Z
A group of functionalized triamine chelants and their derivatives that form complexes with radioactive metal ions are disclosed. The complexes can be covalently attached to a protein or an antibody or antibody fragment and used for therapeutic and/or diagnostic purposes. The chelants are of the formula: ##STR1## wherein n, m, R, R.sup.1, R.sup.2 and L are defined in the specification.
Isolation and Analysis of Optimization Errors MICKEY R. BOYD AND DAVID B. WHALLEY
Whalley, David
features of the optimization viewer include reverse viewing (or undoing) of transformations and the ability, an optimization error isolator is presented that can automatÂ ically determine the first transformation during and after each transformation perÂ formed by the optimizer. One can easily examine the invalid
21st Congress on Learning Disabilities THERE ARE PHONETIC PATTERNS IN VOWEL SUBSTITUTION ERRORS
IN TEXTS WRITTEN BY PERSONS WITH DYSLEXIA Luz Rello1 & Joaquim Llisterri2 1 Universitat Pompeu Fabra, 2 written by Spanish speaking children with dyslexia, establish a typology of errors found in the corpus with dyslexia provides several kinds of relevant information, both from a theoretical and from an applied
Assignment on Algebra for Coding Theory EE512: Error Control Coding
Thangaraj, Andrew
of polynomials with coefficients from Z18. Find a, b Z18 (a = 1) such that 2x(x + 1) = 2x(ax + b) in Z18[xAssignment on Algebra for Coding Theory EE512: Error Control Coding Questions marked (Q) or (F
A Time{Dependent Born{Oppenheimer Approximation with Exponentially Small Error Estimates
Joye, Alain
A Time{Dependent Born{Oppenheimer Approximation with Exponentially Small Error Estimates George A{ Oppenheimer approximation for molecular quantum mechanics. We study molecular systems whose electron masses. The small parameter that governs the approximation is the usual Born{Oppenheimer expansion parameter #15
A Time--Dependent Born--Oppenheimer Approximation with Exponentially Small Error Estimates
Hagedorn, George A.
A Time--Dependent Born--Oppenheimer Approximation with Exponentially Small Error Estimates George A--dependent Born-- Oppenheimer approximation for molecular quantum mechanics. We study molecular systems whose parameter that governs the approximation is the usual Born--Oppenheimer expansion parameter #, where # 4
A Time--Dependent Born--Oppenheimer Approximation with Exponentially Small Error Estimates
A Time--Dependent Born--Oppenheimer Approximation with Exponentially Small Error Estimates George A accurate time--dependent Born-- Oppenheimer approximation for molecular quantum mechanics. We study is the usual Born--Oppenheimer expansion parameter ffl, where ffl 4 is the ratio of the electron mass divided
A Time--Independent Born--Oppenheimer Approximation with Exponentially Accurate Error Estimates
A Time--Independent Born--Oppenheimer Approximation with Exponentially Accurate Error Estimates, we construct quasimodes that are exponentially accurate in the square of the Born--Oppenheimer that governs the approximation is the usual Born--Oppenheimer parameter #, where # 4 is the electron mass
Discretization error estimation and exact solution generation using the method of nearby problems.
Sinclair, Andrew J. (Auburn University Auburn, AL); Raju, Anil (Auburn University Auburn, AL); Kurzen, Matthew J. (Virginia Tech Blacksburg, VA); Roy, Christopher John (Virginia Tech Blacksburg, VA); Phillips, Tyrone S. (Virginia Tech Blacksburg, VA)
2011-10-01T23:59:59.000Z
The Method of Nearby Problems (MNP), a form of defect correction, is examined as a method for generating exact solutions to partial differential equations and as a discretization error estimator. For generating exact solutions, four-dimensional spline fitting procedures were developed and implemented into a MATLAB code for generating spline fits on structured domains with arbitrary levels of continuity between spline zones. For discretization error estimation, MNP/defect correction only requires a single additional numerical solution on the same grid (as compared to Richardson extrapolation which requires additional numerical solutions on systematically-refined grids). When used for error estimation, it was found that continuity between spline zones was not required. A number of cases were examined including 1D and 2D Burgers equation, the 2D compressible Euler equations, and the 2D incompressible Navier-Stokes equations. The discretization error estimation results compared favorably to Richardson extrapolation and had the advantage of only requiring a single grid to be generated.
New Constructions of Non-Adaptive and Error-Tolerance Pooling Designs
Ngo, Hung Q.
probability Â¢ of having a positive. Group testing strategies can also be either adaptive or non in test outcomes. Test errors could be either Â¤Â¦Â¥ Â§ , i.e. a negative pool is identified as positive another column. A Â¡ -disjunct Â©FP matrix S can be used to design a non-adaptive group testing algorithm
New Constructions of NonAdaptive and ErrorTolerance Pooling Designs
Ngo, Hung Q.
fix some probability p of having a positive. Group testing strategies can also be either adaptive to design a nonÂadaptive group testing algorithm on n items by associating the columns with the items). If there are no more than d positives and the test outcomes are errorÂfree, then it is easy to see that the test outcom
Detecting arbitrary quantum errors via stabilizer measurements on a sublattice of the surface code
A. D. Córcoles; Easwar Magesan; Srikanth J. Srinivasan; Andrew W. Cross; M. Steffen; Jay M. Gambetta; Jerry M. Chow
2014-10-23T23:59:59.000Z
To build a fault-tolerant quantum computer, it is necessary to implement a quantum error correcting code. Such codes rely on the ability to extract information about the quantum error syndrome while not destroying the quantum information encoded in the system. Stabilizer codes are attractive solutions to this problem, as they are analogous to classical linear codes, have simple and easily computed encoding networks, and allow efficient syndrome extraction. In these codes, syndrome extraction is performed via multi-qubit stabilizer measurements, which are bit and phase parity checks up to local operations. Previously, stabilizer codes have been realized in nuclei, trapped-ions, and superconducting qubits. However these implementations lack the ability to perform fault-tolerant syndrome extraction which continues to be a challenge for all physical quantum computing systems. Here we experimentally demonstrate a key step towards this problem by using a two-by-two lattice of superconducting qubits to perform syndrome extraction and arbitrary error detection via simultaneous quantum non-demolition stabilizer measurements. This lattice represents a primitive tile for the surface code, which is a promising stabilizer code for scalable quantum computing. Furthermore, we successfully show the preservation of an entangled state in the presence of an arbitrary applied error through high-fidelity syndrome measurement. Our results bolster the promise of employing lattices of superconducting qubits for larger-scale fault-tolerant quantum computing.
Simulating the Transverse Ising Model on a Quantum Computer: Error Correction with the Surface Code
Geller, Michael R.
of a fault- tolerant quantum computer. The surface code approach has one of the highest known tolerable error of the surface code is four orders of magnitude higher than the concatenation code, building a quantum computer implementation, a new approach to building a quantum computer with the surface code (which is a kind
Effects of Spectral Error in Efficiency Measurements of GaInAs-Based Concentrator Solar Cells
Osterwald, C. R.; Wanlass, M. W.; Moriarty, T.; Steiner, M. A.; Emery, K. A.
2014-03-01T23:59:59.000Z
This technical report documents a particular error in efficiency measurements of triple-absorber concentrator solar cells caused by incorrect spectral irradiance -- specifically, one that occurs when the irradiance from unfiltered, pulsed xenon solar simulators into the GaInAs bottom subcell is too high. For cells designed so that the light-generated photocurrents in the three subcells are nearly equal, this condition can cause a large increase in the measured fill factor, which, in turn, causes a significant artificial increase in the efficiency. The error is readily apparent when the data under concentration are compared to measurements with correctly balanced photocurrents, and manifests itself as discontinuities in plots of fill factor and efficiency versus concentration ratio. In this work, we simulate the magnitudes and effects of this error with a device-level model of two concentrator cell designs, and demonstrate how a new Spectrolab, Inc., Model 460 Tunable-High Intensity Pulsed Solar Simulator (T-HIPSS) can mitigate the error.
ADB-Trees: Controlling the Error of Time-Critical Collision Detection Jan Klein Gabriel Zachmann
Behnke, Sven
ADB-Trees: Controlling the Error of Time-Critical Collision Detection Jan Klein Gabriel Zachmann courtesy of VW and BMW) Abstract We present a novel framework for hierarchical col- lision detection by aug- menting AABB trees and present performance mea- surements and comparisons with a very fast previ
Tullos, Desiree
DOWNSTREAM CHANNEL CHANGES AFTER A SMALL DAM REMOVAL: USING AERIAL PHOTOS AND MEASUREMENT ERROR to assess downstream channel changes associated with a small dam removal. The Brownsville Dam, a 2.1 m tall downstream from the dam and in an upstream control reach using aerial photos (19942008) and in the field
Error estimation and anisotropic mesh refinement for 3d laminar aerodynamic flow simulations
Hartmann, Ralf
Error estimation and anisotropic mesh refinement for 3d laminar aerodynamic flow simulations Tobias Leichta,b , Ralf Hartmann,a,b aInstitute of Aerodynamics and Flow Technology, DLR (German Aerospace Center-dimensional laminar aerodynamic flow simulations. The optimal order symmetric interior penalty discontinuous Galerkin
Semyonov, A. A.; Druzhaev, A. A., E-mail: andreydruzhaev@gmail.com; Schukin, N. V. [National Research Nuclear University MEPhI (Russian Federation)
2014-12-15T23:59:59.000Z
A method for reconstructing the neutron field in a reactor with consideration of the spatial correlation of the cross-section value error was analyzed. It was shown that this method is more accurate than the classical approach to reconstruction. An efficient way of using this technique was proposed. The efficiency for the RBMK critical test facility was estimated.
Characterization and removal of errors due to local magnetic anomalies in directional drilling of Geophysics, Colorado School of Mines Summary Directional drilling has evolved over the last few decades utilizes a technique known as magnetic Measurement While Drilling (MWD). Vector measurements of geomagnetic
A nonideal error-field response model for strongly shaped tokamak R. Fitzpatrick
Fitzpatrick, Richard
damping at Alfvén and/or sound wave resonances inside the plasma. The nonresonant component magnetic flux-surfaces.7 Such chains severely degrade global energy confinement.8 Fortunately, the highly the relationship between the har- monic content of an error-field and the associated locking torque that is exerted
Automated Diagnosis of Product-line Configuration Errors on Feature Models
Schmidt, Douglas C.
Automated Diagnosis of Product-line Configuration Errors on Feature Models Jules White and Doulas Feature models are widely used to model software product-line (SPL) variability. SPL variants are config Introduction Current trends and challenges. Software Product- Lines (SPLs) are a technique for creating
Error Analysis of Heat Transfer for Finned-Tube Heat-Exchanger Text-Board
Chen, Y.; Zhang, J.
2006-01-01T23:59:59.000Z
In order to reduce the measurement error of heat transfer in water and air side for finned-tube heat-exchanger as little as possible, and design a heat-exchanger test-board measurement system economically, based on the principle of test-board system...
Ultrasonic thickness measurements on corroded steel members: a statistical analysis of error
Konen, Keith Forman
1999-01-01T23:59:59.000Z
with measuring the wall thickness of a corroded tubular member and 2) determining how the strength calculations are affected by an error in a wall thickness measurement. This thesis is based on the first phase of a research project funded by Mineral Management...
GRAVITY ERROR COMPENSATION USING SECOND-ORDER GAUSS-MARKOV PROCESSES
Born, George
AAS 11-502 GRAVITY ERROR COMPENSATION USING SECOND-ORDER GAUSS-MARKOV PROCESSES Jason M. Leonard the use of a second-order Gauss-Markov process to compensate for higher order spherical harmonic gravity an improvement in POD through the use of a second-order Gauss-Markov process (GMP2) for modeling J3 gravity
Allowing Errors in Speech over Wireless LANs Ian Chakeres, Hui Dong, Elizabeth Belding-Royer
Belding-Royer, Elizabeth M.
in call quality. By forcing error-free reception of speech, scarce bandwidth and energy are unnecessarily- dard are experiencing widespread deployment. Currently most devices with WLAN connectivity are laptops. In congested networks, fewer retransmissions reduce the channel usage and result in increased packet delivery
Speech enhancement using a minimum mean-square error short-time spectral modulation magnitude In this paper we investigate the enhancement of speech by applying MMSE short-time spectral magnitude estimation on the quality of enhanced speech, and find that this method works better with speech uncertainty. Finally we
Kunin, Victor; Engelbrektson, Anna; Ochman, Howard; Hugenholtz, Philip
2009-08-01T23:59:59.000Z
Massively parallel pyrosequencing of the small subunit (16S) ribosomal RNA gene has revealed that the extent of rare microbial populations in several environments, the 'rare biosphere', is orders of magnitude higher than previously thought. One important caveat with this method is that sequencing error could artificially inflate diversity estimates. Although the per-base error of 16S rDNA amplicon pyrosequencing has been shown to be as good as or lower than Sanger sequencing, no direct assessments of pyrosequencing errors on diversity estimates have been reported. Using only Escherichia coli MG1655 as a reference template, we find that 16S rDNA diversity is grossly overestimated unless relatively stringent read quality filtering and low clustering thresholds are applied. In particular, the common practice of removing reads with unresolved bases and anomalous read lengths is insufficient to ensure accurate estimates of microbial diversity. Furthermore, common and reproducible homopolymer length errors can result in relatively abundant spurious phylotypes further confounding data interpretation. We suggest that stringent quality-based trimming of 16S pyrotags and clustering thresholds no greater than 97% identity should be used to avoid overestimates of the rare biosphere.
Evaluating the Capability of Compilers and Tools to Detect Serial and Parallel Run-time Errors
Luecke, Glenn R.
, Elizabeth Kleiman, Olga Weiss, Andre Wehe, Melissa Yahya # Iowa State University's High Performance of system software to detect and issue error messages that help programmers quickly fix serial and parallel using the new system software to rted.project@iastate.edu so they can be posted on this web site. II
A Survey of Systems for Detecting Serial Run-Time Errors
Luecke, Glenn R.
Performance Computing Group Glenn R. Luecke, James Coyle, Jim Hoekstra, Marina Kraeva, Ying Li, Olga Taborskaia, and Yanmei Wang {grl, jjc, hoekstra, kraeva, yingli, olga, yanmei}@iastate.edu Revised February-commercial software products to detect serial run-time errors in C and C++ programs, to issue meaningful messages
Time-and Space-efficient Error Calculation for Multiresolution Direct Volume Rendering
Linsen, Lars
Gyulassy1 , Lars Linsen1,2 , and Bernd Hamann1 1 Institute for Data Analysis and Visualization (IDAV such that the error made when visualizing the subset (compared to a visualization of the full data set) is smaller datasets is a common task used in many fields, including medicine, physics, and other sciences. Complexity
Algorithms for BoundedError Correlation of High Dimensional Data in Microarray Experiments
Szpankowski, Wojciech
Algorithms for BoundedÂError Correlation of High Dimensional Data in Microarray Experiments #3 of clustering continuous valued data has been well studied in literature. Its application to microarray analysis are relatively unÂ explored. An instance of analysis of discreteÂattributed data arises in detecting co
Sources of Uncertainty and Error in the Simulation of Flow in Porous Media
New York at Stoney Brook, State University of
that we have observations of oil production from an early time period (the \\past") and we use production is known but the geology which gives rise to it is not. The errors or uncertainty in predictions of future pro- duction depend on geological, physical, and numerical parameters, e.g. the geology
Error Control Coding in Low-Power Wireless Sensor Networks: When is ECC
Howard, Sheryl
Error Control Coding in Low-Power Wireless Sensor Networks: When is ECC Energy-Efficient? Sheryl L. Interference is not considered; it would lower dCR. Analog decoders are shown to be the most energy-efficient at the decoder outweighs the transmitted power savings due to using ECC, then ECC would not be energy-efficient
Improving the Accuracy of Industrial Robots by offline Compensation of Joints Errors
Paris-Sud XI, UniversitÃ© de
level pose accuracy and to achieve a good path tracking. Unfortunately industrial robots are designed are concidered [4]. In [5], authors have worked on modeling the Cartesian compliance of an industrial robotImproving the Accuracy of Industrial Robots by offline Compensation of Joints Errors Adel Olabi
Judd, Kevin
1973; Oran and Boris 1987; Murray 1989; Gershenfeld 1999). Weather and climate prediction models, which to the initial conditions, which is a major source of uncertainty in Numerical Weather Prediction (NWP; eTime Step Sensitivity of Nonlinear Atmospheric Models: Numerical Convergence, Truncation Error
Bounds on the Error Probability of ML Decoding for Block and Turbo-Block Codes
Sason, Igal
Bounds on the Error Probability of ML Decoding for Block and Turbo-Block Codes Igal Sason 32000, Israel March 1999 Abstract The performance of either structured or random turbo-block codes on the other. We focus here on uniformly interleaved and parallel concatenated turbo-Hamming codes
Avoiding, Finding and Fixing Spreadsheet Errors -A Survey of Automated Approaches for Spreadsheet QA
Jannach, Dietmar
reviews and audits, to better tool support for the user during the spreadsheet development processAvoiding, Finding and Fixing Spreadsheet Errors - A Survey of Automated Approaches for Spreadsheet Graz, Austria Abstract Spreadsheet programs can be found everywhere in organizations and they are used
Error Bounds for the Large Time Step Glimm Scheme Applied to Scalar Conservation Laws 1
for the Glimm scheme under the restriction of Courant numbers up to 1=2. 1. Introduction The large time step solution of (1) generated from the LTS Glimm scheme, #1;x the spatial mesh length and #1;t the time stepError Bounds for the Large Time Step Glimm Scheme Applied to Scalar Conservation Laws 1 Jinyang
Using Energy-Efficient Overlays to Reduce Packet Error Rates in Wireless Ad-Hoc Networks
Khan, Bilal
the problem of how to balance Western Michigan University, MI. John Jay College of Criminal Justice, City1 Using Energy-Efficient Overlays to Reduce Packet Error Rates in Wireless Ad-Hoc Networks A. Al-Fuqaha G. Ben Brahim M. Guizani B. Khan Abstract-- In this paper we present new energy-efficient tech
Impact of forecasting error on the performance of capacitated multi-item production systems
Xie, Jinxing
Impact of forecasting error on the performance of capacitated multi-item production systems Jinxing multi-item production system under demand uncertainty and a rolling time horizon. The output from parameters, thus improving the performance of production systems. q 2004 Elsevier Ltd. All rights reserved
Mutlu, Onur
Characterizing Application Memory Error Vulnerability to Optimize Datacenter Cost via Heterogeneous--Memory devices represent a key component of datacenter total cost of ownership (TCO), and techniques used, we make three main con- tributions to enable highly-reliable servers at low datacenter cost. First
Improving End of Life Care: an Information Systems Approach to Reducing Medical Errors
Kopec, Danny
, quality of care, end of life, electronic health records Introduction This research will evaluate the waysImproving End of Life Care: an Information Systems Approach to Reducing Medical Errors TAMANG S. a, CUNY, USA b Department of Research, Metropolitan Jewish Health System, NY, USA Abstract. Chronic
Renaut, Rosemary
describe the performance of a solid oxide fuel cell requires the solution of an inverse problem. Two at the electrodeelectrolyte interfaces of solid oxide fuel cells (SOFC) is investigated physically using ElectrochemicalStability and error analysis of the polarization estimation inverse problem for solid oxide fuel
A review of the theory of Coriolis flowmeter measurement errors due to entrained particles
Basse, Nils Plesner
is provided in Table 1. The measurement errors due to compressibility increase with decreasing speed of sound,12]. Nomenclature: a fluid is either a liquid or a gas. A particle can be either a solid or a fluid (gas bubble or liquid droplet). To date, the published bubble theory has dealt with zero particle density combined
Proving the Absence of RunTime Errors in SafetyCritical Avionics Code
Cousot, Patrick
, timetriggered, realtime, safety critical, embedded software as found in earth transportation, nuclearProving the Absence of RunTime Errors in SafetyCritical Avionics Code Patrick Cousot École is not acceptable in safety and mission crit ical applications. An avenue is therefore opened for formal methods
Particle-induced bit errors in high performance fiber optic data links for satellite data management
Marshall, P.W.; Carts, M.A. (Naval Research Lab., Washington, DC (United States) SFA, Inc., Landover, MD (United States)); Dale, C.J. (Naval Research Lab., Washington, DC (United States)); LaBel, K.A. (NASA Goddard Space Flight Center, Greenbelt, MD (United States))
1994-12-01T23:59:59.000Z
Experimental test methods and analysis tools are demonstrated to assess particle-induced bit errors on fiber optic link receivers for satellites. Susceptibility to direct ionization from low LET particles is quantified by analyzing proton and helium ion data as a function of particle LET. Existing single event analysis approaches are shown to apply, with appropriate modifications, to the regime of temporally (rather than spatially) distributed bits, even though the sensitivity to single events exceeds conventional memory technologies by orders of magnitude. The cross-section LET dependence follows a Weibull distribution at data rates from 200 to 1,000 Mbps and at various incident optical power levels. The LET threshold for errors is shown, through both experiment and modeling, to be 0 in all cases. The error cross-section exhibits a strong inverse dependence on received optical power in the LET range where most orbital single events would occur, thus indicating that errors can be minimized by operating links with higher incident optical power. Also, an analytic model is described which incorporates the appropriate physical characteristics of the link as well as the optical and receiver electrical characteristics. Results indicate appropriate steps to assure suitable link performance even in severe particle orbits.
Modeling conically scanning lidar error in complex terrain with WAsP
the vertical wind speed W changes linearly with the downwind position x. This could crudely mimic the flow over to be homogeneous in order to deduce the horizontal wind speed. However, in mountainous or complex terrain this assumption is not valid implying an erroneous wind speed. The magnitude of this error is measured
On the Characterization and Optimization of On-Chip Cache Reliability against Soft Errors
Ziavras, Sotirios G.
--Soft errors induced by energetic particle strikes in on-chip cache memories have become an increasing challenge in designing new generation reliable microprocessors. Previous efforts have exploited information. In this paper, we propose a new framework for conducting comprehensive studies and characterization
Vibrotactile Feedback in Steering Wheel Reduces Navigation Errors during GPS-Guided Car Driving
Basdogan, Cagatay
vibration motors are mounted onto the steering wheel of a driving simulator and driving experiments-based car navigation system to improve the navigation performance of a driver. In [5], vibration motors were auditory noise and distraction exist in the environment, the navigation errors (making a wrong turn
Error analysis due to laser beams misalignment of a double laser self-mixing velocimeter
Tanios, Bendy; Bony, Francis; Bosch, Thierry [CNRS, LAAS, 7 avenue du colonel Roche, F-31400 Toulouse (France) and Univ de Toulouse, UPS, LAAS, F-31400 Toulouse (France); CNRS, LAAS, 7 avenue du colonel Roche, F-31400 Toulouse, France and Univ de Toulouse, INP, LAAS, F-31400 Toulouse (France)
2012-06-13T23:59:59.000Z
In this paper, we present a self-mixing double-head laser diode velocimeter. Analyzes are performed to evaluate the sensitivity to misalignment for this setup and calculate errors due to this misalignment. The analyses and calculations are verified by experimental results.
hal-00119494,version1-10Dec2006 Error structures and parameter estimation
Boyer, Edmond
probabilistic approach we have to know the law of the pair (C, C) or equivalently the law of C and the conditional law of C given C. Thus, the study of error transmission is associated to the calculus of images of probability measures. Unfortunately, the knowledge of the law of C given C by means of experiment
Methods to Speed Up Error Back-Propagation Learning Algorithm DILIP SARKAR
Sarkar, Dilip
Methods to Speed Up Error Back-Propagation Learning Algorithm DILIP SARKAR University of Miami's address: D. Sarkar, Department of Mathematics and Computer Science, University of Miami, Coral Gables, FL Surveys. Vol. 27, No. 4, December 1995 #12;520 q Dilip Sarkar I CONTENTS 1. 2 3. 4. 5. INTRODUCTION
Quantifying Errors Associated with Satellite Sampling of Offshore Wind S.C. Pryor1,2
1 Quantifying Errors Associated with Satellite Sampling of Offshore Wind Speeds S.C. Pryor1,2 , R, Bloomington, IN47405, USA. Tel: 1-812-855-5155. Fax: 1-812-855-1661 Email: spryor@indiana.edu 2 Dept. of Wind an attractive proposition for measuring wind speeds over the oceans because in principle they also offer
Two-point derivative dispersion relations
Erasmo Ferreira; Javier Sesma
2014-03-24T23:59:59.000Z
A new derivation is given for the representation, under certain conditions, of the integral dispersion relations of scattering theory through local forms. The resulting expressions have been obtained through an independent procedure to construct the real part, and consist of new mathematical structures of double infinite summations of derivatives. In this new form the derivatives are calculated at the generic value of the energy $E$ and separately at the reference point $E=m$ that is the lower limit of the integration. This new form may be more interesting in certain circumstances and directly shows the origin of the difficulties in convergence that were present in the old truncated forms called standard-DDR. For all cases in which the reductions of the double to single sums were obtained in our previous work, leading to explicit demonstration of convergence, these new expressions are seen to be identical to the previous ones. We present, as a glossary, the most simplified explicit results for the DDR's in the cases of imaginary amplitudes of forms $(E/m)^\\lambda[\\ln (E/m)]^n$, that cover the cases of practical interest in particle physics phenomenology at high energies. We explicitly study the expressions for the cases with $\\lambda$ negative odd integers, that require identification of cancelation of singularities, and provide the corresponding final results.
Mathur, Anuj
1994-01-01T23:59:59.000Z
shown in Fig. 5. 1a, a = ?, p = 2 (2cr & p). 122 3' Table 5. 4a. Components of the energy-norm of the error in ~s. Square domain shown in Fig. 5. 2, smooth solution u(z?z, ) = A sin xz, sinh zzs, A=10s, p ? 2 127 Table 5. 4b. Pollution.... Pollution-effect for meshes refined locally in an interior subdomain. 4 Convex domain shown in Fig. 5. 7, a = ?, p = 2 (2cr & p) 131 xx Page Table 5. 5b. Pollution-efi'ect for meshes refined locally in an interior subdomain. Pollution factors p...
Doerry, Armin W. (Albuquerque, NM); Heard, Freddie E. (Albuquerque, NM); Cordaro, J. Thomas (Albuquerque, NM)
2010-07-20T23:59:59.000Z
Motion measurement errors that extend beyond the range resolution of a synthetic aperture radar (SAR) can be corrected by effectively decreasing the range resolution of the SAR in order to permit measurement of the error. Range profiles can be compared across the slow-time dimension of the input data in order to estimate the error. Once the error has been determined, appropriate frequency and phase correction can be applied to the uncompressed input data, after which range and azimuth compression can be performed to produce a desired SAR image.
Transformation of spatial and perturbation derivatives of travel time
Cerveny, Vlastislav
Transformation of spatial and perturbation derivatives of travel time at a general interface and perturbation parameters. We derive the explicit equations for transforming these travelÂtime derivatives Hamiltonian function and are applicable to the transformation of travelÂtime derivatives in both isotropic
Derivation of an Applied Nonlinear Schroedinger Equation.
Pitts, Todd Alan; Laine, Mark Richard; Schwarz, Jens; Rambo, Patrick K.; Karelitz, David B.
2015-01-01T23:59:59.000Z
We derive from first principles a mathematical physics model useful for understanding nonlinear optical propagation (including filamentation). All assumptions necessary for the development are clearly explained. We include the Kerr effect, Raman scattering, and ionization (as well as linear and nonlinear shock, diffraction and dispersion). We explain the phenomenological sub-models and each assumption required to arrive at a complete and consistent theoretical description. The development includes the relationship between shock and ionization and demonstrates why inclusion of Drude model impedance effects alters the nature of the shock operator. Unclassified Unlimited Release
Cationically polymerizable monomers derived from renewable sources
Crivello, J.V.
1992-10-01T23:59:59.000Z
The objectives of this project are to design and synthesize novel monomers which orginate from renewable biological sources and to carry out their rapid, efficient, pollution-free and energy efficient cationic polymerization to useful products under the influence of ultraviolet light or heat. A summary of the results of the past year's research on cationically polymerizable monomers derived from renewable sources is presented. Three major areas of investigation corresponding to the different classes of naturally occurring starting materials were investigated; epoxidized terpenes and natural rubber and vinyl ethers from alcohols and carbohydrates.
Formal Derivation of Concurrent Garbage Collectors
Pavlovic, Dusko; Smith, Douglas R
2010-01-01T23:59:59.000Z
Concurrent garbage collectors are notoriously difficult to implement correctly. Previous approaches to the issue of producing correct collectors have mainly been based on posit-and-prove verification or on the application of domain-specific templates and transformations. We show how to derive the upper reaches of a family of concurrent garbage collectors by refinement from a formal specification, emphasizing the application of domain-independent design theories and transformations. A key contribution is an extension to the classical lattice-theoretic fixpoint theorems to account for the dynamics of concurrent mutation and collection.
NOAA ERL Technical Memorandum ERL GLERL-64 MINIMIZING LONG-TERM WIND SET-UP ERRORS IN
ADMINISTRATION Malcolm Baldrige, Anthony J. Calio, Secretary Administrator Environmental Research Laboratories Thiessen network directional unit set-up error in mean Lake Erie level for 2- through 16-gage networks . . . . . . . . . . . 6 Figure 4. --Number of gages required to limit error in spatial-optimum network for Lake Erie
A multi-site analysis of random error2 in tower-based measurements of carbon and energy fluxes3
Forest Service, 271 Mast Road, Durham, NH 03824 USA.25 #12;RANDOM ERRORS IN ENERGY AND CO2 FLUX1 A multi-site analysis of random error2 in tower-based measurements of carbon and energy fluxes3 4 Forest Service, 271 Mast Road, Durham, NH 03824, USA.11 3 LI-COR Biosciences, Inc., 4421 Superior Street
Kumar, Vipin
Integrative Biomarker Discovery for Breast Cancer Metastasis from Gene Expression and Protein Interaction Data Using Error-tolerant Pattern Mining Rohit Gupta Smita Agrawal+ Navneet Rao Ze Tian rohit- ers. In this paper, we propose a novel error-tolerant pattern mining approach for integrated analysis
Paris-Sud XI, UniversitÃ© de
Estimation of errors in the inverse modeling of accidental release of atmospheric pollutant difficulty when inverting the source term of an atmospheric tracer dispersion problem is the estimation of the prior errors: those of the atmospheric transport model, those ascribed to the representativity
Chen, Yong
or applying an estimation method that is robust to the error structure assumption in modelling the dynamicsCan a more realistic model error structure improve the parameter estimation in modelling the dynamics of ®sh populations? Y. Chena,* , J.E. Paloheimob a Fisheries Conservation Chair Program, Fisheries
Paris-Sud XI, UniversitÃ© de
The Inventory inaccuracy issue under a multiplicative error structure Selma Khader*, Yacine Rekik of events: Let Q be the quantity that the inventory manager orders from his supply system. He will receive of errors, , i.e. what is physically available in inventory, may not be equal to , i.e. what the information
Kalman-predictive-proportional-integral-derivative (KPPID)
Fluerasu, A.; Sutton, M. (McGill)
2004-12-17T23:59:59.000Z
With third generation synchrotron X-ray sources, it is possible to acquire detailed structural information about the system under study with time resolution orders of magnitude faster than was possible a few years ago. These advances have generated many new challenges for changing and controlling the state of the system on very short time scales, in a uniform and controlled manner. For our particular X-ray experiments on crystallization or order-disorder phase transitions in metallic alloys, we need to change the sample temperature by hundreds of degrees as fast as possible while avoiding over or under shooting. To achieve this, we designed and implemented a computer-controlled temperature tracking system which combines standard Proportional-Integral-Derivative (PID) feedback, thermal modeling and finite difference thermal calculations (feedforward), and Kalman filtering of the temperature readings in order to reduce the noise. The resulting Kalman-Predictive-Proportional-Integral-Derivative (KPPID) algorithm allows us to obtain accurate control, to minimize the response time and to avoid over/under shooting, even in systems with inherently noisy temperature readings and time delays. The KPPID temperature controller was successfully implemented at the Advanced Photon Source at Argonne National Laboratories and was used to perform coherent and time-resolved X-ray diffraction experiments.
Primer on electricity futures and other derivatives
Stoft, S.; Belden, T.; Goldman, C.; Pickle, S.
1998-01-01T23:59:59.000Z
Increased competition in bulk power and retail electricity markets is likely to lower electricity prices, but will also result in greater price volatility as the industry moves away from administratively determined, cost-based rates and encourages market-driven prices. Price volatility introduces new risks for generators, consumers, and marketers. Electricity futures and other derivatives can help each of these market participants manage, or hedge, price risks in a competitive electricity market. Futures contracts are legally binding and negotiable contracts that call for the future delivery of a commodity. In most cases, physical delivery does not take place, and the futures contract is closed by buying or selling a futures contract on or near the delivery date. Other electric rate derivatives include options, price swaps, basis swaps, and forward contracts. This report is intended as a primer for public utility commissioners and their staff on futures and other financial instruments used to manage price risks. The report also explores some of the difficult choices facing regulators as they attempt to develop policies in this area.
de Boer, Gijs; Collins, William D.; Menon, Surabi; Long, Charles N.
2011-12-02T23:59:59.000Z
Measurements from ground-based cloud radar, high spectral resolution lidar and microwave radiometer are used in conjunction with a column version of the Rapid Radiative Transfer Model (RRTMG) and radiosonde measurements to derive the surface radiative properties under mixed-phase cloud conditions. These clouds were observed during the United States Department of Energy (US DOE) Atmospheric Radiation Measurement (ARM) Mixed-Phase Arctic Clouds Experiment (M-PACE) between September and November of 2004. In total, sixteen half hour time periods are reviewed due to their coincidence with radiosonde launches. Cloud liquid (ice) water paths are found to range between 11.0-366.4 (0.5-114.1) gm-2, and cloud physical thicknesses fall between 286-2075 m. Combined with temperature and hydrometeor size estimates, this information is used to calculate surface radiative flux densities using RRTMG, which are demonstrated to generally agree with measured flux densities from surface-based radiometric instrumentation. Errors in longwave flux density estimates are found to be largest for thin clouds, while shortwave flux density errors are generally largest for thicker clouds. A sensitivity study is performed to understand the impact of retrieval assumptions and uncertainties on derived surface radiation estimates. Cloud radiative forcing is calculated for all profiles, illustrating longwave dominance during this time of year, with net cloud forcing generally between 50 and 90 Wm-2.
Saeki, Hiroshi, E-mail: saeki@spring8.or.jp; Magome, Tamotsu, E-mail: saeki@spring8.or.jp [Japan Synchrotron Radiation Research Institute, SPring-8, Kohto 1-1-1, Sayo, Hyogo 679-5198 (Japan)
2014-10-06T23:59:59.000Z
To compensate pressure-measurement errors caused by a synchrotron radiation environment, a precise method using a hot-cathode-ionization-gauge head with correcting electrode, was developed and tested in a simulation experiment with excess electrons in the SPring-8 storage ring. This precise method to improve the measurement accuracy, can correctly reduce the pressure-measurement errors caused by electrons originating from the external environment, and originating from the primary gauge filament influenced by spatial conditions of the installed vacuum-gauge head. As the result of the simulation experiment to confirm the performance reducing the errors caused by the external environment, the pressure-measurement error using this method was approximately less than several percent in the pressure range from 10{sup ?5} Pa to 10{sup ?8} Pa. After the experiment, to confirm the performance reducing the error caused by spatial conditions, an additional experiment was carried out using a sleeve and showed that the improved function was available.
Simulating the Transverse Ising Model on a Quantum Computer: Error Correction with the Surface Code
Hao You; Michael R. Geller; P. C. Stancil
2013-03-29T23:59:59.000Z
We estimate the resource requirements for the quantum simulation of the ground state energy of the one dimensional quantum transverse Ising model (TIM), based on the surface code implementation of a fault tolerant quantum computer. The surface code approach has one of the highest known tolerable error rates (1%) which makes it currently one of the most practical quantum computing schemes. Compared to results of the same model using the concatenated Steane code, the current results indicate that the simulation time is comparable but the number of physical qubits for the surface code is 2-3 orders of magnitude larger than that of the concatenation code. Considering that the error threshold requirements of the surface code is four orders of magnitude higher than the concatenation code, building a quantum computer with a surface code implementation appears more promising given current physical hardware capabilities.
THE APPLICATION OF FIXED AND RANDOM ERROR TO SHIPPER/RECEIVER DIFFERENCES
B. G. SCOTT
2001-06-01T23:59:59.000Z
In order to determine if a statistically significant difference exists between shipper and receiver measurements, a statistical combination of the shipper's and receiver's limit-of-error (LOE) is calculated to determine the shipper/receiver limit-of-error, LOES/R. The shipper's and receiver's LOE may possess random and systematic components. Depending on the interpretation of the systematic and random components, the determination of the LOES/R can be performed by several different calculational methods. These calculational methods and their associated underlying assumptions are reviewed in the context of the LANL shipper receiver program. This paper, by presenting the assumptions that form the basis of a site-specific shipper/receiver difference calculation, can assist those individuals responsible for calculating LOES/R.
Discussion on common errors in analyzing sea level accelerations, solar trends and global warming
Scafetta, Nicola
2013-01-01T23:59:59.000Z
Errors in applying regression models and wavelet filters used to analyze geophysical signals are discussed: (1) multidecadal natural oscillations (e.g. the quasi 60-year Atlantic Multidecadal Oscillation (AMO), North Atlantic Oscillation (NAO) and Pacific Decadal Oscillation (PDO)) need to be taken into account for properly quantifying anomalous accelerations in tide gauge records such as in New York City; (2) uncertainties and multicollinearity among climate forcing functions prevent a proper evaluation of the solar contribution to the 20th century global surface temperature warming using overloaded linear regression models during the 1900-2000 period alone; (3) when periodic wavelet filters, which require that a record is pre-processed with a reflection methodology, are improperly applied to decompose non-stationary solar and climatic time series, Gibbs boundary artifacts emerge yielding misleading physical interpretations. By correcting these errors and using optimized regression models that reduce multico...
Effect of Field Errors in Muon Collider IR Magnets on Beam Dynamics
Alexahin, Y.; Gianfelice-Wendt, E.; Kapin, V.V.; /Fermilab
2012-05-01T23:59:59.000Z
In order to achieve peak luminosity of a Muon Collider (MC) in the 10{sup 35} cm{sup -2}s{sup -1} range very small values of beta-function at the interaction point (IP) are necessary ({beta}* {le} 1 cm) while the distance from IP to the first quadrupole can not be made shorter than {approx}6 m as dictated by the necessity of detector protection from backgrounds. In the result the beta-function at the final focus quadrupoles can reach 100 km making beam dynamics very sensitive to all kind of errors. In the present report we consider the effects on momentum acceptance and dynamic aperture of multipole field errors in the body of IR dipoles as well as of fringe-fields in both dipoles and quadrupoles in the ase of 1.5 TeV (c.o.m.) MC. Analysis shows these effects to be strong but correctable with dedicated multipole correctors.
Analysis of error in using fractured gas well type curves for constant pressure production
Schkade, David Wayne
1987-01-01T23:59:59.000Z
of normalized time and normalized cumulative production is a large improvement over using a constant evaluation pressure. 0 imens ion less cumulative production type curves are particularly useful in modeling production for economic projections, such as re... of MASTER OF SCIENCE May 1987 Major Subject: Petroleum Engineering ANALYSIS OF ERROR IN USING FRACTURED GAS WELL TYPE CURVES FOR CONSTANT PRESSURE PRDDUCTION A Thesis by DAVID WAYNE SCHKADE Approved as to style and content by: S. A. Ho lditch...
1997-2001 by M. Kostic Ch.5: Uncertainty/Error Analysis
Kostic, Milivoje M.
1 Â©1997-2001 by M. Kostic Ch.5: Uncertainty/Error Analysis Â· Introduction Â· Bias and Precision Summation/Propagation (Expanded Combined Uncertainty) Â· Problem 5-30 Â©1997-2001 by M. Kostic Ch.5) at corresponding Probability (%P) Remember: u = d%P = t,%PS (@ %P); z=t=d/S #12;2 Â©1997-2001 by M. Kostic Bias
Zollanvari, Amin
2012-02-14T23:59:59.000Z
, Aniruddha Datta Guy L. Curry Head of Department, Costas N. Georghiades December 2010 Major Subject: Electrical Engineering iii ABSTRACT Analytic Study of Performance of Error Estimators for Linear Discriminant Analysis with Applications in Genomics... : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 133 x LIST OF TABLES TABLE Page I Minimum sample size, n, (n0 = n1 = n) for desired (n;0:5) in univariate case. : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : : 67 II Genes selected using the validity-goodness model selection...
Delusions and Prediction Error: Clarifying the roles of behavioral and brain responses
Corlett, Philip; Fletcher, Paul
2015-01-03T23:59:59.000Z
(Kapur, 2003; Maher, 1974, 1988) or abnormal inferences acting on normal experiences (Campbell, 2001; Currie, 2000). It has been argued that neither deficit alone could account for a delusional belief and that two factors – both abnormal perception... , the delusion has formed as the best way to account for a noisy and uncertain prediction error [and one that perhaps has an unjustly elevated level of precision (Adams, Stephan, Brown, Frith, & Friston, 2013; Corlett et al., 2010; Fletcher & Frith, 2008...
Generalized Bhattacharyya and Chernoff upper bounds on Bayes error using quasi-arithmetic means
Nielsen, Frank
1 = w1(c11 + c21) and a2 = w2(c12 + c22). The identity simplifies for probability of error Pe to Pe to the total variation met- ric distance TV(p, q) = 1 2 |p(x) - q(x)|dx by Be = a1+a2 2 - TV(a1p1, a2p2) with a
Zhang, J.; Hodge, B. M.; Florita, A.
2013-05-01T23:59:59.000Z
Wind and solar power generations differ from conventional energy generation because of the variable and uncertain nature of their power output. This variability and uncertainty can have significant impacts on grid operations. Thus, short-term forecasting of wind and solar generation is uniquely helpful for power system operations to balance supply and demand in an electricity system. This paper investigates the correlation between wind and solar power forecasting errors.
Wind Power Forecasting Error Frequency Analyses for Operational Power System Studies: Preprint
Florita, A.; Hodge, B. M.; Milligan, M.
2012-08-01T23:59:59.000Z
The examination of wind power forecasting errors is crucial for optimal unit commitment and economic dispatch of power systems with significant wind power penetrations. This scheduling process includes both renewable and nonrenewable generators, and the incorporation of wind power forecasts will become increasingly important as wind fleets constitute a larger portion of generation portfolios. This research considers the Western Wind and Solar Integration Study database of wind power forecasts and numerical actualizations. This database comprises more than 30,000 locations spread over the western United States, with a total wind power capacity of 960 GW. Error analyses for individual sites and for specific balancing areas are performed using the database, quantifying the fit to theoretical distributions through goodness-of-fit metrics. Insights into wind-power forecasting error distributions are established for various levels of temporal and spatial resolution, contrasts made among the frequency distribution alternatives, and recommendations put forth for harnessing the results. Empirical data are used to produce more realistic site-level forecasts than previously employed, such that higher resolution operational studies are possible. This research feeds into a larger work of renewable integration through the links wind power forecasting has with various operational issues, such as stochastic unit commitment and flexible reserve level determination.
Theoretical inputs and errors in the new hadronic currents in TAUOLA
Roig, P.; Nugent, I. M.; Przedzinski, T.; Shekhovtsova, O.; Was, Z. [Grup de Fisica Teorica, Institut de Fisica d'Altes Energies, Universitat Autonoma de Barcelona, E-08193 Bellaterra, Barcelona (Spain); RWTH Aachen University, III. Physikalisches Institut B, Aachen (Germany); Faculty of Physics, Astronomy and Applied Computer Science, Jagellonian University, Reymonta 4, 30-059 Cracow, Poland and Institute of Nuclear Physics, PAN, Cracow, ul. Radzikowskiego 152 (Poland); IFIC, Universitat de Valencia-CSIC, Apt. Correus 22085, E-46071, Valencia (Spain); CERN PH-TH, CH-1211 Geneva 23, Switzerland and Institute of Nuclear Physics, PAN, Cracow, ul. Radzikowskiego 152 (Poland)
2012-10-23T23:59:59.000Z
The new hadronic currents implemented in the TAUOLA library are obtained in the unified and consistent framework of Resonance Chiral Theory: a Lagrangian approach in which the resonances exchanged in the hadronic tau decays are active degrees of freedom included in a way that reproduces the low-energy results of Chiral Perturbation Theory. The short-distance QCD constraints on the imaginary part of the spin-one correlators yield relations among the couplings that render the theory predictive. In this communication, the obtaining of the two- and three-meson form factors is sketched. One of the criticisms to our framework is that the error may be as large as 1/3, since it is a realization of the large-N{sub C} limit of QCD in a meson theory. A number of arguments are given which disfavor that claim pointing to smaller errors, which would explain the phenomenological success of our description in these decays. Finally, other minor sources of error and current improvements of the code are discussed.
Stephen Pankavich; Zeina Shreif; Yinglong Miao; Peter Ortoleva
2010-03-08T23:59:59.000Z
The kinetics of the self-assembly of nanocomponents into a virus, nanocapsule, or other composite structure is analyzed via a multiscale approach. The objective is to achieve predictability and to preserve key atomic-scale features that underlie the formation and stability of the composite structures. We start with an all-atom description, the Liouville equation, and the order parameters characterizing nanoscale features of the system. An equation of Smoluchowski type for the stochastic dynamics of the order parameters is derived from the Liouville equation via a multiscale perturbation technique. The self-assembly of composite structures from nanocomponents with internal atomic structure is analyzed and growth rates are derived. Applications include the assembly of a viral capsid from capsomers, a ribosome from its major subunits, and composite materials from fibers and nanoparticles. Our approach overcomes errors in other coarse-graining methods which neglect the influence of the nanoscale configuration on the atomistic fluctuations. We account for the effect of order parameters on the statistics of the atomistic fluctuations which contribute to the entropic and average forces driving order parameter evolution. This approach enables an efficient algorithm for computer simulation of self-assembly, whereas other methods severely limit the timestep due to the separation of diffusional and complexing characteristic times. Given that our approach does not require recalibration with each new application, it provides a way to estimate assembly rates and thereby facilitate the discovery of self-assembly pathways and kinetic dead-end structures.
Scrap tire derived fuel: Markets and issues
Serumgard, J. [Scrap Tire Management Council, Washington, DC (United States)
1997-12-01T23:59:59.000Z
More than 250 million scrap tires are generated annually in the United States and their proper management continues to be a solid waste management concern. Sound markets for scrap tires are growing and are consuming an ever increasing percentage of annual generation, with market capacity reaching more than 75% of annual generation in 1996. Of the three major markets - fuel, civil engineering applications, and ground rubber markets - the use of tires as a fuel is by far the largest market. The major fuel users include cement kilns, pulp and paper mills, electrical generation facilities, and some industrial facilities. Current issues that may impact the tire fuel market include continued public concern over the use of tires as fuels, the new EPA PM 2.5 standard, possible additional Clean Air emissions standards, access to adequate supplies of scrap tires, quality of processed tire derived fuel, and the possibility of creating a commodity market through the development of ASTM TDF standards.
Generalized Holographic Superconductors with Higher Derivative Couplings
Anshuman Dey; Subhash Mahapatra; Tapobrata Sarkar
2014-06-13T23:59:59.000Z
We introduce and study generalized holographic superconductors with higher derivative couplings between the field strength tensor and a complex scalar field, in four dimensional AdS black hole backgrounds. We study this theory in the probe limit, as well as with backreaction. There are multiple tuning parameters in the theory, and with two non-zero parameters, we show that the theory has a rich phase structure, and in particular, the transition from the normal to the superconducting phase can be tuned to be of first order or of second order within a window of one of these. This is established numerically as well as by computing the free energy of the boundary theory. We further present analytical results for the critical temperature of the model, and compare these with numerical analysis. Optical properties of this system are also studied numerically in the probe limit, and our results show evidence for negative refraction at low frequencies.
Chaotic inflation in higher derivative gravity theories
Myrzakul, Shynaray; Sebastiani, Lorenzo
2015-01-01T23:59:59.000Z
In this paper, we investigate chaotic inflation from scalar field subjected to potential in the framework of $f(R^2, P, Q)$-gravity, where we add a correction to Einstein's gravity based on a function of the square of the Ricci scalar $R^2$, the contraction of the Ricci tensor $P$, and the contraction of the Riemann tensor $Q$. The Gauss-Bonnet case is also discussed. We give the general formalism of inflation, deriving the slow-roll parameters, the $e$-folds number, and the spectral indexes. Several explicit examples are furnished, namely we will consider the cases of massive scalar field and scalar field with quartic potential and some power-law function of the curvature invariants under investigation in the gravitational action of the theory. Viable inflation according with observations is analyzed.
Cationically polymerizable monomers derived from renewable sources
Crivello, J.V.
1991-10-01T23:59:59.000Z
The objective of this project is to make use of products obtained from renewable plant sources as monomers for the direct production of polymers which can be used for a wide range of plastic applications. In this report is described progress in the synthesis and polymerization of cationically polymerizable monomers and oligomers derived from botanical oils, terpenes, natural rubber, and lignin. Nine different botanical oils were obtained from various sources, characterized and then epoxidized. Their photopolymerization was carried out using cationic photoinitiators and the mechanical properties of the resulting polymers characterized. Preliminary biodegradation studies are being conducted on the photopolymerized films from several of these oils. Limonene was cationically polymerized to give dimers and the dimers epoxidized to yield highly reactive monomers suitable for coatings, inks and adhesives. The direct phase transfer epoxidation of squalene and natural rubber was carried out. The modified rubbers undergo facile photocrosslinking in the presence of onium salts to give crosslinked elastomers. 12 refs., 3 figs., 10 tabs.
Derivation of quantum probabilities from deterministic evolution
T. G. Philbin
2014-10-15T23:59:59.000Z
The predictions of quantum mechanics are probabilistic. Quantum probabilities are extracted using a postulate of the theory called the Born rule, the status of which is central to the "measurement problem" of quantum mechanics. Efforts to justify the Born rule from other physical principles, and thus elucidate the measurement process, have involved lengthy statistical or information-theoretic arguments. Here we show that Bohm's deterministic formulation of quantum mechanics allows the Born rule for measurements on a single system to be derived, without any statistical assumptions. We solve a simple example where the creation of an ensemble of identical quantum states, together with position measurements on those states, are described by Bohm's quantum dynamics. The calculated measurement outcomes agree with the Born-rule probabilities, which are thus a consequence of deterministic evolution. Our results demonstrate that quantum probabilities can emerge from simple dynamical laws alone, and they support the view that there is no underlying indeterminism in quantum phenomena.
A general entry to C7-borono indole derivatives
Fenton, Owen S. (Owen Shea)
2012-01-01T23:59:59.000Z
The development of a methodology to access C7 pinacolatoboron substituted indole derivatives is described. It has been applied to indole, tryptophan, and tryptamine derivatives. Further functionalization to a C7 phenolic ...
Mathematics with DERIVE as Didactical Tool Wolfram Koepf
Koepf, Wolfram
Mathematics with DERIVE as Didactical Tool Wolfram Koepf University Kassel koepf@mathematik.uni-kassel.de 1 Introduction In this talk, I would like to show you how DERIVE can be used as a didactical tool
Resurgence, Stokes phenomenon and alien derivatives for levelone linear
Resurgence, Stokes phenomenon and alien derivatives for levelÂone linear di#erential systems. Mich of alien derivatives. We make explicit the StokesÂRamis matrices as functions of the connection constants, summability, resurgence, alien derivatives. AMS classification: 34M03, 34M30, 34M35, 34M40. 1 #12; Loday
Biomass-Derived Energy Products and Co-Products Market
Biomass-Derived Energy Products and Co-Products Market This report identifies the bio-fuels and co & Earth Science & Technology Â University of Hawai`i at Manoa #12;Biomass-Derived Energy Products and Co agency thereof. #12;Biomass Derived Energy Products and Co- Products Market and Off-take Study Hawaii
If ( ) and ( ), such that ( ) [ ( )], then [ ( )] ( ) Derivative of a composite ...
Bailey, Charlotte M
2014-10-21T23:59:59.000Z
Derivative of a composite function = derivative of the 'outer' function. (with respect to or ( ))times the deriv. y f u. u g x. y f u. f g x dy dy du dy or. f g x g x dx du dx.
A Methodology for the Derivation of Parallel Programs
Goodman, Joy
A Methodology for the Derivation of Parallel Programs Joy Goodman Department of Computer Science, University of Glasgow Abstract. I am currently developing a methodology for deriving paralÂ lel programs from equational reasoning, a more efficient parallel program in a variety of languages and styles can be derived
Topolnjak, Rajko [Department of Radiation Oncology, Netherlands Cancer Institute-Antoni van Leeuwenhoek Hospital, Amsterdam (Netherlands); Sonke, Jan-Jakob, E-mail: j.sonke@nki.n [Department of Radiation Oncology, Netherlands Cancer Institute-Antoni van Leeuwenhoek Hospital, Amsterdam (Netherlands); Nijkamp, Jasper; Rasch, Coen; Minkema, Danny; Remeijer, Peter; Vliet-Vroegindeweij, Corine van [Department of Radiation Oncology, Netherlands Cancer Institute-Antoni van Leeuwenhoek Hospital, Amsterdam (Netherlands)
2010-11-15T23:59:59.000Z
Purpose: To quantify the differences in setup errors measured with the cone-beam computed tomography (CBCT) and electronic portal image devices (EPID) in breast cancer patients. Methods and Materials: Repeat CBCT scan were acquired for routine offline setup verification in 20 breast cancer patients. During the CBCT imaging fractions, EPID images of the treatment beams were recorded. Registrations of the bony anatomy for CBCT to planning CT and EPID to digitally reconstructed-radiographs (DRRs) were compared. In addition, similar measurements of an anthropomorphic thorax phantom were acquired. Bland-Altman and linear regression analysis were performed for clinical and phantom registrations. Systematic and random setup errors were quantified for CBCT and EPID-driven correction protocols in the EPID coordinate system (U, V), with V parallel to the cranial-caudal axis and U perpendicular to V and the central beam axis. Results: Bland-Altman analysis of clinical EPID and CBCT registrations yielded 4 to 6-mm limits of agreement, indicating that both methods were not compatible. The EPID-based setup errors were smaller than the CBCT-based setup errors. Phantom measurements showed that CBCT accurately measures setup error whereas EPID underestimates setup errors in the cranial-caudal direction. In the clinical measurements, the residual bony anatomy setup errors after offline CBCT-based corrections were {Sigma}{sub U} = 1.4 mm, {Sigma}{sub V} = 1.7 mm, and {sigma}{sub U} = 2.6 mm, {sigma}{sub V} = 3.1 mm. Residual setup errors of EPID driven corrections corrected for underestimation were estimated at {Sigma}{sub U} = 2.2mm, {Sigma}{sub V} = 3.3 mm, and {sigma}{sub U} = 2.9 mm, {sigma}{sub V} = 2.9 mm. Conclusion: EPID registration underestimated the actual bony anatomy setup error in breast cancer patients by 20% to 50%. Using CBCT decreased setup uncertainties significantly.
Bard, D.; Chang, C.; Kahn, S. M.; Gilmore, K.; Marshall, S. [KIPAC, Stanford University, 452 Lomita Mall, Stanford, CA 94309 (United States); Kratochvil, J. M.; Huffenberger, K. M. [Department of Physics, University of Miami, Coral Gables, FL 33124 (United States); May, M. [Physics Department, Brookhaven National Laboratory, Upton, NY 11973 (United States); AlSayyad, Y.; Connolly, A.; Gibson, R. R.; Jones, L.; Krughoff, S. [Department of Astronomy, University of Washington, Seattle, WA 98195 (United States); Ahmad, Z.; Bankert, J.; Grace, E.; Hannel, M.; Lorenz, S. [Department of Physics, Purdue University, West Lafayette, IN 47907 (United States); Haiman, Z.; Jernigan, J. G., E-mail: djbard@slac.stanford.edu [Department of Astronomy and Astrophysics, Columbia University, New York, NY 10027 (United States); and others
2013-09-01T23:59:59.000Z
We study the effect of galaxy shape measurement errors on predicted cosmological constraints from the statistics of shear peak counts with the Large Synoptic Survey Telescope (LSST). We use the LSST Image Simulator in combination with cosmological N-body simulations to model realistic shear maps for different cosmological models. We include both galaxy shape noise and, for the first time, measurement errors on galaxy shapes. We find that the measurement errors considered have relatively little impact on the constraining power of shear peak counts for LSST.
Feasibility of neuro-morphic computing to emulate error-conflict based decision making.
Branch, Darren W.
2009-09-01T23:59:59.000Z
A key aspect of decision making is determining when errors or conflicts exist in information and knowing whether to continue or terminate an action. Understanding the error-conflict processing is crucial in order to emulate higher brain functions in hardware and software systems. Specific brain regions, most notably the anterior cingulate cortex (ACC) are known to respond to the presence of conflicts in information by assigning a value to an action. Essentially, this conflict signal triggers strategic adjustments in cognitive control, which serve to prevent further conflict. The most probable mechanism is the ACC reports and discriminates different types of feedback, both positive and negative, that relate to different adaptations. Unique cells called spindle neurons that are primarily found in the ACC (layer Vb) are known to be responsible for cognitive dissonance (disambiguation between alternatives). Thus, the ACC through a specific set of cells likely plays a central role in the ability of humans to make difficult decisions and solve challenging problems in the midst of conflicting information. In addition to dealing with cognitive dissonance, decision making in high consequence scenarios also relies on the integration of multiple sets of information (sensory, reward, emotion, etc.). Thus, a second area of interest for this proposal lies in the corticostriatal networks that serve as an integration region for multiple cognitive inputs. In order to engineer neurological decision making processes in silicon devices, we will determine the key cells, inputs, and outputs of conflict/error detection in the ACC region. The second goal is understand in vitro models of corticostriatal networks and the impact of physical deficits on decision making, specifically in stressful scenarios with conflicting streams of data from multiple inputs. We will elucidate the mechanisms of cognitive data integration in order to implement a future corticostriatal-like network in silicon devices for improved decision processing.
H. Boche; J. Noetzel
2014-05-07T23:59:59.000Z
This work is motivated by a quite general question: Under which circumstances are the capacities of information transmission systems continuous? The research is explicitly carried out on arbitrarily varying quantum channels (AVQCs). We give an explicit example that answers the recent question whether the transmission of messages over AVQCs can benefit from distribution of randomness between the legitimate sender and receiver in the affirmative. The specific class of channels introduced in that example is then extended to show that the deterministic capacity does have discontinuity points, while that behaviour is, at the same time, not generic: We show that it is continuous around its positivity points. This is in stark contrast to the randomness-assisted capacity, which is always continuous in the channel. Our results imply that the deterministic message transmission capacity of an AVQC can be discontinuous only in points where it is zero, while the randomness assisted capacity is nonzero. Apart from the zero-error capacities, this is the first result that shows a discontinuity of a capacity for a large class of quantum channels. The continuity of the respective capacity for memoryless quantum channels had, among others, been listed as an open problem on the problem page of the ITP Hannover for about six years before it was proven to be continuous. We also quantify the interplay between the distribution of finite amounts of randomness between the legitimate sender and receiver, the (nonzero) decoding error with respect to the average error criterion that can be achieved over a finite number of channel uses and the number of messages that can be sent. This part of our results also applies to entanglement- and strong subspace transmission. In addition, we give a new sufficient criterion for the entanglement transmission capacity with randomness assistance to vanish.
Lee, Sang Hyun, 1973-
2006-01-01T23:59:59.000Z
Construction projects are uncertain and complex in nature. One of the major driving forces that may account for these characteristics is iterative cycles caused by errors and changes. Errors and changes worsen project ...
A Time-Dependent Born-Oppenheimer Approximation with Exponentially Small Error Estimates
George A. Hagedorn; Alain Joye
2000-05-03T23:59:59.000Z
We present the construction of an exponentially accurate time-dependent Born-Oppenheimer approximation for molecular quantum mechanics. We study molecular systems whose electron masses are held fixed and whose nuclear masses are proportional to $\\epsilon^{-4}$, where $\\epsilon$ is a small expansion parameter. By optimal truncation of an asymptotic expansion, we construct approximate solutions to the time-dependent Schr\\"odinger equation that agree with exact normalized solutions up to errors whose norms are bounded by $\\ds C \\exp(-\\gamma/\\epsilon^2)$, for some C and $\\gamma>0$.
Full accuracy differential GPS C/A code pseudorange error correction concept
Desai, Rajendra Laxman
1983-01-01T23:59:59.000Z
, hf A8M University, 1983. L2] S . P . Teasl ey, W. M . Hoover, C . R . Johnson, NDi ffenenti al GPS Navigation, h IEEE Plans 80 Position, Navi gatt on, and Symposium Record, pp. 9-16, Atlantic City, NJ, 6-9 December 1982. [3] J . r&. Li gon, NDi... OF SCIENCE December, 1983 Major Subject: Electrical Engineering FULL ACCURACY DIFFEREIITIAL GPS C/A CODE PSUEDORANGE ERROR CORRECTION CONCEPT A Thesis by RAJENDRA LAXHAN DESAI Approved as to style and content by: Dr. P. S. Noe ( Chairman...
High performance single-error-correcting quantum codes for amplitude damping
Peter W. Shor; Graeme Smith; John A. Smolin; Bei Zeng
2009-07-29T23:59:59.000Z
We construct families of high performance quantum amplitude damping codes. All of our codes are nonadditive and most modestly outperform the best possible additive codes in terms of encoded dimension. One family is built from nonlinear error-correcting codes for classical asymmetric channels, with which we systematically construct quantum amplitude damping codes with parameters better than any prior construction known for any block length n > 7 except n=2^r-1. We generalize this construction to employ classical codes over GF(3) with which we numerically obtain better performing codes up to length 14. Because the resulting codes are of the codeword stabilized (CWS) type, easy encoding and decoding circuits are available.
RSE Table 4.1 Relative Standard Errors for Table 4.1
U.S. Energy Information Administration (EIA) Indexed Site
AFDC Printable Version Share this resource Send a link to EERE: Alternative Fuels Data Center Home Page to someone by E-mail Share EERE: Alternative Fuels Data Center Home Page on Facebook Tweet about EERE: Alternative Fuels Data Center Home Page on Twitter Bookmark EERE: Alternative Fuels Data Center Home Page on Google Bookmark EERE: Alternative Fuels Data CenterFranconia,(Million Barrels) Crude Oil Reserves in NonproducingAdditions to Capacity on theThousand7.End1 Relative Standard Errors
ERROR REDUCTION IN DUCT LEAKAGE TESTING THROUGH DATA CROSS-CHECKS
ANDREWS, J.W.
1998-12-31T23:59:59.000Z
One way to reduce uncertainty in scientific measurement is to devise a protocol in which more quantities are measured than are absolutely required, so that the result is over constrained. This report develops a method for so combining data from two different tests for air leakage in residential duct systems. An algorithm, which depends on the uncertainty estimates for the measured quantities, optimizes the use of the excess data. In many cases it can significantly reduce the error bar on at least one of the two measured duct leakage rates (supply or return), and it provides a rational method of reconciling any conflicting results from the two leakage tests.
Panel Unit Root Tests in the Presence of a Multifactor Error Structure
Pesaran, M Hashem; Smith, L Vanessa; Yamagata, Takashi
Analysis and Policy (CFAP), University of Cambridge cUniversity of York December 2007 Abstract This paper extends the cross sectionally augmented panel unit root test proposed by Pesaran (2007) to the case of a multifactor error structure. The basic idea... = eit#0; pit is the real exchange rate, eqit is the real equity price index, gdpit is real output, rLit is the quarterly long-term interest rate and poilt is the nominal oil price.5 The 32 countries considered are: Argentina, Australia, Austria, Belgium...
Sayil, Selahattin
IEEE TRANSACTIONS ON NUCLEAR SCIENCE, VOL. 57, NO. 6, DECEMBER 2010 3553 Soft Error and Soft Delay of a circuit increases. The increasing use of portable devices and higher demand on battery life has made power
Ding, Meixia
2009-05-15T23:59:59.000Z
to Student Errors and Difficulties (TRED) in teaching equivalent fractions with an eye on students’ cognitive gains as the assessment of teaching effects. This research used a qualitative paradigm. Classroom videos concerning equivalent fractions from six...
Sudarshan, Raghunathan, 1978-
2005-01-01T23:59:59.000Z
We propose a simple and unified approach for a posteriori error estimation and adaptive mesh refinement in finite element analysis using multiresolution signal processing principles. Given a sequence of nested discretizations ...
Lun-Shin Yao; Dan Hughes
2007-04-26T23:59:59.000Z
Comment on "Time Step Sensitivity of Nonlinear Atmospheric Models: Numerical Convergence, Truncation Error Growth, and Ensemble Design" Teixeira et al.
Johnson, Eric William
1988-01-01T23:59:59.000Z
SIMULATION OF A STOL AIRLIFTER IN WIND SHEAR, USING TOTAL ENERGY AND GLIDESLOPE ANGULAR ERROR METHODS FOR GLIDEPATH CONTROL A Thesis by ERIC WILLIAM JOHNSON Submitted to the Graduate College of Texas A&M University in partial fulfillment... of the requirements for the degree of MASTER OF SCIENCE May 1988 Major Subject: Aerospace Engineering SIMULATION OF A STOL AIRLIFTER IN WIND SHEAR, USING TOTAL ENERGY AND GLIDESLOPE ANGULAR ERROR METHODS FOR GLIDEPATH CONTROL A Thesis by ERIC WILLIAM JOHNSON...
Davidson, R. L.; Earle, G. D.; Heelis, R. A. [William B. Hanson Center for Space Sciences, University of Texas at Dallas, 800 W. Campbell Road, WT15, Richardson, Texas 75080 (United States); Klenzing, J. H. [Space Weather Laboratory/Code 674, Goddard Space Flight Center, Greenbelt, Maryland 20771 (United States)
2010-08-15T23:59:59.000Z
Planar retarding potential analyzers (RPAs) have been utilized numerous times on high profile missions such as the Communications/Navigation Outage Forecast System and the Defense Meteorological Satellite Program to measure plasma composition, temperature, density, and the velocity component perpendicular to the plane of the instrument aperture. These instruments use biased grids to approximate ideal biased planes. These grids introduce perturbations in the electric potential distribution inside the instrument and when unaccounted for cause errors in the measured plasma parameters. Traditionally, the grids utilized in RPAs have been made of fine wires woven into a mesh. Previous studies on the errors caused by grids in RPAs have approximated woven grids with a truly flat grid. Using a commercial ion optics software package, errors in inferred parameters caused by both woven and flat grids are examined. A flat grid geometry shows the smallest temperature and density errors, while the double thick flat grid displays minimal errors for velocities over the temperature and velocity range used. Wire thickness along the dominant flow direction is found to be a critical design parameter in regard to errors in all three inferred plasma parameters. The results shown for each case provide valuable design guidelines for future RPA development.
A theoretical analysis of the systematic errors in the Red Clump distance to the LMC
Maurizio Salaris; Susan Percival; Leo Girardi
2003-07-17T23:59:59.000Z
We present a detailed analysis of the uncertainty on the theoretical population corrections to the LMC Red Clump (RC) absolute magnitude, by employing a population synthesis algorithm to simulate theoretically the photometric and spectroscopic properties of RC stars, under various assumptions about the LMC Star Formation Rate (SFR) and Age Metallicity Relationship (AMR). A comparison of the outcome of our simulations with observations of evolved low-intermediate mass stars in the LMC allows one to select the combinations of SFR and AMR that bracket the real LMC star formation history, and to estimate the systematic error on the associated RC population corrections. The most accurate estimate of the LMC distance modulus from the RC method (adopting the OGLE-II reddening maps for the LMC) is obtained from the K-band magnitude, and provides (m-M)_{0, LMC}=18.47 +/-0.01(random) +0.05/-0.06(systematic). Distances obtained from the I-band, or from the multicolour RC technique which determines at the same time reddening and distance, both agree (albeit with a slightly larger error bar) with this value.
Errors induced in triaxial stress tensor calculations using incorrect lattice parameters
Ruud, C.O. [Pennsylvania State Univ., University Park, PA (United States). Materials Research Lab.; Kozaczek, K.J. [Oak Ridge National Lab., TN (United States)
1994-06-01T23:59:59.000Z
A number of researchers have proposed that for some metallic alloys, an elaborate procedure is necessary in order to improve the accuracy of the measured triaxial stress tensor. Others have been concerned that the uncertainties in establishing the precise zero-stress lattice parameter of an alloyed and/or cold worked engineering metal could cause significantly more error than would result in ignoring the triaxial stress state and assuming the plane stress condition. This paper illustrates the effect of uncertainties in the zero-stress lattice parameters on the calculated triaxial stress state for zero stress powders of three common engineering alloys, i.e., 1010 steel, 304 stainless steel, and 2024 aluminum. Also, errors due to the incorrect lattice spacing in experimental stress analysis are presented for three examples, i.e., a silicon powder, 304 gainless steel cylinder and a diamond. For cases where the plane strain assumption is justified, the uncertainties due to the stress free lattice parameter can be reduced by a simple measurement.
Nanoscale Reinforced, Polymer Derived Ceramic Matrix Coatings
Rajendra Bordia
2009-07-31T23:59:59.000Z
The goal of this project was to explore and develop a novel class of nanoscale reinforced ceramic coatings for high temperature (600-1000 C) corrosion protection of metallic components in a coal-fired environment. It was focused on developing coatings that are easy to process and low cost. The approach was to use high-yield preceramic polymers loaded with nano-size fillers. The complex interplay of the particles in the polymer, their role in controlling shrinkage and phase evolution during thermal treatment, resulting densification and microstructural evolution, mechanical properties and effectiveness as corrosion protection coatings were investigated. Fe-and Ni-based alloys currently used in coal-fired environments do not possess the requisite corrosion and oxidation resistance for next generation of advanced power systems. One example of this is the power plants that use ultra supercritical steam as the working fluid. The increase in thermal efficiency of the plant and decrease in pollutant emissions are only possible by changing the properties of steam from supercritical to ultra supercritical. However, the conditions, 650 C and 34.5 MPa, are too severe and result in higher rate of corrosion due to higher metal temperatures. Coating the metallic components with ceramics that are resistant to corrosion, oxidation and erosion, is an economical and immediate solution to this problem. Good high temperature corrosion protection ceramic coatings for metallic structures must have a set of properties that are difficult to achieve using established processing techniques. The required properties include ease of coating complex shapes, low processing temperatures, thermal expansion match with metallic structures and good mechanical and chemical properties. Nanoscale reinforced composite coatings in which the matrix is derived from preceramic polymers have the potential to meet these requirements. The research was focused on developing suitable material systems and processing techniques for these coatings. In addition, we investigated the effect of microstructure on the mechanical properties and oxidation protection ability of the coatings. Coatings were developed to provide oxidation protection to both ferritic and austentic alloys and Ni-based alloys. The coatings that we developed are based on low viscosity pre-ceramic polymers. Thus they can be easily applied to any shape by using a variety of techniques including dip-coating, spray-coating and painting. The polymers are loaded with a variety of nanoparticles. The nanoparticles have two primary roles: control of the final composition and phases (and hence the properties); and control of the shrinkage during thermal decomposition of the polymer. Thus the selection of the nanoparticles was the most critical aspect of this project. Based on the results of the processing studies, the performance of selected coatings in oxidizing conditions (both static and cyclic) was investigated.
Massive Gravity from Higher Derivative Gravity with Boundary Conditions
Minjoon Park; Lorenzo Sorbo
2012-10-29T23:59:59.000Z
With an appropriate choice of parameters, a higher derivative theory of gravity can describe a normal massive sector and a ghost massless sector. We show that, when defined on an asymptotically de Sitter spacetime with Dirichlet boundary conditions, such a higher derivative gravity can provide a framework for a unitary theory of massive gravity in four spacetime dimensions. The resulting theory is free not only of higher derivative ghosts but also of the Boulware-Deser mode.
P wave velocity variations in the Coso region, California, derived...
Jump to: navigation, search OpenEI Reference LibraryAdd to library Journal Article: P wave velocity variations in the Coso region, California, derived from local earthquake...
Explore the Genetic Frontier: Labeling Foods Derived from Biotechnology
Vestal, Andy; Hawkins, Carole
2002-10-19T23:59:59.000Z
Today's food labels provide consumers with nutrition information. This publication discusses the content of labels on foods derived from biotechnology and the agencies that regulate such labeling....
Derivative-free Robust Optimization for Circuit Design
2013-09-27T23:59:59.000Z
Keywords Robust Optimization · Derivative-free Methods · Circuit Design ... uncertainty in real-world applications. ... no first order information is available (
Benzene Dimer: Dynamic Structure and Thermodynamics Derived from...
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
Benzene Dimer: Dynamic Structure and Thermodynamics Derived from On-the-Fly ab initio DFT-D Molecular Dynamic Simulations. Benzene Dimer: Dynamic Structure and Thermodynamics...
MA 16020 – EXAM FORMULAS THE SECOND DERIVATIVE TEST ...
2014-08-12T23:59:59.000Z
MA 16020 – EXAM FORMULAS. THE SECOND DERIVATIVE TEST. Suppose f is a function of two variables x and y, and that all the second-order partial ...
Bio-Derived Liquids to Hydrogen Distributed Reforming Working...
Office of Energy Efficiency and Renewable Energy (EERE) Indexed Site
Working Group (BILIWG), Hydrogen Separation and Purification Working Group (PURIWG) & Hydrogen Production Technical Team Bio-Derived Liquids to Hydrogen Distributed Reforming...
Graphene and its derivatives : fabrication and Raman spectroscopy study.
Cong, Chunxiao.
2012-01-01T23:59:59.000Z
??This thesis presents results on fabrication and Raman spectroscopy studies of graphene and its derivates. The works can be divided into two parts as follows.… (more)
Bio-Derived Liquids to Hydrogen Distributed Reforming Targets
Broader source: Energy.gov [DOE]
Presentation by Arlene Anderson at the October 24, 2006 Bio-Derived Liquids to Hydrogen Distributed Reforming Working Group Kick-Off Meeting.
Optimization Online - Exploiting derivative-free local searches in ...
Giampaolo Liuzzi
2014-02-28T23:59:59.000Z
Feb 28, 2014 ... Exploiting derivative-free local searches in DIRECT-type algorithms for global optimization. Giampaolo Liuzzi(giampaolo.liuzzi ***at*** ...
Optimization Online - Derivative-free Methods for Mixed-Integer ...
G. Liuzzi
2014-03-07T23:59:59.000Z
Mar 7, 2014 ... Derivative-free Methods for Mixed-Integer Constrained Optimization Problems. G. Liuzzi(giampaolo.liuzzi ***at*** iasi.cnr.it) S. Lucidi(lucidi ...
Bio-Derived Liquids to Hydrogen Distributed Reforming Targets...
Broader source: Energy.gov (indexed) [DOE]
Development Manager, U.S. DOE Office of Energy Efficiency and Renewable Energy Hydrogen, Fuel Cells and Infrastructure Technologies Program Bio-Derived Liquids to Hydrogen...
Optical pattern recognition architecture implementing the mean-square error correlation algorithm
Molley, Perry A. (Albuquerque, NM)
1991-01-01T23:59:59.000Z
An optical architecture implementing the mean-square error correlation algorithm, MSE=.SIGMA.[I-R].sup.2 for discriminating the presence of a reference image R in an input image scene I by computing the mean-square-error between a time-varying reference image signal s.sub.1 (t) and a time-varying input image signal s.sub.2 (t) includes a laser diode light source which is temporally modulated by a double-sideband suppressed-carrier source modulation signal I.sub.1 (t) having the form I.sub.1 (t)=A.sub.1 [1+.sqroot.2m.sub.1 s.sub.1 (t)cos (2.pi.f.sub.o t)] and the modulated light output from the laser diode source is diffracted by an acousto-optic deflector. The resultant intensity of the +1 diffracted order from the acousto-optic device is given by: I.sub.2 (t)=A.sub.2 [+2m.sub.2.sup.2 s.sub.2.sup.2 (t)-2.sqroot.2m.sub.2 (t) cos (2.pi.f.sub.o t] The time integration of the two signals I.sub.1 (t) and I.sub.2 (t) on the CCD deflector plane produces the result R(.tau.) of the mean-square error having the form: R(.tau.)=A.sub.1 A.sub.2 {[T]+[2m.sub.2.sup.2.multidot..intg.s.sub.2.sup.2 (t-.tau.)dt]-[2m.sub.1 m.sub.2 cos (2.tau.f.sub.o .tau.).multidot..intg.s.sub.1 (t)s.sub.2 (t-.tau.)dt]} where: s.sub.1 (t) is the signal input to the diode modulation source: s.sub.2 (t) is the signal input to the AOD modulation source; A.sub.1 is the light intensity; A.sub.2 is the diffraction efficiency; m.sub.1 and m.sub.2 are constants that determine the signal-to-bias ratio; f.sub.o is the frequency offset between the oscillator at f.sub.c and the modulation at f.sub.c +f.sub.o ; and a.sub.o and a.sub.1 are constant chosen to bias the diode source and the acousto-optic deflector into their respective linear operating regions so that the diode source exhibits a linear intensity characteristic and the AOD exhibits a linear amplitude characteristic.
Numerical errors in the computation of subfilter scalar variance in large eddy simulations
Raman, Venkat
Mechanics, The University of Texas at Austin, Austin, Texas 78712, USA 2 Center for Turbulence Research of specific model forms. Like other subfilter quantities, such as kinetic energy, subfilter variance can-rule manipulations used to derive convection, diffusion, and production terms associated with the square
Visulization of Time-Varying Multiresolution Date Using Error-Based Temporal-Spatial Resuse
Nuber, C; LaMar, E; Hamann, B; Joy, K
2002-04-22T23:59:59.000Z
In this paper, we report results on exploration of two-dimensional (2D) time varying datasets. We extend the notion of multiresolution spatial data approximation of static datasets to spatio-temporal approximation of time-varying datasets. Time-varying datasets typically do not change ''uniformly,'' i.e., some spatial sub-domains can experience only little or no change for extended periods of time. In these sub-domains, we show that approximation error bounds can be met when using sub-domains from other time-steps effectively. We generate a more general approximation scheme where sub-domains may approximate congruent sub-domains from any other time steps. While this incurs an O(T2) overhead, where T is the total number of time-steps, we show significant reduction in data transmission. We also discuss ideas for improvements to reduce overhead.
The Role of Basis Set Superposition Error in Water Addition Reactions to Ln(III) Cations
Kvamme, Brandon; Wander, Matthew C F.; Clark, Aurora E.
2009-06-01T23:59:59.000Z
Our goal in this work was to characterize the importance of basis set superposition error (BSSE) in the reaction energetics of water addition to highly charged metal ions that exhibit strong ion-dipole interactions with water. The gas phase water addition reactions, M(H?O)³+N + H?O ? M(H?O)³+N ? ? [M = La(III), Lu(III); N = 0–8] have been studied, with a particular emphasis on a posteriori methods for calculating BSSE and its constituent energetic components as a function of M(H?O) ³+N cluster size and water basis set. Because of accumulation of BSSE within the subclusters in the reaction series, the successive reaction counterpoise method for determining the BSSE correction interaction energies is advocated.
Tripartite entanglement in qudit stabilizer states and application in quantum error correction
Looi, Shiang Yong; Griffiths, Robert B. [Department of Physics, Carnegie Mellon University, Pittsburgh, Pennsylvania 15213 (United States)
2011-11-15T23:59:59.000Z
Consider a stabilizer state on n qudits, each of dimension D with D being a prime or squarefree integer, divided into three mutually disjoint sets or parts. Generalizing a result of Bravyi et al.[J. Math. Phys. 47, 062106 (2006)] for qubits (D=2), we show that up to local unitaries, the three parts of the state can be written as tensor product of unentangled signle-qudit states, maximally entangled Einstein-Podolsky-Rosen (EPR) pairs, and tripartite Greenberger-Horne-Zeilinger (GHZ) states. We employ this result to obtain a complete characterization of the properties of a class of channels associated with stabilizer error-correcting codes, along with their complementary channels.
Information-preserving structures: A general framework for quantum zero-error information
Blume-Kohout, Robin [Perimeter Institute for Theoretical Physics, 31 Caroline Street North, Waterloo, Ontario N2L 2Y5 (Canada); Ng, Hui Khoon [Institute for Quantum Information, California Institute of Technology, Pasadena, California 91125 (United States); Poulin, David [Department de Physique, Universite de Sherbrooke, Quebec J1K 2R1 (Canada); Viola, Lorenza [Department of Physics and Astronomy, Dartmouth College, 6127 Wilder Laboratory, Hanover, New Hampshire 03755 (United States)
2010-12-15T23:59:59.000Z
Quantum systems carry information. Quantum theory supports at least two distinct kinds of information (classical and quantum), and a variety of different ways to encode and preserve information in physical systems. A system's ability to carry information is constrained and defined by the noise in its dynamics. This paper introduces an operational framework, using information-preserving structures, to classify all the kinds of information that can be perfectly (i.e., with zero error) preserved by quantum dynamics. We prove that every perfectly preserved code has the same structure as a matrix algebra, and that preserved information can always be corrected. We also classify distinct operational criteria for preservation (e.g., 'noiseless','unitarily correctible', etc.) and introduce two natural criteria for measurement-stabilized and unconditionally preserved codes. Finally, for several of these operational criteria, we present efficient (polynomial in the state-space dimension) algorithms to find all of a channel's information-preserving structures.
Acquaviva, Viviana; Gawiser, Eric
2015-01-01T23:59:59.000Z
We seek to improve the accuracy of joint galaxy photometric redshift estimation and spectral energy distribution (SED) fitting. By simulating different sources of uncorrected systematic errors, we demonstrate that if the uncertainties on the photometric redshifts are estimated correctly, so are those on the other SED fitting parameters, such as stellar mass, stellar age, and dust reddening. Furthermore, we find that if the redshift uncertainties are over(under)-estimated, the uncertainties in SED parameters tend to be over(under)-estimated by similar amounts. These results hold even in the presence of severe systematics and provide, for the first time, a mechanism to validate the uncertainties on these parameters via comparison with spectroscopic redshifts. We propose a new technique (annealing) to re-calibrate the joint uncertainties in the photo-z and SED fitting parameters without compromising the performance of the SED fitting + photo-z estimation. This procedure provides a consistent estimation of the mu...
The sensitivity of patient specific IMRT QC to systematic MLC leaf bank offset errors
Rangel, Alejandra; Palte, Gesa; Dunscombe, Peter [Department of Medical Physics, Tom Baker Cancer Centre, 1331-29 Street NW, Calgary, Alberta T2N 4N2, Canada and Department of Physics and Astronomy, University of Calgary, 2500 University Drive North West, Calgary, Alberta T2N 1N4 (Canada); Department of Medical Physics, Tom Baker Cancer Centre, 1331-29 Street NW, Calgary, Alberta T2N 4N2 (Canada); Department of Medical Physics, Tom Baker Cancer Centre, 1331-29 Street NW, Calgary, Alberta T2N 4N2 (Canada); Department of Physics and Astronomy, University of Calgary, 2500 University Drive NW, Calgary, Alberta T2N 1N4 (Canada) and Department of Oncology, Tom Baker Cancer Centre, 1331-29 Street NW, Calgary, Alberta T2N 4N2 (Canada)
2010-07-15T23:59:59.000Z
Purpose: Patient specific IMRT QC is performed routinely in many clinics as a safeguard against errors and inaccuracies which may be introduced during the complex planning, data transfer, and delivery phases of this type of treatment. The purpose of this work is to evaluate the feasibility of detecting systematic errors in MLC leaf bank position with patient specific checks. Methods: 9 head and neck (H and N) and 14 prostate IMRT beams were delivered using MLC files containing systematic offsets ({+-}1 mm in two banks, {+-}0.5 mm in two banks, and 1 mm in one bank of leaves). The beams were measured using both MAPCHECK (Sun Nuclear Corp., Melbourne, FL) and the aS1000 electronic portal imaging device (Varian Medical Systems, Palo Alto, CA). Comparisons with calculated fields, without offsets, were made using commonly adopted criteria including absolute dose (AD) difference, relative dose difference, distance to agreement (DTA), and the gamma index. Results: The criteria most sensitive to systematic leaf bank offsets were the 3% AD, 3 mm DTA for MAPCHECK and the gamma index with 2% AD and 2 mm DTA for the EPID. The criterion based on the relative dose measurements was the least sensitive to MLC offsets. More highly modulated fields, i.e., H and N, showed greater changes in the percentage of passing points due to systematic MLC inaccuracy than prostate fields. Conclusions: None of the techniques or criteria tested is sufficiently sensitive, with the population of IMRT fields, to detect a systematic MLC offset at a clinically significant level on an individual field. Patient specific QC cannot, therefore, substitute for routine QC of the MLC itself.
Kraan, Aafke C., E-mail: aafke.kraan@pi.infn.it [Erasmus MC Daniel den Hoed Cancer Center, Rotterdam (Netherlands); Water, Steven van de; Teguh, David N.; Al-Mamgani, Abrahim [Erasmus MC Daniel den Hoed Cancer Center, Rotterdam (Netherlands); Madden, Tom; Kooy, Hanne M. [Massachusetts General Hospital and Harvard Medical School, Boston, Massachusetts (United States); Heijmen, Ben J.M.; Hoogeman, Mischa S. [Erasmus MC Daniel den Hoed Cancer Center, Rotterdam (Netherlands)
2013-12-01T23:59:59.000Z
Purpose: Setup, range, and anatomical uncertainties influence the dose delivered with intensity modulated proton therapy (IMPT), but clinical quantification of these errors for oropharyngeal cancer is lacking. We quantified these factors and investigated treatment fidelity, that is, robustness, as influenced by adaptive planning and by applying more beam directions. Methods and Materials: We used an in-house treatment planning system with multicriteria optimization of pencil beam energies, directions, and weights to create treatment plans for 3-, 5-, and 7-beam directions for 10 oropharyngeal cancer patients. The dose prescription was a simultaneously integrated boost scheme, prescribing 66 Gy to primary tumor and positive neck levels (clinical target volume-66 Gy; CTV-66 Gy) and 54 Gy to elective neck levels (CTV-54 Gy). Doses were recalculated in 3700 simulations of setup, range, and anatomical uncertainties. Repeat computed tomography (CT) scans were used to evaluate an adaptive planning strategy using nonrigid registration for dose accumulation. Results: For the recalculated 3-beam plans including all treatment uncertainty sources, only 69% (CTV-66 Gy) and 88% (CTV-54 Gy) of the simulations had a dose received by 98% of the target volume (D98%) >95% of the prescription dose. Doses to organs at risk (OARs) showed considerable spread around planned values. Causes for major deviations were mixed. Adaptive planning based on repeat imaging positively affected dose delivery accuracy: in the presence of the other errors, percentages of treatments with D98% >95% increased to 96% (CTV-66 Gy) and 100% (CTV-54 Gy). Plans with more beam directions were not more robust. Conclusions: For oropharyngeal cancer patients, treatment uncertainties can result in significant differences between planned and delivered IMPT doses. Given the mixed causes for major deviations, we advise repeat diagnostic CT scans during treatment, recalculation of the dose, and if required, adaptive planning to improve adequate IMPT dose delivery.
Bounded-Error Quantum State Identification and Exponential Separations in Communication Complexity
Dmitry Gavinsky; Julia Kempe; Oded Regev; Ronald de Wolf
2005-11-02T23:59:59.000Z
We consider the problem of bounded-error quantum state identification: given either state \\alpha_0 or state \\alpha_1, we are required to output `0', `1' or `?' ("don't know"), such that conditioned on outputting `0' or `1', our guess is correct with high probability. The goal is to maximize the probability of not outputting `?'. We prove a direct product theorem: if we're given two such problems, with optimal probabilities a and b, respectively, and the states in the first problem are pure, then the optimal probability for the joint bounded-error state identification problem is O(ab). Our proof is based on semidefinite programming duality and may be of wider interest. Using this result, we present two exponential separations in the simultaneous message passing model of communication complexity. Both are shown in the strongest possible sense. First, we describe a relation that can be computed with O(log n) classical bits of communication in the presence of shared randomness, but needs Omega(n^{1/3}) communication if the parties don't share randomness, even if communication is quantum. This shows the optimality of Yao's recent exponential simulation of shared-randomness protocols by quantum protocols without shared randomness. Second, we describe a relation that can be computed with O(log n) classical bits of communication in the presence of shared entanglement, but needs Omega((n/log n)^{1/3}) communication if the parties share randomness but no entanglement, even if communication is quantum. This is the first example in communication complexity of a situation where entanglement buys you much more than quantum communication does.
Meldi, M.; Sagaut, P. [Institut Jean Le Rond d'Alembert, UMR 7190, Universite Pierre et Marie Curie - Paris 6, F-75005 Paris (France); Salvetti, M. V. [Dipartimento di Ingegneria Aerospaziale, Universita di Pisa, I-56122 Pisa (Italy)
2012-03-15T23:59:59.000Z
A stochastic approach based on generalized polynomial chaos (gPC) is used to quantify the error in large-eddy simulation (LES) of a spatially evolving mixing layer flow and its sensitivity to different simulation parameters, viz., the grid stretching in the streamwise and lateral directions and the subgrid-scale (SGS) Smagorinsky model constant (C{sub S}). The error is evaluated with respect to the results of a highly resolved LES and for different quantities of interest, namely, the mean streamwise velocity, the momentum thickness, and the shear stress. A typical feature of the considered spatially evolving flow is the progressive transition from a laminar regime, highly dependent on the inlet conditions, to a fully developed turbulent one. Therefore, the computational domain is divided in two different zones (inlet dependent and fully turbulent) and the gPC error analysis is carried out for these two zones separately. An optimization of the parameters is also carried out for both these zones. For all the considered quantities, the results point out that the error is mainly governed by the value of the C{sub S} constant. At the end of the inlet-dependent zone, a strong coupling between the normal stretching ratio and the C{sub S} value is observed. The error sensitivity to the parameter values is significantly larger in the inlet-dependent upstream region; however, low-error values can be obtained in this region for all the considered physical quantities by an ad hoc tuning of the parameters. Conversely, in the turbulent regime the error is globally lower and less sensitive to the parameter variations, but it is more difficult to find a set of parameter values leading to optimal results for all the analyzed physical quantities. A similar analysis is also carried out for the dynamic Smagorinsky model, by varying the grid stretching ratios. Comparing the databases generated with the different subgrid-scale models, it is possible to observe that the error cost function computed for the streamwise velocity and for the momentum thickness is not significantly sensitive to the used SGS closure. Conversely, the prediction of the shear stress is much more accurate when using a dynamic subgrid-scale model and the variance of the error is lower in magnitude.
On Termination and Derivation Lengths for Ground Rewrite Systems
Giesl, Juergen
On Termination and Derivation Lengths for Ground Rewrite Systems Dieter Hofbauer 1 UniversitÂ¨at GH@theory.informatik.uniÂkassel.de Abstract. It is shown that for terminating ground term rewrite systems the length of derivations a suitable interpretation into the natural numbers. Terminating ground systems are not necessarily
Multi-Factor Energy Price Models Exotic Derivatives Pricing
Jaimungal, Sebastian
Multi-Factor Energy Price Models and Exotic Derivatives Pricing by Samuel Hikspoors A thesis of Statistics University of Toronto c Copyright by Samuel Hikspoors 2008 #12;Multi-Factor Energy Price Models and practitioners alike recently started to develop the tools of energy derivatives pricing
Faster Valuation of Financial Derivatives S. Paskov \\Lambda
methods with Monte Carlo methods on the valuation of financial derivatives. We use a CollateralizedFaster Valuation of Financial Derivatives S. Paskov \\Lambda Department of Computer Science Columbia financial instruments. Vast sums are spent annually on these methods. Monte Carlo methods use random (or
Intercomparison and validation of continental water level products derived
Paris-Sud XI, UniversitÃ© de
Abstract. Satellite radar altimeter measurements of lake and reservoir water levels complement in situIntercomparison and validation of continental water level products derived from satellite radar water level products derived from satellite radar altimetry Martina Ricko,a Charon M. Birkett,b James A
Advertising in Google Search Deriving a bidding strategy
Boucherie, Richard J.
Advertising in Google Search Deriving a bidding strategy in the biggest auction on earth. Anne This research aims to derive an optimal bidding strategy for keyword auctions in Google Search from at the University of Twente. The project was commis- sioned by Dennis Doubovski, owner of Social Mining, a company
TR-IIS-07-002 Automatic Derivation of
Chen, Sheng-Wei
TR-IIS-07-002 Automatic Derivation of Compositional Rules in Automated Compositional Reasoning Bow-Yaw://www.iis.sinica.edu.tw/LIB/TechReport/tr2007/tr07.html #12;AUTOMATIC DERIVATION OF COMPOSITIONAL RULES IN AUTOMATED COMPOSITIONAL REASONING BOW-YAW algorithm. Consider, for example, the following assume-guarantee rule where M |= P denotes that the system M
15 Software Product Line Engineering with the UML: Deriving Products
Paris-Sud XI, Université de
15 Software Product Line Engineering with the UML: Deriving Products T. Ziadi and J.-M. Jézéquel Abstract Software product line engineering introduces two new dimensions into the traditional engineering of software-based systems: the variability modeling and the product derivation. The variability gathers
Derivation of Locational Marginal Prices for Restructured Wholesale Power Markets
Tesfatsion, Leigh
Derivation of Locational Marginal Prices for Restructured Wholesale Power Markets Haifeng Liu restructured wholesale power markets, the detailed derivation of LMPs as actually used in industry practice Operator (MISO). Keywords: Locational marginal pricing, wholesale power market, AC optimal power flow, DC