National Library of Energy BETA

Sample records for nonsampling errors measures

  1. Errors of Nonobservation

    U.S. Energy Information Administration (EIA) Indexed Site

    Errors of Nonobservation Finally, several potential sources of nonsampling error and bias result from errors of nonobservation. The 1994 MECS represents, in terms of sampling...

  2. Impact of Measurement Error on Synchrophasor Applications

    SciTech Connect (OSTI)

    Liu, Yilu; Gracia, Jose R.; Ewing, Paul D.; Zhao, Jiecheng; Tan, Jin; Wu, Ling; Zhan, Lingwei

    2015-07-01

    Phasor measurement units (PMUs), a type of synchrophasor, are powerful diagnostic tools that can help avert catastrophic failures in the power grid. Because of this, PMU measurement errors are particularly worrisome. This report examines the internal and external factors contributing to PMU phase angle and frequency measurement errors and gives a reasonable explanation for them. It also analyzes the impact of those measurement errors on several synchrophasor applications: event location detection, oscillation detection, islanding detection, and dynamic line rating. The primary finding is that dynamic line rating is more likely to be influenced by measurement error. Other findings include the possibility of reporting nonoscillatory activity as an oscillation as the result of error, failing to detect oscillations submerged by error, and the unlikely impact of error on event location and islanding detection.

  3. Neutron multiplication error in TRU waste measurements

    SciTech Connect (OSTI)

    Veilleux, John [Los Alamos National Laboratory; Stanfield, Sean B [CCP; Wachter, Joe [CCP; Ceo, Bob [CCP

    2009-01-01

    Total Measurement Uncertainty (TMU) in neutron assays of transuranic waste (TRU) are comprised of several components including counting statistics, matrix and source distribution, calibration inaccuracy, background effects, and neutron multiplication error. While a minor component for low plutonium masses, neutron multiplication error is often the major contributor to the TMU for items containing more than 140 g of weapons grade plutonium. Neutron multiplication arises when neutrons from spontaneous fission and other nuclear events induce fissions in other fissile isotopes in the waste, thereby multiplying the overall coincidence neutron response in passive neutron measurements. Since passive neutron counters cannot differentiate between spontaneous and induced fission neutrons, multiplication can lead to positive bias in the measurements. Although neutron multiplication can only result in a positive bias, it has, for the purpose of mathematical simplicity, generally been treated as an error that can lead to either a positive or negative result in the TMU. While the factors that contribute to neutron multiplication include the total mass of fissile nuclides, the presence of moderating material in the matrix, the concentration and geometry of the fissile sources, and other factors; measurement uncertainty is generally determined as a function of the fissile mass in most TMU software calculations because this is the only quantity determined by the passive neutron measurement. Neutron multiplication error has a particularly pernicious consequence for TRU waste analysis because the measured Fissile Gram Equivalent (FGE) plus twice the TMU error must be less than 200 for TRU waste packaged in 55-gal drums and less than 325 for boxed waste. For this reason, large errors due to neutron multiplication can lead to increased rejections of TRU waste containers. This report will attempt to better define the error term due to neutron multiplication and arrive at values that are

  4. Slope Error Measurement Tool for Solar Parabolic Trough Collectors: Preprint

    SciTech Connect (OSTI)

    Stynes, J. K.; Ihas, B.

    2012-04-01

    The National Renewable Energy Laboratory (NREL) has developed an optical measurement tool for parabolic solar collectors that measures the combined errors due to absorber misalignment and reflector slope error. The combined absorber alignment and reflector slope errors are measured using a digital camera to photograph the reflected image of the absorber in the collector. Previous work using the image of the reflection of the absorber finds the reflector slope errors from the reflection of the absorber and an independent measurement of the absorber location. The accuracy of the reflector slope error measurement is thus dependent on the accuracy of the absorber location measurement. By measuring the combined reflector-absorber errors, the uncertainty in the absorber location measurement is eliminated. The related performance merit, the intercept factor, depends on the combined effects of the absorber alignment and reflector slope errors. Measuring the combined effect provides a simpler measurement and a more accurate input to the intercept factor estimate. The minimal equipment and setup required for this measurement technique make it ideal for field measurements.

  5. Pressure Change Measurement Leak Testing Errors

    SciTech Connect (OSTI)

    Pryor, Jeff M; Walker, William C

    2014-01-01

    A pressure change test is a common leak testing method used in construction and Non-Destructive Examination (NDE). The test is known as being a fast, simple, and easy to apply evaluation method. While this method may be fairly quick to conduct and require simple instrumentation, the engineering behind this type of test is more complex than is apparent on the surface. This paper intends to discuss some of the more common errors made during the application of a pressure change test and give the test engineer insight into how to correctly compensate for these factors. The principals discussed here apply to ideal gases such as air or other monoatomic or diatomic gasses; however these same principals can be applied to polyatomic gasses or liquid flow rate with altered formula specific to those types of tests using the same methodology.

  6. Error and uncertainty in Raman thermal conductivity measurements

    DOE Public Access Gateway for Energy & Science Beta (PAGES Beta)

    Thomas Edwin Beechem; Yates, Luke; Graham, Samuel

    2015-04-22

    We investigated error and uncertainty in Raman thermal conductivity measurements via finite element based numerical simulation of two geometries often employed -- Joule-heating of a wire and laser-heating of a suspended wafer. Using this methodology, the accuracy and precision of the Raman-derived thermal conductivity are shown to depend on (1) assumptions within the analytical model used in the deduction of thermal conductivity, (2) uncertainty in the quantification of heat flux and temperature, and (3) the evolution of thermomechanical stress during testing. Apart from the influence of stress, errors of 5% coupled with uncertainties of ±15% are achievable for most materialsmore » under conditions typical of Raman thermometry experiments. Error can increase to >20%, however, for materials having highly temperature dependent thermal conductivities or, in some materials, when thermomechanical stress develops concurrent with the heating. A dimensionless parameter -- termed the Raman stress factor -- is derived to identify when stress effects will induce large levels of error. Together, the results compare the utility of Raman based conductivity measurements relative to more established techniques while at the same time identifying situations where its use is most efficacious.« less

  7. Error and uncertainty in Raman thermal conductivity measurements

    SciTech Connect (OSTI)

    Thomas Edwin Beechem; Yates, Luke; Graham, Samuel

    2015-04-22

    We investigated error and uncertainty in Raman thermal conductivity measurements via finite element based numerical simulation of two geometries often employed -- Joule-heating of a wire and laser-heating of a suspended wafer. Using this methodology, the accuracy and precision of the Raman-derived thermal conductivity are shown to depend on (1) assumptions within the analytical model used in the deduction of thermal conductivity, (2) uncertainty in the quantification of heat flux and temperature, and (3) the evolution of thermomechanical stress during testing. Apart from the influence of stress, errors of 5% coupled with uncertainties of ±15% are achievable for most materials under conditions typical of Raman thermometry experiments. Error can increase to >20%, however, for materials having highly temperature dependent thermal conductivities or, in some materials, when thermomechanical stress develops concurrent with the heating. A dimensionless parameter -- termed the Raman stress factor -- is derived to identify when stress effects will induce large levels of error. Together, the results compare the utility of Raman based conductivity measurements relative to more established techniques while at the same time identifying situations where its use is most efficacious.

  8. A Bayesian Measurment Error Model for Misaligned Radiographic Data

    DOE Public Access Gateway for Energy & Science Beta (PAGES Beta)

    Lennox, Kristin P.; Glascoe, Lee G.

    2013-09-06

    An understanding of the inherent variability in micro-computed tomography (micro-CT) data is essential to tasks such as statistical process control and the validation of radiographic simulation tools. The data present unique challenges to variability analysis due to the relatively low resolution of radiographs, and also due to minor variations from run to run which can result in misalignment or magnification changes between repeated measurements of a sample. Positioning changes artificially inflate the variability of the data in ways that mask true physical phenomena. We present a novel Bayesian nonparametric regression model that incorporates both additive and multiplicative measurement error inmore » addition to heteroscedasticity to address this problem. We also use this model to assess the effects of sample thickness and sample position on measurement variability for an aluminum specimen. Supplementary materials for this article are available online.« less

  9. A Bayesian Measurment Error Model for Misaligned Radiographic Data

    SciTech Connect (OSTI)

    Lennox, Kristin P.; Glascoe, Lee G.

    2013-09-06

    An understanding of the inherent variability in micro-computed tomography (micro-CT) data is essential to tasks such as statistical process control and the validation of radiographic simulation tools. The data present unique challenges to variability analysis due to the relatively low resolution of radiographs, and also due to minor variations from run to run which can result in misalignment or magnification changes between repeated measurements of a sample. Positioning changes artificially inflate the variability of the data in ways that mask true physical phenomena. We present a novel Bayesian nonparametric regression model that incorporates both additive and multiplicative measurement error in addition to heteroscedasticity to address this problem. We also use this model to assess the effects of sample thickness and sample position on measurement variability for an aluminum specimen. Supplementary materials for this article are available online.

  10. The impact of response measurement error on the analysis of designed experiments

    SciTech Connect (OSTI)

    Anderson-Cook, Christine Michaela; Hamada, Michael Scott; Burr, Thomas Lee

    2015-12-21

    This study considers the analysis of designed experiments when there is measurement error in the true response or so-called response measurement error. We consider both additive and multiplicative response measurement errors. Through a simulation study, we investigate the impact of ignoring the response measurement error in the analysis, that is, by using a standard analysis based on t-tests. In addition, we examine the role of repeat measurements in improving the quality of estimation and prediction in the presence of response measurement error. We also study a Bayesian approach that accounts for the response measurement error directly through the specification of the model, and allows including additional information about variability in the analysis. We consider the impact on power, prediction, and optimization. Copyright © 2015 John Wiley & Sons, Ltd.

  11. The impact of response measurement error on the analysis of designed experiments

    DOE Public Access Gateway for Energy & Science Beta (PAGES Beta)

    Anderson-Cook, Christine Michaela; Hamada, Michael Scott; Burr, Thomas Lee

    2015-12-21

    This study considers the analysis of designed experiments when there is measurement error in the true response or so-called response measurement error. We consider both additive and multiplicative response measurement errors. Through a simulation study, we investigate the impact of ignoring the response measurement error in the analysis, that is, by using a standard analysis based on t-tests. In addition, we examine the role of repeat measurements in improving the quality of estimation and prediction in the presence of response measurement error. We also study a Bayesian approach that accounts for the response measurement error directly through the specification ofmore » the model, and allows including additional information about variability in the analysis. We consider the impact on power, prediction, and optimization. Copyright © 2015 John Wiley & Sons, Ltd.« less

  12. Correction of motion measurement errors beyond the range resolution of a synthetic aperture radar

    DOE Patents [OSTI]

    Doerry, Armin W. (Albuquerque, NM); Heard, Freddie E. (Albuquerque, NM); Cordaro, J. Thomas (Albuquerque, NM)

    2008-06-24

    Motion measurement errors that extend beyond the range resolution of a synthetic aperture radar (SAR) can be corrected by effectively decreasing the range resolution of the SAR in order to permit measurement of the error. Range profiles can be compared across the slow-time dimension of the input data in order to estimate the error. Once the error has been determined, appropriate frequency and phase correction can be applied to the uncompressed input data, after which range and azimuth compression can be performed to produce a desired SAR image.

  13. Techniques for reducing error in the calorimetric measurement of low wattage items

    SciTech Connect (OSTI)

    Sedlacek, W.A.; Hildner, S.S.; Camp, K.L.; Cremers, T.L.

    1993-08-01

    The increased need for the measurement of low wattage items with production calorimeters has required the development of techniques to maximize the precision and accuracy of the calorimeter measurements. An error model for calorimetry measurements is presented. This model is used as a basis for optimizing calorimetry measurements through baseline interpolation. The method was applied to the heat measurement of over 100 items and the results compared to chemistry assay and mass spectroscopy.

  14. Sinusoidal Siemens star spatial frequency response measurement errors due to misidentified target centers

    SciTech Connect (OSTI)

    Birch, Gabriel Carisle; Griffin, John Clark

    2015-07-23

    Numerous methods are available to measure the spatial frequency response (SFR) of an optical system. A recent change to the ISO 12233 photography resolution standard includes a sinusoidal Siemens star test target. We take the sinusoidal Siemens star proposed by the ISO 12233 standard, measure system SFR, and perform an analysis of errors induced by incorrectly identifying the center of a test target. We show a closed-form solution for the radial profile intensity measurement given an incorrectly determined center and describe how this error reduces the measured SFR of the system. As a result, using the closed-form solution, we propose a two-step process by which test target centers are corrected and the measured SFR is restored to the nominal, correctly centered values.

  15. Sinusoidal Siemens star spatial frequency response measurement errors due to misidentified target centers

    DOE Public Access Gateway for Energy & Science Beta (PAGES Beta)

    Birch, Gabriel Carisle; Griffin, John Clark

    2015-07-23

    Numerous methods are available to measure the spatial frequency response (SFR) of an optical system. A recent change to the ISO 12233 photography resolution standard includes a sinusoidal Siemens star test target. We take the sinusoidal Siemens star proposed by the ISO 12233 standard, measure system SFR, and perform an analysis of errors induced by incorrectly identifying the center of a test target. We show a closed-form solution for the radial profile intensity measurement given an incorrectly determined center and describe how this error reduces the measured SFR of the system. As a result, using the closed-form solution, we proposemore » a two-step process by which test target centers are corrected and the measured SFR is restored to the nominal, correctly centered values.« less

  16. Effects of Spectral Error in Efficiency Measurements of GaInAs-Based Concentrator Solar Cells

    SciTech Connect (OSTI)

    Osterwald, C. R.; Wanlass, M. W.; Moriarty, T.; Steiner, M. A.; Emery, K. A.

    2014-03-01

    This technical report documents a particular error in efficiency measurements of triple-absorber concentrator solar cells caused by incorrect spectral irradiance -- specifically, one that occurs when the irradiance from unfiltered, pulsed xenon solar simulators into the GaInAs bottom subcell is too high. For cells designed so that the light-generated photocurrents in the three subcells are nearly equal, this condition can cause a large increase in the measured fill factor, which, in turn, causes a significant artificial increase in the efficiency. The error is readily apparent when the data under concentration are compared to measurements with correctly balanced photocurrents, and manifests itself as discontinuities in plots of fill factor and efficiency versus concentration ratio. In this work, we simulate the magnitudes and effects of this error with a device-level model of two concentrator cell designs, and demonstrate how a new Spectrolab, Inc., Model 460 Tunable-High Intensity Pulsed Solar Simulator (T-HIPSS) can mitigate the error.

  17. X:\\L6046\\Data_Publication\\Pma\\current\\ventura\\pma.vp

    U.S. Energy Information Administration (EIA) Indexed Site

    a complete enumeration has the same nonsampling errors as the sample survey. The sampling error, or standard error of the estimate, is a measure of the variability among the...

  18. Explanatory Notes

    U.S. Energy Information Administration (EIA) Indexed Site

    complete enumera- tion has the same nonsampling errors as the sample survey. The sampling error, or standard error of the estimate, is a measure of the variability among the...

  19. X:\\Data_Publication\\Pma\\current\\ventura\\pma00.vp

    U.S. Energy Information Administration (EIA) Indexed Site

    a complete enumeration has the same nonsampling errors as the sample survey. The sampling error, or standard error of the estimate, is a measure of the variability among the...

  20. Accuracy of the European solar water heater test procedure. Part 1: Measurement errors and parameter estimates

    SciTech Connect (OSTI)

    Rabl, A.; Leide, B. ); Carvalho, M.J.; Collares-Pereira, M. ); Bourges, B.

    1991-01-01

    The Collector and System Testing Group (CSTG) of the European Community has developed a procedure for testing the performance of solar water heaters. This procedure treats a solar water heater as a black box with input-output parameters that are determined by all-day tests. In the present study the authors carry out a systematic analysis of the accuracy of this procedure, in order to answer the question: what tolerances should one impose for the measurements and how many days of testing should one demand under what meteorological conditions, in order to be able to quarantee a specified maximum error for the long term performance The methodology is applicable to other test procedures as well. The present paper (Part 1) examines the measurement tolerances of the current version of the procedure and derives a priori estimates of the errors of the parameters; these errors are then compared with the regression results of the Round Robin test series. The companion paper (Part 2) evaluates the consequences for the accuracy of the long term performance prediction. The authors conclude that the CSTG test procedure makes it possible to predict the long term performance with standard errors around 5% for sunny climates (10% for cloudy climates). The apparent precision of individual test sequences is deceptive because of large systematic discrepancies between different sequences. Better results could be obtained by imposing tighter control on the constancy of the cold water supply temperature and on the environment of the test, the latter by enforcing the recommendation for the ventilation of the collector.

  1. Decreasing range resolution of a SAR image to permit correction of motion measurement errors beyond the SAR range resolution

    DOE Patents [OSTI]

    Doerry, Armin W. (Albuquerque, NM); Heard, Freddie E. (Albuquerque, NM); Cordaro, J. Thomas (Albuquerque, NM)

    2010-07-20

    Motion measurement errors that extend beyond the range resolution of a synthetic aperture radar (SAR) can be corrected by effectively decreasing the range resolution of the SAR in order to permit measurement of the error. Range profiles can be compared across the slow-time dimension of the input data in order to estimate the error. Once the error has been determined, appropriate frequency and phase correction can be applied to the uncompressed input data, after which range and azimuth compression can be performed to produce a desired SAR image.

  2. Quantifying error of lidar and sodar Doppler beam swinging measurements of wind turbine wakes using computational fluid dynamics

    SciTech Connect (OSTI)

    Lundquist, J. K.; Churchfield, M. J.; Lee, S.; Clifton, A.

    2015-02-23

    Wind-profiling lidars are now regularly used in boundary-layer meteorology and in applications such as wind energy and air quality. Lidar wind profilers exploit the Doppler shift of laser light backscattered from particulates carried by the wind to measure a line-of-sight (LOS) velocity. The Doppler beam swinging (DBS) technique, used by many commercial systems, considers measurements of this LOS velocity in multiple radial directions in order to estimate horizontal and vertical winds. The method relies on the assumption of homogeneous flow across the region sampled by the beams. Using such a system in inhomogeneous flow, such as wind turbine wakes or complex terrain, will result in errors.

    To quantify the errors expected from such violation of the assumption of horizontal homogeneity, we simulate inhomogeneous flow in the atmospheric boundary layer, notably stably stratified flow past a wind turbine, with a mean wind speed of 6.5 m s-1 at the turbine hub-height of 80 m. This slightly stable case results in 15° of wind direction change across the turbine rotor disk. The resulting flow field is sampled in the same fashion that a lidar samples the atmosphere with the DBS approach, including the lidar range weighting function, enabling quantification of the error in the DBS observations. The observations from the instruments located upwind have small errors, which are ameliorated with time averaging. However, the downwind observations, particularly within the first two rotor diameters downwind from the wind turbine, suffer from errors due to the heterogeneity of the wind turbine wake. Errors in the stream-wise component of the flow approach 30% of the hub-height inflow wind speed close to the rotor disk. Errors in the cross-stream and vertical velocity components are also significant: cross-stream component errors are on the order of 15% of the hub-height inflow wind speed (1.0 m s−1) and errors in the vertical velocity measurement

  3. Quantifying error of lidar and sodar Doppler beam swinging measurements of wind turbine wakes using computational fluid dynamics

    DOE Public Access Gateway for Energy & Science Beta (PAGES Beta)

    Lundquist, J. K.; Churchfield, M. J.; Lee, S.; Clifton, A.

    2015-02-23

    Wind-profiling lidars are now regularly used in boundary-layer meteorology and in applications such as wind energy and air quality. Lidar wind profilers exploit the Doppler shift of laser light backscattered from particulates carried by the wind to measure a line-of-sight (LOS) velocity. The Doppler beam swinging (DBS) technique, used by many commercial systems, considers measurements of this LOS velocity in multiple radial directions in order to estimate horizontal and vertical winds. The method relies on the assumption of homogeneous flow across the region sampled by the beams. Using such a system in inhomogeneous flow, such as wind turbine wakes ormore » complex terrain, will result in errors. To quantify the errors expected from such violation of the assumption of horizontal homogeneity, we simulate inhomogeneous flow in the atmospheric boundary layer, notably stably stratified flow past a wind turbine, with a mean wind speed of 6.5 m s-1 at the turbine hub-height of 80 m. This slightly stable case results in 15° of wind direction change across the turbine rotor disk. The resulting flow field is sampled in the same fashion that a lidar samples the atmosphere with the DBS approach, including the lidar range weighting function, enabling quantification of the error in the DBS observations. The observations from the instruments located upwind have small errors, which are ameliorated with time averaging. However, the downwind observations, particularly within the first two rotor diameters downwind from the wind turbine, suffer from errors due to the heterogeneity of the wind turbine wake. Errors in the stream-wise component of the flow approach 30% of the hub-height inflow wind speed close to the rotor disk. Errors in the cross-stream and vertical velocity components are also significant: cross-stream component errors are on the order of 15% of the hub-height inflow wind speed (1.0 m s−1) and errors in the vertical velocity measurement exceed the actual

  4. A numerical study of geometry dependent errors in velocity, temperature, and density measurements from single grid planar retarding potential analyzers

    SciTech Connect (OSTI)

    Davidson, R. L.; Earle, G. D.; Heelis, R. A.; Klenzing, J. H.

    2010-08-15

    Planar retarding potential analyzers (RPAs) have been utilized numerous times on high profile missions such as the Communications/Navigation Outage Forecast System and the Defense Meteorological Satellite Program to measure plasma composition, temperature, density, and the velocity component perpendicular to the plane of the instrument aperture. These instruments use biased grids to approximate ideal biased planes. These grids introduce perturbations in the electric potential distribution inside the instrument and when unaccounted for cause errors in the measured plasma parameters. Traditionally, the grids utilized in RPAs have been made of fine wires woven into a mesh. Previous studies on the errors caused by grids in RPAs have approximated woven grids with a truly flat grid. Using a commercial ion optics software package, errors in inferred parameters caused by both woven and flat grids are examined. A flat grid geometry shows the smallest temperature and density errors, while the double thick flat grid displays minimal errors for velocities over the temperature and velocity range used. Wire thickness along the dominant flow direction is found to be a critical design parameter in regard to errors in all three inferred plasma parameters. The results shown for each case provide valuable design guidelines for future RPA development.

  5. The Measure of Human Error: Direct and Indirect Performance Shaping Factors

    SciTech Connect (OSTI)

    Ronald L. Boring; Candice D. Griffith; Jeffrey C. Joe

    2007-08-01

    The goal of performance shaping factors (PSFs) is to provide measures to account for human performance. PSFs fall into two categoriesdirect and indirect measures of human performance. While some PSFs such as time to complete a task are directly measurable, other PSFs, such as fitness for duty, can only be measured indirectly through other measures and PSFs, such as through fatigue measures. This paper explores the role of direct and indirect measures in human reliability analysis (HRA) and the implications that measurement theory has on analyses and applications using PSFs. The paper concludes with suggestions for maximizing the reliability and validity of PSFs.

  6. Comparing range data across the slow-time dimension to correct motion measurement errors beyond the range resolution of a synthetic aperture radar

    DOE Patents [OSTI]

    Doerry, Armin W. (Albuquerque, NM); Heard, Freddie E. (Albuquerque, NM); Cordaro, J. Thomas (Albuquerque, NM)

    2010-08-17

    Motion measurement errors that extend beyond the range resolution of a synthetic aperture radar (SAR) can be corrected by effectively decreasing the range resolution of the SAR in order to permit measurement of the error. Range profiles can be compared across the slow-time dimension of the input data in order to estimate the error. Once the error has been determined, appropriate frequency and phase correction can be applied to the uncompressed input data, after which range and azimuth compression can be performed to produce a desired SAR image.

  7. Effects of Spectral Error in Efficiency Measurements of GaInAs...

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    ... This report is available at no cost from the National Renewable Energy Laboratory (NREL) at www.nrel.govpublications. 3.2 Excess Irradiance Estimation Without measurements of the ...

  8. Volumetric apparatus for hydrogen adsorption and diffusion measurements: Sources of systematic error and impact of their experimental resolutions

    SciTech Connect (OSTI)

    Policicchio, Alfonso; Maccallini, Enrico; Kalantzopoulos, Georgios N.; Cataldi, Ugo; Abate, Salvatore; Desiderio, Giovanni; DeltaE s.r.l., c/o Universit della Calabria, Via Pietro Bucci cubo 31D, 87036 Arcavacata di Rende , Italy and CNR-IPCF LiCryL, c/o Universit della Calabria, Via Ponte P. Bucci, Cubo 31C, 87036 Arcavacata di Rende

    2013-10-15

    The development of a volumetric apparatus (also known as a Sieverts apparatus) for accurate and reliable hydrogen adsorption measurement is shown. The instrument minimizes the sources of systematic errors which are mainly due to inner volume calibration, stability and uniformity of the temperatures, precise evaluation of the skeletal volume of the measured samples, and thermodynamical properties of the gas species. A series of hardware and software solutions were designed and introduced in the apparatus, which we will indicate as f-PcT, in order to deal with these aspects. The results are represented in terms of an accurate evaluation of the equilibrium and dynamical characteristics of the molecular hydrogen adsorption on two well-known porous media. The contribution of each experimental solution to the error propagation of the adsorbed moles is assessed. The developed volumetric apparatus for gas storage capacity measurements allows an accurate evaluation over a 4 order-of-magnitude pressure range (from 1 kPa to 8 MPa) and in temperatures ranging between 77 K and 470 K. The acquired results are in good agreement with the values reported in the literature.

  9. Error abstractions

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Error and fault abstractions Mattan Erez UT Austin *Who should care about faults and errors? *Ideally, only system cares about masked faults? - Assuming application bugs are not...

  10. In Situ Validation of a Correction for Time-Lag and Bias Errors in Vaisala RS80-H Radiosonde Humidity Measurements

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    In Situ Validation of a Correction for Time-Lag and Bias Errors in Vaisala RS80-H Radiosonde Humidity Measurements L. M. Miloshevich National Center for Atmospheric Research Boulder, Colorado H. Vömel and S. J. Oltmans National Oceanic and Atmospheric Administration Boulder, Colorado A. Paukkunen Vaisala Oy Helsinki, Finland Introduction Radiosonde relative humidity (RH) measurements are fundamentally important to Atmospheric Radiation Measurement (ARM) Program goals because they are used in a

  11. Tuning the narrow-band beam position monitor sampling clock to remove the aliasing errors in APS storage ring orbit measurements.

    SciTech Connect (OSTI)

    Sun, X.; Singh, O. )

    2007-01-01

    The Advanced Photon Source storage ring employs a real-time orbit correction system to reduce orbit motion up to 50 Hz. This system uses up to 142 narrow-band rf beam position monitors (Nbbpms) in a correction algorithm by sampling at a frequency of 1.53 kHz. Several Nbbpms exhibit aliasing errors in orbit measurements, rendering these Nbbpms unusable in real-time orbit feedback. The aliasing errors are caused by beating effects of the internal sampling clocks with various other processing clocks residing within the BPM electronics. A programmable external clock has been employed to move the aliasing errors out of the active frequency band of the real-time feedback system (RTFB) and rms beam motion calculation. This paper discusses the process of tuning and provides test results.

  12. Statistical and systematic errors in the measurement of weak-lensing Minkowski functionals: Application to the Canada-France-Hawaii Lensing Survey

    SciTech Connect (OSTI)

    Shirasaki, Masato; Yoshida, Naoki

    2014-05-01

    The measurement of cosmic shear using weak gravitational lensing is a challenging task that involves a number of complicated procedures. We study in detail the systematic errors in the measurement of weak-lensing Minkowski Functionals (MFs). Specifically, we focus on systematics associated with galaxy shape measurements, photometric redshift errors, and shear calibration correction. We first generate mock weak-lensing catalogs that directly incorporate the actual observational characteristics of the Canada-France-Hawaii Lensing Survey (CFHTLenS). We then perform a Fisher analysis using the large set of mock catalogs for various cosmological models. We find that the statistical error associated with the observational effects degrades the cosmological parameter constraints by a factor of a few. The Subaru Hyper Suprime-Cam (HSC) survey with a sky coverage of ?1400 deg{sup 2} will constrain the dark energy equation of the state parameter with an error of ?w {sub 0} ? 0.25 by the lensing MFs alone, but biases induced by the systematics can be comparable to the 1? error. We conclude that the lensing MFs are powerful statistics beyond the two-point statistics only if well-calibrated measurement of both the redshifts and the shapes of source galaxies is performed. Finally, we analyze the CFHTLenS data to explore the ability of the MFs to break degeneracies between a few cosmological parameters. Using a combined analysis of the MFs and the shear correlation function, we derive the matter density ?{sub m0}=0.256{sub 0.046}{sup 0.054}.

  13. EIA - Sorry! Unexpected Error

    U.S. Energy Information Administration (EIA) Indexed Site

    Cold Fusion Error Unexpected Error Sorry An error was encountered. This error could be due to scheduled maintenance. Information about the error has been routed to the appropriate...

  14. EIA - Sorry! Unexpected Error

    Gasoline and Diesel Fuel Update (EIA)

    Cold Fusion Error Unexpected Error Sorry An error was encountered. This error could be due to scheduled maintenance. Information about the error has been routed to the appropriate ...

  15. Error Page

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    script writes out the header html. We are sorry to report that an error has occurred. Internal identifier for doc type not found. Return to RevCom | Return to Web Portal Need help? Email Technical Support. This site managed by the Office of Management / US Department of Energy Directives | Regulations | Technical Standards | Reference Library | DOE Forms | About Us | Privacy & Security Notice This script breaks up the email address to avoid spam

  16. Error detection method

    DOE Patents [OSTI]

    Olson, Eric J.

    2013-06-11

    An apparatus, program product, and method that run an algorithm on a hardware based processor, generate a hardware error as a result of running the algorithm, generate an algorithm output for the algorithm, compare the algorithm output to another output for the algorithm, and detect the hardware error from the comparison. The algorithm is designed to cause the hardware based processor to heat to a degree that increases the likelihood of hardware errors to manifest, and the hardware error is observable in the algorithm output. As such, electronic components may be sufficiently heated and/or sufficiently stressed to create better conditions for generating hardware errors, and the output of the algorithm may be compared at the end of the run to detect a hardware error that occurred anywhere during the run that may otherwise not be detected by traditional methodologies (e.g., due to cooling, insufficient heat and/or stress, etc.).

  17. Trouble Shooting and Error Messages

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    ... Check the error code of your application. error obtaining user credentials system Resubmit. Contact consultants for repeated problems. nemgnierrorhandler(): a transaction error ...

  18. Error handling strategies in multiphase inverse modeling

    SciTech Connect (OSTI)

    Finsterle, S.; Zhang, Y.

    2010-12-01

    Parameter estimation by inverse modeling involves the repeated evaluation of a function of residuals. These residuals represent both errors in the model and errors in the data. In practical applications of inverse modeling of multiphase flow and transport, the error structure of the final residuals often significantly deviates from the statistical assumptions that underlie standard maximum likelihood estimation using the least-squares method. Large random or systematic errors are likely to lead to convergence problems, biased parameter estimates, misleading uncertainty measures, or poor predictive capabilities of the calibrated model. The multiphase inverse modeling code iTOUGH2 supports strategies that identify and mitigate the impact of systematic or non-normal error structures. We discuss these approaches and provide an overview of the error handling features implemented in iTOUGH2.

  19. An estimate of the error caused by the elongation of the wavelength in a focused beam in free-space electromagnetic parameters measurement

    SciTech Connect (OSTI)

    Zhang, Yunpeng; Li, En Guo, Gaofeng; Xu, Jiadi; Wang, Chao

    2014-09-15

    A pair of spot-focusing horn lens antenna is the key component in a free-space measurement system. The electromagnetic constitutive parameters of a planar sample are determined using transmitted and reflected electromagnetic beams. These parameters are obtained from the measured scattering parameters by the microwave network analyzer, thickness of the sample, and wavelength of a focused beam on the sample. Free-space techniques introduced by most papers consider the focused wavelength as the free-space wavelength. But in fact, the incident wave projected by a lens into the sample approximates a Gaussian beam, thus, there has an elongation of the wavelength in the focused beam and this elongation should be taken into consideration in dielectric and magnetic measurement. In this paper, elongation of the wavelength has been analyzed and measured. Measurement results show that the focused wavelength in the vicinity of the focus has an elongation of 1%5% relative to the free-space wavelength. Elongation's influence on the measurement result of the permittivity and permeability has been investigated. Numerical analyses show that the elongation of the focused wavelength can cause the increase of the measured value of the permeability relative to traditionally measured value, but for the permittivity, it is affected by several parameters and may increase or decrease relative to traditionally measured value.

  20. runtime error message: "readControlMsg: System returned error...

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    readControlMsg: System returned error Connection timed out on TCP socket fd" runtime error message: "readControlMsg: System returned error Connection timed out on TCP socket fd"...

  1. Trouble Shooting and Error Messages

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    ... Check the error code of your application. error obtaining user credentials system Resubmit. Contact consultants for repeated problems. NERSC and Cray are working on this issue. ...

  2. Trouble Shooting and Error Messages

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    not be a problem. Check the error code of your application. error obtaining user credentials system Resubmit. Contact consultants for repeated problems. Last edited: 2015-01-16 ...

  3. Modular error embedding

    DOE Patents [OSTI]

    Sandford, II, Maxwell T.; Handel, Theodore G.; Ettinger, J. Mark

    1999-01-01

    A method of embedding auxiliary information into the digital representation of host data containing noise in the low-order bits. The method applies to digital data representing analog signals, for example digital images. The method reduces the error introduced by other methods that replace the low-order bits with auxiliary information. By a substantially reverse process, the embedded auxiliary data can be retrieved easily by an authorized user through use of a digital key. The modular error embedding method includes a process to permute the order in which the host data values are processed. The method doubles the amount of auxiliary information that can be added to host data values, in comparison with bit-replacement methods for high bit-rate coding. The invention preserves human perception of the meaning and content of the host data, permitting the addition of auxiliary data in the amount of 50% or greater of the original host data.

  4. Error 404 - Document not found

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    govErrors ERROR 404 - URL Not Found We are sorry but the URL that you have requested cannot be found or it is linked to a file that no longer exists. Please check the spelling or...

  5. Trouble Shooting and Error Messages

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Trouble Shooting and Error Messages Trouble Shooting and Error Messages Error Messages Message or Symptom Fault Recommendation job hit wallclock time limit user or system Submit job for longer time or start job from last checkpoint and resubmit. If your job hung and produced no output contact consultants. received node failed or halted event for nid xxxx system resubmit the job error with width parameters to aprun user Make sure #PBS -l mppwidth value matches aprun -n value new values for

  6. Register file soft error recovery

    DOE Patents [OSTI]

    Fleischer, Bruce M.; Fox, Thomas W.; Wait, Charles D.; Muff, Adam J.; Watson, III, Alfred T.

    2013-10-15

    Register file soft error recovery including a system that includes a first register file and a second register file that mirrors the first register file. The system also includes an arithmetic pipeline for receiving data read from the first register file, and error detection circuitry to detect whether the data read from the first register file includes corrupted data. The system further includes error recovery circuitry to insert an error recovery instruction into the arithmetic pipeline in response to detecting the corrupted data. The inserted error recovery instruction replaces the corrupted data in the first register file with a copy of the data from the second register file.

  7. Confidence limits and their errors

    SciTech Connect (OSTI)

    Rajendran Raja

    2002-03-22

    Confidence limits are common place in physics analysis. Great care must be taken in their calculation and use especially in cases of limited statistics. We introduce the concept of statistical errors of confidence limits and argue that not only should limits be calculated but also their errors in order to represent the results of the analysis to the fullest. We show that comparison of two different limits from two different experiments becomes easier when their errors are also quoted. Use of errors of confidence limits will lead to abatement of the debate on which method is best suited to calculate confidence limits.

  8. Error localization in RHIC by fitting difference orbits

    SciTech Connect (OSTI)

    Liu C.; Minty, M.; Ptitsyn, V.

    2012-05-20

    The presence of realistic errors in an accelerator or in the model used to describe the accelerator are such that a measurement of the beam trajectory may deviate from prediction. Comparison of measurements to model can be used to detect such errors. To do so the initial conditions (phase space parameters at any point) must be determined which can be achieved by fitting the difference orbit compared to model prediction using only a few beam position measurements. Using these initial conditions, the fitted orbit can be propagated along the beam line based on the optics model. Measurement and model will agree up to the point of an error. The error source can be better localized by additionally fitting the difference orbit using downstream BPMs and back-propagating the solution. If one dominating error source exist in the machine, the fitted orbit will deviate from the difference orbit at the same point.

  9. Error studies for SNS Linac. Part 1: Transverse errors

    SciTech Connect (OSTI)

    Crandall, K.R.

    1998-12-31

    The SNS linac consist of a radio-frequency quadrupole (RFQ), a drift-tube linac (DTL), a coupled-cavity drift-tube linac (CCDTL) and a coupled-cavity linac (CCL). The RFQ and DTL are operated at 402.5 MHz; the CCDTL and CCL are operated at 805 MHz. Between the RFQ and DTL is a medium-energy beam-transport system (MEBT). This error study is concerned with the DTL, CCDTL and CCL, and each will be analyzed separately. In fact, the CCL is divided into two sections, and each of these will be analyzed separately. The types of errors considered here are those that affect the transverse characteristics of the beam. The errors that cause the beam center to be displaced from the linac axis are quad displacements and quad tilts. The errors that cause mismatches are quad gradient errors and quad rotations (roll).

  10. Error 404 - Document not found

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    govErrors ERROR 404 - URL Not Found We are sorry but the URL that you have requested cannot be found or it is linked to a file that no longer exists. Please check the spelling or send e-mail to WWW Administrator

  11. Trouble Shooting and Error Messages

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Trouble Shooting and Error Messages Trouble Shooting and Error Messages Error Messages Message or Symptom Fault Recommendation job hit wallclock time limit user or system Submit job for longer time or start job from last checkpoint and resubmit. If your job hung and produced no output contact consultants. received node failed or halted event for nid xxxx system One of the compute nodes assigned to the job failed. Resubmit the job PtlNIInit failed : PTL_NOT_REGISTERED user The executable is from

  12. Monte Carlo analysis of localization errors in magnetoencephalography

    SciTech Connect (OSTI)

    Medvick, P.A.; Lewis, P.S.; Aine, C.; Flynn, E.R.

    1989-01-01

    In magnetoencephalography (MEG), the magnetic fields created by electrical activity in the brain are measured on the surface of the skull. To determine the location of the activity, the measured field is fit to an assumed source generator model, such as a current dipole, by minimizing chi-square. For current dipoles and other nonlinear source models, the fit is performed by an iterative least squares procedure such as the Levenberg-Marquardt algorithm. Once the fit has been computed, analysis of the resulting value of chi-square can determine whether the assumed source model is adequate to account for the measurements. If the source model is adequate, then the effect of measurement error on the fitted model parameters must be analyzed. Although these kinds of simulation studies can provide a rough idea of the effect that measurement error can be expected to have on source localization, they cannot provide detailed enough information to determine the effects that the errors in a particular measurement situation will produce. In this work, we introduce and describe the use of Monte Carlo-based techniques to analyze model fitting errors for real data. Given the details of the measurement setup and a statistical description of the measurement errors, these techniques determine the effects the errors have on the fitted model parameters. The effects can then be summarized in various ways such as parameter variances/covariances or multidimensional confidence regions. 8 refs., 3 figs.

  13. error | netl.doe.gov

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    error Sorry, there is no www.netl.doe.gov web page that matches your request. It may be possible that you typed the address incorrectly. Connect to National Energy Technology...

  14. Error Reduction for Weigh-In-Motion

    SciTech Connect (OSTI)

    Hively, Lee M; Abercrombie, Robert K; Scudiere, Matthew B; Sheldon, Frederick T

    2009-01-01

    Federal and State agencies need certifiable vehicle weights for various applications, such as highway inspections, border security, check points, and port entries. ORNL weigh-in-motion (WIM) technology was previously unable to provide certifiable weights, due to natural oscillations, such as vehicle bouncing and rocking. Recent ORNL work demonstrated a novel filter to remove these oscillations. This work shows further filtering improvements to enable certifiable weight measurements (error < 0.1%) for a higher traffic volume with less effort (elimination of redundant weighing).

  15. Error Reduction in Weigh-In-Motion

    Energy Science and Technology Software Center (OSTI)

    2007-09-21

    Federal and State agencies need certifiable vehicle weights for various applications, such as highway inspections, border security, check points, and port entries. ORNL weigh-in-motion (WIM) technology was previously unable to provide certifiable weights, due to natural oscillations, such as vehicle bounding and rocking. Recent ORNL work demonstrated a novel filter to remove these oscillations. This work shows further filtering improvements to enable certifiable weight measurements (error < 0.1%) for a higher traffic volume with lessmore » effort (elimination of redundant weighing)« less

  16. Using doppler radar images to estimate aircraft navigational heading error

    DOE Patents [OSTI]

    Doerry, Armin W.; Jordan, Jay D.; Kim, Theodore J.

    2012-07-03

    A yaw angle error of a motion measurement system carried on an aircraft for navigation is estimated from Doppler radar images captured using the aircraft. At least two radar pulses aimed at respectively different physical locations in a targeted area are transmitted from a radar antenna carried on the aircraft. At least two Doppler radar images that respectively correspond to the at least two transmitted radar pulses are produced. These images are used to produce an estimate of the yaw angle error.

  17. A Possible Calorimetric Error in Heavy Water Electrolysis on Platinum

    SciTech Connect (OSTI)

    Shanahan, K.L.

    2001-03-16

    A systematic error in mass flow calorimetry calibration procedures potentially capable of explaining most positive excess power measurements is described. Data recently interpreted as providing evidence of the Pons-Fleischmann effect with a platinum cathode are reinterpreted with the opposite conclusion. This indicates it is premature to conclude platinum displays a Pons and Fleischmann effect, and places the requirement to evaluate the error's magnitude on all mass flow calorimetric experiments.

  18. WIPP Field Practices: Common Errors and Innovative Solutions | Department

    Office of Energy Efficiency and Renewable Energy (EERE) Indexed Site

    of Energy WIPP Field Practices: Common Errors and Innovative Solutions WIPP Field Practices: Common Errors and Innovative Solutions What to do when approaching an unfamiliar house for weatherization, with hidden air leakage and a multitude of mysteries? This webinar focuses on the Dos and Don'ts of WIPP weatherization, and is guaranteed to be an hour well spent looking over photographs that show detail and perspective on air sealing, blocking, venting, and other weatherization measures. This

  19. Catastrophic photometric redshift errors: Weak-lensing survey requirements

    DOE Public Access Gateway for Energy & Science Beta (PAGES Beta)

    Bernstein, Gary; Huterer, Dragan

    2010-01-11

    We study the sensitivity of weak lensing surveys to the effects of catastrophic redshift errors - cases where the true redshift is misestimated by a significant amount. To compute the biases in cosmological parameters, we adopt an efficient linearized analysis where the redshift errors are directly related to shifts in the weak lensing convergence power spectra. We estimate the number Nspec of unbiased spectroscopic redshifts needed to determine the catastrophic error rate well enough that biases in cosmological parameters are below statistical errors of weak lensing tomography. While the straightforward estimate of Nspec is ~106 we find that using onlymore » the photometric redshifts with z ≤ 2.5 leads to a drastic reduction in Nspec to ~ 30,000 while negligibly increasing statistical errors in dark energy parameters. Therefore, the size of spectroscopic survey needed to control catastrophic errors is similar to that previously deemed necessary to constrain the core of the zs – zp distribution. We also study the efficacy of the recent proposal to measure redshift errors by cross-correlation between the photo-z and spectroscopic samples. We find that this method requires ~ 10% a priori knowledge of the bias and stochasticity of the outlier population, and is also easily confounded by lensing magnification bias. In conclusion, the cross-correlation method is therefore unlikely to supplant the need for a complete spectroscopic redshift survey of the source population.« less

  20. Error Rate Comparison during Polymerase Chain Reaction by DNA Polymerase

    DOE Public Access Gateway for Energy & Science Beta (PAGES Beta)

    McInerney, Peter; Adams, Paul; Hadi, Masood Z.

    2014-01-01

    As larger-scale cloning projects become more prevalent, there is an increasing need for comparisons among high fidelity DNA polymerases used for PCR amplification. All polymerases marketed for PCR applications are tested for fidelity properties (i.e., error rate determination) by vendors, and numerous literature reports have addressed PCR enzyme fidelity. Nonetheless, it is often difficult to make direct comparisons among different enzymes due to numerous methodological and analytical differences from study to study. We have measured the error rates for 6 DNA polymerases commonly used in PCR applications, including 3 polymerases typically used for cloning applications requiring high fidelity. Errormore » rate measurement values reported here were obtained by direct sequencing of cloned PCR products. The strategy employed here allows interrogation of error rate across a very large DNA sequence space, since 94 unique DNA targets were used as templates for PCR cloning. The six enzymes included in the study, Taq polymerase, AccuPrime-Taq High Fidelity, KOD Hot Start, cloned Pfu polymerase, Phusion Hot Start, and Pwo polymerase, we find the lowest error rates with Pfu , Phusion, and Pwo polymerases. Error rates are comparable for these 3 enzymes and are >10x lower than the error rate observed with Taq polymerase. Mutation spectra are reported, with the 3 high fidelity enzymes displaying broadly similar types of mutations. For these enzymes, transition mutations predominate, with little bias observed for type of transition.« less

  1. Measurement

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    magnetic fluctuation-induced particle flux "invited... a... W. X. Ding, D. L. Brower, and T. Y. Yates Department of Physics and Astronomy, University of California-Los Angeles, Los Angeles, California 90095, USA ͑Presented 13 May 2008; received 12 May 2008; accepted 16 May 2008; published online 31 October 2008͒ Magnetic field fluctuation-induced particle transport has been directly measured in the high-temperature core of the MST reversed field pinch plasma. Measurement of radial

  2. Finite Bandwidth Related Errors in Noise Parameter Determination of PHEMTs

    SciTech Connect (OSTI)

    Wiatr, Wojciech

    2005-08-25

    We analyze errors in the determination of the four noise parameters due to finite measurement bandwidth and the delay time in the source circuit. The errors are especially large when characterizing low-noise microwave transistors at low microwave frequencies. They result from the spectral noise density variation across the measuring receiver band, due to resonant interaction of the highly mismatched transistor input with the source termination. We show also effects of virtual de-correlation of transistor's noise waves due to finite delay time at the input.

  3. Field errors in hybrid insertion devices

    SciTech Connect (OSTI)

    Schlueter, R.D.

    1995-02-01

    Hybrid magnet theory as applied to the error analyses used in the design of Advanced Light Source (ALS) insertion devices is reviewed. Sources of field errors in hybrid insertion devices are discussed.

  4. Reducing collective quantum state rotation errors with reversible dephasing

    SciTech Connect (OSTI)

    Cox, Kevin C.; Norcia, Matthew A.; Weiner, Joshua M.; Bohnet, Justin G.; Thompson, James K.

    2014-12-29

    We demonstrate that reversible dephasing via inhomogeneous broadening can greatly reduce collective quantum state rotation errors, and observe the suppression of rotation errors by more than 21?dB in the context of collective population measurements of the spin states of an ensemble of 2.110{sup 5} laser cooled and trapped {sup 87}Rb atoms. The large reduction in rotation noise enables direct resolution of spin state populations 13(1) dB below the fundamental quantum projection noise limit. Further, the spin state measurement projects the system into an entangled state with 9.5(5) dB of directly observed spectroscopic enhancement (squeezing) relative to the standard quantum limit, whereas no enhancement would have been obtained without the suppression of rotation errors.

  5. Measurement

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Neutrino Induced, Charged Current, Charged Pion Production by Michael Joseph Wilking B.Ch.E., University of Minnesota, 2001 M.S., University of Colorado, 2007 A thesis submitted to the Faculty of the Graduate School of the University of Colorado in partial fulfillment of the requirements for the degree of Doctor of Philosophy Department of Physics 2009 This thesis entitled: Measurement of Neutrino Induced, Charged Current, Charged Pion Production written by Michael Joseph Wilking has been

  6. Clover: Compiler directed lightweight soft error resilience

    SciTech Connect (OSTI)

    Liu, Qingrui; Lee, Dongyoon; Jung, Changhee; Tiwari, Devesh

    2015-05-01

    This paper presents Clover, a compiler directed soft error detection and recovery scheme for lightweight soft error resilience. The compiler carefully generates soft error tolerant code based on idem-potent processing without explicit checkpoint. During program execution, Clover relies on a small number of acoustic wave detectors deployed in the processor to identify soft errors by sensing the wave made by a particle strike. To cope with DUE (detected unrecoverable errors) caused by the sensing latency of error detection, Clover leverages a novel selective instruction duplication technique called tail-DMR (dual modular redundancy). Once a soft error is detected by either the sensor or the tail-DMR, Clover takes care of the error as in the case of exception handling. To recover from the error, Clover simply redirects program control to the beginning of the code region where the error is detected. Lastly, the experiment results demonstrate that the average runtime overhead is only 26%, which is a 75% reduction compared to that of the state-of-the-art soft error resilience technique.

  7. Clover: Compiler directed lightweight soft error resilience

    DOE Public Access Gateway for Energy & Science Beta (PAGES Beta)

    Liu, Qingrui; Lee, Dongyoon; Jung, Changhee; Tiwari, Devesh

    2015-05-01

    This paper presents Clover, a compiler directed soft error detection and recovery scheme for lightweight soft error resilience. The compiler carefully generates soft error tolerant code based on idem-potent processing without explicit checkpoint. During program execution, Clover relies on a small number of acoustic wave detectors deployed in the processor to identify soft errors by sensing the wave made by a particle strike. To cope with DUE (detected unrecoverable errors) caused by the sensing latency of error detection, Clover leverages a novel selective instruction duplication technique called tail-DMR (dual modular redundancy). Once a soft error is detected by either themore » sensor or the tail-DMR, Clover takes care of the error as in the case of exception handling. To recover from the error, Clover simply redirects program control to the beginning of the code region where the error is detected. Lastly, the experiment results demonstrate that the average runtime overhead is only 26%, which is a 75% reduction compared to that of the state-of-the-art soft error resilience technique.« less

  8. Approximate error conjugation gradient minimization methods

    DOE Patents [OSTI]

    Kallman, Jeffrey S

    2013-05-21

    In one embodiment, a method includes selecting a subset of rays from a set of all rays to use in an error calculation for a constrained conjugate gradient minimization problem, calculating an approximate error using the subset of rays, and calculating a minimum in a conjugate gradient direction based on the approximate error. In another embodiment, a system includes a processor for executing logic, logic for selecting a subset of rays from a set of all rays to use in an error calculation for a constrained conjugate gradient minimization problem, logic for calculating an approximate error using the subset of rays, and logic for calculating a minimum in a conjugate gradient direction based on the approximate error. In other embodiments, computer program products, methods, and systems are described capable of using approximate error in constrained conjugate gradient minimization problems.

  9. Measurement

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    core velocity fluctuations and the dynamo in a reversed-field pinch * D. J. Den Hartog, †,a) J. T. Chapman, b) D. Craig, G. Fiksel, P. W. Fontana, S. C. Prager, and J. S. Sarff Department of Physics, University of Wisconsin-Madison, 1150 University Avenue, Madison, Wisconsin 53706 ͑Received 16 November 1998; accepted 20 January 1999͒ Plasma flow velocity fluctuations have been directly measured in the high-temperature magnetically confined plasma in the Madison Symmetric Torus ͑MST͒

  10. Measurement

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    1 H( 7 Be, 8 B)γ cross section by Ryan P. Fitzgerald A dissertation submitted to the faculty of the University of North Carolina at Chapel Hill in partial fulfillment of the requirements for the degree of Doctor of Philosophy in the Department of Physics & Astronomy. Chapel Hill 2005 Approved: A. E. Champagne, Advisor J. C. Blackmon, Reader C. Iliadis, Reader ABSTRACT Ryan P. Fitzgerald: Measurement of the 1 H( 7 Be, 8 B)γ cross section (Under the Direction of A. E. Champagne) The fusion

  11. Measurement

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    interpretation of micro benchmark and application energy use on the Cray XC30 Brian Austin, and Nicholas J. Wright ⇤ August 29, 2014 Abstract Understanding patterns of application energy use is key to reaching future HPC e ciency goals. We have measured the sensitivity of en- ergy use to CPU frequency for several microbenchmarks and applications on a Cray XC30. First order fits to the performance and power data are su cient to describe the energy used by these applications. Exam- ination of

  12. Measurement

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Electronegative Contaminants and Drift Electron Lifetime in the MicroBooNE Experiment The MicroBooNE Collaboration May 19, 2016 Abstract High-purity liquid argon is critical for the operation of a liquid argon time projec- tion chamber (LArTPC). At MicroBooNE, we have achieved an electron drift lifetime of at least 6 ms without evacuation of the detector vessel. Measurements of the elec- tronegative contaminants oxygen and water are described and shown as the gas and liquid argon stages of

  13. The contour method cutting assumption: error minimization and correction

    SciTech Connect (OSTI)

    Prime, Michael B; Kastengren, Alan L

    2010-01-01

    The recently developed contour method can measure 2-D, cross-sectional residual-stress map. A part is cut in two using a precise and low-stress cutting technique such as electric discharge machining. The contours of the new surfaces created by the cut, which will not be flat if residual stresses are relaxed by the cutting, are then measured and used to calculate the original residual stresses. The precise nature of the assumption about the cut is presented theoretically and is evaluated experimentally. Simply assuming a flat cut is overly restrictive and misleading. The critical assumption is that the width of the cut, when measured in the original, undeformed configuration of the body is constant. Stresses at the cut tip during cutting cause the material to deform, which causes errors. The effect of such cutting errors on the measured stresses is presented. The important parameters are quantified. Experimental procedures for minimizing these errors are presented. An iterative finite element procedure to correct for the errors is also presented. The correction procedure is demonstrated on experimental data from a steel beam that was plastically bent to put in a known profile of residual stresses.

  14. Scalable error correction in distributed ion trap computers

    SciTech Connect (OSTI)

    Oi, Daniel K. L.; Devitt, Simon J.; Hollenberg, Lloyd C. L.

    2006-11-15

    A major challenge for quantum computation in ion trap systems is scalable integration of error correction and fault tolerance. We analyze a distributed architecture with rapid high-fidelity local control within nodes and entangled links between nodes alleviating long-distance transport. We demonstrate fault-tolerant operator measurements which are used for error correction and nonlocal gates. This scheme is readily applied to linear ion traps which cannot be scaled up beyond a few ions per individual trap but which have access to a probabilistic entanglement mechanism. A proof-of-concept system is presented which is within the reach of current experiment.

  15. Posters The Impacts of Data Error and Model Resolution

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    7 Posters The Impacts of Data Error and Model Resolution on the Result of Variational Data Assimilation S. Yang and Q. Xu Cooperative Institute of Mesoscale Meteorological Studies University of Oklahoma Norman, Oklahoma Introduction The representativeness and accuracy of the measurements or estimates of the lateral boundary fluxes and surface fluxes are crucial for the single-column model and budget studies of climatic variables over Atmospheric Radiation Measurement (ARM) sites. Since the

  16. Error field and magnetic diagnostic modeling for W7-X

    SciTech Connect (OSTI)

    Lazerson, Sam A.; Gates, David A.; NEILSON, GEORGE H.; OTTE, M.; Bozhenkov, S.; Pedersen, T. S.; GEIGER, J.; LORE, J.

    2014-07-01

    The prediction, detection, and compensation of error fields for the W7-X device will play a key role in achieving a high beta (Β = 5%), steady state (30 minute pulse) operating regime utilizing the island divertor system [1]. Additionally, detection and control of the equilibrium magnetic structure in the scrape-off layer will be necessary in the long-pulse campaign as bootstrapcurrent evolution may result in poor edge magnetic structure [2]. An SVD analysis of the magnetic diagnostics set indicates an ability to measure the toroidal current and stored energy, while profile variations go undetected in the magnetic diagnostics. An additional set of magnetic diagnostics is proposed which improves the ability to constrain the equilibrium current and pressure profiles. However, even with the ability to accurately measure equilibrium parameters, the presence of error fields can modify both the plasma response and diverter magnetic field structures in unfavorable ways. Vacuum flux surface mapping experiments allow for direct measurement of these modifications to magnetic structure. The ability to conduct such an experiment is a unique feature of stellarators. The trim coils may then be used to forward model the effect of an applied n = 1 error field. This allows the determination of lower limits for the detection of error field amplitude and phase using flux surface mapping. *Research supported by the U.S. DOE under Contract No. DE-AC02-09CH11466 with Princeton University.

  17. Group representations, error bases and quantum codes

    SciTech Connect (OSTI)

    Knill, E

    1996-01-01

    This report continues the discussion of unitary error bases and quantum codes. Nice error bases are characterized in terms of the existence of certain characters in a group. A general construction for error bases which are non-abelian over the center is given. The method for obtaining codes due to Calderbank et al. is generalized and expressed purely in representation theoretic terms. The significance of the inertia subgroup both for constructing codes and obtaining the set of transversally implementable operations is demonstrated.

  18. Linux Kernel Error Detection and Correction

    Energy Science and Technology Software Center (OSTI)

    2007-04-11

    EDAC-utils consists fo a library and set of utilities for retrieving statistics from the Linux Kernel Error Detection and Correction (EDAC) drivers.

  19. runtime error message: "readControlMsg: System returned error Connection

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    timed out on TCP socket fd" readControlMsg: System returned error Connection timed out on TCP socket fd" runtime error message: "readControlMsg: System returned error Connection timed out on TCP socket fd" June 30, 2015 Symptom User jobs with sinlge or multiple apruns in a batch script may get this run time error: "readControlMsg: System returned error Connection timed out on TCP socket fd". This problem is intermittent, sometimes resubmit works. This error

  20. Shape error analysis for reflective nano focusing optics

    SciTech Connect (OSTI)

    Modi, Mohammed H.; Idir, Mourad

    2010-06-23

    Focusing performance of reflective x-ray optics is determined by surface figure accuracy. Any surface imperfection present on such optics introduces a phase error in the outgoing wave fields. Therefore converging beam at the focal spot will differ from the desired performance. Effect of these errors on focusing performance can be calculated by wave optical approach considering a coherent wave field illumination of optical elements. We have developed a wave optics simulator using Fresnel-Kirchhoff diffraction integral to calculate the mirror pupil function. Both analytically calculated and measured surface topography data can be taken as an aberration source to outgoing wave fields. Simulations are performed to study the effect of surface height fluctuations on focusing performances over wide frequency range in high, mid and low frequency band. The results using real shape profile measured with long trace profilometer (LTP) suggest that the shape error of {lambda}/4 PV (peak to valley) is tolerable to achieve diffraction limited performance. It is desirable to remove shape error of very low frequency as 0.1 mm{sup -1} which otherwise will generate beam waist or satellite peaks. All other frequencies above this limit will not affect the focused beam profile but only caused a loss in intensity.

  1. The role of variation, error, and complexity in manufacturing defects

    SciTech Connect (OSTI)

    Hinckley, C.M.; Barkan, P.

    1994-03-01

    Variation in component properties and dimensions is a widely recognized factor in product defects which can be quantified and controlled by Statistical Process Control methodologies. Our studies have shown, however, that traditional statistical methods are ineffective in characterizing and controlling defects caused by error. The distinction between error and variation becomes increasingly important as the target defect rates approach extremely low values. Motorola data substantiates our thesis that defect rates in the range of several parts per million can only be achieved when traditional methods for controlling variation are combined with methods that specifically focus on eliminating defects due to error. Complexity in the product design, manufacturing processes, or assembly increases the likelihood of defects due to both variation and error. Thus complexity is also a root cause of defects. Until now, the absence of a sound correlation between defects and complexity has obscured the importance of this relationship. We have shown that assembly complexity can be quantified using Design for Assembly (DFA) analysis. High levels of correlation have been found between our complexity measures and defect data covering tens of millions of assembly operations in two widely different industries. The availability of an easily determined measure of complexity, combined with these correlations, permits rapid estimation of the relative defect rates for alternate design concepts. This should prove to be a powerful tool since it can guide design improvement at an early stage when concepts are most readily modified.

  2. Wind Power Forecasting Error Distributions over Multiple Timescales (Presentation)

    SciTech Connect (OSTI)

    Hodge, B. M.; Milligan, M.

    2011-07-01

    This presentation presents some statistical analysis of wind power forecast errors and error distributions, with examples using ERCOT data.

  3. NOx Measurement Errors in Ammonia-Containing Exhaust | Department...

    Broader source: Energy.gov (indexed) [DOE]

    Presentation given at DEER 2006, August 20-24, 2006, Detroit, Michigan. Sponsored by the U.S. DOE's EERE FreedomCar and Fuel Partnership and 21st Century Truck Programs. ...

  4. Error and Uncertainty in Raman Thermal Conductivity Measurements

    Office of Scientific and Technical Information (OSTI)

    ... This ability not only provides experimental simplification but, more importantly, 26 is imperative for the probing of 2D-solids where any contact with surrounding materials (e.g., ...

  5. SU-E-J-235: Varian Portal Dosimetry Accuracy at Detecting Simulated Delivery Errors

    SciTech Connect (OSTI)

    Gordon, J; Bellon, M; Barton, K; Gulam, M; Chetty, I

    2014-06-01

    Purpose: To use receiver operating characteristic (ROC) analysis to quantify the Varian Portal Dosimetry (VPD) application's ability to detect delivery errors in IMRT fields. Methods: EPID and VPD were calibrated/commissioned using vendor-recommended procedures. Five clinical plans comprising 56 modulated fields were analyzed using VPD. Treatment sites were: pelvis, prostate, brain, orbit, and base of tongue. Delivery was on a Varian Trilogy linear accelerator at 6MV using a Millenium120 multi-leaf collimator. Image pairs (VPD-predicted and measured) were exported in dicom format. Each detection test imported an image pair into Matlab, optionally inserted a simulated error (rectangular region with intensity raised or lowered) into the measured image, performed 3%/3mm gamma analysis, and saved the gamma distribution. For a given error, 56 negative tests (without error) were performed, one per 56 image pairs. Also, 560 positive tests (with error) with randomly selected image pairs and randomly selected in-field error location. Images were classified as errored (or error-free) if percent pixels with γ<κ was < (or ≥) τ. (Conventionally, κ=1 and τ=90%.) A ROC curve was generated from the 616 tests by varying τ. For a range of κ and τ, true/false positive/negative rates were calculated. This procedure was repeated for inserted errors of different sizes. VPD was considered to reliably detect an error if images were correctly classified as errored or error-free at least 95% of the time, for some κ+τ combination. Results: 20mm{sup 2} errors with intensity altered by ≥20% could be reliably detected, as could 10mm{sup 2} errors with intensity was altered by ≥50%. Errors with smaller size or intensity change could not be reliably detected. Conclusion: Varian Portal Dosimetry using 3%/3mm gamma analysis is capable of reliably detecting only those fluence errors that exceed the stated sizes. Images containing smaller errors can pass mathematical analysis, though

  6. Error recovery to enable error-free message transfer between nodes of a computer network

    DOE Patents [OSTI]

    Blumrich, Matthias A.; Coteus, Paul W.; Chen, Dong; Gara, Alan; Giampapa, Mark E.; Heidelberger, Philip; Hoenicke, Dirk; Takken, Todd; Steinmacher-Burow, Burkhard; Vranas, Pavlos M.

    2016-01-26

    An error-recovery method to enable error-free message transfer between nodes of a computer network. A first node of the network sends a packet to a second node of the network over a link between the nodes, and the first node keeps a copy of the packet on a sending end of the link until the first node receives acknowledgment from the second node that the packet was received without error. The second node tests the packet to determine if the packet is error free. If the packet is not error free, the second node sets a flag to mark the packet as corrupt. The second node returns acknowledgement to the first node specifying whether the packet was received with or without error. When the packet is received with error, the link is returned to a known state and the packet is sent again to the second node.

  7. Verification of unfold error estimates in the unfold operator code

    SciTech Connect (OSTI)

    Fehl, D.L.; Biggs, F.

    1997-01-01

    Spectral unfolding is an inverse mathematical operation that attempts to obtain spectral source information from a set of response functions and data measurements. Several unfold algorithms have appeared over the past 30 years; among them is the unfold operator (UFO) code written at Sandia National Laboratories. In addition to an unfolded spectrum, the UFO code also estimates the unfold uncertainty (error) induced by estimated random uncertainties in the data. In UFO the unfold uncertainty is obtained from the error matrix. This built-in estimate has now been compared to error estimates obtained by running the code in a Monte Carlo fashion with prescribed data distributions (Gaussian deviates). In the test problem studied, data were simulated from an arbitrarily chosen blackbody spectrum (10 keV) and a set of overlapping response functions. The data were assumed to have an imprecision of 5{percent} (standard deviation). One hundred random data sets were generated. The built-in estimate of unfold uncertainty agreed with the Monte Carlo estimate to within the statistical resolution of this relatively small sample size (95{percent} confidence level). A possible 10{percent} bias between the two methods was unresolved. The Monte Carlo technique is also useful in underdetermined problems, for which the error matrix method does not apply. UFO has been applied to the diagnosis of low energy x rays emitted by Z-pinch and ion-beam driven hohlraums. {copyright} {ital 1997 American Institute of Physics.}

  8. Quantum error-correcting codes and devices

    DOE Patents [OSTI]

    Gottesman, Daniel

    2000-10-03

    A method of forming quantum error-correcting codes by first forming a stabilizer for a Hilbert space. A quantum information processing device can be formed to implement such quantum codes.

  9. Evaluating operating system vulnerability to memory errors.

    SciTech Connect (OSTI)

    Ferreira, Kurt Brian; Bridges, Patrick G.; Pedretti, Kevin Thomas Tauke; Mueller, Frank; Fiala, David; Brightwell, Ronald Brian

    2012-05-01

    Reliability is of great concern to the scalability of extreme-scale systems. Of particular concern are soft errors in main memory, which are a leading cause of failures on current systems and are predicted to be the leading cause on future systems. While great effort has gone into designing algorithms and applications that can continue to make progress in the presence of these errors without restarting, the most critical software running on a node, the operating system (OS), is currently left relatively unprotected. OS resiliency is of particular importance because, though this software typically represents a small footprint of a compute node's physical memory, recent studies show more memory errors in this region of memory than the remainder of the system. In this paper, we investigate the soft error vulnerability of two operating systems used in current and future high-performance computing systems: Kitten, the lightweight kernel developed at Sandia National Laboratories, and CLE, a high-performance Linux-based operating system developed by Cray. For each of these platforms, we outline major structures and subsystems that are vulnerable to soft errors and describe methods that could be used to reconstruct damaged state. Our results show the Kitten lightweight operating system may be an easier target to harden against memory errors due to its smaller memory footprint, largely deterministic state, and simpler system structure.

  10. Measurements of plutonium

    SciTech Connect (OSTI)

    Larsen, R.P. )

    1989-11-01

    Based on reviews of the early and recent literature concerning comparative measurements of plutonium, sources of measurement error are discussed. This paper focuses on those related to mass spectrometric isotope dilution.

  11. Superdense coding interleaved with forward error correction

    DOE Public Access Gateway for Energy & Science Beta (PAGES Beta)

    Humble, Travis S.; Sadlier, Ronald J.

    2016-05-12

    Superdense coding promises increased classical capacity and communication security but this advantage may be undermined by noise in the quantum channel. We present a numerical study of how forward error correction (FEC) applied to the encoded classical message can be used to mitigate against quantum channel noise. By studying the bit error rate under different FEC codes, we identify the unique role that burst errors play in superdense coding, and we show how these can be mitigated against by interleaving the FEC codewords prior to transmission. As a result, we conclude that classical FEC with interleaving is a useful methodmore » to improve the performance in near-term demonstrations of superdense coding.« less

  12. Laser Phase Errors in Seeded FELs

    SciTech Connect (OSTI)

    Ratner, D.; Fry, A.; Stupakov, G.; White, W.; /SLAC

    2012-03-28

    Harmonic seeding of free electron lasers has attracted significant attention from the promise of transform-limited pulses in the soft X-ray region. Harmonic multiplication schemes extend seeding to shorter wavelengths, but also amplify the spectral phase errors of the initial seed laser, and may degrade the pulse quality. In this paper we consider the effect of seed laser phase errors in high gain harmonic generation and echo-enabled harmonic generation. We use simulations to confirm analytical results for the case of linearly chirped seed lasers, and extend the results for arbitrary seed laser envelope and phase.

  13. The Impact of Soil Sampling Errors on Variable Rate Fertilization

    SciTech Connect (OSTI)

    R. L. Hoskinson; R C. Rope; L G. Blackwood; R D. Lee; R K. Fink

    2004-07-01

    Variable rate fertilization of an agricultural field is done taking into account spatial variability in the soils characteristics. Most often, spatial variability in the soils fertility is the primary characteristic used to determine the differences in fertilizers applied from one point to the next. For several years the Idaho National Engineering and Environmental Laboratory (INEEL) has been developing a Decision Support System for Agriculture (DSS4Ag) to determine the economically optimum recipe of various fertilizers to apply at each site in a field, based on existing soil fertility at the site, predicted yield of the crop that would result (and a predicted harvest-time market price), and the current costs and compositions of the fertilizers to be applied. Typically, soil is sampled at selected points within a field, the soil samples are analyzed in a lab, and the lab-measured soil fertility of the point samples is used for spatial interpolation, in some statistical manner, to determine the soil fertility at all other points in the field. Then a decision tool determines the fertilizers to apply at each point. Our research was conducted to measure the impact on the variable rate fertilization recipe caused by variability in the measurement of the soils fertility at the sampling points. The variability could be laboratory analytical errors or errors from variation in the sample collection method. The results show that for many of the fertility parameters, laboratory measurement error variance exceeds the estimated variability of the fertility measure across grid locations. These errors resulted in DSS4Ag fertilizer recipe recommended application rates that differed by up to 138 pounds of urea per acre, with half the field differing by more than 57 pounds of urea per acre. For potash the difference in application rate was up to 895 pounds per acre and over half the field differed by more than 242 pounds of potash per acre. Urea and potash differences accounted

  14. Intel C++ compiler error: stl_iterator_base_types.h

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    C++ compiler error: stliteratorbasetypes.h Intel C++ compiler error: stliteratorbasetypes.h December 7, 2015 by Scott French Because the system-supplied version of GCC is...

  15. Error estimates for fission neutron outputs (Conference) | SciTech...

    Office of Scientific and Technical Information (OSTI)

    Error estimates for fission neutron outputs Citation Details In-Document Search Title: Error estimates for fission neutron outputs You are accessing a document from the...

  16. Internal compiler error for function pointer with identically...

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Internal compiler error for function pointer with identically named arguments Internal compiler error for function pointer with identically named arguments June 9, 2015 by Scott...

  17. V-235: Cisco Mobility Services Engine Configuration Error Lets...

    Office of Energy Efficiency and Renewable Energy (EERE) Indexed Site

    5: Cisco Mobility Services Engine Configuration Error Lets Remote Users Login Anonymously V-235: Cisco Mobility Services Engine Configuration Error Lets Remote Users Login ...

  18. Error Estimation for Fault Tolerance in Numerical Integration...

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Error Estimation for Fault Tolerance in Numerical Integration Solvers Event Sponsor: ... In numerical integration solvers, approximation error can be estimated at a low cost. We ...

  19. A posteriori error analysis of parameterized linear systems using...

    Office of Scientific and Technical Information (OSTI)

    Journal Article: A posteriori error analysis of parameterized linear systems using spectral methods. Citation Details In-Document Search Title: A posteriori error analysis of ...

  20. Table 1b. Relative Standard Errors for Effective, Occupied, and...

    U.S. Energy Information Administration (EIA) Indexed Site

    b.Relative Standard Errors Table 1b. Relative Standard Errors for Effective Occupied, and Vacant Square Footage, 1992 Building Characteristics All Buildings (thousand) Total...

  1. Accounting for Model Error in the Calibration of Physical Models...

    Office of Scientific and Technical Information (OSTI)

    Accounting for Model Error in the Calibration of Physical Models. Citation Details In-Document Search Title: Accounting for Model Error in the Calibration of Physical Models. ...

  2. Table 2b. Relative Standard Errors for Electricity Consumption...

    U.S. Energy Information Administration (EIA) Indexed Site

    2b. Relative Standard Errors for Electricity Table 2b. Relative Standard Errors for Electricity Consumption and Electricity Intensities, per Square Foot, Specific to Occupied and...

  3. Error Analysis in Nuclear Density Functional Theory (Journal...

    Office of Scientific and Technical Information (OSTI)

    Error Analysis in Nuclear Density Functional Theory Citation Details In-Document Search Title: Error Analysis in Nuclear Density Functional Theory Authors: Schunck, N ; McDonnell,...

  4. Error Analysis in Nuclear Density Functional Theory (Journal...

    Office of Scientific and Technical Information (OSTI)

    Error Analysis in Nuclear Density Functional Theory Citation Details In-Document Search Title: Error Analysis in Nuclear Density Functional Theory You are accessing a document...

  5. Raman Thermometry: Comparing Methods to Minimize Error. (Conference...

    Office of Scientific and Technical Information (OSTI)

    Raman Thermometry: Comparing Methods to Minimize Error. Citation Details In-Document Search Title: Raman Thermometry: Comparing Methods to Minimize Error. Abstract not provided....

  6. Precise method of compensating radiation-induced errors in a hot-cathode-ionization gauge with correcting electrode

    SciTech Connect (OSTI)

    Saeki, Hiroshi Magome, Tamotsu

    2014-10-06

    To compensate pressure-measurement errors caused by a synchrotron radiation environment, a precise method using a hot-cathode-ionization-gauge head with correcting electrode, was developed and tested in a simulation experiment with excess electrons in the SPring-8 storage ring. This precise method to improve the measurement accuracy, can correctly reduce the pressure-measurement errors caused by electrons originating from the external environment, and originating from the primary gauge filament influenced by spatial conditions of the installed vacuum-gauge head. As the result of the simulation experiment to confirm the performance reducing the errors caused by the external environment, the pressure-measurement error using this method was approximately less than several percent in the pressure range from 10{sup ?5} Pa to 10{sup ?8} Pa. After the experiment, to confirm the performance reducing the error caused by spatial conditions, an additional experiment was carried out using a sleeve and showed that the improved function was available.

  7. Quantifying the Effect of Lidar Turbulence Error on Wind Power Prediction

    SciTech Connect (OSTI)

    Newman, Jennifer F.; Clifton, Andrew

    2016-01-01

    Currently, cup anemometers on meteorological towers are used to measure wind speeds and turbulence intensity to make decisions about wind turbine class and site suitability; however, as modern turbine hub heights increase and wind energy expands to complex and remote sites, it becomes more difficult and costly to install meteorological towers at potential sites. As a result, remote-sensing devices (e.g., lidars) are now commonly used by wind farm managers and researchers to estimate the flow field at heights spanned by a turbine. Although lidars can accurately estimate mean wind speeds and wind directions, there is still a large amount of uncertainty surrounding the measurement of turbulence using these devices. Errors in lidar turbulence estimates are caused by a variety of factors, including instrument noise, volume averaging, and variance contamination, in which the magnitude of these factors is highly dependent on measurement height and atmospheric stability. As turbulence has a large impact on wind power production, errors in turbulence measurements will translate into errors in wind power prediction. The impact of using lidars rather than cup anemometers for wind power prediction must be understood if lidars are to be considered a viable alternative to cup anemometers.In this poster, the sensitivity of power prediction error to typical lidar turbulence measurement errors is assessed. Turbulence estimates from a vertically profiling WINDCUBE v2 lidar are compared to high-resolution sonic anemometer measurements at field sites in Oklahoma and Colorado to determine the degree of lidar turbulence error that can be expected under different atmospheric conditions. These errors are then incorporated into a power prediction model to estimate the sensitivity of power prediction error to turbulence measurement error. Power prediction models, including the standard binning method and a random forest method, were developed using data from the aeroelastic simulator FAST

  8. Analysis of Solar Two Heliostat Tracking Error Sources

    SciTech Connect (OSTI)

    Jones, S.A.; Stone, K.W.

    1999-01-28

    This paper explores the geometrical errors that reduce heliostat tracking accuracy at Solar Two. The basic heliostat control architecture is described. Then, the three dominant error sources are described and their effect on heliostat tracking is visually illustrated. The strategy currently used to minimize, but not truly correct, these error sources is also shown. Finally, a novel approach to minimizing error is presented.

  9. Distribution of Wind Power Forecasting Errors from Operational Systems (Presentation)

    SciTech Connect (OSTI)

    Hodge, B. M.; Ela, E.; Milligan, M.

    2011-10-01

    This presentation offers new data and statistical analysis of wind power forecasting errors in operational systems.

  10. WIPP Weatherization: Common Errors and Innovative Solutions Presentation

    Broader source: Energy.gov [DOE]

    This presentation contains information on WIPP Weatherization: Common Errors and Innovative Solutions.

  11. Progress in Understanding Error-field Physics in NSTX Spherical Torus Plasmas

    SciTech Connect (OSTI)

    E. Menard, R.E. Bell, D.A. Gates, S.P. Gerhardt, J.-K. Park, S.A. Sabbagh, J.W. Berkery, A. Egan, J. Kallman, S.M. Kaye, B. LeBlanc, Y.Q. Liu, A. Sontag, D. Swanson, H. Yuh, W. Zhu and the NSTX Research Team

    2010-05-19

    The low aspect ratio, low magnetic field, and wide range of plasma beta of NSTX plasmas provide new insight into the origins and effects of magnetic field errors. An extensive array of magnetic sensors has been used to analyze error fields, to measure error field amplification, and to detect resistive wall modes in real time. The measured normalized error-field threshold for the onset of locked modes shows a linear scaling with plasma density, a weak to inverse dependence on toroidal field, and a positive scaling with magnetic shear. These results extrapolate to a favorable error field threshold for ITER. For these low-beta locked-mode plasmas, perturbed equilibrium calculations find that the plasma response must be included to explain the empirically determined optimal correction of NSTX error fields. In high-beta NSTX plasmas exceeding the n=1 no-wall stability limit where the RWM is stabilized by plasma rotation, active suppression of n=1 amplified error fields and the correction of recently discovered intrinsic n=3 error fields have led to sustained high rotation and record durations free of low-frequency core MHD activity. For sustained rotational stabilization of the n=1 RWM, both the rotation threshold and magnitude of the amplification are important. At fixed normalized dissipation, kinetic damping models predict rotation thresholds for RWM stabilization to scale nearly linearly with particle orbit frequency. Studies for NSTX find that orbit frequencies computed in general geometry can deviate significantly from those computed in the high aspect ratio and circular plasma cross-section limit, and these differences can strongly influence the predicted RWM stability. The measured and predicted RWM stability is found to be very sensitive to the E B rotation profile near the plasma edge, and the measured critical rotation for the RWM is approximately a factor of two higher than predicted by the MARS-F code using the semi-kinetic damping model.

  12. Errors in response calculations for beams

    SciTech Connect (OSTI)

    Wada, H.; Wurburton, G.B.

    1985-05-01

    When the finite element method is used to idealize a structure, its dynamic response can be determined from the governing matrix equation by the normal mode method or by one of the many approximate direct integration methods. In either method the approximate data of the finite element idealization are used, but further assumptions are introduced by the direct integration scheme. It is the purpose of this paper to study these errors for a simple structure. The transient flexural vibrations of a uniform cantilever beam, which is subjected to a transverse force at the free end, are determined by the Laplace transform method. Comparable responses are obtained for a finite element idealization of the beam, using the normal mode and Newmark average acceleration methods; the errors associated with the approximate methods are studied. If accuracy has priority and the quantity of data is small, the normal mode method is recommended; however, if the quantity of data is large, the Newmark method is useful.

  13. Detecting Soft Errors in Stencil based Computations

    SciTech Connect (OSTI)

    Sharma, V.; Gopalkrishnan, G.; Bronevetsky, G.

    2015-05-06

    Given the growing emphasis on system resilience, it is important to develop software-level error detectors that help trap hardware-level faults with reasonable accuracy while minimizing false alarms as well as the performance overhead introduced. We present a technique that approaches this idea by taking stencil computations as our target, and synthesizing detectors based on machine learning. In particular, we employ linear regression to generate computationally inexpensive models which form the basis for error detection. Our technique has been incorporated into a new open-source library called SORREL. In addition to reporting encouraging experimental results, we demonstrate techniques that help reduce the size of training data. We also discuss the efficacy of various detectors synthesized, as well as our future plans.

  14. Redundancy and Error Resilience in Boolean Networks

    SciTech Connect (OSTI)

    Peixoto, Tiago P.

    2010-01-29

    We consider the effect of noise in sparse Boolean networks with redundant functions. We show that they always exhibit a nonzero error level, and the dynamics undergoes a phase transition from nonergodicity to ergodicity, as a function of noise, after which the system is no longer capable of preserving a memory of its initial state. We obtain upper bounds on the critical value of noise for networks of different sparsity.

  15. Systematic errors in long baseline oscillation experiments

    SciTech Connect (OSTI)

    Harris, Deborah A.; /Fermilab

    2006-02-01

    This article gives a brief overview of long baseline neutrino experiments and their goals, and then describes the different kinds of systematic errors that are encountered in these experiments. Particular attention is paid to the uncertainties that come about because of imperfect knowledge of neutrino cross sections and more generally how neutrinos interact in nuclei. Near detectors are planned for most of these experiments, and the extent to which certain uncertainties can be reduced by the presence of near detectors is also discussed.

  16. Improving Memory Error Handling Using Linux

    SciTech Connect (OSTI)

    Carlton, Michael Andrew; Blanchard, Sean P.; Debardeleben, Nathan A.

    2014-07-25

    As supercomputers continue to get faster and more powerful in the future, they will also have more nodes. If nothing is done, then the amount of memory in supercomputer clusters will soon grow large enough that memory failures will be unmanageable to deal with by manually replacing memory DIMMs. "Improving Memory Error Handling Using Linux" is a process oriented method to solve this problem by using the Linux kernel to disable (offline) faulty memory pages containing bad addresses, preventing them from being used again by a process. The process of offlining memory pages simplifies error handling and results in reducing both hardware and manpower costs required to run Los Alamos National Laboratory (LANL) clusters. This process will be necessary for the future of supercomputing to allow the development of exascale computers. It will not be feasible without memory error handling to manually replace the number of DIMMs that will fail daily on a machine consisting of 32-128 petabytes of memory. Testing reveals the process of offlining memory pages works and is relatively simple to use. As more and more testing is conducted, the entire process will be automated within the high-performance computing (HPC) monitoring software, Zenoss, at LANL.

  17. Common Errors and Innovative Solutions Transcript | Department of Energy

    Office of Energy Efficiency and Renewable Energy (EERE) Indexed Site

    Common Errors and Innovative Solutions Transcript Common Errors and Innovative Solutions Transcript An example of case studies, mainly by showing photos of errors and good examples, then discussing the purpose of the home energy professional guidelines and certification. There may be more examples of what not to do only because these were good learning opportunities. common_errors_innovative_solutions.doc (41.5 KB) More Documents & Publications WIPP Weatherization: Common Errors and

  18. An effective approach for the minimization of errors in capacitance-voltage carrier profiling of quantum structures

    SciTech Connect (OSTI)

    Biswas, Dipankar Panda, Siddhartha

    2014-04-07

    Experimental capacitance–voltage (C-V) profiling of semiconductor heterojunctions and quantum wells has remained ever important and relevant. The apparent carrier distributions (ACDs) thus obtained reveal the carrier depletions, carrier peaks and their positions, in and around the quantum structures. Inevitable errors, encountered in such measurements, are the deviations of the peak concentrations of the ACDs and their positions, from the actual carrier peaks obtained from quantum mechanical computations with the fundamental parameters. In spite of the very wide use of the C-V method, comprehensive discussions on the qualitative and quantitative nature of the errors remain wanting. The errors are dependent on the fundamental parameters, the temperature of measurements, the Debye length, and the series resistance. In this paper, the errors have been studied with doping concentration, band offset, and temperature. From this study, a rough estimate may be drawn about the error. It is seen that the error in the position of the ACD peak decreases at higher doping, higher band offset, and lower temperature, whereas the error in the peak concentration changes in a strange fashion. A completely new method is introduced, for derivation of the carrier profiles from C-V measurements on quantum structures to minimize errors which are inevitable in the conventional formulation.

  19. Resolved: "error while loading shared libraries: libalpslli.so...

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    "error while loading shared libraries: libalpslli.so.0" with serial codes on login nodes Resolved: "error while loading shared libraries: libalpslli.so.0" with serial codes on...

  20. MPI errors from cray-mpich/7.3.0

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    MPI errors from cray-mpich7.3.0 MPI errors from cray-mpich7.3.0 January 6, 2016 by Ankit Bhagatwala A change in the MPICH2 library that now strictly enforces non-overlapping...

  1. Verification of unfold error estimates in the UFO code

    SciTech Connect (OSTI)

    Fehl, D.L.; Biggs, F.

    1996-07-01

    Spectral unfolding is an inverse mathematical operation which attempts to obtain spectral source information from a set of tabulated response functions and data measurements. Several unfold algorithms have appeared over the past 30 years; among them is the UFO (UnFold Operator) code. In addition to an unfolded spectrum, UFO also estimates the unfold uncertainty (error) induced by running the code in a Monte Carlo fashion with prescribed data distributions (Gaussian deviates). In the problem studied, data were simulated from an arbitrarily chosen blackbody spectrum (10 keV) and a set of overlapping response functions. The data were assumed to have an imprecision of 5% (standard deviation). 100 random data sets were generated. The built-in estimate of unfold uncertainty agreed with the Monte Carlo estimate to within the statistical resolution of this relatively small sample size (95% confidence level). A possible 10% bias between the two methods was unresolved. The Monte Carlo technique is also useful in underdetemined problems, for which the error matrix method does not apply. UFO has been applied to the diagnosis of low energy x rays emitted by Z-Pinch and ion-beam driven hohlraums.

  2. Structural power flow measurement

    SciTech Connect (OSTI)

    Falter, K.J.; Keltie, R.F.

    1988-12-01

    Previous investigations of structural power flow through beam-like structures resulted in some unexplained anomalies in the calculated data. In order to develop structural power flow measurement as a viable technique for machine tool design, the causes of these anomalies needed to be found. Once found, techniques for eliminating the errors could be developed. Error sources were found in the experimental apparatus itself as well as in the instrumentation. Although flexural waves are the carriers of power in the experimental apparatus, at some frequencies longitudinal waves were excited which were picked up by the accelerometers and altered power measurements. Errors were found in the phase and gain response of the sensors and amplifiers used for measurement. A transfer function correction technique was employed to compensate for these instrumentation errors.

  3. Locked modes and magnetic field errors in MST

    SciTech Connect (OSTI)

    Almagri, A.F.; Assadi, S.; Prager, S.C.; Sarff, J.S.; Kerst, D.W.

    1992-06-01

    In the MST reversed field pinch magnetic oscillations become stationary (locked) in the lab frame as a result of a process involving interactions between the modes, sawteeth, and field errors. Several helical modes become phase locked to each other to form a rotating localized disturbance, the disturbance locks to an impulsive field error generated at a sawtooth crash, the error fields grow monotonically after locking (perhaps due to an unstable interaction between the modes and field error), and over the tens of milliseconds of growth confinement degrades and the discharge eventually terminates. Field error control has been partially successful in eliminating locking.

  4. Analysis of Errors in a Special Perturbations Satellite Orbit Propagator

    SciTech Connect (OSTI)

    Beckerman, M.; Jones, J.P.

    1999-02-01

    We performed an analysis of error densities for the Special Perturbations orbit propagator using data for 29 satellites in orbits of interest to Space Shuttle and International Space Station collision avoidance. We find that the along-track errors predominate. These errors increase monotonically over each 36-hour prediction interval. The predicted positions in the along-track direction progressively either leap ahead of or lag behind the actual positions. Unlike the along-track errors the radial and cross-track errors oscillate about their nearly zero mean values. As the number of observations per fit interval decline the along-track prediction errors, and amplitudes of the radial and cross-track errors, increase.

  5. A technique for human error analysis (ATHEANA)

    SciTech Connect (OSTI)

    Cooper, S.E.; Ramey-Smith, A.M.; Wreathall, J.; Parry, G.W.

    1996-05-01

    Probabilistic risk assessment (PRA) has become an important tool in the nuclear power industry, both for the Nuclear Regulatory Commission (NRC) and the operating utilities. Human reliability analysis (HRA) is a critical element of PRA; however, limitations in the analysis of human actions in PRAs have long been recognized as a constraint when using PRA. A multidisciplinary HRA framework has been developed with the objective of providing a structured approach for analyzing operating experience and understanding nuclear plant safety, human error, and the underlying factors that affect them. The concepts of the framework have matured into a rudimentary working HRA method. A trial application of the method has demonstrated that it is possible to identify potentially significant human failure events from actual operating experience which are not generally included in current PRAs, as well as to identify associated performance shaping factors and plant conditions that have an observable impact on the frequency of core damage. A general process was developed, albeit in preliminary form, that addresses the iterative steps of defining human failure events and estimating their probabilities using search schemes. Additionally, a knowledge- base was developed which describes the links between performance shaping factors and resulting unsafe actions.

  6. Scheme for precise correction of orbit variation caused by dipole error field of insertion device

    SciTech Connect (OSTI)

    Nakatani, T.; Agui, A.; Aoyagi, H.; Matsushita, T.; Takao, M.; Takeuchi, M.; Yoshigoe, A.; Tanaka, H.

    2005-05-15

    We developed a scheme for precisely correcting the orbit variation caused by a dipole error field of an insertion device (ID) in a storage ring and investigated its performance. The key point for achieving the precise correction is to extract the variation of the beam orbit caused by the change of the ID error field from the observed variation. We periodically change parameters such as the gap and phase of the specified ID with a mirror-symmetric pattern over the measurement period to modulate the variation. The orbit variation is measured using conventional wide-frequency-band detectors and then the induced variation is extracted precisely through averaging and filtering procedures. Furthermore, the mirror-symmetric pattern enables us to independently extract the orbit variations caused by a static error field and by a dynamic one, e.g., an error field induced by the dynamical change of the ID gap or phase parameter. We built a time synchronization measurement system with a sampling rate of 100 Hz and applied the scheme to the correction of the orbit variation caused by the error field of an APPLE-2-type undulator installed in the SPring-8 storage ring. The result shows that the developed scheme markedly improves the correction performance and suppresses the orbit variation caused by the ID error field down to the order of submicron. This scheme is applicable not only to the correction of the orbit variation caused by a special ID, the gap or phase of which is periodically changed during an experiment, but also to the correction of the orbit variation caused by a conventional ID which is used with a fixed gap and phase.

  7. Analysis of Cloud Variability and Sampling Errors in Surface and Satellite Mesurements

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Analysis of Cloud Variability and Sampling Errors in Surface and Satellite Measurements Z. Li, M. C. Cribb, and F.-L. Chang Earth System Science Interdisciplinary Center University of Maryland College Park, Maryland A. P. Trishchenko and Y. Luo Canada Centre for Remote Sensing Ottawa, Ontario, Canada Introduction Radiation measurements have been widely employed for evaluating cloud parameterization schemes and model simulation results. As the most comprehensive program aiming to improve cloud

  8. Polaractivation for classical zero-error capacity of qudit channels

    SciTech Connect (OSTI)

    Gyongyosi, Laszlo; Imre, Sandor

    2014-12-04

    We introduce a new phenomenon for zero-error transmission of classical information over quantum channels that initially were not able for zero-error classical communication. The effect is called polaractivation, and the result is similar to the superactivation effect. We use the Choi-Jamiolkowski isomorphism and the Schmidt-theorem to prove the polaractivation of classical zero-error capacity and define the polaractivator channel coding scheme.

  9. Internal compiler error for function pointer with identically named

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    arguments Internal compiler error for function pointer with identically named arguments Internal compiler error for function pointer with identically named arguments June 9, 2015 by Scott French, NERSC USG Status: Bug 21435 reported to PGI For pgcc versions after 12.x (up through 12.9 is fine, but 13.x and 14.x are not), you may observe an internal compiler error associated with function pointer prototypes when named arguments are used. Specifically, if a function pointer type is defined

  10. WIPP Weatherization: Common Errors and Innovative Solutions Presentati...

    Office of Energy Efficiency and Renewable Energy (EERE) Indexed Site

    More Documents & Publications Common Errors and Innovative Solutions Transcript Building ... America Best Practices Series: Volume 12. Energy Renovations-Insulation: A Guide for ...

  11. Output-Based Error Estimation and Adaptation for Uncertainty...

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Output-Based Error Estimation and Adaptation for Uncertainty Quantification Isaac M. Asher and Krzysztof J. Fidkowski University of Michigan US National Congress on Computational...

  12. Platform-Independent Method for Detecting Errors in Metagenomic...

    Office of Scientific and Technical Information (OSTI)

    Title: Platform-Independent Method for Detecting Errors in Metagenomic Sequencing Data: DRISEE Authors: Keegan, K. P. ; Trimble, W. L. ; Wilkening, J. ; Wilke, A. ; Harrison, T. ; ...

  13. Detecting and correcting hard errors in a memory array

    DOE Patents [OSTI]

    Kalamatianos, John; John, Johnsy Kanjirapallil; Gelinas, Robert; Sridharan, Vilas K.; Nevius, Phillip E.

    2015-11-19

    Hard errors in the memory array can be detected and corrected in real-time using reusable entries in an error status buffer. Data may be rewritten to a portion of a memory array and a register in response to a first error in data read from the portion of the memory array. The rewritten data may then be written from the register to an entry of an error status buffer in response to the rewritten data read from the register differing from the rewritten data read from the portion of the memory array.

  14. Info-Gap Analysis of Truncation Errors in Numerical Simulations...

    Office of Scientific and Technical Information (OSTI)

    Title: Info-Gap Analysis of Truncation Errors in Numerical Simulations. Authors: Kamm, James R. ; Witkowski, Walter R. ; Rider, William J. ; Trucano, Timothy Guy ; Ben-Haim, Yakov. ...

  15. Info-Gap Analysis of Numerical Truncation Errors. (Conference...

    Office of Scientific and Technical Information (OSTI)

    Title: Info-Gap Analysis of Numerical Truncation Errors. Authors: Kamm, James R. ; Witkowski, Walter R. ; Rider, William J. ; Trucano, Timothy Guy ; Ben-Haim, Yakov. Publication ...

  16. Table 6b. Relative Standard Errors for Total Electricity Consumption...

    U.S. Energy Information Administration (EIA) Indexed Site

    b. Relative Standard Errors for Total Electricity Consumption per Effective Occupied Square Foot, 1992 Building Characteristics All Buildings Using Electricity (thousand) Total...

  17. Accounting for Model Error in the Calibration of Physical Models

    Office of Scientific and Technical Information (OSTI)

    ... model error term in locations where key modeling assumptions and approximations are made ... to represent the truth o In this context, the data has no noise o Discrepancy ...

  18. Handling Model Error in the Calibration of Physical Models

    Office of Scientific and Technical Information (OSTI)

    ... model error term in locations where key modeling assumptions and approximations are made ... to represent the truth o In this context, the data has no noise o Discrepancy ...

  19. Confirmation of standard error analysis techniques applied to...

    Office of Scientific and Technical Information (OSTI)

    reported parameter errors are not reliable in many EXAFS studies in the literature. ... Country of Publication: United States Language: English Subject: 75; ABSORPTION; ACCURACY; ...

  20. U-058: Apache Struts Conversion Error OGNL Expression Injection...

    Office of Energy Efficiency and Renewable Energy (EERE) Indexed Site

    in Apache Struts. A remote user can execute arbitrary commands on the target system. PLATFORM: Apache Struts 2.x ABSTRACT: Apache Struts Conversion Error OGNL Expression...

  1. Gross error detection and stage efficiency estimation in a separation process

    SciTech Connect (OSTI)

    Serth, R.W.; Srikanth, B. . Dept. of Chemical and Natural Gas Engineering); Maronga, S.J. . Dept. of Chemical and Process Engineering)

    1993-10-01

    Accurate process models are required for optimization and control in chemical plants and petroleum refineries. These models involve various equipment parameters, such as stage efficiencies in distillation columns, the values of which must be determined by fitting the models to process data. Since the data contain random and systematic measurement errors, some of which may be large (gross errors), they must be reconciled to obtain reliable estimates of equipment parameters. The problem thus involves parameter estimation coupled with gross error detection and data reconciliation. MacDonald and Howat (1988) studied the above problem for a single-stage flash distillation process. Their analysis was based on the definition of stage efficiency due to Hausen, which has some significant disadvantages in this context, as discussed below. In addition, they considered only data sets which contained no gross errors. The purpose of this article is to extend the above work by considering alternative definitions of state efficiency and efficiency estimation in the presence of gross errors.

  2. Errors in determination of soil water content using time-domain reflectometry caused by soil compaction around wave guides

    SciTech Connect (OSTI)

    Ghezzehei, T.A.

    2008-05-29

    Application of time domain reflectometry (TDR) in soil hydrology often involves the conversion of TDR-measured dielectric permittivity to water content using universal calibration equations (empirical or physically based). Deviations of soil-specific calibrations from the universal calibrations have been noted and are usually attributed to peculiar composition of soil constituents, such as high content of clay and/or organic matter. Although it is recognized that soil disturbance by TDR waveguides may have impact on measurement errors, to our knowledge, there has not been any quantification of this effect. In this paper, we introduce a method that estimates this error by combining two models: one that describes soil compaction around cylindrical objects and another that translates change in bulk density to evolution of soil water retention characteristics. Our analysis indicates that the compaction pattern depends on the mechanical properties of the soil at the time of installation. The relative error in water content measurement depends on the compaction pattern as well as the water content and water retention properties of the soil. Illustrative calculations based on measured soil mechanical and hydrologic properties from the literature indicate that the measurement errors of using a standard three-prong TDR waveguide could be up to 10%. We also show that the error scales linearly with the ratio of rod radius to the interradius spacing.

  3. Assessment of Systematic Chromatic Errors that Impact Sub-1% Photometric Precision in Large-Area Sky Surveys

    SciTech Connect (OSTI)

    Li, T.S.; et al.

    2016-01-01

    Meeting the science goals for many current and future ground-based optical large-area sky surveys requires that the calibrated broadband photometry is stable in time and uniform over the sky to 1% precision or better. Past surveys have achieved photometric precision of 1-2% by calibrating the survey's stellar photometry with repeated measurements of a large number of stars observed in multiple epochs. The calibration techniques employed by these surveys only consider the relative frame-by-frame photometric zeropoint offset and the focal plane position-dependent illumination corrections, which are independent of the source color. However, variations in the wavelength dependence of the atmospheric transmission and the instrumental throughput induce source color-dependent systematic errors. These systematic errors must also be considered to achieve the most precise photometric measurements. In this paper, we examine such systematic chromatic errors using photometry from the Dark Energy Survey (DES) as an example. We define a natural magnitude system for DES and calculate the systematic errors on stellar magnitudes, when the atmospheric transmission and instrumental throughput deviate from the natural system. We conclude that the systematic chromatic errors caused by the change of airmass in each exposure, the change of the precipitable water vapor and aerosol in the atmosphere over time, and the non-uniformity of instrumental throughput over the focal plane, can be up to 2% in some bandpasses. We compare the calculated systematic chromatic errors with the observed DES data. For the test sample data, we correct these errors using measurements of the atmospheric transmission and instrumental throughput. The residual after correction is less than 0.3%. We also find that the errors for non-stellar objects are redshift-dependent and can be larger than those for stars at certain redshifts.

  4. Balancing aggregation and smoothing errors in inverse models

    DOE Public Access Gateway for Energy & Science Beta (PAGES Beta)

    Turner, A. J.; Jacob, D. J.

    2015-01-13

    Inverse models use observations of a system (observation vector) to quantify the variables driving that system (state vector) by statistical optimization. When the observation vector is large, such as with satellite data, selecting a suitable dimension for the state vector is a challenge. A state vector that is too large cannot be effectively constrained by the observations, leading to smoothing error. However, reducing the dimension of the state vector leads to aggregation error as prior relationships between state vector elements are imposed rather than optimized. Here we present a method for quantifying aggregation and smoothing errors as a function ofmore » state vector dimension, so that a suitable dimension can be selected by minimizing the combined error. Reducing the state vector within the aggregation error constraints can have the added advantage of enabling analytical solution to the inverse problem with full error characterization. We compare three methods for reducing the dimension of the state vector from its native resolution: (1) merging adjacent elements (grid coarsening), (2) clustering with principal component analysis (PCA), and (3) applying a Gaussian mixture model (GMM) with Gaussian pdfs as state vector elements on which the native-resolution state vector elements are projected using radial basis functions (RBFs). The GMM method leads to somewhat lower aggregation error than the other methods, but more importantly it retains resolution of major local features in the state vector while smoothing weak and broad features.« less

  5. Balancing aggregation and smoothing errors in inverse models

    DOE Public Access Gateway for Energy & Science Beta (PAGES Beta)

    Turner, A. J.; Jacob, D. J.

    2015-06-30

    Inverse models use observations of a system (observation vector) to quantify the variables driving that system (state vector) by statistical optimization. When the observation vector is large, such as with satellite data, selecting a suitable dimension for the state vector is a challenge. A state vector that is too large cannot be effectively constrained by the observations, leading to smoothing error. However, reducing the dimension of the state vector leads to aggregation error as prior relationships between state vector elements are imposed rather than optimized. Here we present a method for quantifying aggregation and smoothing errors as a function ofmore » state vector dimension, so that a suitable dimension can be selected by minimizing the combined error. Reducing the state vector within the aggregation error constraints can have the added advantage of enabling analytical solution to the inverse problem with full error characterization. We compare three methods for reducing the dimension of the state vector from its native resolution: (1) merging adjacent elements (grid coarsening), (2) clustering with principal component analysis (PCA), and (3) applying a Gaussian mixture model (GMM) with Gaussian pdfs as state vector elements on which the native-resolution state vector elements are projected using radial basis functions (RBFs). The GMM method leads to somewhat lower aggregation error than the other methods, but more importantly it retains resolution of major local features in the state vector while smoothing weak and broad features.« less

  6. Link error from craype/2.5.0

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Link error from craype/2.5.0 Link error from craype/2.5.0 January 13, 2016 by Woo-Sun Yang If you build a code using a file called 'configure' with craype/2.5.0, Cray build-tools assumes that you want to use the 'native' link mode (e.g., gcc defaults to dynamic linking), by adding '-Wl,-rpath=/opt/intel/composer_xe_2015/compiler/lib/intel64 -lintlc'. This creates a link error: /usr/bin/ld: cannot find -lintlc A temporary work around is to swap the default craype (2.5.0) with an older or newer

  7. Wind Power Forecasting Error Distributions: An International Comparison; Preprint

    SciTech Connect (OSTI)

    Hodge, B. M.; Lew, D.; Milligan, M.; Holttinen, H.; Sillanpaa, S.; Gomez-Lazaro, E.; Scharff, R.; Soder, L.; Larsen, X. G.; Giebel, G.; Flynn, D.; Dobschinski, J.

    2012-09-01

    Wind power forecasting is expected to be an important enabler for greater penetration of wind power into electricity systems. Because no wind forecasting system is perfect, a thorough understanding of the errors that do occur can be critical to system operation functions, such as the setting of operating reserve levels. This paper provides an international comparison of the distribution of wind power forecasting errors from operational systems, based on real forecast data. The paper concludes with an assessment of similarities and differences between the errors observed in different locations.

  8. Intel C++ compiler error: stl_iterator_base_types.h

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    C++ compiler error: stl_iterator_base_types.h Intel C++ compiler error: stl_iterator_base_types.h December 7, 2015 by Scott French Because the system-supplied version of GCC is relatively old (4.3.4) it is common practice to load the gcc module on our Cray systems when C++11 support is required under the Intel C++ compilers. While this works as expected under the GCC 4.8 and 4.9 series compilers, the 5.x series can cause Intel C++ compile-time errors similar to the following:

  9. Error Detection, Factorization and Correction for Multi-View Scene Reconstruction from Aerial Imagery

    SciTech Connect (OSTI)

    Hess-Flores, M

    2011-11-10

    Scene reconstruction from video sequences has become a prominent computer vision research area in recent years, due to its large number of applications in fields such as security, robotics and virtual reality. Despite recent progress in this field, there are still a number of issues that manifest as incomplete, incorrect or computationally-expensive reconstructions. The engine behind achieving reconstruction is the matching of features between images, where common conditions such as occlusions, lighting changes and texture-less regions can all affect matching accuracy. Subsequent processes that rely on matching accuracy, such as camera parameter estimation, structure computation and non-linear parameter optimization, are also vulnerable to additional sources of error, such as degeneracies and mathematical instability. Detection and correction of errors, along with robustness in parameter solvers, are a must in order to achieve a very accurate final scene reconstruction. However, error detection is in general difficult due to the lack of ground-truth information about the given scene, such as the absolute position of scene points or GPS/IMU coordinates for the camera(s) viewing the scene. In this dissertation, methods are presented for the detection, factorization and correction of error sources present in all stages of a scene reconstruction pipeline from video, in the absence of ground-truth knowledge. Two main applications are discussed. The first set of algorithms derive total structural error measurements after an initial scene structure computation and factorize errors into those related to the underlying feature matching process and those related to camera parameter estimation. A brute-force local correction of inaccurate feature matches is presented, as well as an improved conditioning scheme for non-linear parameter optimization which applies weights on input parameters in proportion to estimated camera parameter errors. Another application is in

  10. Servo control booster system for minimizing following error

    DOE Patents [OSTI]

    Wise, William L.

    1985-01-01

    A closed-loop feedback-controlled servo system is disclosed which reduces command-to-response error to the system's position feedback resolution least increment, .DELTA.S.sub.R, on a continuous real-time basis for all operating speeds. The servo system employs a second position feedback control loop on a by exception basis, when the command-to-response error .gtoreq..DELTA.S.sub.R, to produce precise position correction signals. When the command-to-response error is less than .DELTA.S.sub.R, control automatically reverts to conventional control means as the second position feedback control loop is disconnected, becoming transparent to conventional servo control means. By operating the second unique position feedback control loop used herein at the appropriate clocking rate, command-to-response error may be reduced to the position feedback resolution least increment. The present system may be utilized in combination with a tachometer loop for increased stability.

  11. Error and tolerance studies for the SSC Linac

    SciTech Connect (OSTI)

    Raparia, D.; Chang, Chu Rui; Guy, F.; Hurd, J.W.; Funk, W.; Crandall, K.R.

    1993-05-01

    This paper summarizes error and tolerance studies for the SSC Linac. These studies also include higher-order multipoles. The codes used in these simulations are PARMTEQ, PARMILA, CCLDYN, PARTRACE, and CCLTRACE.

  12. Advisory on the reporting error in the combined propane stocks...

    Annual Energy Outlook [U.S. Energy Information Administration (EIA)]

    Advisory on the reporting error in the combined propane stocks for PADDs 4 and 5 Release Date: June 12, 2013 The U.S. Energy Information Administration issued the following...

  13. Quantification of the effects of dependence on human error probabiliti...

    Office of Scientific and Technical Information (OSTI)

    In estimating the probabilities of human error in the performance of a series of tasks in a nuclear power plant, the situation-specific characteristics of the series must be ...

  14. Wind Power Forecasting Error Distributions over Multiple Timescales: Preprint

    SciTech Connect (OSTI)

    Hodge, B. M.; Milligan, M.

    2011-03-01

    In this paper, we examine the shape of the persistence model error distribution for ten different wind plants in the ERCOT system over multiple timescales. Comparisons are made between the experimental distribution shape and that of the normal distribution.

  15. Visio-Error&OmissionNoClouds.vsd

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Error/Omission Process Process Owner: Department Managers, Corporate Projects and Facilities Projects February 7, 2008 KEY Responsibilities *A/E - Architectural/Engineering Firm *SCR - Sandia Contracting Representative *SDR - Sandia Delegated Representative *E&OB - Errors & Omissions Board * PM - Project Manager * REQ - Requester Facilities Projects Line Item Projects Review Design Findings and Begin Discovery PM Cost Impact? Yes Cost Impact <3% of ICAA? Yes Yes Take Out of Project

  16. Compiler-Assisted Detection of Transient Memory Errors

    SciTech Connect (OSTI)

    Tavarageri, Sanket; Krishnamoorthy, Sriram; Sadayappan, Ponnuswamy

    2014-06-09

    The probability of bit flips in hardware memory systems is projected to increase significantly as memory systems continue to scale in size and complexity. Effective hardware-based error detection and correction requires that the complete data path, involving all parts of the memory system, be protected with sufficient redundancy. First, this may be costly to employ on commodity computing platforms and second, even on high-end systems, protection against multi-bit errors may be lacking. Therefore, augmenting hardware error detection schemes with software techniques is of consider- able interest. In this paper, we consider software-level mechanisms to comprehensively detect transient memory faults. We develop novel compile-time algorithms to instrument application programs with checksum computation codes so as to detect memory errors. Unlike prior approaches that employ checksums on computational and architectural state, our scheme verifies every data access and works by tracking variables as they are produced and consumed. Experimental evaluation demonstrates that the proposed comprehensive error detection solution is viable as a completely software-only scheme. We also demonstrate that with limited hardware support, overheads of error detection can be further reduced.

  17. Minimising the error in eigenvalue calculations involving the Boltzmann transport equation using goal-based adaptivity on unstructured meshes

    SciTech Connect (OSTI)

    Goffin, Mark A.; Baker, Christopher M.J.; Buchan, Andrew G.; Pain, Christopher C.; Eaton, Matthew D.; Smith, Paul N.

    2013-06-01

    This article presents a method for goal-based anisotropic adaptive methods for the finite element method applied to the Boltzmann transport equation. The neutron multiplication factor, k{sub eff}, is used as the goal of the adaptive procedure. The anisotropic adaptive algorithm requires error measures for k{sub eff} with directional dependence. General error estimators are derived for any given functional of the flux and applied to k{sub eff} to acquire the driving force for the adaptive procedure. The error estimators require the solution of an appropriately formed dual equation. Forward and dual error indicators are calculated by weighting the Hessian of each solution with the dual and forward residual respectively. The Hessian is used as an approximation of the interpolation error in the solution which gives rise to the directional dependence. The two indicators are combined to form a single error metric that is used to adapt the finite element mesh. The residual is approximated using a novel technique arising from the sub-grid scale finite element discretisation. Two adaptive routes are demonstrated: (i) a single mesh is used to solve all energy groups, and (ii) a different mesh is used to solve each energy group. The second method aims to capture the benefit from representing the flux from each energy group on a specifically optimised mesh. The k{sub eff} goal-based adaptive method was applied to three examples which illustrate the superior accuracy in criticality problems that can be obtained.

  18. A High-Precision Instrument for Mapping of Rotational Errors in Rotary Stages

    DOE Public Access Gateway for Energy & Science Beta (PAGES Beta)

    Xu, W.; Lauer, K.; Chu, Y.; Nazaretski, E.

    2014-11-02

    A rotational stage is a key component of every X-ray instrument capable of providing tomographic or diffraction measurements. To perform accurate three-dimensional reconstructions, runout errors due to imperfect rotation (e.g. circle of confusion) must be quantified and corrected. A dedicated instrument capable of full characterization and circle of confusion mapping in rotary stages down to the sub-10 nm level has been developed. A high-stability design, with an array of five capacitive sensors, allows simultaneous measurements of wobble, radial and axial displacements. The developed instrument has been used for characterization of two mechanical stages which are part of an X-ray microscope.

  19. Economic penalties of problems and errors in solar energy systems

    SciTech Connect (OSTI)

    Raman, K.; Sparkes, H.R.

    1983-01-01

    Experience with a large number of installed solar energy systems in the HUD Solar Program has shown that a variety of problems and design/installation errors have occurred in many solar systems, sometimes resulting in substantial additional costs for repair and/or replacement. In this paper, the effect of problems and errors on the economics of solar energy systems is examined. A method is outlined for doing this in terms of selected economic indicators. The method is illustrated by a simple example of a residential solar DHW system. An example of an installed, instrumented solar energy system in the HUD Solar Program is then discussed. Detailed results are given for the effects of the problems and errors on the cash flow, cost of delivered heat, discounted payback period, and life-cycle cost of the solar energy system. Conclusions are drawn regarding the most suitable economic indicators for showing the effects of problems and errors in solar energy systems. A method is outlined for deciding on the maximum justifiable expenditure for maintenance on a solar energy system with problems or errors.

  20. Calculation of the Johann error for spherically bent x-ray imaging crystal spectrometers

    SciTech Connect (OSTI)

    Wang, E.; Beiersdorfer, P.; Gu, M.; Bitter, M.; Delgado-Aparicio, L.; Hill, K. W.; Reinke, M.; Rice, J. E.; Podpaly, Y.

    2010-10-15

    New x-ray imaging crystal spectrometers, currently operating on Alcator C-Mod, NSTX, EAST, and KSTAR, record spectral lines of highly charged ions, such as Ar{sup 16+}, from multiple sightlines to obtain profiles of ion temperature and of toroidal plasma rotation velocity from Doppler measurements. In the present work, we describe a new data analysis routine, which accounts for the specific geometry of the sightlines of a curved-crystal spectrometer and includes corrections for the Johann error to facilitate the tomographic inversion. Such corrections are important to distinguish velocity induced Doppler shifts from instrumental line shifts caused by the Johann error. The importance of this correction is demonstrated using data from Alcator C-Mod.

  1. SU-E-P-36: Evaluation of MLC Positioning Errors in Dynamic IMRT Treatments by Analyzing Dynalog Files

    SciTech Connect (OSTI)

    Olasolo, J; Pellejero, S; Gracia, M; Gallardo, N; Martin, M; Lozares, S; Maneru, F; Bragado, L; Miquelez, S; Rubio, A

    2015-06-15

    Purpose: To assess the accuracy of MLC positioning in Varian linear accelerator, in dynamic IMRT technique, from the analysis of dynalog files generated by the MLC controller. Methods: In Clinac accelerators (pre-TrueBeam technology), control system has an approximately 50ms delay (one control cycle time). Then, the system compares the measured position to the planned position corresponding to the next control cycle. As it has been confirmed by Varian technical support, this effect causes that measured positions appear in dynalogs one cycle out of phase with respect to the planned positions. Around 9000 dynalogs have been analyzed, coming from the three linear accelerators of our center (one Trilogy and two Clinac 21EX) equipped with a Millennium 120 MLC. In order to compare our results to recent publications, leaf positioning errors (RMS and 95th percentile) are calculated with and without delay effect. Dynalogs have been analyzed using a in-house Matlab software. Results: The RMS errors were 0.341, 0.339 and 0.348mm for each Linac; being the average error 0.343 mm. The 95th percentiles of the error were 0.617, 0.607 and 0.625; with an average of 0.617mm. A recent multi-institution study carried out by Kerns et al. found a mean leaf RMS error of 0.32mm and a 95th percentile error value of 0.64mm.Without delay effect, mean leaf RMS errors obtained were 0.040, 0.042 and 0.038mm for each treatment machine; being the average 0.040mm. The 95th percentile error values obtained were 0.057, 0.058 and 0.054 mm, with an average of 0.056mm. Conclusion: Results obtained for the mean leaf RMS error and the mean 95th percentile were consistent with the multi-institution study. Calculated error statistics with delay effect are significantly larger due to the speed proportional and systematic leaf offset. Consequently it is proposed to correct this effect in dynalogs analysis to determine the MLC performance.

  2. When soft controls get slippery: User interfaces and human error

    SciTech Connect (OSTI)

    Stubler, W.F.; O`Hara, J.M.

    1998-12-01

    Many types of products and systems that have traditionally featured physical control devices are now being designed with soft controls--input formats appearing on computer-based display devices and operated by a variety of input devices. A review of complex human-machine systems found that soft controls are particularly prone to some types of errors and may affect overall system performance and safety. This paper discusses the application of design approaches for reducing the likelihood of these errors and for enhancing usability, user satisfaction, and system performance and safety.

  3. Laser Phase Errors in Seeded Free Electron Lasers

    SciTech Connect (OSTI)

    Ratner, D.; Fry, A.; Stupakov, G.; White, W.; /SLAC

    2012-04-17

    Harmonic seeding of free electron lasers has attracted significant attention as a method for producing transform-limited pulses in the soft x-ray region. Harmonic multiplication schemes extend seeding to shorter wavelengths, but also amplify the spectral phase errors of the initial seed laser, and may degrade the pulse quality and impede production of transform-limited pulses. In this paper we consider the effect of seed laser phase errors in high gain harmonic generation and echo-enabled harmonic generation. We use simulations to confirm analytical results for the case of linearly chirped seed lasers, and extend the results for arbitrary seed laser envelope and phase.

  4. MPI errors from cray-mpich/7.3.0

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    MPI errors from cray-mpich/7.3.0 MPI errors from cray-mpich/7.3.0 January 6, 2016 by Ankit Bhagatwala A change in the MPICH2 library that now strictly enforces non-overlapping buffers in MPI collectives may cause some MPI applications that use overlapping buffers to fail at runtime. As an example, one of the routines affected is MPI_ALLGATHER. There are several possible fixes. The cleanest one is to specify MPI_IN_PLACE instead of the address of the send buffer for cases where sendbuf and

  5. JLab SRF Cavity Fabrication Errors, Consequences and Lessons Learned

    SciTech Connect (OSTI)

    Frank Marhauser

    2011-09-01

    Today, elliptical superconducting RF (SRF) cavities are preferably made from deep-drawn niobium sheets as pursued at Jefferson Laboratory (JLab). The fabrication of a cavity incorporates various cavity cell machining, trimming and electron beam welding (EBW) steps as well as surface chemistry that add to forming errors creating geometrical deviations of the cavity shape from its design. An analysis of in-house built cavities over the last years revealed significant errors in cavity production. Past fabrication flaws are described and lessons learned applied successfully to the most recent in-house series production of multi-cell cavities.

  6. V-172: ISC BIND RUNTIME_CHECK Error Lets Remote Users Deny Service...

    Office of Energy Efficiency and Renewable Energy (EERE) Indexed Site

    ISC BIND RUNTIMECHECK Error Lets Remote Users Deny Service Against Recursive Resolvers V-172: ISC BIND RUNTIMECHECK Error Lets Remote Users Deny Service Against Recursive...

  7. Measurement uncertainty relations

    SciTech Connect (OSTI)

    Busch, Paul; Lahti, Pekka; Werner, Reinhard F.

    2014-04-15

    Measurement uncertainty relations are quantitative bounds on the errors in an approximate joint measurement of two observables. They can be seen as a generalization of the error/disturbance tradeoff first discussed heuristically by Heisenberg. Here we prove such relations for the case of two canonically conjugate observables like position and momentum, and establish a close connection with the more familiar preparation uncertainty relations constraining the sharpness of the distributions of the two observables in the same state. Both sets of relations are generalized to means of order ? rather than the usual quadratic means, and we show that the optimal constants are the same for preparation and for measurement uncertainty. The constants are determined numerically and compared with some bounds in the literature. In both cases, the near-saturation of the inequalities entails that the state (resp. observable) is uniformly close to a minimizing one.

  8. Contributions to Human Errors and Breaches in National Security Applications.

    SciTech Connect (OSTI)

    Pond, D. J.; Houghton, F. K.; Gilmore, W. E.

    2002-01-01

    Los Alamos National Laboratory has recognized that security infractions are often the consequence of various types of human errors (e.g., mistakes, lapses, slips) and/or breaches (i.e., deliberate deviations from policies or required procedures with no intention to bring about an adverse security consequence) and therefore has established an error reduction program based in part on the techniques used to mitigate hazard and accident potentials. One cornerstone of this program, definition of the situational and personal factors that increase the likelihood of employee errors and breaches, is detailed here. This information can be used retrospectively (as in accident investigations) to support and guide inquiries into security incidents or prospectively (as in hazard assessments) to guide efforts to reduce the likelihood of error/incident occurrence. Both approaches provide the foundation for targeted interventions to reduce the influence of these factors and for the formation of subsequent 'lessons learned.' Overall security is enhanced not only by reducing the inadvertent releases of classified information but also by reducing the security and safeguards resources devoted to them, thereby allowing these resources to be concentrated on acts of malevolence.

  9. Error Detection and Correction LDMS Plugin Version 1.0

    SciTech Connect (OSTI)

    Shoga, Kathleen; Allan, Ben

    2015-11-02

    Sandia's Lightweight Distributed Metric Service (LDMS) is a data collection and transport system used at Livermore Computing to gather performance data across the center. While Sandia has a set of plugins available, they do not include all the data we need to capture. The ECAC plugin that we have developed enables collection of the Error Detection and Correction (EDAC) counters.

  10. Servo control booster system for minimizing following error

    DOE Patents [OSTI]

    Wise, W.L.

    1979-07-26

    A closed-loop feedback-controlled servo system is disclosed which reduces command-to-response error to the system's position feedback resolution least increment, ..delta..S/sub R/, on a continuous real-time basis, for all operational times of consequence and for all operating speeds. The servo system employs a second position feedback control loop on a by exception basis, when the command-to-response error greater than or equal to ..delta..S/sub R/, to produce precise position correction signals. When the command-to-response error is less than ..delta..S/sub R/, control automatically reverts to conventional control means as the second position feedback control loop is disconnected, becoming transparent to conventional servo control means. By operating the second unique position feedback control loop used herein at the appropriate clocking rate, command-to-response error may be reduced to the position feedback resolution least increment. The present system may be utilized in combination with a tachometer loop for increased stability.

  11. Types of Possible Survey Errors in Estimates Published in the Weekly Natural Gas Storage Report

    Reports and Publications (EIA)

    2016-01-01

    This document lists types of potential errors in EIA estimates published in the WNGSR. Survey errors are an unavoidable aspect of data collection. Error is inherent in all collected data, regardless of the source of the data and the care and competence of data collectors. The type and extent of error depends on the type and characteristics of the survey.

  12. ARM - Measurements

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    govMeasurementsAerosols

  13. ARM - Measurements

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    govMeasurementsRadiometric

  14. MPI Runtime Error Detection with MUST: Advances in Deadlock Detection

    DOE Public Access Gateway for Energy & Science Beta (PAGES Beta)

    Hilbrich, Tobias; Protze, Joachim; Schulz, Martin; de Supinski, Bronis R.; Müller, Matthias S.

    2013-01-01

    The widely used Message Passing Interface (MPI) is complex and rich. As a result, application developers require automated tools to avoid and to detect MPI programming errors. We present the Marmot Umpire Scalable Tool (MUST) that detects such errors with significantly increased scalability. We present improvements to our graph-based deadlock detection approach for MPI, which cover future MPI extensions. Our enhancements also check complex MPI constructs that no previous graph-based detection approach handled correctly. Finally, we present optimizations for the processing of MPI operations that reduce runtime deadlock detection overheads. Existing approaches often require 𝒪( p ) analysis timemore » per MPI operation, for p processes. We empirically observe that our improvements lead to sub-linear or better analysis time per operation for a wide range of real world applications.« less

  15. Comparison of Wind Power and Load Forecasting Error Distributions: Preprint

    SciTech Connect (OSTI)

    Hodge, B. M.; Florita, A.; Orwig, K.; Lew, D.; Milligan, M.

    2012-07-01

    The introduction of large amounts of variable and uncertain power sources, such as wind power, into the electricity grid presents a number of challenges for system operations. One issue involves the uncertainty associated with scheduling power that wind will supply in future timeframes. However, this is not an entirely new challenge; load is also variable and uncertain, and is strongly influenced by weather patterns. In this work we make a comparison between the day-ahead forecasting errors encountered in wind power forecasting and load forecasting. The study examines the distribution of errors from operational forecasting systems in two different Independent System Operator (ISO) regions for both wind power and load forecasts at the day-ahead timeframe. The day-ahead timescale is critical in power system operations because it serves the unit commitment function for slow-starting conventional generators.

  16. Method and system for reducing errors in vehicle weighing systems

    DOE Patents [OSTI]

    Hively, Lee M.; Abercrombie, Robert K.

    2010-08-24

    A method and system (10, 23) for determining vehicle weight to a precision of <0.1%, uses a plurality of weight sensing elements (23), a computer (10) for reading in weighing data for a vehicle (25) and produces a dataset representing the total weight of a vehicle via programming (40-53) that is executable by the computer (10) for (a) providing a plurality of mode parameters that characterize each oscillatory mode in the data due to movement of the vehicle during weighing, (b) by determining the oscillatory mode at which there is a minimum error in the weighing data; (c) processing the weighing data to remove that dynamical oscillation from the weighing data; and (d) repeating steps (a)-(c) until the error in the set of weighing data is <0.1% in the vehicle weight.

  17. Some aspects of statistical modeling of human-error probability

    SciTech Connect (OSTI)

    Prairie, R. R.

    1982-01-01

    Human reliability analyses (HRA) are often performed as part of risk assessment and reliability projects. Recent events in nuclear power have shown the potential importance of the human element. There are several on-going efforts in the US and elsewhere with the purpose of modeling human error such that the human contribution can be incorporated into an overall risk assessment associated with one or more aspects of nuclear power. An effort that is described here uses the HRA (event tree) to quantify and model the human contribution to risk. As an example, risk analyses are being prepared on several nuclear power plants as part of the Interim Reliability Assessment Program (IREP). In this process the risk analyst selects the elements of his fault tree that could be contributed to by human error. He then solicits the HF analyst to do a HRA on this element.

  18. Runtime Detection of C-Style Errors in UPC Code

    SciTech Connect (OSTI)

    Pirkelbauer, P; Liao, C; Panas, T; Quinlan, D

    2011-09-29

    Unified Parallel C (UPC) extends the C programming language (ISO C 99) with explicit parallel programming support for the partitioned global address space (PGAS), which provides a global memory space with localized partitions to each thread. Like its ancestor C, UPC is a low-level language that emphasizes code efficiency over safety. The absence of dynamic (and static) safety checks allows programmer oversights and software flaws that can be hard to spot. In this paper, we present an extension of a dynamic analysis tool, ROSE-Code Instrumentation and Runtime Monitor (ROSECIRM), for UPC to help programmers find C-style errors involving the global address space. Built on top of the ROSE source-to-source compiler infrastructure, the tool instruments source files with code that monitors operations and keeps track of changes to the system state. The resulting code is linked to a runtime monitor that observes the program execution and finds software defects. We describe the extensions to ROSE-CIRM that were necessary to support UPC. We discuss complications that arise from parallel code and our solutions. We test ROSE-CIRM against a runtime error detection test suite, and present performance results obtained from running error-free codes. ROSE-CIRM is released as part of the ROSE compiler under a BSD-style open source license.

  19. Radar range measurements in the atmosphere.

    SciTech Connect (OSTI)

    Doerry, Armin Walter

    2013-02-01

    The earth's atmosphere affects the velocity of propagation of microwave signals. This imparts a range error to radar range measurements that assume the typical simplistic model for propagation velocity. This range error is a function of atmospheric constituents, such as water vapor, as well as the geometry of the radar data collection, notably altitude and range. Models are presented for calculating atmospheric effects on radar range measurements, and compared against more elaborate atmospheric models.

  20. SU-E-T-374: Sensitivity of ArcCHECK to Tomotherapy Delivery Errors: Dependence On Analysis Technique

    SciTech Connect (OSTI)

    Templeton, A; Chu, J; Turian, J

    2014-06-01

    Purpose: ArcCHECK (Sun Nuclear) is a cylindrical diode array detector allowing three-dimensional sampling of dose, particularly useful in treatment delivery QA of helical tomotherapy. Gamma passing rate is a common method of analyzing results from diode arrays, but is less intuitive in 3D with complex measured dose distributions. This study explores the sensitivity of gamma passing rate to choice of analysis technique in the context of its ability to detect errors introduced into the treatment delivery. Methods: Nine treatment plans were altered to introduce errors in: couch speed, gantry/sonogram synchronization, and leaf open time. Each plan was then delivered to ArcCHECK in each of the following arrangements: offset, when the high dose area of the plan is delivered to the side of the phantom so that some diode measurements will be on the order of the prescription dose, and centered, when the high dose is in the center of the phantom where an ion chamber measurement may be acquired, but the diode measurements are in the mid to low-dose region at the periphery of the plan. Gamma analysis was performed at 3%/3mm tolerance and both global and local gamma criteria. The threshold of detectability for each error type was calculated as the magnitude at which the gamma passing rate drops below 90%. Results: Global gamma criteria reduced the sensitivity in the offset arrangement (from 2.3% to 4.5%, 8 to 21, and 3ms to 8ms for couch-speed decrease, gantry-error, and leaf-opening increase, respectively). The centered arrangement detected changes at 3.3%, 5, and 4ms with smaller variation. Conclusion: Each arrangement has advantages; offsetting allows more sampling of the higher dose region, while centering allows an ion chamber measurement and potentially better use of tools such as 3DVH, at the cost of positioning more of the diodes in the sometimes noisy mid-dose region.

  1. Solution mass measurement

    SciTech Connect (OSTI)

    Ford, W.; Marshall, R.S.; Osborn, L.C.; Picard, R.; Thomas, C.C. Jr.

    1982-07-01

    This report describes the efforts to develop and demonstrate a solution mass measurement system for use at the Los Alamos Plutonium Facility. Because of inaccuracy of load cell measurements, our major effort was directed towards the pneumatic bubbler tube. The differential pressure between the air inlet to the bubbler tube and the glovebox interior is measured and is proportional to the solution mass in the tank. An inexpensive, reliable pressure transducer system for measuring solution mass in vertical, cylindrical tanks was developed, tested, and evaluated in a laboratory test bed. The system can withstand the over- and underpressures resulting from solution transfer operations and can prevent solution backup into the measurement pressure transducer during transfers. Drifts, noise, quantization error, and other effects limit the accuracy to 30 g. A transportable calibration system using a precision machined tank, pneumatic bubbler tubes, and a Ruska DDR 6000 electromanometer was designed, fabricated, tested, and evaluated. Resolution of the system is +-3.5 g out of 50 kg. The calibration error is 5 g, using room-temperature water as the calibrating fluid. Future efforts will be directed towards in-plant test and evaluation of the tank measurement systems. 16 figures, 3 tables.

  2. Shared dosimetry error in epidemiological dose-response analyses

    SciTech Connect (OSTI)

    Stram, Daniel O.; Preston, Dale L.; Sokolnikov, Mikhail; Napier, Bruce; Kopecky, Kenneth J.; Boice, John; Beck, Harold; Till, John; Bouville, Andre; Zeeb, Hajo

    2015-03-23

    Radiation dose reconstruction systems for large-scale epidemiological studies are sophisticated both in providing estimates of dose and in representing dosimetry uncertainty. For example, a computer program was used by the Hanford Thyroid Disease Study to provide 100 realizations of possible dose to study participants. The variation in realizations reflected the range of possible dose for each cohort member consistent with the data on dose determinates in the cohort. Another example is the Mayak Worker Dosimetry System 2013 which estimates both external and internal exposures and provides multiple realizations of "possible" dose history to workers given dose determinants. This paper takes up the problem of dealing with complex dosimetry systems that provide multiple realizations of dose in an epidemiologic analysis. In this paper we derive expected scores and the information matrix for a model used widely in radiation epidemiology, namely the linear excess relative risk (ERR) model that allows for a linear dose response (risk in relation to radiation) and distinguishes between modifiers of background rates and of the excess risk due to exposure. We show that treating the mean dose for each individual (calculated by averaging over the realizations) as if it was true dose (ignoring both shared and unshared dosimetry errors) gives asymptotically unbiased estimates (i.e. the score has expectation zero) and valid tests of the null hypothesis that the ERR slope ? is zero. Although the score is unbiased the information matrix (and hence the standard errors of the estimate of ?) is biased for ??0 when ignoring errors in dose estimates, and we show how to adjust the information matrix to remove this bias, using the multiple realizations of dose. The use of these methods in the context of several studies including, the Mayak Worker Cohort, and the U.S. Atomic Veterans Study, is discussed.

  3. Shared dosimetry error in epidemiological dose-response analyses

    DOE Public Access Gateway for Energy & Science Beta (PAGES Beta)

    Stram, Daniel O.; Preston, Dale L.; Sokolnikov, Mikhail; Napier, Bruce; Kopecky, Kenneth J.; Boice, John; Beck, Harold; Till, John; Bouville, Andre; Zeeb, Hajo

    2015-03-23

    Radiation dose reconstruction systems for large-scale epidemiological studies are sophisticated both in providing estimates of dose and in representing dosimetry uncertainty. For example, a computer program was used by the Hanford Thyroid Disease Study to provide 100 realizations of possible dose to study participants. The variation in realizations reflected the range of possible dose for each cohort member consistent with the data on dose determinates in the cohort. Another example is the Mayak Worker Dosimetry System 2013 which estimates both external and internal exposures and provides multiple realizations of "possible" dose history to workers given dose determinants. This paper takesmore » up the problem of dealing with complex dosimetry systems that provide multiple realizations of dose in an epidemiologic analysis. In this paper we derive expected scores and the information matrix for a model used widely in radiation epidemiology, namely the linear excess relative risk (ERR) model that allows for a linear dose response (risk in relation to radiation) and distinguishes between modifiers of background rates and of the excess risk due to exposure. We show that treating the mean dose for each individual (calculated by averaging over the realizations) as if it was true dose (ignoring both shared and unshared dosimetry errors) gives asymptotically unbiased estimates (i.e. the score has expectation zero) and valid tests of the null hypothesis that the ERR slope β is zero. Although the score is unbiased the information matrix (and hence the standard errors of the estimate of β) is biased for β≠0 when ignoring errors in dose estimates, and we show how to adjust the information matrix to remove this bias, using the multiple realizations of dose. The use of these methods in the context of several studies including, the Mayak Worker Cohort, and the U.S. Atomic Veterans Study, is discussed.« less

  4. Unconventional Rotor Power Response to Yaw Error Variations

    DOE Public Access Gateway for Energy & Science Beta (PAGES Beta)

    Schreck, S. J.; Schepers, J. G.

    2014-12-16

    Continued inquiry into rotor and blade aerodynamics remains crucial for achieving accurate, reliable prediction of wind turbine power performance under yawed conditions. To exploit key advantages conferred by controlled inflow conditions, we used EU-JOULE DATA Project and UAE Phase VI experimental data to characterize rotor power production under yawed conditions. Anomalies in rotor power variation with yaw error were observed, and the underlying fluid dynamic interactions were isolated. Unlike currently recognized influences caused by angled inflow and skewed wake, which may be considered potential flow interactions, these anomalies were linked to pronounced viscous and unsteady effects.

  5. Error-field penetration in reversed magnetic shear configurations

    SciTech Connect (OSTI)

    Wang, H. H.; Wang, Z. X.; Wang, X. Q. [MOE Key Laboratory of Materials Modification by Beams of the Ministry of Education, School of Physics and Optoelectronic Engineering, Dalian University of Technology, Dalian 116024 (China)] [MOE Key Laboratory of Materials Modification by Beams of the Ministry of Education, School of Physics and Optoelectronic Engineering, Dalian University of Technology, Dalian 116024 (China); Wang, X. G. [School of Physics, Peking University, Beijing 100871 (China)] [School of Physics, Peking University, Beijing 100871 (China)

    2013-06-15

    Error-field penetration in reversed magnetic shear (RMS) configurations is numerically investigated by using a two-dimensional resistive magnetohydrodynamic model in slab geometry. To explore different dynamic processes in locked modes, three equilibrium states are adopted. Stable, marginal, and unstable current profiles for double tearing modes are designed by varying the current intensity between two resonant surfaces separated by a certain distance. Further, the dynamic characteristics of locked modes in the three RMS states are identified, and the relevant physics mechanisms are elucidated. The scaling behavior of critical perturbation value with initial plasma velocity is numerically obtained, which obeys previously established relevant analytical theory in the viscoresistive regime.

  6. ARM - Measurements

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    govMeasurementsCloud Properties

  7. ARM - Measurements

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    govMeasurementsSurface Properties

  8. FlipSphere: A Software-based DRAM Error Detection and Correction...

    Office of Scientific and Technical Information (OSTI)

    FlipSphere: A Software-based DRAM Error Detection and Correction Library for HPC. Citation Details In-Document Search Title: FlipSphere: A Software-based DRAM Error Detection and ...

  9. V-194: Citrix XenServer Memory Management Error Lets Local Administrat...

    Office of Energy Efficiency and Renewable Energy (EERE) Indexed Site

    XenServer Memory Management Error Lets Local Administrative Users on the Guest Gain Access on the Host V-194: Citrix XenServer Memory Management Error Lets Local Administrative...

  10. From the Lab to the real world : sources of error in UF {sub 6} gas enrichment monitoring

    SciTech Connect (OSTI)

    Lombardi, Marcie L.

    2012-03-01

    Safeguarding uranium enrichment facilities is a serious concern for the International Atomic Energy Agency (IAEA). Safeguards methods have changed over the years, most recently switching to an improved safeguards model that calls for new technologies to help keep up with the increasing size and complexity of today’s gas centrifuge enrichment plants (GCEPs). One of the primary goals of the IAEA is to detect the production of uranium at levels greater than those an enrichment facility may have declared. In order to accomplish this goal, new enrichment monitors need to be as accurate as possible. This dissertation will look at the Advanced Enrichment Monitor (AEM), a new enrichment monitor designed at Los Alamos National Laboratory. Specifically explored are various factors that could potentially contribute to errors in a final enrichment determination delivered by the AEM. There are many factors that can cause errors in the determination of uranium hexafluoride (UF{sub 6}) gas enrichment, especially during the period when the enrichment is being measured in an operating GCEP. To measure enrichment using the AEM, a passive 186-keV (kiloelectronvolt) measurement is used to determine the {sup 235}U content in the gas, and a transmission measurement or a gas pressure reading is used to determine the total uranium content. A transmission spectrum is generated using an x-ray tube and a “notch” filter. In this dissertation, changes that could occur in the detection efficiency and the transmission errors that could result from variations in pipe-wall thickness will be explored. Additional factors that could contribute to errors in enrichment measurement will also be examined, including changes in the gas pressure, ambient and UF{sub 6} temperature, instrumental errors, and the effects of uranium deposits on the inside of the pipe walls will be considered. The sensitivity of the enrichment calculation to these various parameters will then be evaluated. Previously, UF

  11. Resolved: "error while loading shared libraries: libalpslli.so.0" with

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    serial codes on login nodes "error while loading shared libraries: libalpslli.so.0" with serial codes on login nodes Resolved: "error while loading shared libraries: libalpslli.so.0" with serial codes on login nodes December 13, 2013 by Helen He Symptom: Dynamic executables built with compiler wrappers running directly on the external login nodes are getting the following error message: % ftn -dynamic -o testf testf.f % ./testf ./testf: error while loading shared

  12. T-719:Apache mod_proxy_ajp HTTP Processing Error Lets Remote Users Deny Service

    Broader source: Energy.gov [DOE]

    A remote user can cause the backend server to remain in an error state until the retry timeout expires.

  13. Temperature Measurements in the Magnetic Measurement Facility

    SciTech Connect (OSTI)

    Wolf, Zachary

    2010-12-13

    Several key LCLS undulator parameter values depend strongly on temperature primarily because of the permanent magnet material the undulators are constructed with. The undulators will be tuned to have specific parameter values in the Magnetic Measurement Facility (MMF). Consequently, it is necessary for the temperature of the MMF to remain fairly constant. Requirements on undulator temperature have been established. When in use, the undulator temperature will be in the range 20.0 {+-} 0.2 C. In the MMF, the undulator tuning will be done at 20.0 {+-} 0.1 C. For special studies, the MMF temperature set point can be changed to a value between 18 C and 23 C with stability of {+-}0.1 C. In order to ensure that the MMF temperature requirements are met, the MMF must have a system to measure temperatures. The accuracy of the MMF temperature measurement system must be better than the {+-}0.1 C undulator tuning temperature tolerance, and is taken to be {+-}0.01 C. The temperature measurement system for the MMF is under construction. It is similar to a prototype system we built two years ago in the Sector 10 alignment lab at SLAC. At that time, our goal was to measure the lab temperature to {+-}0.1 C. The system has worked well for two years and has maintained its accuracy. For the MMF system, we propose better sensors and a more extensive calibration program to achieve the factor of 10 increase in accuracy. In this note we describe the measurement system under construction. We motivate our choice of system components and give an overview of the system. Most of the software for the system has been written and will be discussed. We discuss error sources in temperature measurements and show how these errors have been dealt with. The calibration system is described in detail. All the LCLS undulators must be tuned in the Magnetic Measurement Facility at the same temperature to within {+-}0.1 C. In order to ensure this, we are building a system to measure the temperature of the

  14. Error field penetration and locking to the backward propagating wave

    SciTech Connect (OSTI)

    Finn, John M.; Cole, Andrew J.; Brennan, Dylan P.

    2015-12-30

    In this letter we investigate error field penetration, or locking, behavior in plasmas having stable tearing modes with finite real frequencies wr in the plasma frame. In particular, we address the fact that locking can drive a significant equilibrium flow. We show that this occurs at a velocity slightly above v = wr/k, corresponding to the interaction with a backward propagating tearing mode in the plasma frame. Results are discussed for a few typical tearing mode regimes, including a new derivation showing that the existence of real frequencies occurs for viscoresistive tearing modes, in an analysis including the effects of pressure gradient, curvature and parallel dynamics. The general result of locking to a finite velocity flow is applicable to a wide range of tearing mode regimes, indeed any regime where real frequencies occur.

  15. Coordinated joint motion control system with position error correction

    DOE Patents [OSTI]

    Danko, George L.

    2016-04-05

    Disclosed are an articulated hydraulic machine supporting, control system and control method for same. The articulated hydraulic machine has an end effector for performing useful work. The control system is capable of controlling the end effector for automated movement along a preselected trajectory. The control system has a position error correction system to correct discrepancies between an actual end effector trajectory and a desired end effector trajectory. The correction system can employ one or more absolute position signals provided by one or more acceleration sensors supported by one or more movable machine elements. Good trajectory positioning and repeatability can be obtained. A two joystick controller system is enabled, which can in some cases facilitate the operator's task and enhance their work quality and productivity.

  16. Coordinated joint motion control system with position error correction

    DOE Patents [OSTI]

    Danko, George

    2011-11-22

    Disclosed are an articulated hydraulic machine supporting, control system and control method for same. The articulated hydraulic machine has an end effector for performing useful work. The control system is capable of controlling the end effector for automated movement along a preselected trajectory. The control system has a position error correction system to correct discrepancies between an actual end effector trajectory and a desired end effector trajectory. The correction system can employ one or more absolute position signals provided by one or more acceleration sensors supported by one or more movable machine elements. Good trajectory positioning and repeatability can be obtained. A two-joystick controller system is enabled, which can in some cases facilitate the operator's task and enhance their work quality and productivity.

  17. Error field penetration and locking to the backward propagating wave

    DOE Public Access Gateway for Energy & Science Beta (PAGES Beta)

    Finn, John M.; Cole, Andrew J.; Brennan, Dylan P.

    2015-12-30

    In this letter we investigate error field penetration, or locking, behavior in plasmas having stable tearing modes with finite real frequencies wr in the plasma frame. In particular, we address the fact that locking can drive a significant equilibrium flow. We show that this occurs at a velocity slightly above v = wr/k, corresponding to the interaction with a backward propagating tearing mode in the plasma frame. Results are discussed for a few typical tearing mode regimes, including a new derivation showing that the existence of real frequencies occurs for viscoresistive tearing modes, in an analysis including the effects ofmore » pressure gradient, curvature and parallel dynamics. The general result of locking to a finite velocity flow is applicable to a wide range of tearing mode regimes, indeed any regime where real frequencies occur.« less

  18. TU-C-BRE-05: Clinical Implications of AAA Commissioning Errors and Ability of Common Commissioning ' Credentialing Procedures to Detect Them

    SciTech Connect (OSTI)

    McVicker, A; Oldham, M; Yin, F; Adamson, J

    2014-06-15

    Purpose: To test the ability of the TG-119 commissioning process and RPC credentialing to detect errors in the commissioning process for a commercial Treatment Planning System (TPS). Methods: We introduced commissioning errors into the commissioning process for the Anisotropic Analytical Algorithm (AAA) within the Eclipse TPS. We included errors in Dosimetric Leaf Gap (DLG), electron contamination, flattening filter material, and beam profile measurement with an inappropriately large farmer chamber (simulated using sliding window smoothing of profiles). We then evaluated the clinical impact of these errors on clinical intensity modulated radiation therapy (IMRT) plans (head and neck, low and intermediate risk prostate, mesothelioma, and scalp) by looking at PTV D99, and mean and max OAR dose. Finally, for errors with substantial clinical impact we determined sensitivity of the RPC IMRT film analysis at the midpoint between PTV and OAR using a 4mm distance to agreement metric, and of a 7% TLD dose comparison. We also determined sensitivity of the 3 dose planes of the TG-119 C-shape IMRT phantom using gamma criteria of 3% 3mm. Results: The largest clinical impact came from large changes in the DLG with a change of 1mm resulting in up to a 5% change in the primary PTV D99. This resulted in a discrepancy in the RPC TLDs in the PTVs and OARs of 7.1% and 13.6% respectively, which would have resulted in detection. While use of incorrect flattening filter caused only subtle errors (<1%) in clinical plans, the effect was most pronounced for the RPC TLDs in the OARs (>6%). Conclusion: The AAA commissioning process within the Eclipse TPS is surprisingly robust to user error. When errors do occur, the RPC and TG-119 commissioning credentialing criteria are effective at detecting them; however OAR TLDs are the most sensitive despite the RPC currently excluding them from analysis.

  19. SUMP MEASURING SYSTEM

    SciTech Connect (OSTI)

    Vrettos, N; Athneal Marzolf, A; Casandra Robinson, C; James Fiscus, J; Daniel Krementz, D; Thomas Nance, T

    2007-11-26

    The process sumps in H-Canyon at the Savannah River Site (SRS) collect leaks from process tanks and jumpers. To prevent build-up of fissile material the sumps are frequently flushed which generates liquid waste and is prone to human error. The development of inserts filled with a neutron poison will allow a reduction in the frequency of flushing. Due to concrete deterioration and deformation of the sump liners the current dimensions of the sumps are unknown. Knowledge of these dimensions is necessary for development of the inserts. To solve this problem a remote Sump Measurement System was designed, fabricated, and tested to aid development of the sump inserts.

  20. Adaptive error covariances estimation methods for ensemble Kalman filters

    SciTech Connect (OSTI)

    Zhen, Yicun; Harlim, John

    2015-08-01

    This paper presents a computationally fast algorithm for estimating, both, the system and observation noise covariances of nonlinear dynamics, that can be used in an ensemble Kalman filtering framework. The new method is a modification of Belanger's recursive method, to avoid an expensive computational cost in inverting error covariance matrices of product of innovation processes of different lags when the number of observations becomes large. When we use only product of innovation processes up to one-lag, the computational cost is indeed comparable to a recently proposed method by Berry–Sauer's. However, our method is more flexible since it allows for using information from product of innovation processes of more than one-lag. Extensive numerical comparisons between the proposed method and both the original Belanger's and Berry–Sauer's schemes are shown in various examples, ranging from low-dimensional linear and nonlinear systems of SDEs and 40-dimensional stochastically forced Lorenz-96 model. Our numerical results suggest that the proposed scheme is as accurate as the original Belanger's scheme on low-dimensional problems and has a wider range of more accurate estimates compared to Berry–Sauer's method on L-96 example.

  1. Spectral characteristics of background error covariance and multiscale data assimilation

    DOE Public Access Gateway for Energy & Science Beta (PAGES Beta)

    Li, Zhijin; Cheng, Xiaoping; Gustafson, Jr., William I.; Vogelmann, Andrew M.

    2016-05-17

    The steady increase of the spatial resolutions of numerical atmospheric and oceanic circulation models has occurred over the past decades. Horizontal grid spacing down to the order of 1 km is now often used to resolve cloud systems in the atmosphere and sub-mesoscale circulation systems in the ocean. These fine resolution models encompass a wide range of temporal and spatial scales, across which dynamical and statistical properties vary. In particular, dynamic flow systems at small scales can be spatially localized and temporarily intermittent. Difficulties of current data assimilation algorithms for such fine resolution models are numerically and theoretically examined. Ourmore » analysis shows that the background error correlation length scale is larger than 75 km for streamfunctions and is larger than 25 km for water vapor mixing ratios, even for a 2-km resolution model. A theoretical analysis suggests that such correlation length scales prevent the currently used data assimilation schemes from constraining spatial scales smaller than 150 km for streamfunctions and 50 km for water vapor mixing ratios. Moreover, our results highlight the need to fundamentally modify currently used data assimilation algorithms for assimilating high-resolution observations into the aforementioned fine resolution models. Lastly, within the framework of four-dimensional variational data assimilation, a multiscale methodology based on scale decomposition is suggested and challenges are discussed.« less

  2. In vivo enzyme activity in inborn errors of metabolism

    SciTech Connect (OSTI)

    Thompson, G.N.; Walter, J.H.; Leonard, J.V.; Halliday, D. )

    1990-08-01

    Low-dose continuous infusions of (2H5)phenylalanine, (1-13C)propionate, and (1-13C)leucine were used to quantitate phenylalanine hydroxylation in phenylketonuria (PKU, four subjects), propionate oxidation in methylmalonic acidaemia (MMA, four subjects), and propionic acidaemia (PA, four subjects) and leucine oxidation in maple syrup urine disease (MSUD, four subjects). In vivo enzyme activity in PKU, MMA, and PA subjects was similar to or in excess of that in adult controls (range of phenylalanine hydroxylation in PKU, 3.7 to 6.5 mumol/kg/h, control 3.2 to 7.9, n = 7; propionate oxidation in MMA, 15.2 to 64.8 mumol/kg/h, and in PA, 11.1 to 36.0, control 5.1 to 19.0, n = 5). By contrast, in vivo leucine oxidation was undetectable in three of the four MSUD subjects (less than 0.5 mumol/kg/h) and negligible in the remaining subject (2 mumol/kg/h, control 10.4 to 15.7, n = 6). These results suggest that significant substrate removal can be achieved in some inborn metabolic errors either through stimulation of residual enzyme activity in defective enzyme systems or by activation of alternate metabolic pathways. Both possibilities almost certainly depend on gross elevation of substrate concentrations. By contrast, only minimal in vivo oxidation of leucine appears possible in MSUD.

  3. SU-E-I-83: Error Analysis of Multi-Modality Image-Based Volumes of Rodent Solid Tumors Using a Preclinical Multi-Modality QA Phantom

    SciTech Connect (OSTI)

    Lee, Y; Fullerton, G; Goins, B

    2015-06-15

    Purpose: In our previous study a preclinical multi-modality quality assurance (QA) phantom that contains five tumor-simulating test objects with 2, 4, 7, 10 and 14 mm diameters was developed for accurate tumor size measurement by researchers during cancer drug development and testing. This study analyzed the errors during tumor volume measurement from preclinical magnetic resonance (MR), micro-computed tomography (micro- CT) and ultrasound (US) images acquired in a rodent tumor model using the preclinical multi-modality QA phantom. Methods: Using preclinical 7-Tesla MR, US and micro-CT scanners, images were acquired of subcutaneous SCC4 tumor xenografts in nude rats (3–4 rats per group; 5 groups) along with the QA phantom using the same imaging protocols. After tumors were excised, in-air micro-CT imaging was performed to determine reference tumor volume. Volumes measured for the rat tumors and phantom test objects were calculated using formula V = (π/6)*a*b*c where a, b and c are the maximum diameters in three perpendicular dimensions determined by the three imaging modalities. Then linear regression analysis was performed to compare image-based tumor volumes with the reference tumor volume and known test object volume for the rats and the phantom respectively. Results: The slopes of regression lines for in-vivo tumor volumes measured by three imaging modalities were 1.021, 1.101 and 0.862 for MRI, micro-CT and US respectively. For phantom, the slopes were 0.9485, 0.9971 and 0.9734 for MRI, micro-CT and US respectively. Conclusion: For both animal and phantom studies, random and systematic errors were observed. Random errors were observer-dependent and systematic errors were mainly due to selected imaging protocols and/or measurement method. In the animal study, there were additional systematic errors attributed to ellipsoidal assumption for tumor shape. The systematic errors measured using the QA phantom need to be taken into account to reduce measurement

  4. A two-dimensional matrix correction for off-axis portal dose prediction errors

    SciTech Connect (OSTI)

    Bailey, Daniel W.; Kumaraswamy, Lalith; Bakhtiari, Mohammad; Podgorsak, Matthew B.

    2013-05-15

    Purpose: This study presents a follow-up to a modified calibration procedure for portal dosimetry published by Bailey et al. ['An effective correction algorithm for off-axis portal dosimetry errors,' Med. Phys. 36, 4089-4094 (2009)]. A commercial portal dose prediction system exhibits disagreement of up to 15% (calibrated units) between measured and predicted images as off-axis distance increases. The previous modified calibration procedure accounts for these off-axis effects in most regions of the detecting surface, but is limited by the simplistic assumption of radial symmetry. Methods: We find that a two-dimensional (2D) matrix correction, applied to each calibrated image, accounts for off-axis prediction errors in all regions of the detecting surface, including those still problematic after the radial correction is performed. The correction matrix is calculated by quantitative comparison of predicted and measured images that span the entire detecting surface. The correction matrix was verified for dose-linearity, and its effectiveness was verified on a number of test fields. The 2D correction was employed to retrospectively examine 22 off-axis, asymmetric electronic-compensation breast fields, five intensity-modulated brain fields (moderate-high modulation) manipulated for far off-axis delivery, and 29 intensity-modulated clinical fields of varying complexity in the central portion of the detecting surface. Results: Employing the matrix correction to the off-axis test fields and clinical fields, predicted vs measured portal dose agreement improves by up to 15%, producing up to 10% better agreement than the radial correction in some areas of the detecting surface. Gamma evaluation analyses (3 mm, 3% global, 10% dose threshold) of predicted vs measured portal dose images demonstrate pass rate improvement of up to 75% with the matrix correction, producing pass rates that are up to 30% higher than those resulting from the radial correction technique alone. As in

  5. SU-E-T-195: Gantry Angle Dependency of MLC Leaf Position Error

    SciTech Connect (OSTI)

    Ju, S; Hong, C; Kim, M; Chung, K; Kim, J; Han, Y; Ahn, S; Chung, S; Shin, E; Shin, J; Kim, H; Kim, D; Choi, D

    2014-06-01

    Purpose: The aim of this study was to investigate the gantry angle dependency of the multileaf collimator (MLC) leaf position error. Methods: An automatic MLC quality assurance system (AutoMLCQA) was developed to evaluate the gantry angle dependency of the MLC leaf position error using an electronic portal imaging device (EPID). To eliminate the EPID position error due to gantry rotation, we designed a reference maker (RM) that could be inserted into the wedge mount. After setting up the EPID, a reference image was taken of the RM using an open field. Next, an EPID-based picket-fence test (PFT) was performed without the RM. These procedures were repeated at every 45° intervals of the gantry angle. A total of eight reference images and PFT image sets were analyzed using in-house software. The average MLC leaf position error was calculated at five pickets (-10, -5, 0, 5, and 10 cm) in accordance with general PFT guidelines using in-house software. This test was carried out for four linear accelerators. Results: The average MLC leaf position errors were within the set criterion of <1 mm (actual errors ranged from -0.7 to 0.8 mm) for all gantry angles, but significant gantry angle dependency was observed in all machines. The error was smaller at a gantry angle of 0° but increased toward the positive direction with gantry angle increments in the clockwise direction. The error reached a maximum value at a gantry angle of 90° and then gradually decreased until 180°. In the counter-clockwise rotation of the gantry, the same pattern of error was observed but the error increased in the negative direction. Conclusion: The AutoMLCQA system was useful to evaluate the MLC leaf position error for various gantry angles without the EPID position error. The Gantry angle dependency should be considered during MLC leaf position error analysis.

  6. GREAT3 results - I. Systematic errors in shear estimation and the impact of real galaxy morphology

    SciTech Connect (OSTI)

    Mandelbaum, Rachel; Rowe, Barnaby; Armstrong, Robert; Bard, Deborah; Bertin, Emmanuel; Bosch, James; Boutigny, Dominique; Courbin, Frederic; Dawson, William A.; Donnarumma, Annamaria; Fenech Conti, Ian; Gavazzi, Raphael; Gentile, Marc; Gill, Mandeep S. S.; Hogg, David W.; Huff, Eric M.; Jee, M. James; Kacprzak, Tomasz; Kilbinger, Martin; Kuntzer, Thibault; Lang, Dustin; Luo, Wentao; March, Marisa C.; Marshall, Philip J.; Meyers, Joshua E.; Miller, Lance; Miyatake, Hironao; Nakajima, Reiko; Ngole Mboula, Fred Maurice; Nurbaeva, Guldariya; Okura, Yuki; Paulin-Henriksson, Stephane; Rhodes, Jason; Schneider, Michael D.; Shan, Huanyuan; Sheldon, Erin S.; Simet, Melanie; Starck, Jean -Luc; Sureau, Florent; Tewes, Malte; Zarb Adami, Kristian; Zhang, Jun; Zuntz, Joe

    2015-05-11

    The study present first results from the third GRavitational lEnsing Accuracy Testing (GREAT3) challenge, the third in a sequence of challenges for testing methods of inferring weak gravitational lensing shear distortions from simulated galaxy images. GREAT3 was divided into experiments to test three specific questions, and included simulated space- and ground-based data with constant or cosmologically varying shear fields. The simplest (control) experiment included parametric galaxies with a realistic distribution of signal-to-noise, size, and ellipticity, and a complex point spread function (PSF). The other experiments tested the additional impact of realistic galaxy morphology, multiple exposure imaging, and the uncertainty about a spatially varying PSF; the last two questions will be explored in Paper II. The 24 participating teams competed to estimate lensing shears to within systematic error tolerances for upcoming Stage-IV dark energy surveys, making 1525 submissions overall. GREAT3 saw considerable variety and innovation in the types of methods applied. Several teams now meet or exceed the targets in many of the tests conducted (to within the statistical errors). We conclude that the presence of realistic galaxy morphology in simulations changes shear calibration biases by ~1 per cent for a wide range of methods. Other effects such as truncation biases due to finite galaxy postage stamps, and the impact of galaxy type as measured by the Sérsic index, are quantified for the first time. Our results generalize previous studies regarding sensitivities to galaxy size and signal-to-noise, and to PSF properties such as seeing and defocus. Almost all methods’ results support the simple model in which additive shear biases depend linearly on PSF ellipticity.

  7. GREAT3 results - I. Systematic errors in shear estimation and the impact of real galaxy morphology

    DOE Public Access Gateway for Energy & Science Beta (PAGES Beta)

    Mandelbaum, Rachel; Rowe, Barnaby; Armstrong, Robert; Bard, Deborah; Bertin, Emmanuel; Bosch, James; Boutigny, Dominique; Courbin, Frederic; Dawson, William A.; Donnarumma, Annamaria; et al

    2015-05-11

    The study present first results from the third GRavitational lEnsing Accuracy Testing (GREAT3) challenge, the third in a sequence of challenges for testing methods of inferring weak gravitational lensing shear distortions from simulated galaxy images. GREAT3 was divided into experiments to test three specific questions, and included simulated space- and ground-based data with constant or cosmologically varying shear fields. The simplest (control) experiment included parametric galaxies with a realistic distribution of signal-to-noise, size, and ellipticity, and a complex point spread function (PSF). The other experiments tested the additional impact of realistic galaxy morphology, multiple exposure imaging, and the uncertainty aboutmore » a spatially varying PSF; the last two questions will be explored in Paper II. The 24 participating teams competed to estimate lensing shears to within systematic error tolerances for upcoming Stage-IV dark energy surveys, making 1525 submissions overall. GREAT3 saw considerable variety and innovation in the types of methods applied. Several teams now meet or exceed the targets in many of the tests conducted (to within the statistical errors). We conclude that the presence of realistic galaxy morphology in simulations changes shear calibration biases by ~1 per cent for a wide range of methods. Other effects such as truncation biases due to finite galaxy postage stamps, and the impact of galaxy type as measured by the Sérsic index, are quantified for the first time. Our results generalize previous studies regarding sensitivities to galaxy size and signal-to-noise, and to PSF properties such as seeing and defocus. Almost all methods’ results support the simple model in which additive shear biases depend linearly on PSF ellipticity.« less

  8. Towards eliminating systematic errors caused by the experimental conditions in Biochemical Methane Potential (BMP) tests

    SciTech Connect (OSTI)

    Strömberg, Sten; Nistor, Mihaela; Liu, Jing

    2014-11-15

    Highlights: • The evaluated factors introduce significant systematic errors (10–38%) in BMP tests. • Ambient temperature (T) has the most substantial impact (∼10%) at low altitude. • Ambient pressure (p) has the most substantial impact (∼68%) at high altitude. • Continuous monitoring of T and p is not necessary for kinetic calculations. - Abstract: The Biochemical Methane Potential (BMP) test is increasingly recognised as a tool for selecting and pricing biomass material for production of biogas. However, the results for the same substrate often differ between laboratories and much work to standardise such tests is still needed. In the current study, the effects from four environmental factors (i.e. ambient temperature and pressure, water vapour content and initial gas composition of the reactor headspace) on the degradation kinetics and the determined methane potential were evaluated with a 2{sup 4} full factorial design. Four substrates, with different biodegradation profiles, were investigated and the ambient temperature was found to be the most significant contributor to errors in the methane potential. Concerning the kinetics of the process, the environmental factors’ impact on the calculated rate constants was negligible. The impact of the environmental factors on the kinetic parameters and methane potential from performing a BMP test at different geographical locations around the world was simulated by adjusting the data according to the ambient temperature and pressure of some chosen model sites. The largest effect on the methane potential was registered from tests performed at high altitudes due to a low ambient pressure. The results from this study illustrate the importance of considering the environmental factors’ influence on volumetric gas measurement in BMP tests. This is essential to achieve trustworthy and standardised results that can be used by researchers and end users from all over the world.

  9. GREAT3 results - I. Systematic errors in shear estimation and the impact of real galaxy morphology

    SciTech Connect (OSTI)

    Mandelbaum, Rachel; Rowe, Barnaby; Armstrong, Robert; Bard, Deborah; Bertin, Emmanuel; Bosch, James; Boutigny, Dominique; Courbin, Frederic; Dawson, William A.; Donnarumma, Annamaria; Fenech Conti, Ian; Gavazzi, Raphael; Gentile, Marc; Gill, Mandeep S. S.; Hogg, David W.; Huff, Eric M.; Jee, M. James; Kacprzak, Tomasz; Kilbinger, Martin; Kuntzer, Thibault; Lang, Dustin; Luo, Wentao; March, Marisa C.; Marshall, Philip J.; Meyers, Joshua E.; Miller, Lance; Miyatake, Hironao; Nakajima, Reiko; Ngole Mboula, Fred Maurice; Nurbaeva, Guldariya; Okura, Yuki; Paulin-Henriksson, Stephane; Rhodes, Jason; Schneider, Michael D.; Shan, Huanyuan; Sheldon, Erin S.; Simet, Melanie; Starck, Jean -Luc; Sureau, Florent; Tewes, Malte; Zarb Adami, Kristian; Zhang, Jun; Zuntz, Joe

    2015-05-11

    The study present first results from the third GRavitational lEnsing Accuracy Testing (GREAT3) challenge, the third in a sequence of challenges for testing methods of inferring weak gravitational lensing shear distortions from simulated galaxy images. GREAT3 was divided into experiments to test three specific questions, and included simulated space- and ground-based data with constant or cosmologically varying shear fields. The simplest (control) experiment included parametric galaxies with a realistic distribution of signal-to-noise, size, and ellipticity, and a complex point spread function (PSF). The other experiments tested the additional impact of realistic galaxy morphology, multiple exposure imaging, and the uncertainty about a spatially varying PSF; the last two questions will be explored in Paper II. The 24 participating teams competed to estimate lensing shears to within systematic error tolerances for upcoming Stage-IV dark energy surveys, making 1525 submissions overall. GREAT3 saw considerable variety and innovation in the types of methods applied. Several teams now meet or exceed the targets in many of the tests conducted (to within the statistical errors). We conclude that the presence of realistic galaxy morphology in simulations changes shear calibration biases by ~1 per cent for a wide range of methods. Other effects such as truncation biases due to finite galaxy postage stamps, and the impact of galaxy type as measured by the Srsic index, are quantified for the first time. Our results generalize previous studies regarding sensitivities to galaxy size and signal-to-noise, and to PSF properties such as seeing and defocus. Almost all methods results support the simple model in which additive shear biases depend linearly on PSF ellipticity.

  10. Spectral Characteristics of Background Error Covariance and Multiscale Data Assimilation: Background Error Covariance and Multiscale Data Assimilation

    DOE Public Access Gateway for Energy & Science Beta (PAGES Beta)

    Li, Zhijin; Cheng, Xiaoping; Gustafson, William I.; Vogelmann, Andrew M.

    2016-05-17

    The steady increase of the spatial resolutions of numerical atmospheric and oceanic circulation models has occurred over the past decades. Horizontal grid spacing down to the order of 1 km is now often used to resolve cloud systems in the atmosphere and sub-mesoscale circulation systems in the ocean. These fine resolution models encompass a wide range of temporal and spatial scales, across which dynamical and statistical properties vary. In particular, dynamic flow systems at small scales can be spatially localized and temporarily intermittent. Difficulties of current data assimilation algorithms for such fine resolution models are numerically and theoretically examined. Ourmore » analysis shows that the background error correlation length scale is larger than 75 km for streamfunctions and is larger than 25 km for water vapor mixing ratios, even for a 2-km resolution model. A theoretical analysis suggests that such correlation length scales prevent the currently used data assimilation schemes from constraining spatial scales smaller than 150 km for streamfunctions and 50 km for water vapor mixing ratios. Moreover, our results highlight the need to fundamentally modify currently used data assimilation algorithms for assimilating high-resolution observations into the aforementioned fine resolution models. Within the framework of four-dimensional variational data assimilation, a multiscale methodology based on scale decomposition is suggested and challenges are discussed.« less

  11. Nonlocal reactive transport with physical and chemical heterogeneity: Localization errors

    SciTech Connect (OSTI)

    Cushman, J.H.; Hu, B.X.; Deng, F.W.

    1995-09-01

    The origin of nonlocality in {open_quotes}macroscale{close_quotes} models for subsurface chemical transport is illustrated. It is argued that media that are either nonperiodic (e.g., media with evolving heterogeneity) or periodic viewed on a scale wherein a unit cell is discernible must display some nonlocality in the mean. A metaphysical argument suggests that owing to the scarcity of information on natural scales of heterogeneity and on scales of observation associated with an instrument window, constitutive theories for the mean concentration should at the outset of any modeling effort always be considered nonlocal. The intuitive appeal to nonlocality is reinforced with an analytical derivation of the constitutive theory for a conservative tracer without appeal to any mathematical approximations. Comparisons are made between the fully nonlocal (FNL), nonlocal in time (NLT), and fully localized (FL) theories. For conservative transport, there is little difference between the first-order FL and FNL models for spatial moments up to and including the third. However, for conservative transport the first-order NLT model differs significantly from the FNL model in the third spatial moments. For reactive transport, all spatial moments differ between the FNL and FL models. The second transverse-horizontal and third longitudinal-horizontal moments for the NLT model differ from the FNL model. These results suggest that localized first-order transport models for conservative tracers are reasonable if only lower-order moments are desired. However, when the chemical reacts with its environment, the localization approximation can lead to significant error in all moments, and a FNL model will in general be required for accurate simulation. 18 refs., 9 figs., 1 tab.

  12. Method and apparatus for detecting timing errors in a system oscillator

    DOE Patents [OSTI]

    Gliebe, Ronald J.; Kramer, William R.

    1993-01-01

    A method of detecting timing errors in a system oscillator for an electronic device, such as a power supply, includes the step of comparing a system oscillator signal with a delayed generated signal and generating a signal representative of the timing error when the system oscillator signal is not identical to the delayed signal. An LED indicates to an operator that a timing error has occurred. A hardware circuit implements the above-identified method.

  13. Estimation of organic carbon blank values and error structures of the speciation trends network data for source apportionment

    SciTech Connect (OSTI)

    Eugene Kim; Philip K. Hopke; Youjun Qin

    2005-08-01

    Because the particulate organic carbon (OC) concentrations reported in U.S. Environment Protection Agency Speciation Trends Network (STN) data were not blank corrected, the OC blank concentrations were estimated using the intercept in particulate matter {lt} 2.5 {mu}m in aerodynamic diameter (PM2.5) regression against OC concentrations. The estimated OC blank concentrations ranged from 1 to 2.4 {mu}g/m{sup 3} showing higher values in urban areas for the 13 monitoring sites in the northeastern United States. In the STN data, several different samplers and analyzers are used, and various instruments show different method detection limit (MDL) values, as well as errors. A comprehensive set of error structures that would be used for numerous source apportionment studies of STN data was estimated by comparing a limited set of measured concentrations and their associated uncertainties. To examine the estimated error structures and investigate the appropriate MDL values, PM2.5 samples collected at a STN site in Burlington, VT, were analyzed through the application of the positive matrix factorization. A total of 323 samples that were collected between December 2000 and December 2003 and 49 species based on several variable selection criteria were used, and eight sources were successfully identified in this study with the estimated error structures and min values among different MDL values from the five instruments: secondary sulfate aerosol (41%) identified as the result of emissions from coal-fired power plants, secondary nitrate aerosol (20%), airborne soil (15%), gasoline vehicle emissions (7%), diesel emissions (7%), aged sea salt (4%), copper smelting (3%), and ferrous smelting (2%). Time series plots of contributions from airborne soil indicate that the highly elevated impacts from this source were likely caused primarily by dust storms.

  14. A global conformance quality model. A new strategic tool for minimizing defects caused by variation, error, and complexity

    SciTech Connect (OSTI)

    Hinckley, C.M.

    1994-01-01

    The performance of Japanese products in the marketplace points to the dominant role of quality in product competition. Our focus is motivated by the tremendous pressure to improve conformance quality by reducing defects to previously unimaginable limits in the range of 1 to 10 parts per million. Toward this end, we have developed a new model of conformance quality that addresses each of the three principle defect sources: (1) Variation, (2) Human Error, and (3) Complexity. Although the role of variation in conformance quality is well documented, errors occur so infrequently that their significance is not well known. We have shown that statistical methods are not useful in characterizing and controlling errors, the most common source of defects. Excessive complexity is also a root source of defects, since it increases errors and variation defects. A missing link in the defining a global model has been the lack of a sound correlation between complexity and defects. We have used Design for Assembly (DFA) methods to quantify assembly complexity and have shown that assembly times can be described in terms of the Pareto distribution in a clear exception to the Central Limit Theorem. Within individual companies we have found defects to be highly correlated with DFA measures of complexity in broad studies covering tens of millions of assembly operations. Applying the global concepts, we predicted that Motorola`s Six Sigma method would only reduce defects by roughly a factor of two rather than orders of magnitude, a prediction confirmed by Motorola`s data. We have also shown that the potential defects rates of product concepts can be compared in the earliest stages of development. The global Conformance Quality Model has demonstrated that the best strategy for improvement depends upon the quality control strengths and weaknesses.

  15. V-172: ISC BIND RUNTIME_CHECK Error Lets Remote Users Deny Service Against Recursive Resolvers

    Broader source: Energy.gov [DOE]

    A defect exists which allows an attacker to crash a BIND 9 recursive resolver with a RUNTIME_CHECK error in resolver.c

  16. Table 4b. Relative Standard Errors for Total Fuel Oil Consumption...

    Gasoline and Diesel Fuel Update (EIA)

    4b. Relative Standard Errors for Total Fuel Oil Consumption per Effective Occupied Square Foot, 1992 Building Characteristics All Buildings Using Fuel Oil (thousand) Total Fuel Oil...

  17. Absolute beam emittance measurements at RHIC using ionization profile monitors

    SciTech Connect (OSTI)

    Minty, M.; Connolly, R; Liu, C.; Summers, T.; Tepikian, S.

    2014-08-15

    In the past, comparisons between emittance measurements obtained using ionization profile monitors, Vernier scans (using as input the measured rates from the zero degree counters, or ZDCs), the polarimeters and the Schottky detectors evidenced significant variations of up to 100%. In this report we present studies of the RHIC ionization profile monitors (IPMs). After identifying and correcting for two systematic instrumental errors in the beam size measurements, we present experimental results showing that the remaining dominant error in beam emittance measurements at RHIC using the IPMs was imprecise knowledge of the local beta functions. After removal of the systematic errors and implementation of measured beta functions, precise emittance measurements result. Also, consistency between the emittances measured by the IPMs and those derived from the ZDCs was demonstrated.

  18. Compilation error with cray-petsc/3.6.1.0

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Compilation error with cray-petsc3.6.1.0 Compilation error with cray-petsc3.6.1.0 January 5, 2016 The current default cray-petsc module, cray-petsc3.6.1.0, does not work with...

  19. The cce/8.3.0 C++ compiler may run into a linking error on Edison

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    The cce8.3.0 C++ compiler may run into a linking error on Edison The cce8.3.0 C++ compiler may run into a linking error on Edison July 1, 2014 You may run into the following...

  20. Potential Hydraulic Modelling Errors Associated with Rheological Data Extrapolation in Laminar Flow

    SciTech Connect (OSTI)

    Shadday, Martin A., Jr.

    1997-03-20

    The potential errors associated with the modelling of flows of non-Newtonian slurries through pipes, due to inadequate rheological models and extrapolation outside of the ranges of data bases, are demonstrated. The behaviors of both dilatant and pseudoplastic fluids with yield stresses, and the errors associated with treating them as Bingham plastics, are investigated.

  1. A Case for Soft Error Detection and Correction in Computational Chemistry

    SciTech Connect (OSTI)

    van Dam, Hubertus JJ; Vishnu, Abhinav; De Jong, Wibe A.

    2013-09-10

    High performance computing platforms are expected to deliver 10(18) floating operations per second by the year 2022 through the deployment of millions of cores. Even if every core is highly reliable the sheer number of the them will mean that the mean time between failures will become so short that most applications runs will suffer at least one fault. In particular soft errors caused by intermittent incorrect behavior of the hardware are a concern as they lead to silent data corruption. In this paper we investigate the impact of soft errors on optimization algorithms using Hartree-Fock as a particular example. Optimization algorithms iteratively reduce the error in the initial guess to reach the intended solution. Therefore they may intuitively appear to be resilient to soft errors. Our results show that this is true for soft errors of small magnitudes but not for large errors. We suggest error detection and correction mechanisms for different classes of data structures. The results obtained with these mechanisms indicate that we can correct more than 95% of the soft errors at moderate increases in the computational cost.

  2. Measurement of fenestration net energy performance: Considerations leading to development of a Mobile Window Thermal Test (MoWitt) facility

    SciTech Connect (OSTI)

    Klems, J.H.

    1988-08-01

    The authors present a detailed consideration of the energy flows entering a building space and the effect of random measurement errors on determining fenestration performance. Estimates of error magnitudes are made for a passive test cell; they show that a more accurate test facility is needed for reliable measurements on fenestration systems with thermal resistance 2-10 times that of single glazing or with shading coefficients less than 0.7. A test facility of this type, built at Lawrence Berkeley Laboratory, is described. The effect of random errors in this facility is discussed and computer calculations of its performance are presented. The discussion shows that, for any measurement facility, random errors are most serious in nighttime measurements, and systematic errors are most important in daytime measurements. It is concluded that, for this facility, errors from both sources should be small.

  3. Methods and apparatus using commutative error detection values for fault isolation in multiple node computers

    DOE Patents [OSTI]

    Almasi, Gheorghe [Ardsley, NY; Blumrich, Matthias Augustin [Ridgefield, CT; Chen, Dong [Croton-On-Hudson, NY; Coteus, Paul [Yorktown, NY; Gara, Alan [Mount Kisco, NY; Giampapa, Mark E. [Irvington, NY; Heidelberger, Philip [Cortlandt Manor, NY; Hoenicke, Dirk I. [Ossining, NY; Singh, Sarabjeet [Mississauga, CA; Steinmacher-Burow, Burkhard D. [Wernau, DE; Takken, Todd [Brewster, NY; Vranas, Pavlos [Bedford Hills, NY

    2008-06-03

    Methods and apparatus perform fault isolation in multiple node computing systems using commutative error detection values for--example, checksums--to identify and to isolate faulty nodes. When information associated with a reproducible portion of a computer program is injected into a network by a node, a commutative error detection value is calculated. At intervals, node fault detection apparatus associated with the multiple node computer system retrieve commutative error detection values associated with the node and stores them in memory. When the computer program is executed again by the multiple node computer system, new commutative error detection values are created and stored in memory. The node fault detection apparatus identifies faulty nodes by comparing commutative error detection values associated with reproducible portions of the application program generated by a particular node from different runs of the application program. Differences in values indicate a possible faulty node.

  4. Summary of QM02 Measurements

    SciTech Connect (OSTI)

    Fisher, Andrew

    2010-12-03

    This note summarizes both the beam-based and various laboratory measurements of quadrupole magnets, units 387 and 428, used for QM02 in the LCLS Injector. These were undertaken because of a consistent discrepancy between accelerator model predictions and beam observations which seemed to indicate a weak QM02. A report 'QM02 Strength Measurement', by Welch and Wu, describes the discrepancy and beam-based measurements on unit 387. Subsequently, unit 387 was replaced by unit 428, refinements were made to analysis of the beam-based measurements were made, and additional magnetic measurements were made on unit 387 in the lab. These new results are summarized in this note. The principle results are: (1) Laboratory measurements of integrated gradient for the same magnet, or for different magnets of the same type, are all within 1% of each other at gradients of interest. These cases cover three independent types of measurements, disassembly/reassembly of the units, and extended periods of time between measurements. (2) Standardization, or lack thereof, can cause integrated gradient errors of approximately 0.2 kG, which can amount to a few percent of the strength of the magnet depending on the setting. (3) Model-independent beam-based measurements indicate the magnets are actually weaker than expected by about 2 percent, but these measurements are subject to the uncertainty of the BPMS1 location. (4) The standardization cycle is effective. (5) The stainless steel BPM vacuum chamber inside the magnets has no significant effect on the beam. The discrepancy between the accelerator model predictions and the actual orbit response is not resolved, but the evidence points away from magnetic strength errors as the source. Differences between the model locations and effective locations of BPM's is a possible culprit. This idea is explored in Section 2. Table 1 summarizes the measurement activities that were performed on the units. The dates listed correspond to the date when the

  5. Atmospheric Dispersion Effects in Weak Lensing Measurements

    SciTech Connect (OSTI)

    Plazas, Andrés Alejandro; Bernstein, Gary

    2012-10-01

    The wavelength dependence of atmospheric refraction causes elongation of finite-bandwidth images along the elevation vector, which produces spurious signals in weak gravitational lensing shear measurements unless this atmospheric dispersion is calibrated and removed to high precision. Because astrometric solutions and PSF characteristics are typically calibrated from stellar images, differences between the reference stars' spectra and the galaxies' spectra will leave residual errors in both the astrometric positions (dr) and in the second moment (width) of the wavelength-averaged PSF (dv) for galaxies.We estimate the level of dv that will induce spurious weak lensing signals in PSF-corrected galaxy shapes that exceed the statistical errors of the DES and the LSST cosmic-shear experiments. We also estimate the dr signals that will produce unacceptable spurious distortions after stacking of exposures taken at different airmasses and hour angles. We also calculate the errors in the griz bands, and find that dispersion systematics, uncorrected, are up to 6 and 2 times larger in g and r bands,respectively, than the requirements for the DES error budget, but can be safely ignored in i and z bands. For the LSST requirements, the factors are about 30, 10, and 3 in g, r, and i bands,respectively. We find that a simple correction linear in galaxy color is accurate enough to reduce dispersion shear systematics to insignificant levels in the r band for DES and i band for LSST,but still as much as 5 times than the requirements for LSST r-band observations. More complex corrections will likely be able to reduce the systematic cosmic-shear errors below statistical errors for LSST r band. But g-band effects remain large enough that it seems likely that induced systematics will dominate the statistical errors of both surveys, and cosmic-shear measurements should rely on the redder bands.

  6. Atmospheric Dispersion Effects in Weak Lensing Measurements

    DOE Public Access Gateway for Energy & Science Beta (PAGES Beta)

    Plazas, Andrés Alejandro; Bernstein, Gary

    2012-10-01

    The wavelength dependence of atmospheric refraction causes elongation of finite-bandwidth images along the elevation vector, which produces spurious signals in weak gravitational lensing shear measurements unless this atmospheric dispersion is calibrated and removed to high precision. Because astrometric solutions and PSF characteristics are typically calibrated from stellar images, differences between the reference stars' spectra and the galaxies' spectra will leave residual errors in both the astrometric positions (dr) and in the second moment (width) of the wavelength-averaged PSF (dv) for galaxies.We estimate the level of dv that will induce spurious weak lensing signals in PSF-corrected galaxy shapes that exceed themore » statistical errors of the DES and the LSST cosmic-shear experiments. We also estimate the dr signals that will produce unacceptable spurious distortions after stacking of exposures taken at different airmasses and hour angles. We also calculate the errors in the griz bands, and find that dispersion systematics, uncorrected, are up to 6 and 2 times larger in g and r bands,respectively, than the requirements for the DES error budget, but can be safely ignored in i and z bands. For the LSST requirements, the factors are about 30, 10, and 3 in g, r, and i bands,respectively. We find that a simple correction linear in galaxy color is accurate enough to reduce dispersion shear systematics to insignificant levels in the r band for DES and i band for LSST,but still as much as 5 times than the requirements for LSST r-band observations. More complex corrections will likely be able to reduce the systematic cosmic-shear errors below statistical errors for LSST r band. But g-band effects remain large enough that it seems likely that induced systematics will dominate the statistical errors of both surveys, and cosmic-shear measurements should rely on the redder bands.« less

  7. Modern Palliative Radiation Treatment: Do Complexity and Workload Contribute to Medical Errors?

    SciTech Connect (OSTI)

    D'Souza, Neil; Odette Cancer Centre, Sunnybrook Health Sciences Centre, Toronto, Ontario ; Holden, Lori; Odette Cancer Centre, Sunnybrook Health Sciences Centre, Toronto, Ontario ; Robson, Sheila; Mah, Kathy; Di Prospero, Lisa; Wong, C. Shun; Chow, Edward; Spayne, Jacqueline; Odette Cancer Centre, Sunnybrook Health Sciences Centre, Toronto, Ontario

    2012-09-01

    Purpose: To examine whether treatment workload and complexity associated with palliative radiation therapy contribute to medical errors. Methods and Materials: In the setting of a large academic health sciences center, patient scheduling and record and verification systems were used to identify patients starting radiation therapy. All records of radiation treatment courses delivered during a 3-month period were retrieved and divided into radical and palliative intent. 'Same day consultation, planning and treatment' was used as a proxy for workload and 'previous treatment' and 'multiple sites' as surrogates for complexity. In addition, all planning and treatment discrepancies (errors and 'near-misses') recorded during the same time frame were reviewed and analyzed. Results: There were 365 new patients treated with 485 courses of palliative radiation therapy. Of those patients, 128 (35%) were same-day consultation, simulation, and treatment patients; 166 (45%) patients had previous treatment; and 94 (26%) patients had treatment to multiple sites. Four near-misses and 4 errors occurred during the audit period, giving an error per course rate of 0.82%. In comparison, there were 10 near-misses and 5 errors associated with 1100 courses of radical treatment during the audit period. This translated into an error rate of 0.45% per course. An association was found between workload and complexity and increased palliative therapy error rates. Conclusions: Increased complexity and workload may have an impact on palliative radiation treatment discrepancies. This information may help guide the necessary recommendations for process improvement for patients who require palliative radiation therapy.

  8. Fractional charge and spin errors in self-consistent Greens function theory

    SciTech Connect (OSTI)

    Phillips, Jordan J. Kananenka, Alexei A.; Zgid, Dominika

    2015-05-21

    We examine fractional charge and spin errors in self-consistent Greens function theory within a second-order approximation (GF2). For GF2, it is known that the summation of diagrams resulting from the self-consistent solution of the Dyson equation removes the divergences pathological to second-order Mller-Plesset (MP2) theory for strong correlations. In the language often used in density functional theory contexts, this means GF2 has a greatly reduced fractional spin error relative to MP2. The natural question then is what effect, if any, does the Dyson summation have on the fractional charge error in GF2? To this end, we generalize our previous implementation of GF2 to open-shell systems and analyze its fractional spin and charge errors. We find that like MP2, GF2 possesses only a very small fractional charge error, and consequently minimal many electron self-interaction error. This shows that GF2 improves on the critical failings of MP2, but without altering the positive features that make it desirable. Furthermore, we find that GF2 has both less fractional charge and fractional spin errors than typical hybrid density functionals as well as random phase approximation with exchange.

  9. ADEPT, a dynamic next generation sequencing data error-detection program with trimming

    DOE Public Access Gateway for Energy & Science Beta (PAGES Beta)

    Feng, Shihai; Lo, Chien-Chi; Li, Po-E; Chain, Patrick S. G.

    2016-02-29

    Illumina is the most widely used next generation sequencing technology and produces millions of short reads that contain errors. These sequencing errors constitute a major problem in applications such as de novo genome assembly, metagenomics analysis and single nucleotide polymorphism discovery. In this study, we present ADEPT, a dynamic error detection method, based on the quality scores of each nucleotide and its neighboring nucleotides, together with their positions within the read and compares this to the position-specific quality score distribution of all bases within the sequencing run. This method greatly improves upon other available methods in terms of the truemore » positive rate of error discovery without affecting the false positive rate, particularly within the middle of reads. We conclude that ADEPT is the only tool to date that dynamically assesses errors within reads by comparing position-specific and neighboring base quality scores with the distribution of quality scores for the dataset being analyzed. The result is a method that is less prone to position-dependent under-prediction, which is one of the most prominent issues in error prediction. The outcome is that ADEPT improves upon prior efforts in identifying true errors, primarily within the middle of reads, while reducing the false positive rate.« less

  10. How Radiation Oncologists Would Disclose Errors: Results of a Survey of Radiation Oncologists and Trainees

    SciTech Connect (OSTI)

    Evans, Suzanne B.; Yu, James B.; Chagpar, Anees

    2012-10-01

    Purpose: To analyze error disclosure attitudes of radiation oncologists and to correlate error disclosure beliefs with survey-assessed disclosure behavior. Methods and Materials: With institutional review board exemption, an anonymous online survey was devised. An email invitation was sent to radiation oncologists (American Society for Radiation Oncology [ASTRO] gold medal winners, program directors and chair persons of academic institutions, and former ASTRO lecturers) and residents. A disclosure score was calculated based on the number or full, partial, or no disclosure responses chosen to the vignette-based questions, and correlation was attempted with attitudes toward error disclosure. Results: The survey received 176 responses: 94.8% of respondents considered themselves more likely to disclose in the setting of a serious medical error; 72.7% of respondents did not feel it mattered who was responsible for the error in deciding to disclose, and 3.9% felt more likely to disclose if someone else was responsible; 38.0% of respondents felt that disclosure increased the likelihood of a lawsuit, and 32.4% felt disclosure decreased the likelihood of lawsuit; 71.6% of respondents felt near misses should not be disclosed; 51.7% thought that minor errors should not be disclosed; 64.7% viewed disclosure as an opportunity for forgiveness from the patient; and 44.6% considered the patient's level of confidence in them to be a factor in disclosure. For a scenario that could be considerable, a non-harmful error, 78.9% of respondents would not contact the family. Respondents with high disclosure scores were more likely to feel that disclosure was an opportunity for forgiveness (P=.003) and to have never seen major medical errors (P=.004). Conclusions: The surveyed radiation oncologists chose to respond with full disclosure at a high rate, although ideal disclosure practices were not uniformly adhered to beyond the initial decision to disclose the occurrence of the error.

  11. CORRELATED AND ZONAL ERRORS OF GLOBAL ASTROMETRIC MISSIONS: A SPHERICAL HARMONIC SOLUTION

    SciTech Connect (OSTI)

    Makarov, V. V.; Dorland, B. N.; Gaume, R. A.; Hennessy, G. S.; Berghea, C. T.; Dudik, R. P.; Schmitt, H. R.

    2012-07-15

    We propose a computer-efficient and accurate method of estimating spatially correlated errors in astrometric positions, parallaxes, and proper motions obtained by space- and ground-based astrometry missions. In our method, the simulated observational equations are set up and solved for the coefficients of scalar and vector spherical harmonics representing the output errors rather than for individual objects in the output catalog. Both accidental and systematic correlated errors of astrometric parameters can be accurately estimated. The method is demonstrated on the example of the JMAPS mission, but can be used for other projects in space astrometry, such as SIM or JASMINE.

  12. Error propagation equations for estimating the uncertainty in high-speed wind tunnel test results

    SciTech Connect (OSTI)

    Clark, E.L.

    1994-07-01

    Error propagation equations, based on the Taylor series model, are derived for the nondimensional ratios and coefficients most often encountered in high-speed wind tunnel testing. These include pressure ratio and coefficient, static force and moment coefficients, dynamic stability coefficients, and calibration Mach number. The error equations contain partial derivatives, denoted as sensitivity coefficients, which define the influence of free-steam Mach number, M{infinity}, on various aerodynamic ratios. To facilitate use of the error equations, sensitivity coefficients are derived and evaluated for five fundamental aerodynamic ratios which relate free-steam test conditions to a reference condition.

  13. T-545: RealPlayer Heap Corruption Error in 'vidplin.dll' Lets Remote Users

    Office of Energy Efficiency and Renewable Energy (EERE) Indexed Site

    Execute Arbitrary Code | Department of Energy 5: RealPlayer Heap Corruption Error in 'vidplin.dll' Lets Remote Users Execute Arbitrary Code T-545: RealPlayer Heap Corruption Error in 'vidplin.dll' Lets Remote Users Execute Arbitrary Code January 28, 2011 - 7:21am Addthis PROBLEM: RealPlayer Heap Corruption Error in 'vidplin.dll' Lets Remote Users Execute Arbitrary Code. PLATFORM: RealPlayer 14.0.1 and prior versions ABSTRACT: A vulnerability was reported in RealPlayer. A remote user can

  14. Measurement uncertainty analysis techniques applied to PV performance measurements

    SciTech Connect (OSTI)

    Wells, C.

    1992-10-01

    The purpose of this presentation is to provide a brief introduction to measurement uncertainty analysis, outline how it is done, and illustrate uncertainty analysis with examples drawn from the PV field, with particular emphasis toward its use in PV performance measurements. The uncertainty information we know and state concerning a PV performance measurement or a module test result determines, to a significant extent, the value and quality of that result. What is measurement uncertainty analysis It is an outgrowth of what has commonly been called error analysis. But uncertainty analysis, a more recent development, gives greater insight into measurement processes and tests, experiments, or calibration results. Uncertainty analysis gives us an estimate of the I interval about a measured value or an experiment's final result within which we believe the true value of that quantity will lie. Why should we take the time to perform an uncertainty analysis A rigorous measurement uncertainty analysis: Increases the credibility and value of research results; allows comparisons of results from different labs; helps improve experiment design and identifies where changes are needed to achieve stated objectives (through use of the pre-test analysis); plays a significant role in validating measurements and experimental results, and in demonstrating (through the post-test analysis) that valid data have been acquired; reduces the risk of making erroneous decisions; demonstrates quality assurance and quality control measures have been accomplished; define Valid Data as data having known and documented paths of: Origin, including theory; measurements; traceability to measurement standards; computations; uncertainty analysis of results.

  15. Evaluating IMRT and VMAT dose accuracy: Practical examples of failure to detect systematic errors when applying a commonly used metric and action levels

    SciTech Connect (OSTI)

    Nelms, Benjamin E.; Chan, Maria F.; Jarry, Genevive; Lemire, Matthieu; Lowden, John; Hampton, Carnell

    2013-11-15

    Purpose: This study (1) examines a variety of real-world cases where systematic errors were not detected by widely accepted methods for IMRT/VMAT dosimetric accuracy evaluation, and (2) drills-down to identify failure modes and their corresponding means for detection, diagnosis, and mitigation. The primary goal of detailing these case studies is to explore different, more sensitive methods and metrics that could be used more effectively for evaluating accuracy of dose algorithms, delivery systems, and QA devices.Methods: The authors present seven real-world case studies representing a variety of combinations of the treatment planning system (TPS), linac, delivery modality, and systematic error type. These case studies are typical to what might be used as part of an IMRT or VMAT commissioning test suite, varying in complexity. Each case study is analyzed according to TG-119 instructions for gamma passing rates and action levels for per-beam and/or composite plan dosimetric QA. Then, each case study is analyzed in-depth with advanced diagnostic methods (dose profile examination, EPID-based measurements, dose difference pattern analysis, 3D measurement-guided dose reconstruction, and dose grid inspection) and more sensitive metrics (2% local normalization/2 mm DTA and estimated DVH comparisons).Results: For these case studies, the conventional 3%/3 mm gamma passing rates exceeded 99% for IMRT per-beam analyses and ranged from 93.9% to 100% for composite plan dose analysis, well above the TG-119 action levels of 90% and 88%, respectively. However, all cases had systematic errors that were detected only by using advanced diagnostic techniques and more sensitive metrics. The systematic errors caused variable but noteworthy impact, including estimated target dose coverage loss of up to 5.5% and local dose deviations up to 31.5%. Types of errors included TPS model settings, algorithm limitations, and modeling and alignment of QA phantoms in the TPS. Most of the errors were

  16. Measurement of the orbit fluctuation caused by an insertion device with the amplitude modulation method

    SciTech Connect (OSTI)

    Nakatani, T.; Agui, A.; Yoshigoe, A.; Matsushita, T.; Takao, M.; Aoyagi, H.; Takeuchi, M.; Tanaka, H.

    2004-05-12

    We have developed a new method to extract only the orbit fluctuation caused by changing magnetic field error of an insertion device (ID). This method consists of two main parts. (i) The orbit fluctuation is measured with modulating the error field of the ID by using the real-time beam position measuring system. (ii) The orbit fluctuation depending on the variation of the error field of the ID is extracted by the filter applying the Wavelet Transform. We call this approach the amplitude modulation method. This analysis technique was applied to measure the orbit fluctuation caused by the error field of APPLE-2 type undulator (ID23) installed in the SPring-8 storage ring. We quantitatively measured two kinds of the orbit fluctuation which are the static term caused by the magnetic field error and the dynamic term caused by the eddy current on the ID23 chamber.

  17. Enabling Predictive Simulation and UQ of Complex Multiphysics PDE Systems by the Development of Goal-Oriented Variational Sensitivity Analysis and A Posteriori Error Estimation Methods

    SciTech Connect (OSTI)

    Ginting, Victor

    2014-03-15

    it was demonstrated that a posteriori analyses in general and in particular one that uses adjoint methods can accurately and efficiently compute numerical error estimates and sensitivity for critical Quantities of Interest (QoIs) that depend on a large number of parameters. Activities include: analysis and implementation of several time integration techniques for solving system of ODEs as typically obtained from spatial discretization of PDE systems; multirate integration methods for ordinary differential equations; formulation and analysis of an iterative multi-discretization Galerkin finite element method for multi-scale reaction-diffusion equations; investigation of an inexpensive postprocessing technique to estimate the error of finite element solution of the second-order quasi-linear elliptic problems measured in some global metrics; investigation of an application of the residual-based a posteriori error estimates to symmetric interior penalty discontinuous Galerkin method for solving a class of second order quasi-linear elliptic problems; a posteriori analysis of explicit time integrations for system of linear ordinary differential equations; derivation of accurate a posteriori goal oriented error estimates for a user-defined quantity of interest for two classes of first and second order IMEX schemes for advection-diffusion-reaction problems; Postprocessing finite element solution; and A Bayesian Framework for Uncertain Quantification of Porous Media Flows.

  18. V-228: RealPlayer Buffer Overflow and Memory Corruption Error...

    Office of Energy Efficiency and Renewable Energy (EERE) Indexed Site

    V-228: RealPlayer Buffer Overflow and Memory Corruption Error Let Remote Users Execute ... Lets Remote Users Execute Arbitrary Code V-049: RealPlayer Buffer Overflow and Invalid ...

  19. "Show Preview" button is not working; gives error | OpenEI Community

    Open Energy Info (EERE)

    "Show Preview" button is not working; gives error Home > Groups > Utility Rate Submitted by Ewilson on 3 January, 2013 - 09:52 1 answer Points: 0 Eric, thanks for reporting this. I...

  20. A Compact Code for Simulations of Quantum Error Correction in Classical Computers

    SciTech Connect (OSTI)

    Nyman, Peter

    2009-03-10

    This study considers implementations of error correction in a simulation language on a classical computer. Error correction will be necessarily in quantum computing and quantum information. We will give some examples of the implementations of some error correction codes. These implementations will be made in a more general quantum simulation language on a classical computer in the language Mathematica. The intention of this research is to develop a programming language that is able to make simulations of all quantum algorithms and error corrections in the same framework. The program code implemented on a classical computer will provide a connection between the mathematical formulation of quantum mechanics and computational methods. This gives us a clear uncomplicated language for the implementations of algorithms.

  1. V-109: Google Chrome WebKit Type Confusion Error Lets Remote...

    Office of Energy Efficiency and Renewable Energy (EERE) Indexed Site

    Google Chrome WebKit Type Confusion Error Lets Remote Users Execute Arbitrary Code PLATFORM: Google Chrome prior to 25.0.1364.160 ABSTRACT: A vulnerability was reported in...

  2. Types of Possible Survey Errors in Estimates Published in the Weekly Natural Gas Storage Report

    Weekly Natural Gas Storage Report (EIA)

    U.S. Energy Information Administration | Types of Possible Survey Errors in Estimates Published in the Weekly Natural Gas Storage Report 1 February 2016 Types of Possible Survey Errors in Estimates Published in the Weekly Natural Gas Storage Report The U.S. Energy Information Administration (EIA) collects and publishes natural gas storage information on a monthly and weekly basis. The Form EIA-191, Monthly Underground Natural Gas Storage Report, is a census survey that collects field-level

  3. V-235: Cisco Mobility Services Engine Configuration Error Lets Remote Users

    Office of Energy Efficiency and Renewable Energy (EERE) Indexed Site

    Login Anonymously | Department of Energy 5: Cisco Mobility Services Engine Configuration Error Lets Remote Users Login Anonymously V-235: Cisco Mobility Services Engine Configuration Error Lets Remote Users Login Anonymously September 5, 2013 - 12:33am Addthis PROBLEM: A vulnerability was reported in Cisco Mobility Services Engine. A remote user can login anonymously. PLATFORM: Cisco Mobility Services Engine ABSTRACT: A vulnerability in Cisco Mobility Services Engine could allow an

  4. Solution-verified reliability analysis and design of bistable MEMS using error estimation and adaptivity.

    SciTech Connect (OSTI)

    Eldred, Michael Scott; Subia, Samuel Ramirez; Neckels, David; Hopkins, Matthew Morgan; Notz, Patrick K.; Adams, Brian M.; Carnes, Brian; Wittwer, Jonathan W.; Bichon, Barron J.; Copps, Kevin D.

    2006-10-01

    This report documents the results for an FY06 ASC Algorithms Level 2 milestone combining error estimation and adaptivity, uncertainty quantification, and probabilistic design capabilities applied to the analysis and design of bistable MEMS. Through the use of error estimation and adaptive mesh refinement, solution verification can be performed in an automated and parameter-adaptive manner. The resulting uncertainty analysis and probabilistic design studies are shown to be more accurate, efficient, reliable, and convenient.

  5. ARM - Measurement -

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    govMeasurements ARM Data Discovery Browse Data Comments? We would love to hear from you! Send us a note below or call us at 1-888-ARM-DATA. Send Measurement : Categories Instruments The above measurement is considered scientifically relevant for the following instruments. Refer to the datastream (netcdf) file headers of each instrument for a list of all available measurements, including those recorded for diagnostic or quality assurance purposes. Field Campaign Instruments UV-MFRSR : Ultraviolet

  6. Measurement of $\

    SciTech Connect (OSTI)

    Acciarri, R.; et al.

    2015-11-03

    The ArgoNeuT collaboration reports the first measurement of neutral current $\\pi^{0}$ production in $\

  7. Measuring circuit

    DOE Patents [OSTI]

    Sun, Shan C.; Chaprnka, Anthony G.

    1977-01-11

    An automatic gain control circuit functions to adjust the magnitude of an input signal supplied to a measuring circuit to a level within the dynamic range of the measuring circuit while a log-ratio circuit adjusts the magnitude of the output signal from the measuring circuit to the level of the input signal and optimizes the signal-to-noise ratio performance of the measuring circuit.

  8. Application of ISO-TAG4 to the reporting of limit of error on the inventory difference

    SciTech Connect (OSTI)

    Murdock, C.; Suda, S.

    1993-07-01

    A standard reference does not exist for evaluating and expressing systematic and random uncertainty, thus, there is no basis for comparing measurement uncertainties at different facilities. Based on recommendations of the International Committee for Weights and Measures, the National Center for Standards and Certification Information, which is responsible for information on standardization programs and related activities, has published ISO-TAG4, Guide to the Expression of Uncertainty in Measurement (1993). The guide establishes general rules for evaluating and expressing uncertainty in physical measurements by presenting definitions, basic concepts and examples. it focuses on the methods of evaluating uncertainty components rather than categorizing the components, thus avoiding the ambiguity encountered when categorizing uncertainty components as ``random`` and ``systematic.`` This paper presents an overview of the terms specific to the guide, including standard and combined standard uncertainty, Type A and Type B evaluation, expanded uncertainty, and coverage factor. It illustrates Type A and Type B evaluation of random and systematic errors in forms relating to nuclear material accountability work. This guide could be adapted by the MC&A community.

  9. Procedures for using expert judgment to estimate human-error probabilities in nuclear power plant operations. [PWR; BWR

    SciTech Connect (OSTI)

    Seaver, D.A.; Stillwell, W.G.

    1983-03-01

    This report describes and evaluates several procedures for using expert judgment to estimate human-error probabilities (HEPs) in nuclear power plant operations. These HEPs are currently needed for several purposes, particularly for probabilistic risk assessments. Data do not exist for estimating these HEPs, so expert judgment can provide these estimates in a timely manner. Five judgmental procedures are described here: paired comparisons, ranking and rating, direct numerical estimation, indirect numerical estimation and multiattribute utility measurement. These procedures are evaluated in terms of several criteria: quality of judgments, difficulty of data collection, empirical support, acceptability, theoretical justification, and data processing. Situational constraints such as the number of experts available, the number of HEPs to be estimated, the time available, the location of the experts, and the resources available are discussed in regard to their implications for selecting a procedure for use.

  10. SU-E-T-377: Inaccurate Positioning Might Introduce Significant MapCheck Calibration Error in Flatten Filter Free Beams

    SciTech Connect (OSTI)

    Wang, S; Chao, C; Chang, J

    2014-06-01

    Purpose: This study investigates the calibration error of detector sensitivity for MapCheck due to inaccurate positioning of the device, which is not taken into account by the current commercial iterative calibration algorithm. We hypothesize the calibration is more vulnerable to the positioning error for the flatten filter free (FFF) beams than the conventional flatten filter flattened beams. Methods: MapCheck2 was calibrated with 10MV conventional and FFF beams, with careful alignment and with 1cm positioning error during calibration, respectively. Open fields of 37cmx37cm were delivered to gauge the impact of resultant calibration errors. The local calibration error was modeled as a detector independent multiplication factor, with which propagation error was estimated with positioning error from 1mm to 1cm. The calibrated sensitivities, without positioning error, were compared between the conventional and FFF beams to evaluate the dependence on the beam type. Results: The 1cm positioning error leads to 0.39% and 5.24% local calibration error in the conventional and FFF beams respectively. After propagating to the edges of MapCheck, the calibration errors become 6.5% and 57.7%, respectively. The propagation error increases almost linearly with respect to the positioning error. The difference of sensitivities between the conventional and FFF beams was small (0.11 ± 0.49%). Conclusion: The results demonstrate that the positioning error is not handled by the current commercial calibration algorithm of MapCheck. Particularly, the calibration errors for the FFF beams are ~9 times greater than those for the conventional beams with identical positioning error, and a small 1mm positioning error might lead to up to 8% calibration error. Since the sensitivities are only slightly dependent of the beam type and the conventional beam is less affected by the positioning error, it is advisable to cross-check the sensitivities between the conventional and FFF beams to detect