Broader source: Energy.gov [DOE]
This section provides guidelines on the application and positioning of the core elements to EERE exhibit graphics. The following core elements must be implemented consistently on all exhibit graphics.
Graphics | Department of Energy
Office of Energy Efficiency and Renewable Energy (EERE) Indexed Site
Graphics Graphics The Visual Media Department consists of a team of graphic artists, photographers and visual librarians working together to provide a wide variety of visual media for the Department of Energy. Graphics Offices Forrestal: Room BH-071, (202) 586-2732 Germantown: Room E-065, (301) 903-5165 Graphics@hq.doe.gov Certificates, Plaques and Special Awards Design Presentations Enhance and Print Your Files Exhibits Posters and Banners Brochures, Pamphlets and Handouts If you have need of
Broader source: Energy.gov [DOE]
For EERE websites and applications, follow these requirements and best practices for designing graphics and developing images. This includes making them Section 508-compliant.
Development of computer graphics
Nuttall, H.E.
1989-07-01
The purpose of this project was to screen and evaluate three graphics packages as to their suitability for displaying concentration contour graphs. The information to be displayed is from computer code simulations describing air-born contaminant transport. The three evaluation programs were MONGO (John Tonry, MIT, Cambridge, MA, 02139), Mathematica (Wolfram Research Inc.), and NCSA Image (National Center for Supercomputing Applications at the University of Illinois at Urbana-Champaign). After a preliminary investigation of each package, NCSA Image appeared to be significantly superior for generating the desired concentration contour graphs. Hence subsequent work and this report describes the implementation and testing of NCSA Image on both an Apple MacII and Sun 4 computers. NCSA Image includes several utilities (Layout, DataScope, HDF, and PalEdit) which were used in this study and installed on Dr. Ted Yamada`s Mac II computer. Dr. Yamada provided two sets of air pollution plume data which were displayed using NCSA Image. Both sets were animated into a sequential expanding plume series.
Contig Graph Tool: A graphical interface for Contig Physical Map assembly
Pecherer, R.M.
1992-01-01
A Contig Physical Map of a chromosome is a collection of DNA clones organized into ordered, overlapping sets called contigs which cover contiguous regions of the chromosome. Contigs may be assembled from a knowledge of the binary overlap relation between all clone pairs in a clone set which covers all or part of the chromosome, and contigs may be positioned along the chromosome by in situ hybridization using unique probes from each contig. Clone overlap is determined experimentally by factoring each clone into restriction fragments that are characterized by size and hybridization probe signals. Clones which overlap therefore share common restriction fragments, making overlap detectable. However, non-uniqueness of restriction fragments and experimental error lead to incorrect determination of the overlap relation and errors in the contig map. The Contig Graph Tool was developed to detect and correct overlap errors using editable visualizations of an abstract graph representation for clones and overlaps. This interactive tool is integrated with an electronic, laboratory notebook and introduces several concepts useful for solving problems with discrete, scientific visualization.
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
Photos and graphics MINERVA_assembly_10-0082-08D MINERvA uses a collider detector design and a high-intensity beam to study neutrino reactions with five different nuclei, creating the first self-contained comparison of interactions in different elements. MINERvA's comprehensive research goals attract scientists from around the globe who specialize in nuclear and particle physics. (Credit: Fermilab) MINERvA's comprehensive research goals attract scientists from around the globe who specialize in
Graphical Contingency Analysis Tool
Energy Science and Technology Software Center (OSTI)
2010-03-02
GCA is a visual analytic tool for power grid contingency analysis to provide more decision support for power grid operations. GCA allows power grid operators to quickly gain situational awareness of power grid by converting large amounts of operational data to graphic domain with a color contoured map; identify system trend and foresee and discern emergencies by performing trending analysis; identify the relationships between system configurations and affected assets by conducting clustering analysis; and identifymore » the best action by interactively evaluate candidate actions.« less
Graphical calculus for Gaussian pure states
Menicucci, Nicolas C.; Flammia, Steven T.; Loock, Peter van
2011-04-15
We provide a unified graphical calculus for all Gaussian pure states, including graph transformation rules for all local and semilocal Gaussian unitary operations, as well as local quadrature measurements. We then use this graphical calculus to analyze continuous-variable (CV) cluster states, the essential resource for one-way quantum computing with CV systems. Current graphical approaches to CV cluster states are only valid in the unphysical limit of infinite squeezing, and the associated graph transformation rules only apply when the initial and final states are of this form. Our formalism applies to all Gaussian pure states and subsumes these rules in a natural way. In addition, the term 'CV graph state' currently has several inequivalent definitions in use. Using this formalism we provide a single unifying definition that encompasses all of them. We provide many examples of how the formalism may be used in the context of CV cluster states: defining the 'closest' CV cluster state to a given Gaussian pure state and quantifying the error in the approximation due to finite squeezing; analyzing the optimality of certain methods of generating CV cluster states; drawing connections between this graphical formalism and bosonic Hamiltonians with Gaussian ground states, including those useful for CV one-way quantum computing; and deriving a graphical measure of bipartite entanglement for certain classes of CV cluster states. We mention other possible applications of this formalism and conclude with a brief note on fault tolerance in CV one-way quantum computing.
Software engineering and graphical programming languages
Jefferson, K.; Porter, T.; West, T.
1997-11-01
This report contains viewgraphs on software engineering and adapting engineering processes to a graphical programming languages.
Multi Platform Graphics Subroutine Library
Energy Science and Technology Software Center (OSTI)
1992-02-21
DIGLIB is a collection of general graphics subroutines. It was designed to be small, reasonably fast, device-independent, and compatible with DEC-supplied operating systems for VAXes, PDP-11s, and LSI-11s, and the DOS operating system for IBM PCs and IBM-compatible machines. The software is readily usable by casual programmers for two-dimensional plotting.
2015-11-04
Gnuforplot Graphics is a Fortran90 program designed to generate two and three dimensional plots of data on a personal computer. The program uses calls to the open source code Gnuplot to generate the plots. Two Fortran90 programs have been written to use the Gnuplot graphics capabilities. The first program, named Plotsetup.f90 reads data from output files created by either the Stadium or LeachXS/Orchestra modeling codes and saves the data in arrays for plotting. This program then calls Gnuforplot which takes the data array along with user specified parameters to set plot specifications and issues Gnuplot commands that generate the screen plots. The user can view the plots and optionally save copies in jpeg format.
Proceedings: Graphics interface '86/Vision interface '86
Not Available
1986-01-01
This book contains papers presented at the Graphics interface conference. The Graphics interface conference is a major forum for the presentation of research in computer graphics and human-computer interaction. Sponsored by the Canadian Man-Computer Communications Society, the conferene draws an international audience with papers presented from France, Great Britain, the United States and Canada. The 1986 meeting was held jointly with Vision Interface '86, sponsored by the Canadian Image Processing and Pattern Recognition Society.
Graphical Contingency Analysis Tool - Energy Innovation Portal
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
The Graphical Contingency Analysis (GCA) Tool is a visual analytic software tool that aids power grid operators and planners to effectively manage complex outage predictions and ...
Crystallographic evidence for simultaneous growth in graphic...
Office of Scientific and Technical Information (OSTI)
Journal Article: Crystallographic evidence for simultaneous growth in graphic granite ... Publication Date: 2015-08-21 OSTI Identifier: 1178828 Resource Type: Journal Article ...
Graphics and Image Standards for the Web
Broader source: Energy.gov [DOE]
For Office of Energy Efficiency and Renewable Energy (EERE) websites and applications, follow these requirements and best practices for designing graphics and developing images. This includes making them Section 508-compliant.
Graphics Services Work Request | The Ames Laboratory
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
Graphics Services Work Request Please fill in the form below. Graphic Designer Grant Luhmann will contact you to discuss your project needs in more detail. Name * Phone Number * E-mail Address * Account to Charge * Request * Needed by * Month Month Jan Feb Mar Apr May Jun Jul Aug Sep Oct Nov Dec Day Day 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 26 27 28 29 30 31 Year Year 2014 2015 2016 2017 2018 Open popup calendar File Upload Files must be less than 2 MB. Allowed file
Graphical methods for evaluating covering arrays
Kim, Youngil; Jang, Dae -Heung; Anderson-Cook, Christine M.
2015-08-10
Covering arrays relax the condition of orthogonal arrays by only requiring that all combination of levels be covered but not requiring that the appearance of all combination of levels be balanced. This allows for a much larger number of factors to be simultaneously considered but at the cost of poorer estimation of the factor effects. To better understand patterns between sets of columns and evaluate the degree of coverage to compare and select between alternative arrays, we suggest several new graphical methods that show some of the patterns of coverage for different designs. These graphical methods for evaluating covering arrays are illustrated with a few examples.
Graphical methods for evaluating covering arrays
Kim, Youngil; Jang, Dae -Heung; Anderson-Cook, Christine M.
2015-08-10
Covering arrays relax the condition of orthogonal arrays by only requiring that all combination of levels be covered but not requiring that the appearance of all combination of levels be balanced. This allows for a much larger number of factors to be simultaneously considered but at the cost of poorer estimation of the factor effects. To better understand patterns between sets of columns and evaluate the degree of coverage to compare and select between alternative arrays, we suggest several new graphical methods that show some of the patterns of coverage for different designs. As a result, these graphical methods for evaluating covering arrays are illustrated with a few examples.
Graphics processing unit-assisted lossless decompression
Loughry, Thomas A.
2016-04-12
Systems and methods for decompressing compressed data that has been compressed by way of a lossless compression algorithm are described herein. In a general embodiment, a graphics processing unit (GPU) is programmed to receive compressed data packets and decompress such packets in parallel. The compressed data packets are compressed representations of an image, and the lossless compression algorithm is a Rice compression algorithm.
Graphical methods for evaluating covering arrays
DOE Public Access Gateway for Energy & Science Beta (PAGES Beta)
Kim, Youngil; Jang, Dae -Heung; Anderson-Cook, Christine M.
2015-08-10
Covering arrays relax the condition of orthogonal arrays by only requiring that all combination of levels be covered but not requiring that the appearance of all combination of levels be balanced. This allows for a much larger number of factors to be simultaneously considered but at the cost of poorer estimation of the factor effects. To better understand patterns between sets of columns and evaluate the degree of coverage to compare and select between alternative arrays, we suggest several new graphical methods that show some of the patterns of coverage for different designs. As a result, these graphical methods formore » evaluating covering arrays are illustrated with a few examples.« less
Parallel processor-based raster graphics system architecture
Littlefield, Richard J.
1990-01-01
An apparatus for generating raster graphics images from the graphics command stream includes a plurality of graphics processors connected in parallel, each adapted to receive any part of the graphics command stream for processing the command stream part into pixel data. The apparatus also includes a frame buffer for mapping the pixel data to pixel locations and an interconnection network for interconnecting the graphics processors to the frame buffer. Through the interconnection network, each graphics processor may access any part of the frame buffer concurrently with another graphics processor accessing any other part of the frame buffer. The plurality of graphics processors can thereby transmit concurrently pixel data to pixel locations in the frame buffer.
GenoGraphics for OpenWindows trademark
Hagstrom, R.; Overbeek, R.; Price, M.; Zawada, D. ); Michaels, G.S.; Taylor, R. . Div. of Computer Research and Technology); Yoshida, Kaoru )
1992-04-01
GenoGraphics is a generic utility for constructing and querying one-dimensional linear plots. The outgrowth of a request from Dr. Cassandra Smith for a tool to facilitate her genome mapping research. GenoGraphics development has benefited from a continued collaboration with her. Written in Sun Microsystem's OpenWindows environment and the BTOL toolkit developed at Argonne National Laboratory. GenoGraphics provides an interactive, intuitive, graphical interface. Its features include: viewing multiple maps simultaneously, zooming, and querying by mouse clicking. By expediting plot generation, GenoGraphics gives the scientist more time to analyze data and a novel means for deducing conclusions.
GenoGraphics for OpenWindows{trademark}
Hagstrom, R.; Overbeek, R.; Price, M.; Zawada, D.; Michaels, G.S.; Taylor, R.; Yoshida, Kaoru
1992-04-01
GenoGraphics is a generic utility for constructing and querying one-dimensional linear plots. The outgrowth of a request from Dr. Cassandra Smith for a tool to facilitate her genome mapping research. GenoGraphics development has benefited from a continued collaboration with her. Written in Sun Microsystem`s OpenWindows environment and the BTOL toolkit developed at Argonne National Laboratory. GenoGraphics provides an interactive, intuitive, graphical interface. Its features include: viewing multiple maps simultaneously, zooming, and querying by mouse clicking. By expediting plot generation, GenoGraphics gives the scientist more time to analyze data and a novel means for deducing conclusions.
Dead Reckoning Pedometer Graphical User Interface
Energy Science and Technology Software Center (OSTI)
2003-04-26
The Dead Reckoning Pedometer Graphical User Interface (DRP GUI) software is tasked with maturing the technology described in a WSRC patent application. This patent application describes an electronic navigation system that records human foot movements, in three dimensions, for the purpose of determining position, distance, and speed of a walking person. The simiplest form of the apparatus consists of a magnetometer (an instrument that measures magnetic field strength) on one foot and a small permanentmore » magnet on another foot with pressure sensors on each foot. When a person takes a step, the foot will hit the ground and produce a signal on the pressure sensor. This will trigger a reading of the magnetometer so that the relative position of one foot to the other can be calculated. This same process is repeated for each step. The DRP could be very useful for tracking emergency personnel such as firemen, policemen, and paramedics when they travel within a building. Technologies such as global positioning systems to not work within buildings. The goal of the DRP GUI V1.0.0 software is to provide a three-dimensional graphical user interface that will allow WSRC to demonstrate the DRP concepts to potential patent licensees. It is hoped that a partnership will allow WSRC and another company to further develop the DRP technology and software into a commercial product.« less
Scientific Graphical Displays on the Macintosh
Grotch, S.
1994-11-15
In many organizations scientists have ready access to more than one computer, often both a workstation (e.g., SUN, HP, SGI) as well as a Macintosh or other PC. The scientist commonly uses the work station for `number-crunching` and data analysis whereas the Macintosh is relegated to either word processing or serves as a `dumb terminal` to a larger main-frame computer. In an informal poll of my colleagues, very few of them used their Macintoshes for either statistical analysis or for graphical data display. I believe that this state of affairs is particularly unfortunate because over the last few years both the computational capability, and even more so, the software availability for the Macintosh have become quite formidable. In some instances, very powerful tools are now available on the Macintosh that may not exist (or be far too costly) on the so-called `high end` workstations. Many scientists are simply unaware of the wealth of extremely useful, `off-the-shelf` software that already exists on the Macintosh for scientific graphical and statistical analysis.
TMACS test procedure TP003: Graphics. Revision 6
Scanlan, P.K.; Washburn, S.
1994-09-01
The TMACS Software Project Test Procedures translate the project`s acceptance criteria into test steps. Software releases are certified when the affected Test Procedures are successfully performed and the customers authorize installation of these changes. This Test Procedure addresses the graphics requirements of the TMACS. The features to be tested are the data display graphics and the graphic elements that provide for operator control and selection of displays.
TMACS test procedure TP003: Graphics. Revision 5
Scanlan, P.K.
1994-08-25
The TMACS Software Project Test Procedures translate the project`s acceptance criteria into test steps. Software releases are certified when the affected Test Procedures are successfully performed and the customers authorize installation of these changes. This Test Procedure addresses the graphics requirements of the TMACS. The features to be tested are the data display graphics and the graphic elements that provide for operator control and selection of displays.
The multifacet graphically contracted function method. I. Formulation and implementation
Shepard, Ron; Brozell, Scott R.; Gidofalvi, Gergely
2014-08-14
The basic formulation for the multifacet generalization of the graphically contracted function (MFGCF) electronic structure method is presented. The analysis includes the discussion of linear dependency and redundancy of the arc factor parameters, the computation of reduced density matrices, Hamiltonian matrix construction, spin-density matrix construction, the computation of optimization gradients for single-state and state-averaged calculations, graphical wave function analysis, and the efficient computation of configuration state function and Slater determinant expansion coefficients. Timings are given for Hamiltonian matrix element and analytic optimization gradient computations for a range of model problems for full-CI Shavitt graphs, and it is observed that both the energy and the gradient computation scale as O(N{sup 2}n{sup 4}) for N electrons and n orbitals. The important arithmetic operations are within dense matrix-matrix product computational kernels, resulting in a computationally efficient procedure. An initial implementation of the method is used to present applications to several challenging chemical systems, including N{sub 2} dissociation, cubic H{sub 8} dissociation, the symmetric dissociation of H{sub 2}O, and the insertion of Be into H{sub 2}. The results are compared to the exact full-CI values and also to those of the previous single-facet GCF expansion form.
n-dimensional Statistical Inverse Graphical Hydraulic Test Simulator
Energy Science and Technology Software Center (OSTI)
2012-09-12
nSIGHTS (n-dimensional Statistical Inverse Graphical Hydraulic Test Simulator) is a comprehensive well test analysis software package. It provides a user-interface, a well test analysis model and many tools to analyze both field and simulated data. The well test analysis model simulates a single-phase, one-dimensional, radial/non-radial flow regime, with a borehole at the center of the modeled flow system. nSIGHTS solves the radially symmetric n-dimensional forward flow problem using a solver based on a graph-theoretic approach.more » The results of the forward simulation are pressure, and flow rate, given all the input parameters. The parameter estimation portion of nSIGHTS uses a perturbation-based approach to interpret the best-fit well and reservoir parameters, given an observed dataset of pressure and flow rate.« less
EnergyPlus Graphical User Interface
Energy Science and Technology Software Center (OSTI)
2011-01-04
LBNL, Infosys Technologies and Digital Alchemy are developing a free, comprehensive graphical user interface (GUI) that will enable EnergyPlus to be used more easily and effectively by building designers and other professionals, facilitating its widespread adoption. User requirements have been defined through a series of practitioner workshops. A new schematic editor for HVAC systems will be combined with different building envelope geometry generation tools and IFC-based BIM import and export. LBNL and Digital Alchemy havemore » generated a detailed function requirements specification, which is being implemented in software by Infosys, LBNL and and Digital Alchemy. LBNL and practitioner subcontractors will develop a comprehensive set of templates and libraries and will perform extensive testing of the GUI before it is released in Q3 2011. It is planned to use an Open Platfom approach, in which a comprehensive set of well documented Application Programming Interfaces (API's) would be provided to facilitate both the development of third party contributions to the official, standard GUI and the development of derivative works.« less
Real-time graphic displays in Mars. [AVLIS process
Treadway, T.
1985-12-01
Real-time diagnostic data of the AVLIS process is displayed in the form of a two-dimensional plot on selected monitors in the Mars Control Room. Each Mars workstation contains a HP2622 terminal for computer interface to the experiment and a Raster Technologies graphic controller driving a Conrac high resolution color monitor for graphics output.
FASTPLOT: An interface to Microsoft{reg_sign} FORTRAN graphics
Ward, R.C.
1994-03-01
Interface routines to the Microsoft{reg_sign} FORTRAN graphics library (GRAPHICS.LIB) are provided to facilitate development of graphics codes. These routines are collected into the FASTPLOT library (FASTPLOT.LIB). The FASTPLOT routines simplified the development of applications utilizing graphics and add capabilities not available in GRAPHICS.LIB such as plotting histograms, splines, symbols, and error bars. Specifically, these routines were utilized in the development of the mortality data viewing code, MORTVIEW, for the US Environmental Protection Agency. Routines for color imaging, developed for use with the X-ray Computer Tomography (XCT) imaging code, and examples are also provided in the FASTPLOT library. Many example uses of FASTPLOT.LIB are contained in this document to facilitate applications development. The FASTPLOT.LIB library, source, and applications programs are supplied on the accompanying FASTPLOT diskette.
Interactive graphical model building using telepresence and virtual reality
Cooke, C.; Stansfield, S.
1993-10-01
This paper presents a prototype system developed at Sandia National Laboratories to create and verify computer-generated graphical models of remote physical environments. The goal of the system is to create an interface between an operator and a computer vision system so that graphical models can be created interactively. Virtual reality and telepresence are used to allow interaction between the operator, computer, and remote environment. A stereo view of the remote environment is produced by two CCD cameras. The cameras are mounted on a three degree-of-freedom platform which is slaved to a mechanically-tracked, stereoscopic viewing device. This gives the operator a sense of immersion in the physical environment. The stereo video is enhanced by overlaying the graphical model onto it. Overlay of the graphical model onto the stereo video allows visual verification of graphical models. Creation of a graphical model is accomplished by allowing the operator to assist the computer in modeling. The operator controls a 3-D cursor to mark objects to be modeled. The computer then automatically extracts positional and geometric information about the object and creates the graphical model.
High-Throughput Characterization of Porous Materials Using Graphics...
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
High-Throughput Characterization of Porous Materials Using Graphics Processing Units Previous Next List J. Kim, R. L. Martin, O. Rubel, M. Haranczyk, and B. Smit, J. Chem. Theory...
Property:Integrated Display/Graphics | Open Energy Information
DisplayGraphics" Showing 22 pages using this property. A Alden Large Flume + Lab View, Test Point Alden Small Flume + Lab View, Test Point Alden Tow Tank + Lab View, Test Point...
NREL: News - Solar Decathlon Graphics and Communication Results Announced
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
Graphics and Communication Results Announced Tuesday, October 1, 2002 First Place Goes to University of Colorado Washington, D.C.-The University of Colorado at Boulder took first place in the Graphics and Communication results announced today at the U. S. Department of Energy's Solar Decathlon competition on the National Mall in Washington, DC. Auburn University placed second and Crowder College placed third. The Solar Decathlon runs through Oct. 5. With today's third round contest results, and
Interactive statistical-distribution-analysis program utilizing numerical and graphical methods
Glandon, S. R.; Fields, D. E.
1982-04-01
The TERPED/P program is designed to facilitate the quantitative analysis of experimental data, determine the distribution function that best describes the data, and provide graphical representations of the data. This code differs from its predecessors, TEDPED and TERPED, in that a printer-plotter has been added for graphical output flexibility. The addition of the printer-plotter provides TERPED/P with a method of generating graphs that is not dependent on DISSPLA, Integrated Software Systems Corporation's confidential proprietary graphics package. This makes it possible to use TERPED/P on systems not equipped with DISSPLA. In addition, the printer plot is usually produced more rapidly than a high-resolution plot can be generated. Graphical and numerical tests are performed on the data in accordance with the user's assumption of normality or lognormality. Statistical analysis options include computation of the chi-squared statistic and its significance level and the Kolmogorov-Smirnov one-sample test confidence level for data sets of more than 80 points. Plots can be produced on a Calcomp paper plotter, a FR80 film plotter, or a graphics terminal using the high-resolution, DISSPLA-dependent plotter or on a character-type output device by the printer-plotter. The plots are of cumulative probability (abscissa) versus user-defined units (ordinate). The program was developed on a Digital Equipment Corporation (DEC) PDP-10 and consists of 1500 statements. The language used is FORTRAN-10, DEC's extended version of FORTRAN-IV.
Graphic training materials: Your genie in the lamp
Hartley, D.; Stroupe, P.
1995-11-01
In the United States, we have overlooked using illustrated narrative materials (comic books) for training. Illustrated narrative training materials have the following benefits: (1) they promote learning by capitalizing on the visual dependency of the American public; (2) they promote retention by reinforcing the written word with graphic illustrations and with job-related stories; (3) they promote efficient transfer of knowledge to those with limited reading skills and those with limited English comprehension skills; and (4) they increase interest and are read! The Japanese have been successfully using graphic tests for education and training for years. Study comics were developed for mathematics, physics, economics, and multi-volume histories of Japan. Our organization decided to capitalize on the popularity and appeal of comic books and develop a graphic text that teaches the On-the-Job Training (OJT) process and good practices.
Tonopah Test Range EGS graphics tracking display system: HP370
Meyer, R.H.; Bauhs, K.C.
1994-08-01
This report describes the HP370 component of the Enhanced Graphics System (EGS) used at Tonopah Test Range (TTR). Selected Radar data is fed into the computer systems and the resulting tracking symbols are displayed on high-resolution video monitors in real time. These tracking symbols overlay background maps and are used for monitoring/controlling various flight vehicles. This report discusses both the operational aspects and the internal configuration of the HP370 Workstation portion of the EGS system.
Graphical Contingency Analysis for the Nation's Electric Grid
Zhenyu Huang
2011-04-01
PNNL has developed a new tool to manage the electric grid more effectively, helping prevent blackouts and brownouts--and possibly avoiding millions of dollars in fines for system violations. The Graphical Contingency Analysis tool monitors grid performance, shows prioritized lists of problems, provides visualizations of potential consequences, and helps operators identify the most effective courses of action. This technology yields faster, better decisions and a more stable and reliable power grid.
Graphical Contingency Analysis for the Nation's Electric Grid
Zhenyu (Henry) Huang
2012-12-31
PNNL has developed a new tool to manage the electric grid more effectively, helping prevent blackouts and brownouts--and possibly avoiding millions of dollars in fines for system violations. The Graphical Contingency Analysis tool monitors grid performance, shows prioritized lists of problems, provides visualizations of potential consequences, and helps operators identify the most effective courses of action. This technology yields faster, better decisions and a more stable and reliable power grid.
A Module for Graphical Display of Model Results with the CBP Toolbox
Smith, F.
2015-04-21
This report describes work performed by the Savannah River National Laboratory (SRNL) in fiscal year 2014 to add enhanced graphical capabilities to display model results in the Cementitious Barriers Project (CBP) Toolbox. Because Version 2.0 of the CBP Toolbox has just been released, the graphing enhancements described in this report have not yet been integrated into a new version of the Toolbox. Instead they have been tested using a standalone GoldSim model and, while they are substantially complete, may undergo further refinement before full implementation. Nevertheless, this report is issued to document the FY14 development efforts which will provide a basis for further development of the CBP Toolbox.
Top 5 Maps and Interactive Graphics of 2014 | Department of Energy
Office of Energy Efficiency and Renewable Energy (EERE) Indexed Site
Top 5 Maps and Interactive Graphics of 2014 Top 5 Maps and Interactive Graphics of 2014 December 22, 2014 - 12:05pm Addthis Daniel Wood Daniel Wood Data Visualization and...
Top 5 Maps and Interactive Graphics of 2014 | Department of Energy
Office of Environmental Management (EM)
Here are the top five maps and interactive graphics as selected by yours truly. 1) Space-Based Solar Power Our most fun -- and perhaps strangest -- graphic of 2014 (though 2 gives ...
Special Report: Graphics Processing Units Speed Results in Extreme-Scale
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
Supercomputers | Princeton Plasma Physics Lab Special Report: Graphics Processing Units Speed Results in Extreme-Scale Supercomputers American Fusion News Category: U.S. Universities Link: Special Report: Graphics Processing Units Speed Results in Extreme-Scale Supercomputers
Broader source: Energy.gov [DOE]
The Environmental Management Site-Specific Advisory Board recommends that DOE develop graphic representations of waste disposition paths.
Resolution-independent surface rendering using programmable graphics hardware
Loop, Charles T.; Blinn, James Frederick
2008-12-16
Surfaces defined by a Bezier tetrahedron, and in particular quadric surfaces, are rendered on programmable graphics hardware. Pixels are rendered through triangular sides of the tetrahedra and locations on the shapes, as well as surface normals for lighting evaluations, are computed using pixel shader computations. Additionally, vertex shaders are used to aid interpolation over a small number of values as input to the pixel shaders. Through this, rendering of the surfaces is performed independently of viewing resolution, allowing for advanced level-of-detail management. By individually rendering tetrahedrally-defined surfaces which together form complex shapes, the complex shapes can be rendered in their entirety.
Graphical methods for the sensitivity analysis in discriminant analysis
DOE Public Access Gateway for Energy & Science Beta (PAGES Beta)
Kim, Youngil; Anderson-Cook, Christine M.; Dae-Heung, Jang
2015-09-30
Similar to regression, many measures to detect influential data points in discriminant analysis have been developed. Many follow similar principles as the diagnostic measures used in linear regression in the context of discriminant analysis. Here we focus on the impact on the predicted classification posterior probability when a data point is omitted. The new method is intuitive and easily interpretative compared to existing methods. We also propose a graphical display to show the individual movement of the posterior probability of other data points when a specific data point is omitted. This enables the summaries to capture the overall pattern ofmore » the change.« less
Graphical methods for the sensitivity analysis in discriminant analysis
Kim, Youngil; Anderson-Cook, Christine M.; Dae-Heung, Jang
2015-09-30
Similar to regression, many measures to detect influential data points in discriminant analysis have been developed. Many follow similar principles as the diagnostic measures used in linear regression in the context of discriminant analysis. Here we focus on the impact on the predicted classification posterior probability when a data point is omitted. The new method is intuitive and easily interpretative compared to existing methods. We also propose a graphical display to show the individual movement of the posterior probability of other data points when a specific data point is omitted. This enables the summaries to capture the overall pattern of the change.
SBOL Visual: A Graphical Language for Genetic Designs
DOE Public Access Gateway for Energy & Science Beta (PAGES Beta)
Quinn, Jacqueline Y.; Cox, Robert Sidney; Adler, Aaron; Beal, Jacob; Bhatia, Swapnil; Cai, Yizhi; Chen, Joanna; Clancy, Kevin; Galdzicki, Michal; Hillson, Nathan J.; et al
2015-12-03
Synthetic Biology Open Language (SBOL) Visual is a graphical standard for genetic engineering. We report that it consists of symbols representing DNA subsequences, including regulatory elements and DNA assembly features. These symbols can be used to draw illustrations for communication and instruction, and as image assets for computer-aided design. SBOL Visual is a community standard, freely available for personal, academic, and commercial use (Creative Commons CC0 license). We provide prototypical symbol images that have been used in scientific publications and software tools. We encourage users to use and modify them freely, and to join the SBOL Visual community: http://www.sbolstandard.org/visual.
Field Deployable Tritium Assay System Host Graphical User Interface Software
Energy Science and Technology Software Center (OSTI)
1998-05-12
The FDTASHOST software is a Graphical User Interface for the Field Deployable Tritium Assay System (FDTAS - Invention Disclosure SRS-96-09-091 has been submitted). The program runs on the Host computer which is located in the Laboratory and connected to the FDTAS remote field system via a modem over a phone line. The operator receives status information and messages from the Remote system. The operator can enter in commands to be executed by the remote systemmore » using the mouse and a pull down menu.« less
Using Python to Develop Graphical Interfaces to Scientific Data
MacFarland, L; Streletz, G J
1999-09-24
At Lawrence Livermore National Laboratory (LLNL), Python has proven to be a convenient language for the development of graphical user interfaces (GUIs) which allow scientists to view, plot, and analyze scientific data. Two such applications are described in this paper. The first, EOSView, is a browser application for an equation of state data library at LLNL. EOSView is used by scientists throughout the laboratory who use simulation codes that access the data library, or who need equation of state data for other purposes. EOSView provides graphical visualization capabilities, as well as the capability to analyze the data in many different ways. The second application, Zimp, is a GUI that allows interactive use of the Stark Line Shape Database. It is used to access and plot data. The quick construction of Zimp from elements of the EOSView code provides a useful lesson in code reuse, and illustrates how the object-oriented nature of Python facilitates this goal. In general, Python has proven to be an appropriate choice of language for applications of this type for several reasons, including the easy access to GUI functionality provided by Tkinter, the ease with which C functions can be called from Python, and the convenient handling of strings in Python. Moreover, the features of the Python language, combined with the fact that it is interpreted rather than compiled, have allowed for extremely quick prototyping.
Real-time graphic display utility for nuclear safety applications
Yang, S.; Huang, X.; Taylor, J.; Stevens, J.; Gerardis, T.; Hsu, A.; McCreary, T.
2006-07-01
With the increasing interests in the nuclear energy, new nuclear power plants will be constructed and licensed, and older generation ones will be upgraded for assuring continuing operation. The tendency of adopting the latest proven technology and the fact of older parts becoming obsolete have made the upgrades imperative. One of the areas for upgrades is the older CRT display being replaced by the latest graphics displays running under modern real time operating system (RTOS) with safety graded modern computer. HFC has developed a graphic display utility (GDU) under the QNX RTOS. A standard off-the-shelf software with a long history of performance in industrial applications, QNX RTOS used for safety applications has been examined via a commercial dedication process that is consistent with the regulatory guidelines. Through a commercial survey, a design life cycle and an operating history evaluation, and necessary tests dictated by the dedication plan, it is reasonably confirmed that the QNX RTOS was essentially equivalent to what would be expected in the nuclear industry. The developed GDU operates and communicates with the existing equipment through a dedicated serial channel of a flat panel controller (FPC) module. The FPC module drives a flat panel display (FPD) monitor. A touch screen mounted on the FPD serves as the normal operator interface with the FPC/FPD monitor system. The GDU can be used not only for replacing older CRTs but also in new applications. The replacement of the older CRT does not disturb the function of the existing equipment. It not only provides modern proven technology upgrade but also improves human ergonomics. The FPC, which can be used as a standalone controller running with the GDU, is an integrated hardware and software module. It operates as a single board computer within a control system, and applies primarily to the graphics display, targeting, keyboard and mouse. During normal system operation, the GDU has two sources of data
Graphical User Interface Color Display Animation Interaction Tool
Energy Science and Technology Software Center (OSTI)
1999-10-05
The Nuclear Plant Analyzer (NPA) is a highly flexible graphical user interface for displaying the results of a calculation, typically generated by RELAP5 or other code. This display consists of one or more picture, called masks, that mimic the host code input. This mask can be animated to display user-specified code output information mapped as colors, dials, moving arrows, etc., on the mask. The user can also interact with the control systems of the hostmore » input file as the execution progresses, thereby controlling aspects of the calculation. The Computer Visual System (CVS) creates, edits, and animates the the masks for use in the NPA.« less
Graphics processing units accelerated semiclassical initial value representation molecular dynamics
Tamascelli, Dario; Dambrosio, Francesco Saverio; Conte, Riccardo; Ceotto, Michele
2014-05-07
This paper presents a Graphics Processing Units (GPUs) implementation of the Semiclassical Initial Value Representation (SC-IVR) propagator for vibrational molecular spectroscopy calculations. The time-averaging formulation of the SC-IVR for power spectrum calculations is employed. Details about the GPU implementation of the semiclassical code are provided. Four molecules with an increasing number of atoms are considered and the GPU-calculated vibrational frequencies perfectly match the benchmark values. The computational time scaling of two GPUs (NVIDIA Tesla C2075 and Kepler K20), respectively, versus two CPUs (Intel Core i5 and Intel Xeon E5-2687W) and the critical issues related to the GPU implementation are discussed. The resulting reduction in computational time and power consumption is significant and semiclassical GPU calculations are shown to be environment friendly.
An overview of 3-D graphical analysis using DOE-2 hourly simulation data
Haberl, J.S.; MacDonald, M.; Eden, A.
1988-01-01
This paper presents an overview of a 3-D graphical approach for improving the potential of building energy analyses using the DOE-2 computer program. The approach produces 3-D annual profiles from hourly data generated by DOE-2 simulations using a statistical plotting package for specific quantities of interest. The annual profiles of hourly data provide a useful graphical check of voluminous data in a condensed form, allowing several different types of data to be plotted over a year. These profiles provide the user with the opportunity to check simulation results, check for potential problems with user input, provide graphs to customers who may want a simpler presentation, visualize interactions in simulations, and understand where inappropriate modeling conditions may exist in simulations. Future analysis, using such profiles, may allow methods to be developed to check consistency between simulations, check for potential hidden errors in modeling buildings, and better understand how simulations compare with data from real buildings. 22 refs., 23 figs., 1 tab.
User guide for transferring data from NOS (Network Operating System) to Applicon Graphics System
Young, E. G.; Grube, L. K.
1980-08-01
This document is a guide for obtaining data from Sandia Central Computing site that are to be introduced into a drawing file on Applicon Graphics Systems. 1 figure.
EZVIDEO, FORTRAN graphics routines for the IBM AT
Patterson, M.R.; Holdeman, J.T.; Ward, R.C.; Jackson, W.L.
1989-10-01
A set of IBM PC-based FORTRAN plotting routines called EZVIDEO is described in this report. These routines are written in FORTRAN and can be called from FORTRAN programs. EZVIDEO simulates a subset of the well-known DISSPLA graphics calls and makes plots directly on the IBM AT display screen. Screen dumps can also be made to an attached LaserJet or Epson printer to make hard copy without using terminal emulators. More than forty DISSPLA calls are simulated by the EZVIDEO routines. Typical screen plots require about 10 seconds (s), and good hard copy of the screen image on a laser printer requires less than 2 minutes (min). This higher-resolution hard copy is adequate for most purposes because of the enhanced resolution of the screen in the EGA and VGA modes. These EZVIDEO routines give the IB, AT user a stand-alone capability to make useful scientific or engineering plots directly on the AT, using data generated in FORTRAN programs. The routines will also work on the IBM PC or XT in CGA mode, but they require more time and yield less resolution. 7 refs., 4 figs.
Application Protocol, Initial Graphics Exchange Specification (IGES), Layered Electrical Product
O`Connell, L.J.
1994-12-01
An application protocol is an information systems engineering view of a specific product The view represents an agreement on the generic activities needed to design and fabricate the product the agreement on the information needed to support those activities, and the specific constructs of a product data standard for use in transferring some or all of the information required. This application protocol describes the data for electrical and electronic products in terms of a product description standard called the Initial Graphics Exchange Specification (IGES). More specifically, the Layered Electrical Product IGES Application Protocol (AP) specifies the mechanisms for defining and exchanging computer-models and their associated data for those products which have been designed in two dimensional geometry so as to be produced as a series of layers in IGES format The AP defines the appropriateness of the data items for describing the geometry of the various parts of a product (shape and location), the connectivity, and the processing and material characteristics. Excluded is the behavioral requirements which the product was intended to satisfy, except as those requirements have been recorded as design rules or product testing requirements.
Quantitative genetic activity graphical profiles for use in chemical evaluation
Waters, M.D.; Stack, H.F.; Garrett, N.E.; Jackson, M.A.
1990-12-31
A graphic approach, terms a Genetic Activity Profile (GAP), was developed to display a matrix of data on the genetic and related effects of selected chemical agents. The profiles provide a visual overview of the quantitative (doses) and qualitative (test results) data for each chemical. Either the lowest effective dose or highest ineffective dose is recorded for each agent and bioassay. Up to 200 different test systems are represented across the GAP. Bioassay systems are organized according to the phylogeny of the test organisms and the end points of genetic activity. The methodology for producing and evaluating genetic activity profile was developed in collaboration with the International Agency for Research on Cancer (IARC). Data on individual chemicals were compiles by IARC and by the US Environmental Protection Agency (EPA). Data are available on 343 compounds selected from volumes 1-53 of the IARC Monographs and on 115 compounds identified as Superfund Priority Substances. Software to display the GAPs on an IBM-compatible personal computer is available from the authors. Structurally similar compounds frequently display qualitatively and quantitatively similar profiles of genetic activity. Through examination of the patterns of GAPs of pairs and groups of chemicals, it is possible to make more informed decisions regarding the selection of test batteries to be used in evaluation of chemical analogs. GAPs provided useful data for development of weight-of-evidence hazard ranking schemes. Also, some knowledge of the potential genetic activity of complex environmental mixtures may be gained from an assessment of the genetic activity profiles of component chemicals. The fundamental techniques and computer programs devised for the GAP database may be used to develop similar databases in other disciplines. 36 refs., 2 figs.
T-532: Vulnerability in Graphics Rendering Engine Could Allow Remote Code Execution
Broader source: Energy.gov [DOE]
Microsoft is investigating new public reports of a vulnerability in the Windows Graphics Rendering Engine. An attacker who successfully exploited this vulnerability could run arbitrary code in the security context of the logged-on user.
Simergy: Practitioner-Oriented Graphical User Interface for EnergyPlus -
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
Energy Innovation Portal Energy Analysis Energy Analysis Building Energy Efficiency Building Energy Efficiency Find More Like This Return to Search Simergy: Practitioner-Oriented Graphical User Interface for EnergyPlus Lawrence Berkeley National Laboratory Contact LBL About This Technology Technology Marketing Summary Simergy is a graphical user interface (GUI) for the DOE building energy simulation program EnergyPlus for modeling building heating, cooling, lighting, ventilation and other
COREMAP: Graphical user interface for displaying reactor core data in an interactive hexagon map
Muscat, F.L.; Derstine, K.L.
1995-06-01
COREMAP is a Graphical User Interface (GUI) designed to assist users read and check reactor core data from multidimensional neutronic simulation models in color and/or as text in an interactive 2D planar grid of hexagonal subassemblies. COREMAP is a complete GEODST/RUNDESC viewing tool which enables the user to access multi data set files (e.g. planes, moments, energy groups ,... ) and display up to two data sets simultaneously, one as color and the other as text. The user (1) controls color scale characteristics such as type (linear or logarithmic) and range limits, (2) controls the text display based upon conditional statements on data spelling, and value. (3) chooses zoom features such as core map size, number of rings and surrounding subassemblies, and (4) specifies the data selection for supplied popup subwindows which display a selection of data currently off-screen for a selected cell, as a list of data and/or as a graph. COREMAP includes a RUNDESC file editing tool which creates ``proposed`` Run-description files by point and click revisions to subassembly assignments in an existing EBRII Run-description file. COREMAP includes a fully automated printing option which creates high quality PostScript color or greyscale images of the core map independent of the monitor used, e.g. color prints can be generated with a session from a color or monochrome monitor. The automated PostScript output is an alternative to the xgrabsc based printing option. COREMAP includes a plotting option which creates graphs related to a selected cell. The user specifies the X and Y coordinates types (planes, moment, group, flux ,... ) and a parameter, P, when displaying several curves for the specified (X, Y) pair COREMAP supports hexagonal geometry reactor core configurations specified by: the GEODST file and binary Standard Interface Files and the RUNDESC ordering.
Croff, A.G.; Liberman, M.S.; Morrison, G.W.
1982-01-01
Based on the results of ORIGEN2 and a newly developed code called ORMANG, graphical and summary tabular characteristics of spent fuel, high-level waste, and fuel assembly structural material (cladding) waste are presented for a generic pressurized-water reactor (PWR), a liquid-metal fast breeder reactor (LMFBR), and the Fast Flux Test Facility (FFTF). The characteristics include radioactivity, thermal power, and toxicity (water dilution volume). Given are graphs and summary tables containing characteristic totals and the principal nuclide contributors as well as graphs comparing the three reactors for a single material and the three materials for a single reactor.
DOGS: a collection of graphics for support of discrete ordinates codes
Ingersoll, D.T.; Slater, C.O.
1980-03-01
A collection of computer codes called DOGS (Discrete Ordinates Graphics Support) has been developed to assist in the display and presentation of data generated by commonly used discrete ordinates transport codes. The DOGS codes include: EGAD for plotting two-dimensional geometries, ISOPLOT4 for plotting 2-D fluxes in a contour line fashion, FORM for plotting 2-D fluxes in a 3-D surface fashion, ACTUAL for calculating 2-D activities, TOOTH for calculating and plotting space-energy contributon fluxes, and ASPECT for plotting energy spectra. All of the codes use FIDO input formats and DISSPLA graphics software including the DISSPOP post processors.
Graphic presentation of quarterly /sup 90/Sr fallout data, 1954-1982
Larsen, R.J.
1984-01-01
This report graphically presents all of the precipitation and /sup 90/Sr deposition data for all stations operated as part of the Environmental Measurements Laboratory's (EML) global fallout program since the initiation of the program in 1954. 3 references, 179 figures.
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
6, 2009 WIPP Quick Facts (As of 11-5-09) 8,002 Shipments received since opening (7,702 CH and 300 RH) 63,971 Cubic meters of waste disposed (63,828 CH and 143 RH) 124,386 Containers disposed in the underground (124,093 CH and 293 RH) WIPP receives 8000th TRU waste shipment The 8,000th shipment arrives at WIPP late Wednesday night. The two drivers, Phil Godin, left, and Jack Clayton stand next to the truck as it is inspected by Security. The shipment came to WIPP from Los Alamos National
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
2009 WIPP Quick Facts (As of 9-10-09) 7,762 Shipments received since opening (7,487 CH and 275 RH) 62,351 Cubic meters of waste disposed (62,221 CH and 130 RH) 119,958 Containers disposed in the underground (119,688 CH and 270 RH) CBFO hires five new employees Starting a new job is always exciting and a challenge. Five new DOE Carlsbad Field Office employees know the feeling. CBFO has filled five positions with employees that possess a WIPP trucks surpass 9 million loaded miles File photo With
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
1, 2006 WIPP Quick Facts (As of 09-20-06) 5,043 Shipments received since opening 41,877 Cubic meters of waste disposed 82,575 Containers disposed in the underground Storment achieves two computer security certifications Brandon Storment (NCI), WIPP's computer security advisor, has recently received two computer security certifications. On August 1, Storment was certified in Information Security Evaluation Methodology. This certification means that based on National Security Administration
Energy Science and Technology Software Center (OSTI)
2001-06-08
MUSTAFA is a scientific visualization package for visualizing data in the EXODUSII file format. These data files are typically priduced from Sandia's suite of finite element engineering analysis codes.
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
and tribal officials with access to a database can then monitor shipments from their computers. For those interested, please contact Debbie Hofer at 575-302- 3017 or at...
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
The NCAR Command Language (NCL) is a full programming language including looping and conditionals for data selection, manipulation, and display. NCL commands can be executed one at ...
Graphical Environment Tools for Application to Gamma-Ray Energy Tracking Arrays
Todd, Richard A.; Radford, David C.
2013-12-30
Highly segmented, position-sensitive germanium detector systems are being developed for nuclear physics research where traditional electronic signal processing with mixed analog and digital function blocks would be enormously complex and costly. Future systems will be constructed using pipelined processing of high-speed digitized signals as is done in the telecommunications industry. Techniques which provide rapid algorithm and system development for future systems are desirable. This project has used digital signal processing concepts and existing graphical system design tools to develop a set of re-usable modular functions and libraries targeted for the nuclear physics community. Researchers working with complex nuclear detector arrays such as the Gamma-Ray Energy Tracking Array (GRETA) have been able to construct advanced data processing algorithms for implementation in field programmable gate arrays (FPGAs) through application of these library functions using intuitive graphical interfaces.
Future Directions in Computer Graphics and Visualization: From CG&A's Editorial Board
Encarnacao, L. M.; Chuang, Yung-Yu; Stork, Andre; Kasik, David; Rhyne, Theresa-Marie; Avila, Lisa; Kohlhammer, Jorn; LaViola, Joseph; Tory, Melanie; Dill, John; Domik, Gitta; Owen, G. Scott; Wong, Pak C.
2015-01-01
With many new members joining the CG&A editorial board over the past year, and with a renewed commitment to not only document the state of the art in computer graphics research and applications but to anticipate and where possible foster future areas of scientific discourse and industrial practice, we asked editorial and advisory council members about where they see their fields of expertise going. The answers compiled here arent meant to be all encompassing or deterministic when it comes to the opportunities computer graphics and interactive visualization hold for the future. Instead, we aim to accomplish two things: give a more in-depth introduction of members of the editorial board to the CG&A readership and encourage cross-disciplinary discourse toward approaching, complementing, or disputing the visions laid out in this compilation.
Graphical user interfaces for McCellan Nuclear Radiation Center (MNRC).
Brown-VanHoozer, S. A.
1998-08-27
McClellan's Nuclear Radiation Center (MNRC) control console is in the process of being replaced due to spurious scrams, outdated software, and obsolete parts. The intent of the new control console is to eliminate the existing problems by installing a UNIX-based computer system with industry-standard interface software and incorporating human factors during all stages of the graphical user interface (GUI) development and control console design.
Interfacing 1990 US Census TIGER map files with New S graphics software
Rizzardi, M.; Mohr, M.S.; Merrill, D.W.; Selvin, S. |
1992-07-01
In 1990, the United States Bureau of the Census released detailed geographic base files known as TIGER/Line (Topologically Integrated Geographic Encoding and Referencing) which contain detail on the physical features and census tract boundaries of every county in the United States. The TIGER database is attractive for two reasons. First, it is publicly available through the Bureau of the Census on tape or cd-rom for a minimal fee. Second, it contains 24 billion characters of data which describe geographic features of interest to the Census Bureau such as coastlines, hydrography, transportation networks, political boundaries, etc. Unfortunately, the large TIGER database only provides raw alphanumeric data; no utility software, graphical or otherwise, is included. On the other hand New S, a popular statistical software package by AT&T, has easily operated functions that permit advanced graphics in conjunction with data analysis. New S has the ability to plot contours, lines, segments, and points. However, of special interest is the New S function map and its options. Using the map function, which requires polygons as input, census tracts can be quickly selected, plotted, shaded, etc. New S graphics combined with the TIGER database has obvious potential. This paper reports on our efforts to use the TIGER map files with New S, especially to construct census tract maps of counties. While census tract boundaries are inherently polygonal, they are not organized as such in the TIGER database. This conversion of the TIGER ``line`` format into New S ``polygon/polyline`` format is one facet of the work reported here. Also we discuss the selection and extraction of auxiliary geographic information from TIGER files for graphical display using New S.
Jimenez, Edward Steven,
2013-09-01
The goal of this work is to develop a fast computed tomography (CT) reconstruction algorithm based on graphics processing units (GPU) that achieves significant improvement over traditional central processing unit (CPU) based implementations. The main challenge in developing a CT algorithm that is capable of handling very large datasets is parallelizing the algorithm in such a way that data transfer does not hinder performance of the reconstruction algorithm. General Purpose Graphics Processing (GPGPU) is a new technology that the Science and Technology (S&T) community is starting to adopt in many fields where CPU-based computing is the norm. GPGPU programming requires a new approach to algorithm development that utilizes massively multi-threaded environments. Multi-threaded algorithms in general are difficult to optimize since performance bottlenecks occur that are non-existent in single-threaded algorithms such as memory latencies. If an efficient GPU-based CT reconstruction algorithm can be developed; computational times could be improved by a factor of 20. Additionally, cost benefits will be realized as commodity graphics hardware could potentially replace expensive supercomputers and high-end workstations. This project will take advantage of the CUDA programming environment and attempt to parallelize the task in such a way that multiple slices of the reconstruction volume are computed simultaneously. This work will also take advantage of the GPU memory by utilizing asynchronous memory transfers, GPU texture memory, and (when possible) pinned host memory so that the memory transfer bottleneck inherent to GPGPU is amortized. Additionally, this work will take advantage of GPU-specific hardware (i.e. fast texture memory, pixel-pipelines, hardware interpolators, and varying memory hierarchy) that will allow for additional performance improvements.
Computation of Large Covariance Matrices by SAMMY on Graphical Processing Units and Multicore CPUs
Arbanas, Goran [ORNL; Dunn, Michael E [ORNL; Wiarda, Dorothea [ORNL
2011-01-01
Computational power of Graphical Processing Units and multicore CPUs was harnessed by the nuclear data evaluation code SAMMY to speed up computations of large Resonance Parameter Covariance Matrices (RPCMs). This was accomplished by linking SAMMY to vendor-optimized implementations of the matrix-matrix multiplication subroutine of the Basic Linear Algebra Library to compute the most time-consuming step. The U-235 RPCM computed previously using a triple-nested loop was re-computed using the NVIDIA implementation of the subroutine on a single Tesla Fermi Graphical Processing Unit, and also using the Intel's Math Kernel Library implementation on two different multicore CPU systems. A multiplication of two matrices of dimensions 16,000 x 20,000 that had previously taken days, took approximately one minute on the GPU. Similar performance was achieved on a dual six-core CPU system. The magnitude of the speed-up suggests that these, or similar, combinations of hardware and libraries may be useful for large matrix operations in SAMMY. Uniform interfaces of standard linear algebra libraries make them a promising candidate for a programming framework of a new generation of SAMMY for the emerging heterogeneous computing platforms.
Development and New Directions for the RELAP5-3D Graphical Users Interface
Mesina, George Lee
2001-09-01
The direction of development for the RELAP5 Graphical User Interfaces (RGUI) has been extended. In addition to existing plans for displaying all aspects of RELAP5 calculations, the plan now includes plans to display the calculations of a variety of codes including SCDAP, RETRAN and FLUENT. Recent work has included such extensions along with the previously planned and user-requested improvements and extensions. Visualization of heat-structures has been added. Adaptations were made for another computer program, SCDAP-3D, including plant core views. An input model builder for generating RELAP5-3D input files was partially implemented. All these are reported. Plans for future work are also summarized. These include an input processor that transfers steady-state conditions into an input file.
Real-time geo-registration of imagery using COTS graphics processors
Flath, Laurence M.; Kartz, Michael W.
2009-06-30
A method of performing real-time geo-registration of high-resolution digital imagery using existing graphics processing units (GPUs) already found in current personal computers, rather than the main central processing unit (CPU). Digital image data captured by a camera (along with inertial navigation system (INS) data associated with the image data) is transferred to and processed by the GPU to perform the calculations involved in transforming the captured image into a geo-rectified, nadir-looking image. By using the GPU, the order-of-magnitude increase in throughput over conventional software techniques makes real-time geo-registration possible without the significant cost of custom hardware solutions.
Loughry, Thomas A.
2015-02-01
As the volume of data acquired by space-based sensors increases, mission data compression/decompression and forward error correction code processing performance must likewise scale. This competency development effort was explored using the General Purpose Graphics Processing Unit (GPGPU) to accomplish high-rate Rice Decompression and high-rate Reed-Solomon (RS) decoding at the satellite mission ground station. Each algorithm was implemented and benchmarked on a single GPGPU. Distributed processing across one to four GPGPUs was also investigated. The results show that the GPGPU has considerable potential for performing satellite communication Data Signal Processing, with three times or better performance improvements and up to ten times reduction in cost over custom hardware, at least in the case of Rice Decompression and Reed-Solomon Decoding.
SU-E-P-59: A Graphical Interface for XCAT Phantom Configuration, Generation and Processing
Myronakis, M; Cai, W; Dhou, S; Cifter, F; Lewis, J; Hurwitz, M
2015-06-15
Purpose: To design a comprehensive open-source, publicly available, graphical user interface (GUI) to facilitate the configuration, generation, processing and use of the 4D Extended Cardiac-Torso (XCAT) phantom. Methods: The XCAT phantom includes over 9000 anatomical objects as well as respiratory, cardiac and tumor motion. It is widely used for research studies in medical imaging and radiotherapy. The phantom generation process involves the configuration of a text script to parameterize the geometry, motion, and composition of the whole body and objects within it, and to generate simulated PET or CT images. To avoid the need for manual editing or script writing, our MATLAB-based GUI uses slider controls, drop-down lists, buttons and graphical text input to parameterize and process the phantom. Results: Our GUI can be used to: a) generate parameter files; b) generate the voxelized phantom; c) combine the phantom with a lesion; d) display the phantom; e) produce average and maximum intensity images from the phantom output files; f) incorporate irregular patient breathing patterns; and f) generate DICOM files containing phantom images. The GUI provides local help information using tool-tip strings on the currently selected phantom, minimizing the need for external documentation. The DICOM generation feature is intended to simplify the process of importing the phantom images into radiotherapy treatment planning systems or other clinical software. Conclusion: The GUI simplifies and automates the use of the XCAT phantom for imaging-based research projects in medical imaging or radiotherapy. This has the potential to accelerate research conducted with the XCAT phantom, or to ease the learning curve for new users. This tool does not include the XCAT phantom software itself. We would like to acknowledge funding from MRA, Varian Medical Systems Inc.
Stevens, E.J.; McNeilly, G.S.
1994-03-01
The existing National Center for Atmospheric Research (NCAR) code in the Hamburg Oceanic Carbon Cycle Circulation Model and the Hamburg Large-Scale Geostrophic Ocean General Circulation Model was modernized and reduced in size while still producing an equivalent end result. A reduction in the size of the existing code from more than 50,000 lines to approximately 7,500 lines in the new code has made the new code much easier to maintain. The existing code in Hamburg model uses legacy NCAR (including even emulated CALCOMP subrountines) graphics to display graphical output. The new code uses only current (version 3.1) NCAR subrountines.
High-throughput Characterization of Porous Materials Using Graphics Processing Units
Kim, Jihan; Martin, Richard L.; Ruebel, Oliver; Haranczyk, Maciej; Smit, Berend
2012-03-19
We have developed a high-throughput graphics processing units (GPU) code that can characterize a large database of crystalline porous materials. In our algorithm, the GPU is utilized to accelerate energy grid calculations where the grid values represent interactions (i.e., Lennard-Jones + Coulomb potentials) between gas molecules (i.e., CH$_{4}$ and CO$_{2}$) and material's framework atoms. Using a parallel flood fill CPU algorithm, inaccessible regions inside the framework structures are identified and blocked based on their energy profiles. Finally, we compute the Henry coefficients and heats of adsorption through statistical Widom insertion Monte Carlo moves in the domain restricted to the accessible space. The code offers significant speedup over a single core CPU code and allows us to characterize a set of porous materials at least an order of magnitude larger than ones considered in earlier studies. For structures selected from such a prescreening algorithm, full adsorption isotherms can be calculated by conducting multiple grand canonical Monte Carlo simulations concurrently within the GPU.
Fenestration performance analysis using an interactive graphics-based methodology on a microcomputer
Sullivan, R.; Selkowitz, S.
1989-09-01
We show the development and implementation of a new methodology that can be used to evaluate the energy and comfort performance of fenestration in non-residential buildings. The methodology is based on the definition of a fenestration system figure of merit.'' The figure of merit'' is determined by considering five non-dimensional performance indices representing heating energy, cooling energy, cooling energy peak, thermal comfort, and visual comfort. These indices were derived by performing a regression analysis of several thousand hour-by-hour building heat transfer simulations of a prototypical office building module using the DOE-2 simulation program. The regression analysis resulted in a series of simplified algebraic expressions that related fenestration configuration variables to performance parameters. We implemented these equations in a hypermedia'' environment -- one that integrates graphics, sound, animation, and calculation sequences --and created a prototype fenestration performance design tool. Inputs required by the program consist of geographic location, building type, perimeter space, and envelope definition. Outputs are the calculated performance indices for electricity and fuel use, peak electric load, and thermal and visual comfort. 6 refs., 7 figs.
Density-fitted singles and doubles coupled cluster on graphics processing units
Sherrill, David; Sumpter, Bobby G; DePrince, III, A. Eugene
2014-01-01
We adapt an algorithm for singles and doubles coupled cluster (CCSD) that uses density fitting (DF) or Cholesky decomposition (CD) in the construction and contraction of all electron repulsion integrals (ERI s) for use on heterogeneous compute nodes consisting of a multicore CPU and at least one graphics processing unit (GPU). The use of approximate 3-index ERI s ameliorates two of the major difficulties in designing scientific algorithms for GPU s: (i) the extremely limited global memory on the devices and (ii) the overhead associated with data motion across the PCI bus. For the benzene trimer described by an aug-cc-pVDZ basis set, the use of a single NVIDIA Tesla C2070 (Fermi) GPU accelerates a CD-CCSD computation by a factor of 2.1, relative to the multicore CPU-only algorithm that uses 6 highly efficient Intel core i7-3930K CPU cores. The use of two Fermis provides an acceleration of 2.89, which is comparable to that observed when using a single NVIDIA Kepler K20c GPU (2.73).
The application of projected conjugate gradient solvers on graphical processing units
Lin, Youzuo; Renaut, Rosemary
2011-01-26
Graphical processing units introduce the capability for large scale computation at the desktop. Presented numerical results verify that efficiencies and accuracies of basic linear algebra subroutines of all levels when implemented in CUDA and Jacket are comparable. But experimental results demonstrate that the basic linear algebra subroutines of level three offer the greatest potential for improving efficiency of basic numerical algorithms. We consider the solution of the multiple right hand side set of linear equations using Krylov subspace-based solvers. Thus, for the multiple right hand side case, it is more efficient to make use of a block implementation of the conjugate gradient algorithm, rather than to solve each system independently. Jacket is used for the implementation. Furthermore, including projection from one system to another improves efficiency. A relevant example, for which simulated results are provided, is the reconstruction of a three dimensional medical image volume acquired from a positron emission tomography scanner. Efficiency of the reconstruction is improved by using projection across nearby slices.
Not Available
1994-09-01
An application protocol is an information systems engineering view of a specific product. The view represents an agreement on the generic activities needed to design and fabricate the product, the agreement on the information needed to support those activities, and the specific constructs of a product data standard for use in transfering some or all of the information required. This applications protocol describes the data for electrical and electronic products in terms of a product description standard called the Initial Graphics Exchange Specification (IGES). More specifically, the Layered Electrical Product IGES Application Protocol (AP) specifies the mechanisms for defining and exchanging computer-models and their associated data for those products which have been designed in two dimensional geometry so as to be produced as a series of layers in IGES format. The AP defines the appropriateness of the data items for describing the geometry of the various parts of a product (shape and location), the connectivity, and the processing and material characteristics. Excluded is the behavioral requirements which the product was intended to satisfy, except as those requirements have been recorded as design rules or product testing requirements.
Developing extensible lattice-Boltzmann simulationsfor general-purpose graphics-programming units
Walsh, S C; Saar, M O
2011-10-27
Lattice-Boltzmann methods are versatile numerical modeling techniques capable of reproducing a wide variety of fluid-mechanical behavior. These methods are well suited to parallel implementation, particularly on the single-instruction multiple data (SIMD) parallel processing environments found in computer graphics processing units (GPUs). Although more recent programming tools dramatically improve the ease with which GPU programs can be written, the programming environment still lacks the flexibility available to more traditional CPU programs. In particular, it may be difficult to develop modular and extensible programs that require variable on-device functionality with current GPU architectures. This paper describes a process of automatic code generation that overcomes these difficulties for lattice-Boltzmann simulations. It details the development of GPU-based modules for an extensible lattice-Boltzmann simulation package - LBHydra. The performance of the automatically generated code is compared to equivalent purpose written codes for both single-phase, multiple-phase, and multiple-component flows. The flexibility of the new method is demonstrated by simulating a rising, dissolving droplet in a porous medium with user generated lattice-Boltzmann models and subroutines.
Developing extensible lattice-Boltzmann simulators for general-purpose graphics-processing units
Walsh, S C; Saar, M O
2011-12-21
Lattice-Boltzmann methods are versatile numerical modeling techniques capable of reproducing a wide variety of fluid-mechanical behavior. These methods are well suited to parallel implementation, particularly on the single-instruction multiple data (SIMD) parallel processing environments found in computer graphics processing units (GPUs). Although more recent programming tools dramatically improve the ease with which GPU programs can be written, the programming environment still lacks the flexibility available to more traditional CPU programs. In particular, it may be difficult to develop modular and extensible programs that require variable on-device functionality with current GPU architectures. This paper describes a process of automatic code generation that overcomes these difficulties for lattice-Boltzmann simulations. It details the development of GPU-based modules for an extensible lattice-Boltzmann simulation package - LBHydra. The performance of the automatically generated code is compared to equivalent purpose written codes for both single-phase, multiple-phase, and multiple-component flows. The flexibility of the new method is demonstrated by simulating a rising, dissolving droplet in a porous medium with user generated lattice-Boltzmann models and subroutines.
Visualization for Hyper-Heuristics. Front-End Graphical User Interface
Kroenung, Lauren
2015-03-01
Modern society is faced with ever more complex problems, many of which can be formulated as generate-and-test optimization problems. General-purpose optimization algorithms are not well suited for real-world scenarios where many instances of the same problem class need to be repeatedly and efficiently solved because they are not targeted to a particular scenario. Hyper-heuristics automate the design of algorithms to create a custom algorithm for a particular scenario. While such automated design has great advantages, it can often be difficult to understand exactly how a design was derived and why it should be trusted. This project aims to address these issues of usability by creating an easy-to-use graphical user interface (GUI) for hyper-heuristics to support practitioners, as well as scientific visualization of the produced automated designs. My contributions to this project are exhibited in the user-facing portion of the developed system and the detailed scientific visualizations created from back-end data.
The Metropolis Monte Carlo method with CUDA enabled Graphic Processing Units
Hall, Clifford; School of Physics, Astronomy, and Computational Sciences, George Mason University, 4400 University Dr., Fairfax, VA 22030 ; Ji, Weixiao; Blaisten-Barojas, Estela; School of Physics, Astronomy, and Computational Sciences, George Mason University, 4400 University Dr., Fairfax, VA 22030
2014-02-01
We present a CPU–GPU system for runtime acceleration of large molecular simulations using GPU computation and memory swaps. The memory architecture of the GPU can be used both as container for simulation data stored on the graphics card and as floating-point code target, providing an effective means for the manipulation of atomistic or molecular data on the GPU. To fully take advantage of this mechanism, efficient GPU realizations of algorithms used to perform atomistic and molecular simulations are essential. Our system implements a versatile molecular engine, including inter-molecule interactions and orientational variables for performing the Metropolis Monte Carlo (MMC) algorithm, which is one type of Markov chain Monte Carlo. By combining memory objects with floating-point code fragments we have implemented an MMC parallel engine that entirely avoids the communication time of molecular data at runtime. Our runtime acceleration system is a forerunner of a new class of CPU–GPU algorithms exploiting memory concepts combined with threading for avoiding bus bandwidth and communication. The testbed molecular system used here is a condensed phase system of oligopyrrole chains. A benchmark shows a size scaling speedup of 60 for systems with 210,000 pyrrole monomers. Our implementation can easily be combined with MPI to connect in parallel several CPU–GPU duets. -- Highlights: •We parallelize the Metropolis Monte Carlo (MMC) algorithm on one CPU—GPU duet. •The Adaptive Tempering Monte Carlo employs MMC and profits from this CPU—GPU implementation. •Our benchmark shows a size scaling-up speedup of 62 for systems with 225,000 particles. •The testbed involves a polymeric system of oligopyrroles in the condensed phase. •The CPU—GPU parallelization includes dipole—dipole and Mie—Jones classic potentials.
In-Situ Statistical Analysis of Autotune Simulation Data using Graphical Processing Units
Ranjan, Niloo; Sanyal, Jibonananda; New, Joshua Ryan
2013-08-01
Developing accurate building energy simulation models to assist energy efficiency at speed and scale is one of the research goals of the Whole-Building and Community Integration group, which is a part of Building Technologies Research and Integration Center (BTRIC) at Oak Ridge National Laboratory (ORNL). The aim of the Autotune project is to speed up the automated calibration of building energy models to match measured utility or sensor data. The workflow of this project takes input parameters and runs EnergyPlus simulations on Oak Ridge Leadership Computing Facility s (OLCF) computing resources such as Titan, the world s second fastest supercomputer. Multiple simulations run in parallel on nodes having 16 processors each and a Graphics Processing Unit (GPU). Each node produces a 5.7 GB output file comprising 256 files from 64 simulations. Four types of output data covering monthly, daily, hourly, and 15-minute time steps for each annual simulation is produced. A total of 270TB+ of data has been produced. In this project, the simulation data is statistically analyzed in-situ using GPUs while annual simulations are being computed on the traditional processors. Titan, with its recent addition of 18,688 Compute Unified Device Architecture (CUDA) capable NVIDIA GPUs, has greatly extended its capability for massively parallel data processing. CUDA is used along with C/MPI to calculate statistical metrics such as sum, mean, variance, and standard deviation leveraging GPU acceleration. The workflow developed in this project produces statistical summaries of the data which reduces by multiple orders of magnitude the time and amount of data that needs to be stored. These statistical capabilities are anticipated to be useful for sensitivity analysis of EnergyPlus simulations.
Rizzardi, M.; Mohr, M.S.; Merrill, D.W.; Selvin, S. California Univ., Berkeley, CA . School of Public Health)
1992-07-01
In 1990, the United States Bureau of the Census released detailed geographic base files known as TIGER/Line (Topologically Integrated Geographic Encoding and Referencing) which contain detail on the physical features and census tract boundaries of every county in the United States. The TIGER database is attractive for two reasons. First, it is publicly available through the Bureau of the Census on tape or cd-rom for a minimal fee. Second, it contains 24 billion characters of data which describe geographic features of interest to the Census Bureau such as coastlines, hydrography, transportation networks, political boundaries, etc. Unfortunately, the large TIGER database only provides raw alphanumeric data; no utility software, graphical or otherwise, is included. On the other hand New S, a popular statistical software package by AT T, has easily operated functions that permit advanced graphics in conjunction with data analysis. New S has the ability to plot contours, lines, segments, and points. However, of special interest is the New S function map and its options. Using the map function, which requires polygons as input, census tracts can be quickly selected, plotted, shaded, etc. New S graphics combined with the TIGER database has obvious potential. This paper reports on our efforts to use the TIGER map files with New S, especially to construct census tract maps of counties. While census tract boundaries are inherently polygonal, they are not organized as such in the TIGER database. This conversion of the TIGER line'' format into New S polygon/polyline'' format is one facet of the work reported here. Also we discuss the selection and extraction of auxiliary geographic information from TIGER files for graphical display using New S.
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
;
Office of Energy Efficiency and Renewable Energy (EERE) Indexed Site
sealed? No Percent in this location 25% Hot water Fuel Natural gas Efficiency value 0.62 EF For more information on calculation methods, technical terms and units of measure, Score...
Leonard, J.E.
1989-04-01
Two software packages for the petroleum industry are reviewed. The first, Finder, is an integrated exploration and production system using the Oracle relational data base engine. It operates on a microVAX (DEC) platforms with the user interface on an Apple Macintosh. The second program, OpenWorks, will combine petroleum applications with data over existing ''industry standards'' including X-Windows for graphics, TCP/IP for the networking protocols, and some relation model which fits over existing data bases.
Dynamic graph system for a semantic database
Mizell, David
2015-01-27
A method and system in a computer system for dynamically providing a graphical representation of a data store of entries via a matrix interface is disclosed. A dynamic graph system provides a matrix interface that exposes to an application program a graphical representation of data stored in a data store such as a semantic database storing triples. To the application program, the matrix interface represents the graph as a sparse adjacency matrix that is stored in compressed form. Each entry of the data store is considered to represent a link between nodes of the graph. Each entry has a first field and a second field identifying the nodes connected by the link and a third field with a value for the link that connects the identified nodes. The first, second, and third fields represent the rows, column, and elements of the adjacency matrix.
Dynamic graph system for a semantic database
Mizell, David
2016-04-12
A method and system in a computer system for dynamically providing a graphical representation of a data store of entries via a matrix interface is disclosed. A dynamic graph system provides a matrix interface that exposes to an application program a graphical representation of data stored in a data store such as a semantic database storing triples. To the application program, the matrix interface represents the graph as a sparse adjacency matrix that is stored in compressed form. Each entry of the data store is considered to represent a link between nodes of the graph. Each entry has a first field and a second field identifying the nodes connected by the link and a third field with a value for the link that connects the identified nodes. The first, second, and third fields represent the rows, column, and elements of the adjacency matrix.
EMAM, M; Eldib, A; Lin, M; Li, J; Chibani, O; Ma, C
2014-06-01
Purpose: An in-house Monte Carlo based treatment planning system (MC TPS) has been developed for modulated electron radiation therapy (MERT). Our preliminary MERT planning experience called for a more user friendly graphical user interface. The current work aimed to design graphical windows and tools to facilitate the contouring and planning process. Methods: Our In-house GUI MC TPS is built on a set of EGS4 user codes namely MCPLAN and MCBEAM in addition to an in-house optimization code, which was named as MCOPTIM. Patient virtual phantom is constructed using the tomographic images in DICOM format exported from clinical treatment planning systems (TPS). Treatment target volumes and critical structures were usually contoured on clinical TPS and then sent as a structure set file. In our GUI program we developed a visualization tool to allow the planner to visualize the DICOM images and delineate the various structures. We implemented an option in our code for automatic contouring of the patient body and lungs. We also created an interface window displaying a three dimensional representation of the target and also showing a graphical representation of the treatment beams. Results: The new GUI features helped streamline the planning process. The implemented contouring option eliminated the need for performing this step on clinical TPS. The auto detection option for contouring the outer patient body and lungs was tested on patient CTs and it was shown to be accurate as compared to that of clinical TPS. The three dimensional representation of the target and the beams allows better selection of the gantry, collimator and couch angles. Conclusion: An in-house GUI program has been developed for more efficient MERT planning. The application of aiding tools implemented in the program is time saving and gives better control of the planning process.
D'Azevedo, Ed F; Nintcheu Fata, Sylvain
2012-01-01
A collocation boundary element code for solving the three-dimensional Laplace equation, publicly available from \\url{http://www.intetec.org}, has been adapted to run on an Nvidia Tesla general purpose graphics processing unit (GPU). Global matrix assembly and LU factorization of the resulting dense matrix were performed on the GPU. Out-of-core techniques were used to solve problems larger than available GPU memory. The code achieved over eight times speedup in matrix assembly and about 56~Gflops/sec in the LU factorization using only 512~Mbytes of GPU memory. Details of the GPU implementation and comparisons with the standard sequential algorithm are included to illustrate the performance of the GPU code.
Jameson, Michael G; Ohanessian, Lucy; Batumalai, Vikneswary; Patel, Virendra; Holloway, Lois C
2015-06-15
There are a number of different dwell positions and time optimisation options available in the Oncentra® Brachy (Elekta Brachytherapy Solutions, Veenendaal, The Netherlands) brachytherapy treatment planning system. The purpose of this case study was to compare graphical (GRO) and inverse planning by simulated annealing (IPSA) optimisation techniques for interstitial head and neck (HN) and prostate plans considering dosimetry, modelled radiobiology outcome and planning time. Four retrospective brachytherapy patients were chosen for this study, two recurrent HN and two prostatic boosts. Manual GRO and IPSA plans were generated for each patient. Plans were compared using dose–volume histograms (DVH) and dose coverage metrics including; conformity index (CI), homogeneity index (HI) and conformity number (CN). Logit and relative seriality models were used to calculate tumour control probability (TCP) and normal tissue complication probability (NTCP). Approximate planning time was also recorded. There was no significant difference between GRO and IPSA in terms of dose metrics with mean CI of 1.30 and 1.57 (P > 0.05) respectively. IPSA achieved an average HN TCP of 0.32 versus 0.12 for GRO while for prostate there was no significant difference. Mean GRO planning times were greater than 75 min while average IPSA planning times were less than 10 min. Planning times for IPSA were greatly reduced compared to GRO and plans were dosimetrically similar. For this reason, IPSA makes for a useful planning tool in HN and prostate brachytherapy.
An Experiment on Graph Analysis Methodologies for Scenarios
Brothers, Alan J.; Whitney, Paul D.; Wolf, Katherine E.; Kuchar, Olga A.; Chin, George
2005-09-30
Visual graph representations are increasingly used to represent, display, and explore scenarios and the structure of organizations. The graph representations of scenarios are readily understood, and commercial software is available to create and manage these representations. The purpose of the research presented in this paper is to explore whether these graph representations support quantitative assessments of the underlying scenarios. The underlying structure of the scenarios is the information that is being targeted in the experiment and the extent to which the scenarios are similar in content. An experiment was designed that incorporated both the contents of the scenarios and analysts’ graph representations of the scenarios. The scenarios’ content was represented graphically by analysts, and both the structure and the semantics of the graph representation were attempted to be used to understand the content. The structure information was not found to be discriminating for the content of the scenarios in this experiment; but, the semantic information was discriminating.
Home Energy Score Interactive Graphic
Broader source: Energy.gov [DOE]
To see a complete Home Energy Score, including Home Facts and Recommendations, view the Home Energy Score Sample Report.
Snyder, James W.; Hohenstein, Edward G.; Luehr, Nathan; Martínez, Todd J.
2015-10-21
We recently presented an algorithm for state-averaged complete active space self-consistent field (SA-CASSCF) orbital optimization that capitalizes on sparsity in the atomic orbital basis set to reduce the scaling of computational effort with respect to molecular size. Here, we extend those algorithms to calculate the analytic gradient and nonadiabatic coupling vectors for SA-CASSCF. Combining the low computational scaling with acceleration from graphical processing units allows us to perform SA-CASSCF geometry optimizations for molecules with more than 1000 atoms. The new approach will make minimal energy conical intersection searches and nonadiabatic dynamics routine for molecular systems with O(10{sup 2}) atoms.
Fermilab | Graphic Standards at Fermilab | Administrative relationship...
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
communications, depicted at the same level in the document and scaled to a balanced proportion. An example of proper depiction of the Fermilab and DOE logos is the About Fermilab...
Constructing a Cleaner Economy Info Graphic
Broader source: Energy.gov [DOE]
An overview of the impact that the clean energy economy is having on the U.S. construction industry.
Graphic Standards Fermi National Accelerator Laboratory 2014
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
onto your slide, click outside of the text areas on the slide, select Insert > Choose. Search for your file and click Insert. Drag the image to position it to fit on your slide...
Analog graphic display method and apparatus
Kronberg, J.W.
1991-08-13
Disclosed are an apparatus and method for using an output device such as an LED to show the approximate analog level of a variable electrical signal wherein a modulating AC waveform is superimposed either on the signal or a reference voltage, both of which are then fed to a comparator which drives the output device. Said device flashes at a constant perceptible rate with a duty cycle which varies in response to variations in the level of the input signal. The human eye perceives these variations in duty cycle as analogous to variations in the level of the input signal. 21 figures.
GraSPI (Graphical Structured Packing Interface)
Energy Science and Technology Software Center (OSTI)
2004-06-10
GraSPI is a collection of macros (computer programs) written to work in concert with Fluent Inc. software GAMBIT and FLUENT for modeling and design of structured packing columns used in the chemical industry (the application focus is in distillation but other applications such as gas absorbers, and other chemical contactors can also be analyzed). GraSPI is an accessory to GAMBIT and FLUENT that drives the process of complex geometry creation, domain setup, and mesh generation.more » In addition, GraSPI manages automatic flow analysis in the aforementioned domain via either serial or parallel computing using FLUENT. A library of typical commercial structured packing elements is included in GraSPI, so is the capability for user-defined creation of new packings.« less
Fermilab | Graphic Standards at Fermilab | Color palette
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
Color palette PDF Version Fermilab's core color is NAL Blue, developed for Fermilab in the 1970s by the Rustoleum paint company. The laboratory's original color palette was...
O:\\GRAPHICS\\Factsheets\\NEW\\mont
Office of Legacy Management (LM)
9 United States Department of Energy Grand Junction Office Background The original Monticello mill was built in 1942 to provide an additional supply of vanadium during World War ...
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
Methane Hydrates Energy Density refers to the volume of gas stored within a given volume of space in the subsurface. When gas hydrate is dissociated - the gas that is released will expand or contract in response to ambient pressure. When that gas is brought to the surface, it will expand roughly 164x times the volume it occupies in the subsurface. This ratio is independent of the depth at which it occurs, because the rigid hydrate lattice hold methane molecules at a fixed density without regard
Analog graphic display method and apparatus
Kronberg, James W.
1991-01-01
An apparatus and method for using an output device such as an LED to show the approximate analog level of a variable electrical signal wherein a modulating AC waveform is superimposed either on the signal or a reference voltage, both of which are then fed to a comparator which drives the output device. Said device flashes at a constant perceptible rate with a duty cycle which varies in response to variations in the level of the input signal. The human eye perceives these variations in duty cycle as analogous to variations in the level of the input signal.
Sanfilippo, Antonio P.
2005-12-27
Graph theory is a branch of discrete combinatorial mathematics that studies the properties of graphs. The theory was pioneered by the Swiss mathematician Leonhard Euler in the 18th century, commenced its formal development during the second half of the 19th century, and has witnessed substantial growth during the last seventy years, with applications in areas as diverse as engineering, computer science, physics, sociology, chemistry and biology. Graph theory has also had a strong impact in computational linguistics by providing the foundations for the theory of features structures that has emerged as one of the most widely used frameworks for the representation of grammar formalisms.
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
that use data from the PDSF batch scheduler (SGE) to show the utilization of the cluster over the past 24 hours. The graphs were generated with RRDTool and are updated...
Graphical user interface for image acquisition and processing
Goldberg, Kenneth A.
2002-01-01
An event-driven GUI-based image acquisition interface for the IDL programming environment designed for CCD camera control and image acquisition directly into the IDL environment where image manipulation and data analysis can be performed, and a toolbox of real-time analysis applications. Running the image acquisition hardware directly from IDL removes the necessity of first saving images in one program and then importing the data into IDL for analysis in a second step. Bringing the data directly into IDL creates an opportunity for the implementation of IDL image processing and display functions in real-time. program allows control over the available charge coupled device (CCD) detector parameters, data acquisition, file saving and loading, and image manipulation and processing, all from within IDL. The program is built using IDL's widget libraries to control the on-screen display and user interface.
Tensor Algebra Library for NVidia Graphics Processing Units
Energy Science and Technology Software Center (OSTI)
2015-03-16
This is a general purpose math library implementing basic tensor algebra operations on NVidia GPU accelerators. This software is a tensor algebra library that can perform basic tensor algebra operations, including tensor contractions, tensor products, tensor additions, etc., on NVidia GPU accelerators, asynchronously with respect to the CPU host. It supports a simultaneous use of multiple NVidia GPUs. Each asynchronous API function returns a handle which can later be used for querying the completion ofmore » the corresponding tensor algebra operation on a specific GPU. The tensors participating in a particular tensor operation are assumed to be stored in local RAM of a node or GPU RAM. The main research area where this library can be utilized is the quantum many-body theory (e.g., in electronic structure theory).« less
XOP : a graphical user interface for spectral calculations and...
Office of Scientific and Technical Information (OSTI)
Resource Relation: Journal Name: Review of Scientific Instruments; Journal Volume: 66; Journal Issue: 9; Other Information: PBD: Sep 1996; Related Information: Sep. 1996 Research ...
Tensor Algebra Library for NVidia Graphics Processing Units
Liakh, Dmitry
2015-03-16
This is a general purpose math library implementing basic tensor algebra operations on NVidia GPU accelerators. This software is a tensor algebra library that can perform basic tensor algebra operations, including tensor contractions, tensor products, tensor additions, etc., on NVidia GPU accelerators, asynchronously with respect to the CPU host. It supports a simultaneous use of multiple NVidia GPUs. Each asynchronous API function returns a handle which can later be used for querying the completion of the corresponding tensor algebra operation on a specific GPU. The tensors participating in a particular tensor operation are assumed to be stored in local RAM of a node or GPU RAM. The main research area where this library can be utilized is the quantum many-body theory (e.g., in electronic structure theory).
Template:WebServiceGraphic | Open Energy Information
source History View New Pages Recent Changes All Special Pages Semantic SearchQuerying Get Involved Help Apps Datasets Community Login | Sign Up Search Template Edit History...
Molecular Monte Carlo Simulations Using Graphics Processing Units...
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
allocation of the GPU hardware resources. We make comparisons between the GPU and the serial CPU Monte Carlo implementations to assess speedup over conventional microprocessors....
Parallel Block Structured Adaptive Mesh Refinement on Graphics Processing Units
Beckingsale, D. A.; Gaudin, W. P.; Hornung, R. D.; Gunney, B. T.; Gamblin, T.; Herdman, J. A.; Jarvis, S. A.
2014-11-17
Block-structured adaptive mesh refinement is a technique that can be used when solving partial differential equations to reduce the number of zones necessary to achieve the required accuracy in areas of interest. These areas (shock fronts, material interfaces, etc.) are recursively covered with finer mesh patches that are grouped into a hierarchy of refinement levels. Despite the potential for large savings in computational requirements and memory usage without a corresponding reduction in accuracy, AMR adds overhead in managing the mesh hierarchy, adding complex communication and data movement requirements to a simulation. In this paper, we describe the design and implementation of a native GPU-based AMR library, including: the classes used to manage data on a mesh patch, the routines used for transferring data between GPUs on different nodes, and the data-parallel operators developed to coarsen and refine mesh data. We validate the performance and accuracy of our implementation using three test problems and two architectures: an eight-node cluster, and over four thousand nodes of Oak Ridge National Laboratory’s Titan supercomputer. Our GPU-based AMR hydrodynamics code performs up to 4.87× faster than the CPU-based implementation, and has been scaled to over four thousand GPUs using a combination of MPI and CUDA.
GATA: A graphic alignment tool for comparative sequenceanalysis
Nix, David A.; Eisen, Michael B.
2005-01-01
Several problems exist with current methods used to align DNA sequences for comparative sequence analysis. Most dynamic programming algorithms assume that conserved sequence elements are collinear. This assumption appears valid when comparing orthologous protein coding sequences. Functional constraints on proteins provide strong selective pressure against sequence inversions, and minimize sequence duplications and feature shuffling. For non-coding sequences this collinearity assumption is often invalid. For example, enhancers contain clusters of transcription factor binding sites that change in number, orientation, and spacing during evolution yet the enhancer retains its activity. Dotplot analysis is often used to estimate non-coding sequence relatedness. Yet dot plots do not actually align sequences and thus cannot account well for base insertions or deletions. Moreover, they lack an adequate statistical framework for comparing sequence relatedness and are limited to pairwise comparisons. Lastly, dot plots and dynamic programming text outputs fail to provide an intuitive means for visualizing DNA alignments.
Graphical User Interface for Simplified Neutron Transport Calculations
Schwarz, Randolph; Carter, Leland L
2011-07-18
A number of codes perform simple photon physics calculations. The nuclear industry is lacking in similar tools to perform simplified neutron physics shielding calculations. With the increased importance of performing neutron calculations for homeland security applications and defense nuclear nonproliferation tasks, having an efficient method for performing simple neutron transport calculations becomes increasingly important. Codes such as Monte Carlo N-particle (MCNP) can perform the transport calculations; however, the technical details in setting up, running, and interpreting the required simulations are quite complex and typically go beyond the abilities of most users who need a simple answer to a neutron transport calculation. The work documented in this report resulted in the development of the NucWiz program, which can create an MCNP input file for a set of simple geometries, source, and detector configurations. The user selects source, shield, and tally configurations from a set of pre-defined lists, and the software creates a complete MCNP input file that can be optionally run and the results viewed inside NucWiz.
Graphical representation of parallel algorithmic processes. Master's thesis
Williams, E.M.
1990-12-01
Algorithm animation is a visualization method used to enhance understanding of functioning of an algorithm or program. Visualization is used for many purposes, including education, algorithm research, performance analysis, and program debugging. This research applies algorithm animation techniques to programs developed for parallel architectures, with specific on the Intel iPSC/2 hypercube. While both P-time and NP-time algorithms can potentially benefit from using visualization techniques, the set of NP-complete problems provides fertile ground for developing parallel applications, since the combinatoric nature of the problems makes finding the optimum solution impractical. The primary goals for this visualization system are: Data should be displayed as it is generated. The interface to the targe program should be transparent, allowing the animation of existing programs. Flexibility - the system should be able to animate any algorithm. The resulting system incorporates and extends two AFIT products: the AFIT Algorithm Animation Research Facility (AAARF) and the Parallel Resource Analysis Software Environment (PRASE). AAARF is an algorithm animation system developed primarily for sequential programs, but is easily adaptable for use with parallel programs. PRASE is an instrumentation package that extracts system performance data from programs on the Intel hypercubes. Since performance data is an essential part of analyzing any parallel program, views of the performance data are provided as an elementary part of the system. Custom software is designed to interface these systems and to display the program data. The program chosen as the example for this study is a member of the NP-complete problem set; it is a parallel implementation of a general.
SHINES images and graphics gallery | Department of Energy
Broader source: Energy.gov (indexed) [DOE]
EPRI - The EPRI-led collaborative aims to determine the most optimal combination of ... Image: EPRI 4 of 4 This animation shows the two-way power flow between energy generators ...
Wong, Pak C.; Mackey, Patrick S.; Perrine, Kenneth A.; Foote, Harlan P.; Thomas, James J.
2008-12-23
Methods for visualizing a graph by automatically drawing elements of the graph as labels are disclosed. In one embodiment, the method comprises receiving node information and edge information from an input device and/or communication interface, constructing a graph layout based at least in part on that information, wherein the edges are automatically drawn as labels, and displaying the graph on a display device according to the graph layout. In some embodiments, the nodes are automatically drawn as labels instead of, or in addition to, the label-edges.
Energy Science and Technology Software Center (OSTI)
2007-05-22
MpiGraph consists of an MPI application called mpiGraph written in C to measure message bandwidth and an associated crunch_mpiGraph script written in Perl to process the application output into an HTMO report. The mpiGraph application is designed to inspect the health and scalability of a high-performance interconnect while under heavy load. This is useful to detect hardware and software problems in a system, such as slow nodes, links, switches, or contention in switch routing. Itmore » is also useful to characterize how interconnect performance changes with different settings or how one interconnect type compares to another.« less
Lothian, Josh; Powers, Sarah S; Sullivan, Blair D; Baker, Matthew B; Schrock, Jonathan; Poole, Stephen W
2013-12-01
The benchmarking effort within the Extreme Scale Systems Center at Oak Ridge National Laboratory seeks to provide High Performance Computing benchmarks and test suites of interest to the DoD sponsor. The work described in this report is a part of the effort focusing on graph generation. A previously developed benchmark, SystemBurn, allowed the emulation of dierent application behavior profiles within a single framework. To complement this effort, similar capabilities are desired for graph-centric problems. This report examines existing synthetic graph generator implementations in preparation for further study on the properties of their generated synthetic graphs.
Graphs, matrices, and the GraphBLAS: Seven good reasons
Kepner, Jeremy; Bader, David; Buluç, Aydın; Gilbert, John; Mattson, Timothy; Meyerhenke, Henning
2015-01-01
The analysis of graphs has become increasingly important to a wide range of applications. Graph analysis presents a number of unique challenges in the areas of (1) software complexity, (2) data complexity, (3) security, (4) mathematical complexity, (5) theoretical analysis, (6) serial performance, and (7) parallel performance. Implementing graph algorithms using matrix-based approaches provides a number of promising solutions to these challenges. The GraphBLAS standard (istcbigdata.org/GraphBlas) is being developed to bring the potential of matrix based graph algorithms to the broadest possible audience. The GraphBLAS mathematically defines a core set of matrix-based graph operations that can be used to implement a wide class of graph algorithms in a wide range of programming environments. This paper provides an introduction to the GraphBLAS and describes how the GraphBLAS can be used to address many of the challenges associated with analysis of graphs.
Graphs, matrices, and the GraphBLAS: Seven good reasons
DOE Public Access Gateway for Energy & Science Beta (PAGES Beta)
Kepner, Jeremy; Bader, David; Buluç, Aydın; Gilbert, John; Mattson, Timothy; Meyerhenke, Henning
2015-01-01
The analysis of graphs has become increasingly important to a wide range of applications. Graph analysis presents a number of unique challenges in the areas of (1) software complexity, (2) data complexity, (3) security, (4) mathematical complexity, (5) theoretical analysis, (6) serial performance, and (7) parallel performance. Implementing graph algorithms using matrix-based approaches provides a number of promising solutions to these challenges. The GraphBLAS standard (istcbigdata.org/GraphBlas) is being developed to bring the potential of matrix based graph algorithms to the broadest possible audience. The GraphBLAS mathematically defines a core set of matrix-based graph operations that can be used to implementmore » a wide class of graph algorithms in a wide range of programming environments. This paper provides an introduction to the GraphBLAS and describes how the GraphBLAS can be used to address many of the challenges associated with analysis of graphs.« less
Subdominant pseudoultrametric on graphs
Dovgoshei, A A; Petrov, E A
2013-08-31
Let (G,w) be a weighted graph. We find necessary and sufficient conditions under which the weight w:E(G)?R{sup +} can be extended to a pseudoultrametric on V(G), and establish a criterion for the uniqueness of such an extension. We demonstrate that (G,w) is a complete k-partite graph, for k?2, if and only if for any weight that can be extended to a pseudoultrametric, among all such extensions one can find the least pseudoultrametric consistent with w. We give a structural characterization of graphs for which the subdominant pseudoultrametric is an ultrametric for any strictly positive weight that can be extended to a pseudoultrametric. Bibliography: 14 titles.
PyDecay/GraphPhys: A Unified Language and Storage System for Particle Decay Process Descriptions
Dunietz, Jesse N.; /MIT /SLAC
2011-06-22
To ease the tasks of Monte Carlo (MC) simulation and event reconstruction (i.e. inferring particle-decay events from experimental data) for long-term BaBar data preservation and analysis, the following software components have been designed: a language ('GraphPhys') for specifying decay processes, common to both simulation and data analysis, allowing arbitrary parameters on particles, decays, and entire processes; an automated visualization tool to show graphically what decays have been specified; and a searchable database storage mechanism for decay specifications. Unlike HepML, a proposed XML standard for HEP metadata, the specification language is designed not for data interchange between computer systems, but rather for direct manipulation by human beings as well as computers. The components are interoperable: the information parsed from files in the specification language can easily be rendered as an image by the visualization package, and conversion between decay representations was implemented. Several proof-of-concept command-line tools were built based on this framework. Applications include building easier and more efficient interfaces to existing analysis tools for current projects (e.g. BaBar/BESII), providing a framework for analyses in future experimental settings (e.g. LHC/SuperB), and outreach programs that involve giving students access to BaBar data and analysis tools to give them a hands-on feel for scientific analysis.
Energy Science and Technology Software Center (OSTI)
2013-02-19
This library is used in several LLNL projects, including STAT (the Stack Trace Analysis Tool for scalable debugging) and some modules in P^nMPI (a tool MPI tool infrastructure). It can also be used standalone for creating and manipulationg graphs, but its API is primarily tuned to support these other projects
Temporal Representation in Semantic Graphs
Levandoski, J J; Abdulla, G M
2007-08-07
A wide range of knowledge discovery and analysis applications, ranging from business to biological, make use of semantic graphs when modeling relationships and concepts. Most of the semantic graphs used in these applications are assumed to be static pieces of information, meaning temporal evolution of concepts and relationships are not taken into account. Guided by the need for more advanced semantic graph queries involving temporal concepts, this paper surveys the existing work involving temporal representations in semantic graphs.
Simple and Flexible Scene Graph
Energy Science and Technology Software Center (OSTI)
2007-10-01
The system implements a flexible and extensible scene graph for the visualization and analysis of scientific information.
Winlaw, Manda; De Sterck, Hans; Sanders, Geoffrey
2015-10-26
In very simple terms a network can be de ned as a collection of points joined together by lines. Thus, networks can be used to represent connections between entities in a wide variety of elds including engi- neering, science, medicine, and sociology. Many large real-world networks share a surprising number of properties, leading to a strong interest in model development research and techniques for building synthetic networks have been developed, that capture these similarities and replicate real-world graphs. Modeling these real-world networks serves two purposes. First, building models that mimic the patterns and prop- erties of real networks helps to understand the implications of these patterns and helps determine which patterns are important. If we develop a generative process to synthesize real networks we can also examine which growth processes are plausible and which are not. Secondly, high-quality, large-scale network data is often not available, because of economic, legal, technological, or other obstacles [7]. Thus, there are many instances where the systems of interest cannot be represented by a single exemplar network. As one example, consider the eld of cybersecurity, where systems require testing across diverse threat scenarios and validation across diverse network structures. In these cases, where there is no single exemplar network, the systems must instead be modeled as a collection of networks in which the variation among them may be just as important as their common features. By developing processes to build synthetic models, so-called graph generators, we can build synthetic networks that capture both the essential features of a system and realistic variability. Then we can use such synthetic graphs to perform tasks such as simulations, analysis, and decision making. We can also use synthetic graphs to performance test graph analysis algorithms, including clustering algorithms and anomaly detection algorithms.
Recursive Feature Extraction in Graphs
Energy Science and Technology Software Center (OSTI)
2014-08-14
ReFeX extracts recursive topological features from graph data. The input is a graph as a csv file and the output is a csv file containing feature values for each node in the graph. The features are based on topological counts in the neighborhoods of each nodes, as well as recursive summaries of neighbors' features.
Graph Coarsening for Path Finding in Cybersecurity Graphs
Hogan, Emilie A.; Johnson, John R.; Halappanavar, Mahantesh
2013-01-01
n the pass-the-hash attack, hackers repeatedly steal password hashes and move through a computer network with the goal of reaching a computer with high level administrative privileges. In this paper we apply graph coarsening in network graphs for the purpose of detecting hackers using this attack or assessing the risk level of the network's current state. We repeatedly take graph minors, which preserve the existence of paths in the graph, and take powers of the adjacency matrix to count the paths. This allows us to detect the existence of paths as well as find paths that have high risk of being used by adversaries.
Maunz, Peter Lukas Wilhelm; Sterk, Jonathan David; Lobser, Daniel; Parekh, Ojas D.; Ryan-Anderson, Ciaran
2016-01-01
In recent years, advanced network analytics have become increasingly important to na- tional security with applications ranging from cyber security to detection and disruption of ter- rorist networks. While classical computing solutions have received considerable investment, the development of quantum algorithms to address problems, such as data mining of attributed relational graphs, is a largely unexplored space. Recent theoretical work has shown that quan- tum algorithms for graph analysis can be more efficient than their classical counterparts. Here, we have implemented a trapped-ion-based two-qubit quantum information proces- sor to address these goals. Building on Sandia's microfabricated silicon surface ion traps, we have designed, realized and characterized a quantum information processor using the hyperfine qubits encoded in two 171 Yb + ions. We have implemented single qubit gates using resonant microwave radiation and have employed Gate set tomography (GST) to characterize the quan- tum process. For the first time, we were able to prove that the quantum process surpasses the fault tolerance thresholds of some quantum codes by demonstrating a diamond norm distance of less than 1 . 9 x 10 [?] 4 . We used Raman transitions in order to manipulate the trapped ions' motion and realize two-qubit gates. We characterized the implemented motion sensitive and insensitive single qubit processes and achieved a maximal process infidelity of 6 . 5 x 10 [?] 5 . We implemented the two-qubit gate proposed by Molmer and Sorensen and achieved a fidelity of more than 97 . 7%.
A Collection of Features for Semantic Graphs
Eliassi-Rad, T; Fodor, I K; Gallagher, B
2007-05-02
Semantic graphs are commonly used to represent data from one or more data sources. Such graphs extend traditional graphs by imposing types on both nodes and links. This type information defines permissible links among specified nodes and can be represented as a graph commonly referred to as an ontology or schema graph. Figure 1 depicts an ontology graph for data from National Association of Securities Dealers. Each node type and link type may also have a list of attributes. To capture the increased complexity of semantic graphs, concepts derived for standard graphs have to be extended. This document explains briefly features commonly used to characterize graphs, and their extensions to semantic graphs. This document is divided into two sections. Section 2 contains the feature descriptions for static graphs. Section 3 extends the features for semantic graphs that vary over time.
Graph Partitioning and Sequencing Software
Energy Science and Technology Software Center (OSTI)
1995-09-19
Graph partitioning is a fundemental problem in many scientific contexts. CHACO2.0 is a software package designed to partition and sequence graphs. CHACO2.0 allows for recursive application of several methods for finding small edge separators in weighted graphs. These methods include inertial, spectral, Kernighan Lin and multilevel methods in addition to several simpler strategies. Each of these approaches can be used to partition the graph into two, four, or eight pieces at each level of recursion.more » In addition, the Kernighan Lin method can be used to improve partitions generated by any of the other algorithms. CHACO2.0 can also be used to address various graph sequencing problems, with applications to scientific computing, database design, gene sequencing and other problems.« less
Khovanov homology of graph-links
Nikonov, Igor M
2012-08-31
Graph-links arise as the intersection graphs of turning chord diagrams of links. Speaking informally, graph-links provide a combinatorial description of links up to mutations. Many link invariants can be reformulated in the language of graph-links. Khovanov homology, a well-known and useful knot invariant, is defined for graph-links in this paper (in the case of the ground field of characteristic two). Bibliography: 14 titles.
Graph Analytics for Signature Discovery
Hogan, Emilie A.; Johnson, John R.; Halappanavar, Mahantesh; Lo, Chaomei
2013-06-01
Within large amounts of seemingly unstructured data it can be diffcult to find signatures of events. In our work we transform unstructured data into a graph representation. By doing this we expose underlying structure in the data and can take advantage of existing graph analytics capabilities, as well as develop new capabilities. Currently we focus on applications in cybersecurity and communication domains. Within cybersecurity we aim to find signatures for perpetrators using the pass-the-hash attack, and in communications we look for emails or phone calls going up or down a chain of command. In both of these areas, and in many others, the signature we look for is a path with certain temporal properties. In this paper we discuss our methodology for finding these temporal paths within large graphs.
Graph modeling systems and methods
Neergaard, Mike
2015-10-13
An apparatus and a method for vulnerability and reliability modeling are provided. The method generally includes constructing a graph model of a physical network using a computer, the graph model including a plurality of terminating vertices to represent nodes in the physical network, a plurality of edges to represent transmission paths in the physical network, and a non-terminating vertex to represent a non-nodal vulnerability along a transmission path in the physical network. The method additionally includes evaluating the vulnerability and reliability of the physical network using the constructed graph model, wherein the vulnerability and reliability evaluation includes a determination of whether each terminating and non-terminating vertex represents a critical point of failure. The method can be utilized to evaluate wide variety of networks, including power grid infrastructures, communication network topologies, and fluid distribution systems.
RGUI 1.0, New Graphical User Interface for RELAP5-3D
Mesina, George Lee; Galbraith, James Andrew
1999-04-01
With the advent of three-dimensional modeling in nuclear safety analysis codes, the need has arisen for a new display methodology. Currently, analysts either sort through voluminous numerical displays of data at points in a region, or view color coded interpretations of the data on a two-dimensional rendition of the plant. RGUI 1.0 provides 3D capability for displaying data. The 3D isometric hydrodynamic image is built automatically from the input deck without additional input from the user. Standard view change features allow the user to focus on only the important data. Familiar features that are standard to the nuclear industry, such as run, interact, and monitor, are included. RGUI 1.0 reduces the difficulty of analyzing complex three dimensional plants.
RGUI 1.0, New Graphical User Interface for RELAP5-3D
G. L. Mesina; J. Galbraith
1999-04-01
With the advent of three-dimensional modeling in nuclear safety analysis codes, the need has arisen for a new display methodology. Currently, analysts either sort through voluminous numerical displays of data at points in a region, or view color coded interpretations of the data on a two-dimensional rendition of the plant. RGUI 1.0 provides 3D capability for displaying data. The 3D isometric hydrodynamic image is built automatically from the input deck without additional input from the user. Standard view change features allow the user to focus on only the important data. Familiar features that are standard to the nuclear industry, such as run, interact, and monitor, are included. RGUI 1.0 reduces the difficulty of analyzing complex three-dimensional plants.
A graphical interface based model for wind turbine drive train dynamics
Manwell, J.F.; McGowan, J.G.; Abdulwahid, U.; Rogers, A.; McNiff, B.
1996-12-31
This paper presents a summary of a wind turbine drive train dynamics code that has been under development at the University of Massachusetts, under National Renewable Energy Laboratory (NREL) support. The code is intended to be used to assist in the proper design and selection of drive train components. This work summarizes the development of the equations of motion for the model, and discusses the method of solution. In addition, a number of comparisons with analytical solutions and experimental field data are given. The summary includes conclusions and suggestions for future work on the model. 13 refs., 10 figs.
Kotter, Dale K [Shelley, ID; Rohrbaugh, David T [Idaho Falls, ID
2010-09-07
A frequency selective surface (FSS) and associated methods for modeling, analyzing and designing the FSS are disclosed. The FSS includes a pattern of conductive material formed on a substrate to form an array of resonance elements. At least one aspect of the frequency selective surface is determined by defining a frequency range including multiple frequency values, determining a frequency dependent permittivity across the frequency range for the substrate, determining a frequency dependent conductivity across the frequency range for the conductive material, and analyzing the frequency selective surface using a method of moments analysis at each of the multiple frequency values for an incident electromagnetic energy impinging on the frequency selective surface. The frequency dependent permittivity and the frequency dependent conductivity are included in the method of moments analysis.
Whitney, Paul D.; Powers, Mary E.; Chin, George; Johnson, Katherine E.; Kuchar, Olga A.; Sloughter, James M.
2004-06-15
Analysis goals associated with collections of scenarios include: • Summarizing a collection of scenarios – the summarization could take the form of identification of typical patterns of relations within the collection • Retrieve similar scenarios/cases – the retrieval could be based on parts of scenarios; the similar ‘complete’ scenarios could be used to suggest additional lines of investigation • Find “relation paths” between key actors A signature-vector strategy is presented that addresses the above analysis goals. The broad strategy has been applied in various settings.
Microsoft Word - 2014_0305_Graphic_PIC_ThreeMonth_WorkPlan.docx
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
03/05/14 during PIC meeting (Blue titles indicate those topics specifically identified in the 2014 HAB Work Plan) March April May RAD Primer (placeholder) o How workers are protected from exposure? o HSEP lead Public Understanding of... o Involving an informed public on issues like cleanup levels, waste streams and classifications, risk, budget, etc. o Check in with John Howieson & other issue managers on the tank waste classification FAQs (next step: Ken & Liz to set-up IM call) 100-F
Large-scale Nanostructure Simulations from X-ray Scattering Data On Graphics Processor Clusters
Sarje, Abhinav; Pien, Jack; Li, Xiaoye; Chan, Elaine; Chourou, Slim; Hexemer, Alexander; Scholz, Arthur; Kramer, Edward
2012-01-15
X-ray scattering is a valuable tool for measuring the structural properties of materialsused in the design and fabrication of energy-relevant nanodevices (e.g., photovoltaic, energy storage, battery, fuel, and carbon capture andsequestration devices) that are key to the reduction of carbon emissions. Although today's ultra-fast X-ray scattering detectors can provide tremendousinformation on the structural properties of materials, a primary challenge remains in the analyses of the resulting data. We are developing novelhigh-performance computing algorithms, codes, and software tools for the analyses of X-ray scattering data. In this paper we describe two such HPCalgorithm advances. Firstly, we have implemented a flexible and highly efficient Grazing Incidence Small Angle Scattering (GISAXS) simulation code based on theDistorted Wave Born Approximation (DWBA) theory with C++/CUDA/MPI on a cluster of GPUs. Our code can compute the scattered light intensity from any givensample in all directions of space; thus allowing full construction of the GISAXS pattern. Preliminary tests on a single GPU show speedups over 125x compared tothe sequential code, and almost linear speedup when executing across a GPU cluster with 42 nodes, resulting in an additional 40x speedup compared to usingone GPU node. Secondly, for the structural fitting problems in inverse modeling, we have implemented a Reverse Monte Carlo simulation algorithm with C++/CUDAusing one GPU. Since there are large numbers of parameters for fitting in the in X-ray scattering simulation model, the earlier single CPU code required weeks ofruntime. Deploying the AccelerEyes Jacket/Matlab wrapper to use GPU gave around 100x speedup over the pure CPU code. Our further C++/CUDA optimization deliveredan additional 9x speedup.
O:\\GRAPHICS\\FACTSH~1\\MONTICEL\\millsite_restoration_web.pmd
Office of Legacy Management (LM)
July 2001 United States Department of Energy Grand Junction Office The U.S. Department of Energy transferred a former uranium mill tailings site and adjacent government-owned ...
Dr.L: Distributed Recursive (Graph) Layout
Energy Science and Technology Software Center (OSTI)
2007-11-19
Dr. L provides two-dimensional visualizations of very large abstract graph structures. it can be used for data mining applications including biology, scientific literature, and social network analysis. Dr. L is a graph layout program that uses a multilevel force-directed algorithm. A graph is input and drawn using a force-directed algorithm based on simulated annealing. The resulting layout is clustered using a single link algorithm. This clustering is used to produce a coarsened graph (fewer nodes)more » which is then re-drawn. this process is repeated until a sufficiently small graph is produced. The smallest graph is drawn and then used as a basis for drawing the original graph by refining the series of coarsened graphs that were produced. The layout engine can be run in serial or in parallel.« less
API Requirements for Dynamic Graph Prediction
Gallagher, B; Eliassi-Rad, T
2006-10-13
Given a large-scale time-evolving multi-modal and multi-relational complex network (a.k.a., a large-scale dynamic semantic graph), we want to implement algorithms that discover patterns of activities on the graph and learn predictive models of those discovered patterns. This document outlines the application programming interface (API) requirements for fast prototyping of feature extraction, learning, and prediction algorithms on large dynamic semantic graphs. Since our algorithms must operate on large-scale dynamic semantic graphs, we have chosen to use the graph API developed in the CASC Complex Networks Project. This API is supported on the back end by a semantic graph database (developed by Scott Kohn and his team). The advantages of using this API are (i) we have full-control of its development and (ii) the current API meets almost all of the requirements outlined in this document.
Fast generation of sparse random kernel graphs
Hagberg, Aric; Lemons, Nathan; Du, Wen -Bo
2015-09-10
The development of kernel-based inhomogeneous random graphs has provided models that are flexible enough to capture many observed characteristics of real networks, and that are also mathematically tractable. We specify a class of inhomogeneous random graph models, called random kernel graphs, that produces sparse graphs with tunable graph properties, and we develop an efficient generation algorithm to sample random instances from this model. As real-world networks are usually large, it is essential that the run-time of generation algorithms scales better than quadratically in the number of vertices n. We show that for many practical kernels our algorithm runs in time at most ο(n(logn)²). As an example, we show how to generate samples of power-law degree distribution graphs with tunable assortativity.
Fast generation of sparse random kernel graphs
DOE Public Access Gateway for Energy & Science Beta (PAGES Beta)
Hagberg, Aric; Lemons, Nathan; Du, Wen -Bo
2015-09-10
The development of kernel-based inhomogeneous random graphs has provided models that are flexible enough to capture many observed characteristics of real networks, and that are also mathematically tractable. We specify a class of inhomogeneous random graph models, called random kernel graphs, that produces sparse graphs with tunable graph properties, and we develop an efficient generation algorithm to sample random instances from this model. As real-world networks are usually large, it is essential that the run-time of generation algorithms scales better than quadratically in the number of vertices n. We show that for many practical kernels our algorithm runs in timemore » at most ο(n(logn)²). As an example, we show how to generate samples of power-law degree distribution graphs with tunable assortativity.« less
Graph algorithms in the titan toolkit.
McLendon, William Clarence, III; Wylie, Brian Neil
2009-10-01
Graph algorithms are a key component in a wide variety of intelligence analysis activities. The Graph-Based Informatics for Non-Proliferation and Counter-Terrorism project addresses the critical need of making these graph algorithms accessible to Sandia analysts in a manner that is both intuitive and effective. Specifically we describe the design and implementation of an open source toolkit for doing graph analysis, informatics, and visualization that provides Sandia with novel analysis capability for non-proliferation and counter-terrorism.
Enabling Graph Appliance for Genome Assembly
Singh, Rina; Graves, Jeffrey A; Lee, Sangkeun; Sukumar, Sreenivas R; Shankar, Mallikarjun
2015-01-01
In recent years, there has been a huge growth in the amount of genomic data available as reads generated from various genome sequencers. The number of reads generated can be huge, ranging from hundreds to billions of nucleotide, each varying in size. Assembling such large amounts of data is one of the challenging computational problems for both biomedical and data scientists. Most of the genome assemblers developed have used de Bruijn graph techniques. A de Bruijn graph represents a collection of read sequences by billions of vertices and edges, which require large amounts of memory and computational power to store and process. This is the major drawback to de Bruijn graph assembly. Massively parallel, multi-threaded, shared memory systems can be leveraged to overcome some of these issues. The objective of our research is to investigate the feasibility and scalability issues of de Bruijn graph assembly on Cray s Urika-GD system; Urika-GD is a high performance graph appliance with a large shared memory and massively multithreaded custom processor designed for executing SPARQL queries over large-scale RDF data sets. However, to the best of our knowledge, there is no research on representing a de Bruijn graph as an RDF graph or finding Eulerian paths in RDF graphs using SPARQL for potential genome discovery. In this paper, we address the issues involved in representing a de Bruin graphs as RDF graphs and propose an iterative querying approach for finding Eulerian paths in large RDF graphs. We evaluate the performance of our implementation on real world ebola genome datasets and illustrate how genome assembly can be accomplished with Urika-GD using iterative SPARQL queries.
Useful Graphs and Charts - Ion Beams - Radiation Effects Facility...
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
Times 15 MeVu LET vs Range Graph 25 MeVu LET vs Range Graph 40 Mevu LET vs Range Graph Radiation Effects Facility Cyclotron Institute Texas A&M University MS 3366 ...
Graph Mining Meets the Semantic Web
Lee, Sangkeun; Sukumar, Sreenivas R; Lim, Seung-Hwan
2015-01-01
The Resource Description Framework (RDF) and SPARQL Protocol and RDF Query Language (SPARQL) were introduced about a decade ago to enable flexible schema-free data interchange on the Semantic Web. Today, data scientists use the framework as a scalable graph representation for integrating, querying, exploring and analyzing data sets hosted at different sources. With increasing adoption, the need for graph mining capabilities for the Semantic Web has emerged. We address that need through implementation of three popular iterative Graph Mining algorithms (Triangle count, Connected component analysis, and PageRank). We implement these algorithms as SPARQL queries, wrapped within Python scripts. We evaluate the performance of our implementation on 6 real world data sets and show graph mining algorithms (that have a linear-algebra formulation) can indeed be unleashed on data represented as RDF graphs using the SPARQL query interface.
GraphReduce: Processing Large-Scale Graphs on Accelerator-Based Systems
Sengupta, Dipanjan; Song, Shuaiwen; Agarwal, Kapil; Schwan, Karsten
2015-11-15
Recent work on real-world graph analytics has sought to leverage the massive amount of parallelism offered by GPU devices, but challenges remain due to the inherent irregularity of graph algorithms and limitations in GPU-resident memory for storing large graphs. We present GraphReduce, a highly efficient and scalable GPU-based framework that operates on graphs that exceed the device’s internal memory capacity. GraphReduce adopts a combination of edge- and vertex-centric implementations of the Gather-Apply-Scatter programming model and operates on multiple asynchronous GPU streams to fully exploit the high degrees of parallelism in GPUs with efficient graph data movement between the host and device.
GraphReduce: Large-Scale Graph Analytics on Accelerator-Based HPC Systems
Sengupta, Dipanjan; Agarwal, Kapil; Song, Shuaiwen; Schwan, Karsten
2015-09-30
Recent work on real-world graph analytics has sought to leverage the massive amount of parallelism offered by GPU devices, but challenges remain due to the inherent irregularity of graph algorithms and limitations in GPU-resident memory for storing large graphs. We present GraphReduce, a highly efficient and scalable GPU-based framework that operates on graphs that exceed the device’s internal memory capacity. GraphReduce adopts a combination of both edge- and vertex-centric implementations of the Gather-Apply-Scatter programming model and operates on multiple asynchronous GPU streams to fully exploit the high degrees of parallelism in GPUs with efficient graph data movement between the host and the device.
Enabling Graph Mining in RDF Triplestores using SPARQL for Holistic In-situ Graph Analysis
Lee, Sangkeun; Sukumar, Sreenivas R; Hong, Seokyong; Lim, Seung-Hwan
2016-01-01
The graph analysis is now considered as a promising technique to discover useful knowledge in data with a new perspective. We envi- sion that there are two dimensions of graph analysis: OnLine Graph Analytic Processing (OLGAP) and Graph Mining (GM) where each respectively focuses on subgraph pattern matching and automatic knowledge discovery in graph. Moreover, as these two dimensions aim to complementarily solve complex problems, holistic in-situ graph analysis which covers both OLGAP and GM in a single system is critical for minimizing the burdens of operating multiple graph systems and transferring intermediate result-sets between those systems. Nevertheless, most existing graph analysis systems are only capable of one dimension of graph analysis. In this work, we take an approach to enabling GM capabilities (e.g., PageRank, connected-component analysis, node eccentricity, etc.) in RDF triplestores, which are originally developed to store RDF datasets and provide OLGAP capability. More specifically, to achieve our goal, we implemented six representative graph mining algorithms using SPARQL. The approach allows a wide range of available RDF data sets directly applicable for holistic graph analysis within a system. For validation of our approach, we evaluate performance of our implementations with nine real-world datasets and three different computing environments - a laptop computer, an Amazon EC2 instance, and a shared-memory Cray XMT2 URIKA-GD graph-processing appliance. The experimen- tal results show that our implementation can provide promising and scalable performance for real world graph analysis in all tested environments. The developed software is publicly available in an open-source project that we initiated.
Enabling Graph Mining in RDF Triplestores using SPARQL for Holistic In-situ Graph Analysis
DOE Public Access Gateway for Energy & Science Beta (PAGES Beta)
Lee, Sangkeun; Sukumar, Sreenivas R; Hong, Seokyong; Lim, Seung-Hwan
2016-01-01
The graph analysis is now considered as a promising technique to discover useful knowledge in data with a new perspective. We envi- sion that there are two dimensions of graph analysis: OnLine Graph Analytic Processing (OLGAP) and Graph Mining (GM) where each respectively focuses on subgraph pattern matching and automatic knowledge discovery in graph. Moreover, as these two dimensions aim to complementarily solve complex problems, holistic in-situ graph analysis which covers both OLGAP and GM in a single system is critical for minimizing the burdens of operating multiple graph systems and transferring intermediate result-sets between those systems. Nevertheless, most existingmore » graph analysis systems are only capable of one dimension of graph analysis. In this work, we take an approach to enabling GM capabilities (e.g., PageRank, connected-component analysis, node eccentricity, etc.) in RDF triplestores, which are originally developed to store RDF datasets and provide OLGAP capability. More specifically, to achieve our goal, we implemented six representative graph mining algorithms using SPARQL. The approach allows a wide range of available RDF data sets directly applicable for holistic graph analysis within a system. For validation of our approach, we evaluate performance of our implementations with nine real-world datasets and three different computing environments - a laptop computer, an Amazon EC2 instance, and a shared-memory Cray XMT2 URIKA-GD graph-processing appliance. The experimen- tal results show that our implementation can provide promising and scalable performance for real world graph analysis in all tested environments. The developed software is publicly available in an open-source project that we initiated.« less
The MultiThreaded Graph Library (MTGL)
Energy Science and Technology Software Center (OSTI)
2008-07-17
The MultiThreaded Graph Library (MTGL) is a set of header files that implement graph algorithm in such a way that they can run on massively multithreaded architectures. It is based upon the Boost Graph Library, but doesnÃÂÃÂ¢ÃÂÃÂÃÂÃÂt use Boost since the latter doesnÃÂÃÂ¢ÃÂÃÂÃÂÃÂt run well on these architectures.
Bipartite graph partitioning and data clustering
Zha, Hongyuan; He, Xiaofeng; Ding, Chris; Gu, Ming; Simon, Horst D.
2001-05-07
Many data types arising from data mining applications can be modeled as bipartite graphs, examples include terms and documents in a text corpus, customers and purchasing items in market basket analysis and reviewers and movies in a movie recommender system. In this paper, the authors propose a new data clustering method based on partitioning the underlying biopartite graph. The partition is constructed by minimizing a normalized sum of edge weights between unmatched pairs of vertices of the bipartite graph. They show that an approximate solution to the minimization problem can be obtained by computing a partial singular value decomposition (SVD) of the associated edge weight matrix of the bipartite graph. They point out the connection of their clustering algorithm to correspondence analysis used in multivariate analysis. They also briefly discuss the issue of assigning data objects to multiple clusters. In the experimental results, they apply their clustering algorithm to the problem of document clustering to illustrate its effectiveness and efficiency.
Visualization Graph | OpenEI Community
8 August, 2012 - 12:37 New Gapminder Visualizations Added EIA Energy data Gapminder OECD OpenEI SEDS Visualization Graph OpenEI now features some cool new Gapminder...
Fault-tolerant dynamic task graph scheduling
Kurt, Mehmet C.; Krishnamoorthy, Sriram; Agrawal, Kunal; Agrawal, Gagan
2014-11-16
In this paper, we present an approach to fault tolerant execution of dynamic task graphs scheduled using work stealing. In particular, we focus on selective and localized recovery of tasks in the presence of soft faults. We elicit from the user the basic task graph structure in terms of successor and predecessor relationships. The work stealing-based algorithm to schedule such a task graph is augmented to enable recovery when the data and meta-data associated with a task get corrupted. We use this redundancy, and the knowledge of the task graph structure, to selectively recover from faults with low space and time overheads. We show that the fault tolerant design retains the essential properties of the underlying work stealing-based task scheduling algorithm, and that the fault tolerant execution is asymptotically optimal when task re-execution is taken into account. Experimental evaluation demonstrates the low cost of recovery under various fault scenarios.
Bayati Kim Saberi random graph sampler
Energy Science and Technology Software Center (OSTI)
2012-06-05
This software package implements the algorithm from a paper by Bayati, Kim, and Saberi (first reference below) to generate a uniformly random sample of a graph with a prescribed degree distribution.
Accelerating semantic graph databases on commodity clusters
Morari, Alessandro; Castellana, Vito G.; Haglin, David J.; Feo, John T.; Weaver, Jesse R.; Tumeo, Antonino; Villa, Oreste
2013-10-06
We are developing a full software system for accelerating semantic graph databases on commodity cluster that scales to hundreds of nodes while maintaining constant query throughput. Our framework comprises a SPARQL to C++ compiler, a library of parallel graph methods and a custom multithreaded runtime layer, which provides a Partitioned Global Address Space (PGAS) programming model with fork/join parallelism and automatic load balancing over a commodity clusters. We present preliminary results for the compiler and for the runtime.
Graph representation of protein free energy landscape
Li, Minghai; Duan, Mojie; Fan, Jue; Huo, Shuanghong; Han, Li
2013-11-14
The thermodynamics and kinetics of protein folding and protein conformational changes are governed by the underlying free energy landscape. However, the multidimensional nature of the free energy landscape makes it difficult to describe. We propose to use a weighted-graph approach to depict the free energy landscape with the nodes on the graph representing the conformational states and the edge weights reflecting the free energy barriers between the states. Our graph is constructed from a molecular dynamics trajectory and does not involve projecting the multi-dimensional free energy landscape onto a low-dimensional space defined by a few order parameters. The calculation of free energy barriers was based on transition-path theory using the MSMBuilder2 package. We compare our graph with the widely used transition disconnectivity graph (TRDG) which is constructed from the same trajectory and show that our approach gives more accurate description of the free energy landscape than the TRDG approach even though the latter can be organized into a simple tree representation. The weighted-graph is a general approach and can be used on any complex system.
Continuous-time quantum walks on star graphs
Salimi, S.
2009-06-15
In this paper, we investigate continuous-time quantum walk on star graphs. It is shown that quantum central limit theorem for a continuous-time quantum walk on star graphs for N-fold star power graph, which are invariant under the quantum component of adjacency matrix, converges to continuous-time quantum walk on K{sub 2} graphs (complete graph with two vertices) and the probability of observing walk tends to the uniform distribution.
St Germain, Shawn Walter; Farris, Ronald Keith; Thomas, Kenneth David
2015-09-01
The long-term viability of existing nuclear power plants in the United States (U.S.) is dependent upon a number of factors, including maintaining high capacity factors, maintaining nuclear safety, and reducing operating costs, particularly those associated with refueling outages. Refueling outages typically take 20-30 days, and for existing light water NPPs in the U.S., the reactor cannot be in operation during the outage. Furthermore, given that many NPPs generate between $1-1.5 million/day in revenue when in operation, there is considerable interest in shortening the length of refueling outages. Yet refueling outages are highly complex operations, involving multiple concurrent and dependent activities that are somewhat challenging to coordinate; therefore, finding ways to improve refueling outage performance, while maintaining nuclear safety has proven to be difficult. The Advanced Outage Control Center (AOCC) project is a research and development (R&D) demonstration activity under the LWRS Program. LWRS is an R&D program that works closely with industry R&D programs to establish technical foundations for the licensing and managing of long-term, safe, and economical operation of current fleet of NPPs. As such, the LWRS Advanced Outage Control Center project has the goal of improving the management of commercial NPP refueling outages. To accomplish this goal, INL is developing an advanced outage control center (OCC) that is specifically designed to maximize the usefulness of communication and collaboration technologies for outage coordination and problem resolution activities. The overall focus is on developing an AOCC with the following capabilities that enables plant and OCC staff to; Collaborate in real-time to address emergent issues; Effectively communicate outage status to all workers involved in the outage; Effectively communicate discovered conditions in the field to the OCC; Provide real-time work status; Provide automatic pending support notifications; Provide real-time requirements monitoring; Maximize their collective situational awareness to improve decision-making; and Leverage macro data to better support resource allocation. INL has partnered with several commercial NPP utilities to develop a number of advanced outage management technologies. These outage management technologies have focused on both collaborative technologies for control centers and developing mobile technologies for NPP field workers. This report describes recent efforts made in developing a suite of outage technologies to support more effective schedule management. Currently, a master outage schedule is created months in advance using the plant’s existing scheduling software (e.g., Primavera P6). Typically, during the outage, the latest version of the schedule is printed at the beginning of each shift. INL and its partners are developing technologies that will have capabilities such as Automatic Schedule Updating, Automatic Pending Support Notifications, and the ability to allocate and schedule outage support task resources on a sub-hour basis (e.g., outage Micro-Scheduling). The remaining sections of this report describe in more detail the scheduling challenges that occur during outages, how the outage scheduling technologies INL is developing helps address those challenges, and the latest developments on this task (e.g., work accomplished to date and the path forward)
Computer and graphics modeling of heat transfer and phase change in a wall with randomly imbibed PCM
Solomon, A.D.
1989-03-01
We describe the theoretical basis and computer implementation of a simulation code for heat transfer and phase change in a rectangular 2-dimensional region in which PCM has been randomly placed with a preassigned volume fraction.
Communication Graph Generator for Parallel Programs
Energy Science and Technology Software Center (OSTI)
2014-04-08
Graphator is a collection of relatively simple sequential programs that generate communication graphs/matrices for commonly occurring patterns in parallel programs. Currently, there is support for five communication patterns: two-dimensional 4-point stencil, four-dimensional 8-point stencil, all-to-alls over sub-communicators, random near-neighbor communication, and near-neighbor communication.
GraphPrints: Towards a Graph Analytic Method for Network Anomaly Detection
Harshaw, Chris R; Bridges, Robert A; Iannacone, Michael D; Reed, Joel W; Goodall, John R
2016-01-01
This paper introduces a novel graph-analytic approach for detecting anomalies in network flow data called \\textit{GraphPrints}. Building on foundational network-mining techniques, our method represents time slices of traffic as a graph, then counts graphlets\\textemdash small induced subgraphs that describe local topology. By performing outlier detection on the sequence of graphlet counts, anomalous intervals of traffic are identified, and furthermore, individual IPs experiencing abnormal behavior are singled-out. Initial testing of GraphPrints is performed on real network data with an implanted anomaly. Evaluation shows false positive rates bounded by 2.84\\% at the time-interval level, and 0.05\\% at the IP-level with 100\\% true positive rates at both.
Perkins, S.T.; Cullen, D.E. ); Seltzer, S.M. , Gaithersburg, MD . Center for Radiation Research)
1991-11-12
Energy-dependent evaluated electron interaction cross sections and related parameters are presented for elements H through Fm (Z = 1 to 100). Data are given over the energy range from 10 eV to 100 GeV. Cross sections and average energy deposits are presented in tabulated and graphic form. In addition, ionization cross sections and average energy deposits for each shell are presented in graphic form. This information is derived from the Livermore Evaluated Electron Data Library (EEDL) as of July, 1991.
StreamWorks - A system for Dynamic Graph Search
Choudhury, Sutanay; Holder, Larry; Chin, George; Ray, Abhik; Beus, Sherman J.; Feo, John T.
2013-06-11
Acting on time-critical events by processing ever growing social media, news or cyber data streams is a major technical challenge. Many of these data sources can be modeled as multi-relational graphs. Mining and searching for subgraph patterns in a continuous setting requires an efficient approach to incremental graph search. The goal of our work is to enable real-time search capabilities for graph databases. This demonstration will present a dynamic graph query system that leverages the structural and semantic characteristics of the underlying multi-relational graph.
Modular Environment for Graph Research and Analysis with a Persistent
Energy Science and Technology Software Center (OSTI)
2009-11-18
The MEGRAPHS software package provides a front-end to graphs and vectors residing on special-purpose computing resources. It allows these data objects to be instantiated, destroyed, and manipulated. A variety of primitives needed for typical graph analyses are provided. An example program illustrating how MEGRAPHS can be used to implement a PageRank computation is included in the distribution.The MEGRAPHS software package is targeted towards developers of graph algorithms. Programmers using MEGRAPHS would write graph analysis programsmore » in terms of high-level graph and vector operations. These computations are transparently executed on the Cray XMT compute nodes.« less
Knowledge Representation Issues in Semantic Graphs for Relationship Detection
Barthelemy, M; Chow, E; Eliassi-Rad, T
2005-02-02
An important task for Homeland Security is the prediction of threat vulnerabilities, such as through the detection of relationships between seemingly disjoint entities. A structure used for this task is a ''semantic graph'', also known as a ''relational data graph'' or an ''attributed relational graph''. These graphs encode relationships as typed links between a pair of typed nodes. Indeed, semantic graphs are very similar to semantic networks used in AI. The node and link types are related through an ontology graph (also known as a schema). Furthermore, each node has a set of attributes associated with it (e.g., ''age'' may be an attribute of a node of type ''person''). Unfortunately, the selection of types and attributes for both nodes and links depends on human expertise and is somewhat subjective and even arbitrary. This subjectiveness introduces biases into any algorithm that operates on semantic graphs. Here, we raise some knowledge representation issues for semantic graphs and provide some possible solutions using recently developed ideas in the field of complex networks. In particular, we use the concept of transitivity to evaluate the relevance of individual links in the semantic graph for detecting relationships. We also propose new statistical measures for semantic graphs and illustrate these semantic measures on graphs constructed from movies and terrorism data.
Frequent Subgraph Discovery in Large Attributed Streaming Graphs
Ray, Abhik; Holder, Larry; Choudhury, Sutanay
2014-08-13
The problem of finding frequent subgraphs in large dynamic graphs has so far only consid- ered a dynamic graph as being represented by a series of static snapshots taken at various points in time. This representation of a dynamic graph does not lend itself well to real time processing of real world graphs like social networks or internet traffic which consist of a stream of nodes and edges. In this paper we propose an algorithm that discovers the frequent subgraphs present in a graph represented by a stream of labeled nodes and edges. Our algorithm is efficient and consists of tunable parameters that can be tuned by the user to get interesting patterns from various kinds of graph data. In our model updates to the graph arrive in the form of batches which contain new nodes and edges. Our algorithm con- tinuously reports the frequent subgraphs that are estimated to be found in the entire graph as each batch arrives. We evaluate our system using 5 large dynamic graph datasets: the Hetrec 2011 challenge data, Twitter, DBLP and two synthetic. We evaluate our approach against two popular large graph miners, i.e., SUBDUE and GERM. Our experimental re- sults show that we can find the same frequent subgraphs as a non-incremental approach applied to snapshot graphs, and in less time.
Graph processing platforms at scale: practices and experiences
Lim, Seung-Hwan; Lee, Sangkeun; Brown, Tyler C; Sukumar, Sreenivas R; Ganesh, Gautam
2015-01-01
Graph analysis unveils hidden associations of data in many phenomena and artifacts, such as road network, social networks, genomic information, and scientific collaboration. Unfortunately, a wide diversity in the characteristics of graphs and graph operations make it challenging to find a right combination of tools and implementation of algorithms to discover desired knowledge from the target data set. This study presents an extensive empirical study of three representative graph processing platforms: Pegasus, GraphX, and Urika. Each system represents a combination of options in data model, processing paradigm, and infrastructure. We benchmarked each platform using three popular graph operations, degree distribution, connected components, and PageRank over a variety of real-world graphs. Our experiments show that each graph processing platform shows different strength, depending the type of graph operations. While Urika performs the best in non-iterative operations like degree distribution, GraphX outputforms iterative operations like connected components and PageRank. In addition, we discuss challenges to optimize the performance of each platform over large scale real world graphs.
Jargon and Graph Modularity on Twitter
Dowling, Chase P.; Corley, Courtney D.; Farber, Robert M.; Reynolds, William
2013-09-01
The language of conversation is just as dependent upon word choice as it is on who is taking part. Twitter provides an excellent test-bed in which to conduct experiments not only on language usage but on who is using what language with whom. To this end, we combine large scale graph analytical techniques with known socio-linguistic methods. In this article we leverage both expert curated vocabularies and naive mathematical graph analyses to determine if network behavior on Twitter corroborates with the current understanding of language usage. The results reported indicate that, based on networks constructed from user to user communication and communities identified using the Clauset- Newman greedy modularity algorithm we find that more prolific users of these curated vocabularies are concentrated in distinct network communities.
Mining Graphs for Understanding Time-Varying Volumetric Data...
Office of Scientific and Technical Information (OSTI)
SciTech Connect Search Results Journal Article: Mining Graphs for Understanding ... DOE Contract Number: AC02-06CH11357 Resource Type: Journal Article Resource Relation: ...
A Graph Search Heuristic for Shortest Distance Paths
Chow, E
2005-03-24
This paper presents a heuristic for guiding A* search for finding the shortest distance path between two vertices in a connected, undirected, and explicitly stored graph. The heuristic requires a small amount of data to be stored at each vertex. The heuristic has application to quickly detecting relationships between two vertices in a large information or knowledge network. We compare the performance of this heuristic with breadth-first search on graphs with various topological properties. The results show that one or more orders of magnitude improvement in the number of vertices expanded is possible for large graphs, including Poisson random graphs.
Sequoia supercomputer tops Graph 500 | National Nuclear Security...
National Nuclear Security Administration (NNSA)
Lawrence Livermore National Laboratory scientists' search for new ways to solve large complex national security problems led to the top ranking on Graph 500 and new techniques for ...
Two linear time, low overhead algorithms for graph layout
Energy Science and Technology Software Center (OSTI)
2008-01-10
The software comprises two algorithms designed to perform a 2D layout of a graph structure in time linear with respect to the vertices and edges in the graph, whereas most other layout algorithms have a running time that is quadratic with respect to the number of vertices or greater. Although these layout algorithms run in a fraction of the time as their competitors, they provide competitive results when applied to most real-world graphs. These algorithmsmore » also have a low constant running time and small memory footprint, making them useful for small to large graphs.« less
International energy indicators. [Statistical tables and graphs
Bauer, E.K.
1980-05-01
International statistical tables and graphs are given for the following: (1) Iran - Crude Oil Capacity, Production and Shut-in, June 1974-April 1980; (2) Saudi Arabia - Crude Oil Capacity, Production, and Shut-in, March 1974-Apr 1980; (3) OPEC (Ex-Iran and Saudi Arabia) - Capacity, Production and Shut-in, June 1974-March 1980; (4) Non-OPEC Free World and US Production of Crude Oil, January 1973-February 1980; (5) Oil Stocks - Free World, US, Japan, and Europe (Landed, 1973-1st Quarter, 1980); (6) Petroleum Consumption by Industrial Countries, January 1973-December 1979; (7) USSR Crude Oil Production and Exports, January 1974-April 1980; and (8) Free World and US Nuclear Generation Capacity, January 1973-March 1980. Similar statistical tables and graphs included for the United States include: (1) Imports of Crude Oil and Products, January 1973-April 1980; (2) Landed Cost of Saudi Oil in Current and 1974 Dollars, April 1974-January 1980; (3) US Trade in Coal, January 1973-March 1980; (4) Summary of US Merchandise Trade, 1976-March 1980; and (5) US Energy/GNP Ratio, 1947 to 1979.
Scaling Semantic Graph Databases in Size and Performance
Morari, Alessandro; Castellana, Vito G.; Villa, Oreste; Tumeo, Antonino; Weaver, Jesse R.; Haglin, David J.; Choudhury, Sutanay; Feo, John T.
2014-08-06
In this paper we present SGEM, a full software system for accelerating large-scale semantic graph databases on commodity clusters. Unlike current approaches, SGEM addresses semantic graph databases by only employing graph methods at all the levels of the stack. On one hand, this allows exploiting the space efficiency of graph data structures and the inherent parallelism of graph algorithms. These features adapt well to the increasing system memory and core counts of modern commodity clusters. On the other hand, however, these systems are optimized for regular computation and batched data transfers, while graph methods usually are irregular and generate fine-grained data accesses with poor spatial and temporal locality. Our framework comprises a SPARQL to data parallel C compiler, a library of parallel graph methods and a custom, multithreaded runtime system. We introduce our stack, motivate its advantages with respect to other solutions and show how we solved the challenges posed by irregular behaviors. We present the result of our software stack on the Berlin SPARQL benchmarks with datasets up to 10 billion triples (a triple corresponds to a graph edge), demonstrating scaling in dataset size and in performance as more nodes are added to the cluster.
EAGLE: 'EAGLE'Is an' Algorithmic Graph Library for Exploration
2015-01-16
The Resource Description Framework (RDF) and SPARQL Protocol and RDF Query Language (SPARQL) were introduced about a decade ago to enable flexible schema-free data interchange on the Semantic Web. Today data scientists use the framework as a scalable graph representation for integrating, querying, exploring and analyzing data sets hosted at different sources. With increasing adoption, the need for graph mining capabilities for the Semantic Web has emerged. Today there is no tools to conduct "graph mining" on RDF standard data sets. We address that need through implementation of popular iterative Graph Mining algorithms (Triangle count, Connected component analysis, degree distribution, diversity degree, PageRank, etc.). We implement these algorithms as SPARQL queries, wrapped within Python scripts and call our software tool as EAGLE. In RDF style, EAGLE stands for "EAGLE 'Is an' algorithmic graph library for exploration. EAGLE is like 'MATLAB' for 'Linked Data.'
Query optimization for graph analytics on linked data using SPARQL
Hong, Seokyong; Lee, Sangkeun; Lim, Seung -Hwan; Sukumar, Sreenivas R.; Vatsavai, Ranga Raju
2015-07-01
Triplestores that support query languages such as SPARQL are emerging as the preferred and scalable solution to represent data and meta-data as massive heterogeneous graphs using Semantic Web standards. With increasing adoption, the desire to conduct graph-theoretic mining and exploratory analysis has also increased. Addressing that desire, this paper presents a solution that is the marriage of Graph Theory and the Semantic Web. We present software that can analyze Linked Data using graph operations such as counting triangles, finding eccentricity, testing connectedness, and computing PageRank directly on triple stores via the SPARQL interface. We describe the process of optimizing performance of the SPARQL-based implementation of such popular graph algorithms by reducing the space-overhead, simplifying iterative complexity and removing redundant computations by understanding query plans. Our optimized approach shows significant performance gains on triplestores hosted on stand-alone workstations as well as hardware-optimized scalable supercomputers such as the Cray XMT.
Graph facilitates tracking water and gas influx
Gruy, H.J. )
1990-03-26
Graphing the vertical distribution of reservoir volume is an easy method for estimating the acre-ft remaining to be exploited in reservoirs with water or gas encroachment. To evaluate reservoir performance and estimate oil and gas reserves in water-drive reservoirs or oil reservoirs with a gas cap, it is necessary to determine the magnitude of the movement of oil-water and gas-oil contact surfaces. In reviewing reserve estimates and reservoir studies done by others, the authors have found that very few reservoir engineers or geologists have an easy method for tracking the movement of these surfaces and estimating the volumes of oil displaced water encroachment, gas cap expansion, or the volumes of oil lost by wetting the gas cap. The following method evolved from the author's studies of the East Texas field starting in 1942, and it took this form in the early 1950s.
Composing Data Parallel Code for a SPARQL Graph Engine
Castellana, Vito G.; Tumeo, Antonino; Villa, Oreste; Haglin, David J.; Feo, John
2013-09-08
Big data analytics process large amount of data to extract knowledge from them. Semantic databases are big data applications that adopt the Resource Description Framework (RDF) to structure metadata through a graph-based representation. The graph based representation provides several benefits, such as the possibility to perform in memory processing with large amounts of parallelism. SPARQL is a language used to perform queries on RDF-structured data through graph matching. In this paper we present a tool that automatically translates SPARQL queries to parallel graph crawling and graph matching operations. The tool also supports complex SPARQL constructs, which requires more than basic graph matching for their implementation. The tool generates parallel code annotated with OpenMP pragmas for x86 Shared-memory Multiprocessors (SMPs). With respect to commercial database systems such as Virtuoso, our approach reduces memory occupation due to join operations and provides higher performance. We show the scaling of the automatically generated graph-matching code on a 48-core SMP.
Maps: Exploration, Resources, Reserves, and Production - Energy...
Gasoline and Diesel Fuel Update (EIA)
... Pursuant to Section 604 of the Energy Policy and Conservation Act, these maps are one ... button graphic button graphic Greater Green River Basin button graphic button graphic ...
Maps: Exploration, Resources, Reserves, and Production - Energy...
Gasoline and Diesel Fuel Update (EIA)
... Department of the Interior, Department of Energy, and Department of Agriculture study of ... button graphic button graphic Greater Green River Basin button graphic button graphic ...
Energy Perspectives: Trends and Milestones
Reports and Publications (EIA)
2011-01-01
Energy Perspectives is a graphical overview of energy history in the United States. The 43 graphs shown here reveal sweeping trends related to the nation's production, consumption, and trade of energy from 1949 through 2011.
Parallel Algorithms for Graph Optimization using Tree Decompositions
Sullivan, Blair D; Weerapurage, Dinesh P; Groer, Christopher S
2012-06-01
Although many $\\cal{NP}$-hard graph optimization problems can be solved in polynomial time on graphs of bounded tree-width, the adoption of these techniques into mainstream scientific computation has been limited due to the high memory requirements of the necessary dynamic programming tables and excessive runtimes of sequential implementations. This work addresses both challenges by proposing a set of new parallel algorithms for all steps of a tree decomposition-based approach to solve the maximum weighted independent set problem. A hybrid OpenMP/MPI implementation includes a highly scalable parallel dynamic programming algorithm leveraging the MADNESS task-based runtime, and computational results demonstrate scaling. This work enables a significant expansion of the scale of graphs on which exact solutions to maximum weighted independent set can be obtained, and forms a framework for solving additional graph optimization problems with similar techniques.
Data Sources For Emerging Technologies Program MYPP Target Graphs
Broader source: Energy.gov [DOE]
The BTO Emerging Technologies Accomplishments and Outcomes – 2015 page contains graphs on Multi-Year Program Plan R&D targets for certain technologies. This page contains information on data...
Must all charting and graphing code be written in javascript...
Must all charting and graphing code be written in javascript? Home > Groups > Databus In the documentation chapter entitled Developing charts using 3rd party api, we are told that...
EAGLE: 'EAGLE'Is an' Algorithmic Graph Library for Exploration
Energy Science and Technology Software Center (OSTI)
2015-01-16
The Resource Description Framework (RDF) and SPARQL Protocol and RDF Query Language (SPARQL) were introduced about a decade ago to enable flexible schema-free data interchange on the Semantic Web. Today data scientists use the framework as a scalable graph representation for integrating, querying, exploring and analyzing data sets hosted at different sources. With increasing adoption, the need for graph mining capabilities for the Semantic Web has emerged. Today there is no tools to conduct "graphmore » mining" on RDF standard data sets. We address that need through implementation of popular iterative Graph Mining algorithms (Triangle count, Connected component analysis, degree distribution, diversity degree, PageRank, etc.). We implement these algorithms as SPARQL queries, wrapped within Python scripts and call our software tool as EAGLE. In RDF style, EAGLE stands for "EAGLE 'Is an' algorithmic graph library for exploration. EAGLE is like 'MATLAB' for 'Linked Data.'« less
TIFF Image Writer patch for OpenSceneGraph
Energy Science and Technology Software Center (OSTI)
2012-01-05
This software consists of code modifications to the open-source OpenSceneGraph software package to enable the creation of TlFF images containing 16 bit unsigned data. They also allow the user to disable compression and set the DPI tags in the resulting TIFF Images. Some image analysis programs require uncompressed, 16 bit unsigned input data. These code modifications allow programs based on OpenSceneGraph to write out such images, improving connectivity between applications.
Highly Asynchronous VisitOr Queue Graph Toolkit
Energy Science and Technology Software Center (OSTI)
2012-10-01
HAVOQGT is a C++ framework that can be used to create highly parallel graph traversal algorithms. The framework stores the graph and algorithmic data structures on external memory that is typically mapped to high performance locally attached NAND FLASH arrays. The framework supports a vertex-centered visitor programming model. The frameworkd has been used to implement breadth first search, connected components, and single source shortest path.
The peculiar phase structure of random graph bisection
Percus, Allon G; Istrate, Gabriel; Goncalves, Bruno T; Sumi, Robert Z
2008-01-01
The mincut graph bisection problem involves partitioning the n vertices of a graph into disjoint subsets, each containing exactly n/2 vertices, while minimizing the number of 'cut' edges with an endpoint in each subset. When considered over sparse random graphs, the phase structure of the graph bisection problem displays certain familiar properties, but also some surprises. It is known that when the mean degree is below the critical value of 2 log 2, the cutsize is zero with high probability. We study how the minimum cutsize increases with mean degree above this critical threshold, finding a new analytical upper bound that improves considerably upon previous bounds. Combined with recent results on expander graphs, our bound suggests the unusual scenario that random graph bisection is replica symmetric up to and beyond the critical threshold, with a replica symmetry breaking transition possibly taking place above the threshold. An intriguing algorithmic consequence is that although the problem is NP-hard, we can find near-optimal cutsizes (whose ratio to the optimal value approaches 1 asymptotically) in polynomial time for typical instances near the phase transition.
Algorithms and architectures for high performance analysis of semantic graphs.
Hendrickson, Bruce Alan
2005-09-01
Semantic graphs offer one promising avenue for intelligence analysis in homeland security. They provide a mechanism for describing a wide variety of relationships between entities of potential interest. The vertices are nouns of various types, e.g. people, organizations, events, etc. Edges in the graph represent different types of relationships between entities, e.g. 'is friends with', 'belongs-to', etc. Semantic graphs offer a number of potential advantages as a knowledge representation system. They allow information of different kinds, and collected in differing ways, to be combined in a seamless manner. A semantic graph is a very compressed representation of some of relationship information. It has been reported that the semantic graph can be two orders of magnitude smaller than the processed intelligence data. This allows for much larger portions of the data universe to be resident in computer memory. Many intelligence queries that are relevant to the terrorist threat are naturally expressed in the language of semantic graphs. One example is the search for 'interesting' relationships between two individuals or between an individual and an event, which can be phrased as a search for short paths in the graph. Another example is the search for an analyst-specified threat pattern, which can be cast as an instance of subgraph isomorphism. It is important to note than many kinds of analysis are not relationship based, so these are not good candidates for semantic graphs. Thus, a semantic graph should always be used in conjunction with traditional knowledge representation and interface methods. Operations that involve looking for chains of relationships (e.g. friend of a friend) are not efficiently executable in a traditional relational database. However, the semantic graph can be thought of as a pre-join of the database, and it is ideally suited for these kinds of operations. Researchers at Sandia National Laboratories are working to facilitate semantic graph
Mining Large Heterogeneous Graphs using Cray s Urika
Sukumar, Sreenivas R; Bond, Nathaniel A
2013-01-01
Pattern discovery and predictive modeling from seemingly related Big Data represented as massive, ad-hoc, heterogeneous networks (e.g., extremely large graphs with complex, possibly unknown structure) is an outstanding problem in many application domains. To address this problem, we are designing graph-mining algorithms capable of discovering relationship-patterns from such data and using those discovered patterns as features for classification and predictive modeling. Specifically, we are: (i) exploring statistical properties, mechanics and generative models of behavior patterns in heterogeneous information networks, (ii) developing novel, automated and scalable graph-pattern discovery algorithms and (iii) applying our relationship-analytics (data science + network science) expertise to domains spanning healthcare to homeland security.
Fast Search for Dynamic Multi-Relational Graphs
Choudhury, Sutanay; Holder, Larry; Chin, George; Feo, John T.
2013-06-23
Acting on time-critical events by processing ever growing social media or news streams is a major technical challenge. Many of these data sources can be modeled as multi-relational graphs. Continuous queries or techniques to search for rare events that typically arise in monitoring applications have been studied extensively for relational databases. This work is dedicated to answer the question that emerges naturally: how can we efficiently execute a continuous query on a dynamic graph? This paper presents an exact subgraph search algorithm that exploits the temporal characteristics of representative queries for online news or social media monitoring. The algorithm is based on a novel data structure called the that leverages the structural and semantic characteristics of the underlying multi-relational graph. The paper concludes with extensive experimentation on several real-world datasets that demonstrates the validity of this approach.
On the mixing time of geographical threshold graphs
Bradonjic, Milan
2009-01-01
In this paper, we study the mixing time of random graphs generated by the geographical threshold graph (GTG) model, a generalization of random geometric graphs (RGG). In a GTG, nodes are distributed in a Euclidean space, and edges are assigned according to a threshold function involving the distance between nodes as well as randomly chosen node weights. The motivation for analyzing this model is that many real networks (e.g., wireless networks, the Internet, etc.) need to be studied by using a 'richer' stochastic model (which in this case includes both a distance between nodes and weights on the nodes). We specifically study the mixing times of random walks on 2-dimensional GTGs near the connectivity threshold. We provide a set of criteria on the distribution of vertex weights that guarantees that the mixing time is {Theta}(n log n).
Integrated Network Decompositions and Dynamic Programming for Graph Optimization (INDDGO)
Energy Science and Technology Software Center (OSTI)
2012-05-31
The INDDGO software package offers a set of tools for finding exact solutions to graph optimization problems via tree decompositions and dynamic programming algorithms. Currently the framework offers serial and parallel (distributed memory) algorithms for finding tree decompositions and solving the maximum weighted independent set problem. The parallel dynamic programming algorithm is implemented on top of the MADNESS task-based runtime.
In-Memory Graph Databases for Web-Scale Data
Castellana, Vito G.; Morari, Alessandro; Weaver, Jesse R.; Tumeo, Antonino; Haglin, David J.; Villa, Oreste; Feo, John
2015-03-01
RDF databases have emerged as one of the most relevant way for organizing, integrating, and managing expo- nentially growing, often heterogeneous, and not rigidly structured data for a variety of scientific and commercial fields. In this paper we discuss the solutions integrated in GEMS (Graph database Engine for Multithreaded Systems), a software framework for implementing RDF databases on commodity, distributed-memory high-performance clusters. Unlike the majority of current RDF databases, GEMS has been designed from the ground up to primarily employ graph-based methods. This is reflected in all the layers of its stack. The GEMS framework is composed of: a SPARQL-to-C++ compiler, a library of data structures and related methods to access and modify them, and a custom runtime providing lightweight software multithreading, network messages aggregation and a partitioned global address space. We provide an overview of the framework, detailing its component and how they have been closely designed and customized to address issues of graph methods applied to large-scale datasets on clusters. We discuss in details the principles that enable automatic translation of the queries (expressed in SPARQL, the query language of choice for RDF databases) to graph methods, and identify differences with respect to other RDF databases.
STRUCTURAL ANNOTATION OF EM IMAGES BY GRAPH CUT
Chang, Hang; Auer, Manfred; Parvin, Bahram
2009-05-08
Biological images have the potential to reveal complex signatures that may not be amenable to morphological modeling in terms of shape, location, texture, and color. An effective analytical method is to characterize the composition of a specimen based on user-defined patterns of texture and contrast formation. However, such a simple requirement demands an improved model for stability and robustness. Here, an interactive computational model is introduced for learning patterns of interest by example. The learned patterns bound an active contour model in which the traditional gradient descent optimization is replaced by the more efficient optimization of the graph cut methods. First, the energy function is defined according to the curve evolution. Next, a graph is constructed with weighted edges on the energy function and is optimized with the graph cut algorithm. As a result, the method combines the advantages of the level set method and graph cut algorithm, i.e.,"topological" invariance and computational efficiency. The technique is extended to the multi-phase segmentation problem; the method is validated on synthetic images and then applied to specimens imaged by transmission electron microscopy(TEM).
Cyber Graph Queries for Geographically Distributed Data Centers
Berry, Jonathan W.; Collins, Michael; Kearns, Aaron; Phillips, Cynthia A.; Saia, Jared
2015-05-01
We present new algorithms for a distributed model for graph computations motivated by limited information sharing we first discussed in [20]. Two or more independent entities have collected large social graphs. They wish to compute the result of running graph algorithms on the entire set of relationships. Because the information is sensitive or economically valuable, they do not wish to simply combine the information in a single location. We consider two models for computing the solution to graph algorithms in this setting: 1) limited-sharing: the two entities can share only a polylogarithmic size subgraph; 2) low-trust: the entities must not reveal any information beyond the query answer, assuming they are all honest but curious. We believe this model captures realistic constraints on cooperating autonomous data centers. We have algorithms in both setting for s - t connectivity in both models. We also give an algorithm in the low-communication model for finding a planted clique. This is an anomaly- detection problem, finding a subgraph that is larger and denser than expected. For both the low- communication algorithms, we exploit structural properties of social networks to prove perfor- mance bounds better than what is possible for general graphs. For s - t connectivity, we use known properties. For planted clique, we propose a new property: bounded number of triangles per node. This property is based upon evidence from the social science literature. We found that classic examples of social networks do not have the bounded-triangles property. This is because many social networks contain elements that are non-human, such as accounts for a business, or other automated accounts. We describe some initial attempts to distinguish human nodes from automated nodes in social networks based only on topological properties.
AA KRUGER; A GOEL; CP RODRIGUEZ; JS MCCLOY; MJ SCHWEIGER; WW LUKENS; JR, BJ RILEY; D KIM; M LIEZERS; P HRMA
2012-08-13
The immobilization of 99Tc in a suitable host matrix has proved a challenging task for researchers in the nuclear waste community around the world. At the Hanford site in Washington State in the U.S., the total amount of 99Tc in low-activity waste (LAW) is {approx} 1,300 kg and the current strategy is to immobilize the 99Tc in borosilicate glass with vitrification. In this context, the present article reports on the solubility and retention of rhenium, a nonradioactive surrogate for 99Tc, in a LAW sodium borosilicate glass. Due to the radioactive nature of technetium, rhenium was chosen as a simulant because of previously established similarities in ionic radii and other chemical aspects. The glasses containing target Re concentrations varying from 0 to10,000 ppm by mass were synthesized in vacuum-sealed quartz ampoules to minimize the loss of Re by volatilization during melting at 1000 DC. The rhenium was found to be present predominantly as Re7 + in all the glasses as observed by X-ray absorption near-edge structure (XANES). The solubility of Re in borosilicate glasses was determined to be {approx}3,000 ppm (by mass) using inductively coupled plasma-optical emission spectroscopy (ICP-OES). At higher rhenium concentrations, some additional material was retained in the glasses in the form of alkali perrhenate crystalline inclusions detected by X-ray diffraction (XRD) and laser ablation-ICP mass spectrometry (LA-ICP-MS). Assuming justifiably substantial similarities between Re7 + and Tc 7+ behavior in this glass system, these results implied that the processing and immobilization of 99Tc from radioactive wastes should not be limited by the solubility of 99Tc in borosilicate LAW glasses.
LDRD final report : first application of geospatial semantic graphs to SAR image data.
Brost, Randolph C.; McLendon, William Clarence,
2013-01-01
Modeling geospatial information with semantic graphs enables search for sites of interest based on relationships between features, without requiring strong a priori models of feature shape or other intrinsic properties. Geospatial semantic graphs can be constructed from raw sensor data with suitable preprocessing to obtain a discretized representation. This report describes initial work toward extending geospatial semantic graphs to include temporal information, and initial results applying semantic graph techniques to SAR image data. We describe an efficient graph structure that includes geospatial and temporal information, which is designed to support simultaneous spatial and temporal search queries. We also report a preliminary implementation of feature recognition, semantic graph modeling, and graph search based on input SAR data. The report concludes with lessons learned and suggestions for future improvements.
A Graph Analytic Metric for Mitigating Advanced Persistent Threat
Johnson, John R.; Hogan, Emilie A.
2013-06-04
This paper introduces a novel graph analytic metric that can be used to measure the potential vulnerability of a cyber network to specific types of attacks that use lateral movement and privilege escalation such as the well known Pass The Hash, (PTH). The metric is computed from an oriented subgraph of the underlying cyber network induced by selecting only those edges for which a given property holds between the two vertices of the edge. The metric with respect to a select node on the subgraph is defined as the likelihood that the select node is reachable from another arbitrary node in the graph. This metric can be calculated dynamically from the authorization and auditing layers during the network security authorization phase and will potentially enable predictive deterrence against attacks such as PTH.
Codesign Lessons Learned from Implementing Graph Matching on Multithreaded Architectures
Halappanavar, Mahantesh; Pothen, Alex; Azad, Md Ariful; Manne, Fredrik; Langguth, Johannes; Khan, Arif
2015-08-12
Co-design of algorithms and architectures is an effective way to address the performance of irregular applications on multithreaded architectures. We explore the interplay between algorithm design and architectural features using graph matching as a case study. We present the key lessons that we have learnt as a means to influence co-design of algorithms and architecture for execution of data-intensive irregular workloads.
Encoding and analyzing aerial imagery using geospatial semantic graphs
Watson, Jean-Paul; Strip, David R.; McLendon, William C.; Parekh, Ojas D.; Diegert, Carl F.; Martin, Shawn Bryan; Rintoul, Mark Daniel
2014-02-01
While collection capabilities have yielded an ever-increasing volume of aerial imagery, analytic techniques for identifying patterns in and extracting relevant information from this data have seriously lagged. The vast majority of imagery is never examined, due to a combination of the limited bandwidth of human analysts and limitations of existing analysis tools. In this report, we describe an alternative, novel approach to both encoding and analyzing aerial imagery, using the concept of a geospatial semantic graph. The advantages of our approach are twofold. First, intuitive templates can be easily specified in terms of the domain language in which an analyst converses. These templates can be used to automatically and efficiently search large graph databases, for specific patterns of interest. Second, unsupervised machine learning techniques can be applied to automatically identify patterns in the graph databases, exposing recurring motifs in imagery. We illustrate our approach using real-world data for Anne Arundel County, Maryland, and compare the performance of our approach to that of an expert human analyst.
Proximity graphs based multi-scale image segmentation
Skurikhin, Alexei N
2008-01-01
We present a novel multi-scale image segmentation approach based on irregular triangular and polygonal tessellations produced by proximity graphs. Our approach consists of two separate stages: polygonal seeds generation followed by an iterative bottom-up polygon agglomeration into larger chunks. We employ constrained Delaunay triangulation combined with the principles known from the visual perception to extract an initial ,irregular polygonal tessellation of the image. These initial polygons are built upon a triangular mesh composed of irregular sized triangles and their shapes are ad'apted to the image content. We then represent the image as a graph with vertices corresponding to the polygons and edges reflecting polygon relations. The segmentation problem is then formulated as Minimum Spanning Tree extraction. We build a successive fine-to-coarse hierarchy of irregular polygonal grids by an iterative graph contraction constructing Minimum Spanning Tree. The contraction uses local information and merges the polygons bottom-up based on local region-and edge-based characteristics.
New Developments in MadGraph/MadEvent
Alwall, Johan; Artoisenet, Pierre; de Visscher, Simon; Duhr, Claude; Frederix, Rikkert; Herquet, Michel; Mattelaer, Olivier; /IBA, Louvain-la-Neuve
2011-11-08
We here present some recent developments of MadGraph/MadEvent since the latest published version, 4.0. These developments include: Jet matching with Pythia parton showers for both Standard Model and Beyond the Standard Model processes, decay chain functionality, decay width calculation and decay simulation, process generation for the Grid, a package for calculation of quarkonium amplitudes, calculation of Matrix Element weights for experimental events, automatic dipole subtraction for next-to-leading order calculations, and an interface to FeynRules, a package for automatic calculation of Feynman rules and model files from the Lagrangian of any New Physics model.
Constructing compact and effective graphs for recommender systems via node and edge aggregations
DOE Public Access Gateway for Energy & Science Beta (PAGES Beta)
Lee, Sangkeun; Kahng, Minsuk; Lee, Sang-goo
2014-12-10
Exploiting graphs for recommender systems has great potential to flexibly incorporate heterogeneous information for producing better recommendation results. As our baseline approach, we first introduce a naive graph-based recommendation method, which operates with a heterogeneous log-metadata graph constructed from user log and content metadata databases. Although the na ve graph-based recommendation method is simple, it allows us to take advantages of heterogeneous information and shows promising flexibility and recommendation accuracy. However, it often leads to extensive processing time due to the sheer size of the graphs constructed from entire user log and content metadata databases. In this paper, we proposemore » node and edge aggregation approaches to constructing compact and e ective graphs called Factor-Item bipartite graphs by aggregating nodes and edges of a log-metadata graph. Furthermore, experimental results using real world datasets indicate that our approach can significantly reduce the size of graphs exploited for recommender systems without sacrificing the recommendation quality.« less
Discrete Mathematical Approaches to Graph-Based Traffic Analysis
Joslyn, Cliff A.; Cowley, Wendy E.; Hogan, Emilie A.; Olsen, Bryan K.
2014-04-01
Modern cyber defense and anlaytics requires general, formal models of cyber systems. Multi-scale network models are prime candidates for such formalisms, using discrete mathematical methods based in hierarchically-structured directed multigraphs which also include rich sets of labels. An exemplar of an application of such an approach is traffic analysis, that is, observing and analyzing connections between clients, servers, hosts, and actors within IP networks, over time, to identify characteristic or suspicious patterns. Towards that end, NetFlow (or more generically, IPFLOW) data are available from routers and servers which summarize coherent groups of IP packets flowing through the network. In this paper, we consider traffic analysis of Netflow using both basic graph statistics and two new mathematical measures involving labeled degree distributions and time interval overlap measures. We do all of this over the VAST test data set of 96M synthetic Netflow graph edges, against which we can identify characteristic patterns of simulated ground-truth network attacks.
INDDGO: Integrated Network Decomposition & Dynamic programming for Graph Optimization
Groer, Christopher S; Sullivan, Blair D; Weerapurage, Dinesh P
2012-10-01
It is well-known that dynamic programming algorithms can utilize tree decompositions to provide a way to solve some \\emph{NP}-hard problems on graphs where the complexity is polynomial in the number of nodes and edges in the graph, but exponential in the width of the underlying tree decomposition. However, there has been relatively little computational work done to determine the practical utility of such dynamic programming algorithms. We have developed software to construct tree decompositions using various heuristics and have created a fast, memory-efficient dynamic programming implementation for solving maximum weighted independent set. We describe our software and the algorithms we have implemented, focusing on memory saving techniques for the dynamic programming. We compare the running time and memory usage of our implementation with other techniques for solving maximum weighted independent set, including a commercial integer programming solver and a semi-definite programming solver. Our results indicate that it is possible to solve some instances where the underlying decomposition has width much larger than suggested by the literature. For certain types of problems, our dynamic programming code runs several times faster than these other methods.
A Space-Filling Visualization Technique for Multivariate Small World Graphs
Wong, Pak C.; Foote, Harlan P.; Mackey, Patrick S.; Chin, George; Huang, Zhenyu; Thomas, James J.
2012-03-15
We introduce an information visualization technique, known as GreenCurve, for large sparse graphs that exhibit small world properties. Our fractal-based design approach uses spatial cues to approximate the node connections and thus eliminates the links between the nodes in the visualization. The paper describes a sophisticated algorithm to order the neighboring nodes of a large sparse graph by solving the Fiedler vector of its graph Laplacian, and then fold the graph nodes into a space-filling fractal curve based on the Fiedler vector. The result is a highly compact visualization that gives a succinct overview of the graph with guaranteed visibility of every graph node. We show in the paper that the GreenCurve technology is (1) theoretically sustainable by introducing an error estimation metric to measure the fidelity of the new graph representation, (2) empirically rigorous by conducting a usability study to investigate its strengths and weaknesses against the traditional graph layout, and (3) pragmatically feasible by applying it to analyze stressed conditions of the large scale electric power grid on the west coast.
Multi-Level Anomaly Detection on Time-Varying Graph Data
Bridges, Robert A; Collins, John P; Ferragut, Erik M; Laska, Jason A; Sullivan, Blair D
2015-01-01
This work presents a novel modeling and analysis framework for graph sequences which addresses the challenge of detecting and contextualizing anomalies in labelled, streaming graph data. We introduce a generalization of the BTER model of Seshadhri et al. by adding flexibility to community structure, and use this model to perform multi-scale graph anomaly detection. Specifically, probability models describing coarse subgraphs are built by aggregating probabilities at finer levels, and these closely related hierarchical models simultaneously detect deviations from expectation. This technique provides insight into a graph's structure and internal context that may shed light on a detected event. Additionally, this multi-scale analysis facilitates intuitive visualizations by allowing users to narrow focus from an anomalous graph to particular subgraphs or nodes causing the anomaly. For evaluation, two hierarchical anomaly detectors are tested against a baseline Gaussian method on a series of sampled graphs. We demonstrate that our graph statistics-based approach outperforms both a distribution-based detector and the baseline in a labeled setting with community structure, and it accurately detects anomalies in synthetic and real-world datasets at the node, subgraph, and graph levels. To illustrate the accessibility of information made possible via this technique, the anomaly detector and an associated interactive visualization tool are tested on NCAA football data, where teams and conferences that moved within the league are identified with perfect recall, and precision greater than 0.786.
Have Green – A Visual Analytics Framework for Large Semantic Graphs
Wong, Pak C.; Chin, George; Foote, Harlan P.; Mackey, Patrick S.; Thomas, James J.
2006-10-29
A semantic graph is a network of heterogeneous nodes and links annotated with a domain ontology. In intelligence analysis, investigators use semantic graphs to organize concepts and relationships as graph nodes and links in hopes of discovering key trends, patterns, and insights. However, as new information continues to arrive from a multitude of sources, the size and complexity of the semantic graphs will soon overwhelm an investigator's cognitive capacity to carry out significant analyses. We introduce a powerful visual analytics framework designed to enhance investigators--natural analytical capabilities to comprehend and analyze large semantic graphs. The paper describes the overall framework design, presents major development accomplishments to date, and discusses future directions of a new visual analytics system known as Have Green.
A framework for graph-based synthesis, analysis, and visualization of HPC cluster job data.
Mayo, Jackson R.; Kegelmeyer, W. Philip, Jr.; Wong, Matthew H.; Pebay, Philippe Pierre; Gentile, Ann C.; Thompson, David C.; Roe, Diana C.; De Sapio, Vincent; Brandt, James M.
2010-08-01
The monitoring and system analysis of high performance computing (HPC) clusters is of increasing importance to the HPC community. Analysis of HPC job data can be used to characterize system usage and diagnose and examine failure modes and their effects. This analysis is not straightforward, however, due to the complex relationships that exist between jobs. These relationships are based on a number of factors, including shared compute nodes between jobs, proximity of jobs in time, etc. Graph-based techniques represent an approach that is particularly well suited to this problem, and provide an effective technique for discovering important relationships in job queuing and execution data. The efficacy of these techniques is rooted in the use of a semantic graph as a knowledge representation tool. In a semantic graph job data, represented in a combination of numerical and textual forms, can be flexibly processed into edges, with corresponding weights, expressing relationships between jobs, nodes, users, and other relevant entities. This graph-based representation permits formal manipulation by a number of analysis algorithms. This report presents a methodology and software implementation that leverages semantic graph-based techniques for the system-level monitoring and analysis of HPC clusters based on job queuing and execution data. Ontology development and graph synthesis is discussed with respect to the domain of HPC job data. The framework developed automates the synthesis of graphs from a database of job information. It also provides a front end, enabling visualization of the synthesized graphs. Additionally, an analysis engine is incorporated that provides performance analysis, graph-based clustering, and failure prediction capabilities for HPC systems.
Exotic equilibria of Harary graphs and a new minimum degree lower bound for synchronization
Canale, Eduardo A.; Monzn, Pablo
2015-02-15
This work is concerned with stability of equilibria in the homogeneous (equal frequencies) Kuramoto model of weakly coupled oscillators. In 2012 [R. Taylor, J. Phys. A: Math. Theor. 45, 115 (2012)], a sufficient condition for almost global synchronization was found in terms of the minimum degreeorder ratio of the graph. In this work, a new lower bound for this ratio is given. The improvement is achieved by a concrete infinite sequence of regular graphs. Besides, non standard unstable equilibria of the graphs studied in Wiley et al. [Chaos 16, 015103 (2006)] are shown to exist as conjectured in that work.
A weak zero-one law for sequences of random distance graphs
Zhukovskii, Maksim E
2012-07-31
We study zero-one laws for properties of random distance graphs. Properties written in a first-order language are considered. For p(N) such that pN{sup {alpha}}{yields}{infinity} as N{yields}{infinity}, and (1-p)N{sup {alpha}} {yields} {infinity} as N {yields} {infinity} for any {alpha}>0, we succeed in refuting the law. In this connection, we consider a weak zero-one j-law. For this law, we obtain results for random distance graphs which are similar to the assertions concerning the classical zero-one law for random graphs. Bibliography: 18 titles.
Omega: an Overlap-graph de novo Assembler for Meta-genomics
Haider, Bahlul; Ahn, Tae-Hyuk; Bushnell, Brian; Chai, JJ; Copeland, Alex; Pan, Chongle
2014-01-01
Motivation: Metagenomic sequencing allows reconstruction of mi-crobial genomes directly from environmental samples. Omega (overlap-graph metagenome assembler) was developed here for assembling and scaffolding Illumina sequencing data of microbial communities. Results: Omega found overlaps between reads using a prefix/suffix hash table. The overlap graph of reads was simplified by removing transitive edges and trimming small branches. Unitigs were generat-ed based on minimum cost flow analysis of the overlap graph. Obtained unitigs were merged to contigs and scaffolds using mate-pair information. Omega was compared with two de Bruijn graph assemblers, SOAPdenovo and IDBA-UD, using a publically-available Illumina sequencing dataset of a 64-genome mock com-munity. The assembly results were verified by their alignment with reference genomes. The overall performances of the three assem-blers were comparable and each assembler provided best results for a subset of genomes.
al-Saffar, Sinan; Joslyn, Cliff A.; Chappell, Alan R.
2011-07-18
As semantic datasets grow to be very large and divergent, there is a need to identify and exploit their inherent semantic structure for discovery and optimization. Towards that end, we present here a novel methodology to identify the semantic structures inherent in an arbitrary semantic graph dataset. We first present the concept of an extant ontology as a statistical description of the semantic relations present amongst the typed entities modeled in the graph. This serves as a model of the underlying semantic structure to aid in discovery and visualization. We then describe a method of ontological scaling in which the ontology is employed as a hierarchical scaling filter to infer different resolution levels at which the graph structures are to be viewed or analyzed. We illustrate these methods on three large and publicly available semantic datasets containing more than one billion edges each. Keywords-Semantic Web; Visualization; Ontology; Multi-resolution Data Mining;
Wedge sampling for computing clustering coefficients and triangle counts on large graphs
DOE Public Access Gateway for Energy & Science Beta (PAGES Beta)
Seshadhri, C.; Pinar, Ali; Kolda, Tamara G.
2014-05-08
Graphs are used to model interactions in a variety of contexts, and there is a growing need to quickly assess the structure of such graphs. Some of the most useful graph metrics are based on triangles, such as those measuring social cohesion. Despite the importance of these triadic measures, algorithms to compute them can be extremely expensive. We discuss the method of wedge sampling. This versatile technique allows for the fast and accurate approximation of various types of clustering coefficients and triangle counts. Furthermore, these techniques are extensible to counting directed triangles in digraphs. Our methods come with provable andmore » practical time-approximation tradeoffs for all computations. We provide extensive results that show our methods are orders of magnitude faster than the state of the art, while providing nearly the accuracy of full enumeration.« less
Graph of Total Number of Oligos Within Windows of a Sequence
Energy Science and Technology Software Center (OSTI)
1995-11-28
SEQWIN is user-friendly software which graphs the total number of oligos present in a sequence. The sequence is scanned one window at a time; windows can be overlapping. Each bar on the graph represents a single window down the sequence. The user specifies the sequence of interest and a list of oligos as program input. If the sequence is known, locations of specific structure or sequences can be specified and compared with the bars onmore » a graph. The window size, amount of overlap of the windows, number of windows to be considered, and the starting position of the first window used can be adjusted at the user's discretion.« less
Powers, Sarah S.; Lothian, Joshua
2014-12-01
The benchmarking effort within the Extreme Scale Systems Center at Oak Ridge National Laboratory seeks to provide High Performance Computing benchmarks and test suites of interest to the DoD sponsor. The work described in this report is a part of the effort focusing on graph generation. A previously developed benchmark, SystemBurn, allows the emulation of a broad spectrum of application behavior profiles within a single framework. To complement this effort, similar capabilities are desired for graph-centric problems. This report described the in-depth analysis of the generated synthetic graphs' properties at a variety of scales using different generator implementations and examines their applicability to replicating real world datasets.
Human Factors Evaluation of Advanced Electric Power Grid Visualization Tools
Greitzer, Frank L.; Dauenhauer, Peter M.; Wierks, Tamara G.; Podmore, Robin
2009-04-01
This report describes initial human factors evaluation of four visualization tools (Graphical Contingency Analysis, Force Directed Graphs, Phasor State Estimator and Mode Meter/ Mode Shapes) developed by PNNL, and proposed test plans that may be implemented to evaluate their utility in scenario-based experiments.
The Climate at Los Alamos; Are we measurement changes?
Dewart, Jean Marie
2015-04-16
A new report shows new graphic displays of the weather trends in Los Alamos, New Mexico, and at the Los Alamos National Laboratory (LANL). The graphs show trends of average, minimum average, and maximum average temperature for summer and winter months going back decades. Records of summer and winter precipitation are also included in the report.
Aizenman, Michael; Warzel, Simone
2012-09-15
We discuss the dynamical implications of the recent proof that for a quantum particle in a random potential on a regular tree graph absolutely continuous (ac) spectrum occurs non-perturbatively through rare fluctuation-enabled resonances. The main result is spelled in the title.
Risch, John S.; Dowson, Scott T.
2012-03-06
A method of displaying correlations among information objects includes receiving a query against a database; obtaining a query result set; and generating a visualization representing the components of the result set, the visualization including one of a plane and line to represent a data field, nodes representing data values, and links showing correlations among fields and values. Other visualization methods and apparatus are disclosed.
Risch, John S.; Dowson, Scott T.; Hart, Michelle L.; Hatley, Wes L.
2008-05-13
A method of displaying correlations among information objects comprises receiving a query against a database; obtaining a query result set; and generating a visualization representing the components of the result set, the visualization including one of a plane and line to represent a data field, nodes representing data values, and links showing correlations among fields and values. Other visualization methods and apparatus are disclosed.
Exact scattering matrix of graphs in magnetic field and quantum noise
Caudrelier, Vincent; Mintchev, Mihail; Ragoucy, Eric
2014-08-15
We consider arbitrary quantum wire networks modelled by finite, noncompact, connected quantum graphs in the presence of an external magnetic field. We find a general formula for the total scattering matrix of the network in terms of its local scattering properties and its metric structure. This is applied to a quantum ring with N external edges. Connecting the external edges of the ring to heat reservoirs, we study the quantum transport on the graph in ambient magnetic field. We consider two types of dynamics on the ring: the free Schrdinger and the free massless Dirac equations. For each case, a detailed study of the thermal noise is performed analytically. Interestingly enough, in presence of a magnetic field, the standard linear Johnson-Nyquist law for the low temperature behaviour of the thermal noise becomes nonlinear. The precise regime of validity of this effect is discussed and a typical signature of the underlying dynamics is observed.
Quantum dynamics via Planck-scale-stepped action-carrying 'Graph Paths'
Chew, Geoffrey F.
2003-05-05
A divergence-free, parameter-free, path-based discrete-time quantum dynamics is designed to not only enlarge the achievements of general relativity and the standard particle model, by approximations at spacetime scales far above Planck scale while far below Hubble scale, but to allow tackling of hitherto inaccessible questions. ''Path space'' is larger than and precursor to Hilbert-space basis. The wave-function-propagating paths are action-carrying structured graphs-cubic and quartic structured vertices connected by structured ''fermionic'' or ''bosonic'' ''particle'' and ''nonparticle'' arcs. A Planck-scale path step determines the gravitational constant while controlling all graph structure. The basis of the theory's (zero-rest-mass) elementary-particle Hilbert space (which includes neither gravitons nor scalar bosons) resides in particle arcs. Nonparticle arcs within a path are responsible for energy and rest mass.
The d-edge shortest-path problem for a Monge graph
Bein, W.W.; Larmore, L.L.; Park, J.K.
1992-07-14
A complete edge-weighted directed graph on vertices 1,2,...,n that assigns cost c(i,j) to the edge (i,j) is called Monge if its edge costs form a Monge array, i.e., for all i < k and j < l, c[i, j]+c[k,l]{le} < c[i,l]+c[k,j]. One reason Monge graphs are interesting is that shortest paths can be computed quite quickly in such graphs. In particular, Wilber showed that the shortest path from vertex 1 to vertex n of a Monge graph can be computed in O(n) time, and Aggarwal, Klawe, Moran, Shor, and Wilber showed that the shortest d-edge 1-to-n path (i.e., the shortest path among all 1-to-n paths with exactly d edges) can be computed in O(dn) time. This paper`s contribution is a new algorithm for the latter problem. Assuming 0 {le} c[i,j] {le} U and c[i,j + 1] + c[i + 1,j] {minus} c[i,j] {minus} c[i + 1, j + 1] {ge} L > 0 for all i and j, our algorithm runs in O(n(1 + 1g(U/L))) time. Thus, when d {much_gt} 1 + 1g(U/L), our algorithm represents a significant improvement over Aggarwal et al.`s O(dn)-time algorithm. We also present several applications of our algorithm; they include length-limited Huffman coding, finding the maximum-perimeter d-gon inscribed in a given convex n-gon, and a digital-signal-compression problem.
2007 CBECS Large Hospital Building FAQs: 2003-2007 Comparison Graphs
Gasoline and Diesel Fuel Update (EIA)
FAQs: 2003-2007 Comparison Graphs Main Report | Methodology | FAQ | List of Tables CBECS 2007 - Release date: August 17, 2012 Jump to: Figure 1 | Figure 2 | Figure 3 | Figure 4 | Figure 5 Figure 1 Number of Large Hospital Buildings and 95% Confidence Intervals by Census Region, 2003 and 2007 Figure 2 Total Floorspace and 95% Confidence Intervals in Large Hospital Buildings by Census Region, 2003 and 2007 Figure 3 Major Fuel Intensity and 95% Confidence Intervals by Census Region, 2003 and 2007
Madduri, Kamesh; Bader, David A.
2009-02-15
Graph-theoretic abstractions are extensively used to analyze massive data sets. Temporal data streams from socioeconomic interactions, social networking web sites, communication traffic, and scientific computing can be intuitively modeled as graphs. We present the first study of novel high-performance combinatorial techniques for analyzing large-scale information networks, encapsulating dynamic interaction data in the order of billions of entities. We present new data structures to represent dynamic interaction networks, and discuss algorithms for processing parallel insertions and deletions of edges in small-world networks. With these new approaches, we achieve an average performance rate of 25 million structural updates per second and a parallel speedup of nearly28 on a 64-way Sun UltraSPARC T2 multicore processor, for insertions and deletions to a small-world network of 33.5 million vertices and 268 million edges. We also design parallel implementations of fundamental dynamic graph kernels related to connectivity and centrality queries. Our implementations are freely distributed as part of the open-source SNAP (Small-world Network Analysis and Partitioning) complex network analysis framework.
Scaling Graph Community Detection on the Tilera Many-core Architecture
Chavarría-Miranda, Daniel; Halappanavar, Mahantesh; Kalyanaraman, Anantharaman
2014-12-01
In an era when power constraints and data movement are proving to be significant barriers for the application of high-end computing, the Tilera many-core architecture offers a low-power platform exhibiting many important characteristics of future systems, including a large number of simple cores, a sophisticated network-on-chip, and fine-grained control over memory and caching policies. While this emerging architecture has been previously studied for structured compute-intensive kernels, benchmarking the platform for data-bound, irregular applications present significant challenges that have remained unexplored. Community detection is an advanced prototypical graph-theoretic operation with applications in numerous scientific domains including life sciences, cyber security, and power systems. In this work, we explore multiple design strategies toward developing a scalable tool for community detection on the Tilera platform. Using several memory layout and work scheduling techniques we demonstrate speedups of up to 46x on 36 cores of the Tilera TileGX36 platform over the best serial implementation, and also show results that have comparable quality and performance to mainstream x86 platforms. To the best of our knowledge this is the first work addressing graph algorithms on the Tilera platform. This study demonstrates that through careful design space exploration, low-power many-core platforms like Tilera can be effectively exploited for graph algorithms that that embody all the essential characteristics of an irregular application.
Grines, V Z; Pochinka, O V; Kapkaeva, S Kh
2014-10-31
In apaper of Oshemkov and Sharko, three-colour graphs were used to make the topological equivalence of Morse-Smale flows on surfaces obtained by Peixoto more precise. In the present paper, in the language of three-colour graphs equipped with automorphisms, we obtain acomplete (including realization) topological classification of gradient-like cascades on surfaces. Bibliography: 25 titles.
Pattern analysis of directed graphs using DEDICOM: an application to Enron email.
Bader, Brett William; Harshman, Richard A. (University of Western Ontario London, Ontario, Canada); Kolda, Tamara Gibson
2006-12-01
DEDICOM is a linear algebra model for analyzing intrinsically asymmetric relationships, such as trade among nations or the exchange of emails among individuals. DEDICOM decomposes a complex pattern of observed relations among objects into a sum of simpler patterns of inferred relations among latent components of the objects. Three-way DEDICOM is a higher-order extension of the model that incorporates a third mode of the data, such as time, giving it stronger uniqueness properties and consequently enhancing interpretability of solutions. In this paper, we present algorithms for computing these decompositions on large, sparse data as well as a variant for computing an asymmetric nonnegative factorization. When we apply these techniques to adjacency arrays arising from directed graphs with edges labeled by time, we obtain a smaller graph on latent semantic dimensions and gain additional information about their changing relationships over time. We demonstrate these techniques on the Enron email corpus to learn about the social networks and their transient behavior. The mixture of roles assigned to individuals by DEDICOM showed strong correspondence with known job classifications and revealed the patterns of communication between these roles. Changes in the communication pattern over time, e.g., between top executives and the legal department, were also apparent in the solutions.
A Selectivity based approach to Continuous Pattern Detection in Streaming Graphs
Choudhury, Sutanay; Holder, Larry; Chin, George; Agarwal, Khushbu; Feo, John T.
2015-02-02
Cyber security is one of the most significant technical challenges in current times. Detecting adversarial activities, prevention of theft of intellectual properties and customer data is a high priority for corporations and government agencies around the world. Cyber defenders need to analyze massive-scale, high-resolution network flows to identify, categorize, and mitigate attacks involving net- works spanning institutional and national boundaries. Many of the cyber attacks can be described as subgraph patterns, with promi- nent examples being insider infiltrations (path queries), denial of service (parallel paths) and malicious spreads (tree queries). This motivates us to explore subgraph matching on streaming graphs in a continuous setting. The novelty of our work lies in using the subgraph distributional statistics collected from the streaming graph to determine the query processing strategy. We introduce a Lazy Search" algorithm where the search strategy is decided on a vertex-to-vertex basis depending on the likelihood of a match in the vertex neighborhood. We also propose a metric named Relative Selectivity" that is used to se- lect between different query processing strategies. Our experiments performed on real online news, network traffic stream and a syn- thetic social network benchmark demonstrate 10-100x speedups over selectivity agnostic approaches.
ARM - Instructions for Submitting Extended Abstracts
Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)
Graphics should be sent as individual graphic files in their original software and NOT placed within the abstract. Createsize graphics to fit the image area. Make sure that any...