National Library of Energy BETA

Sample records for queries sparql queries

  1. SPARQL Query Form | OpenEI

    Open Energy Info (EERE)

    where a ?Concept LIMIT 100 Display Results As: Auto HTML Spreadsheet XML JSON Javascript NTriples RDFXML Rigorous check of the query Execution timeout, in milliseconds,...

  2. Query optimization for graph analytics on linked data using SPARQL

    SciTech Connect (OSTI)

    Hong, Seokyong; Lee, Sangkeun; Lim, Seung -Hwan; Sukumar, Sreenivas R.; Vatsavai, Ranga Raju

    2015-07-01

    Triplestores that support query languages such as SPARQL are emerging as the preferred and scalable solution to represent data and meta-data as massive heterogeneous graphs using Semantic Web standards. With increasing adoption, the desire to conduct graph-theoretic mining and exploratory analysis has also increased. Addressing that desire, this paper presents a solution that is the marriage of Graph Theory and the Semantic Web. We present software that can analyze Linked Data using graph operations such as counting triangles, finding eccentricity, testing connectedness, and computing PageRank directly on triple stores via the SPARQL interface. We describe the process of optimizing performance of the SPARQL-based implementation of such popular graph algorithms by reducing the space-overhead, simplifying iterative complexity and removing redundant computations by understanding query plans. Our optimized approach shows significant performance gains on triplestores hosted on stand-alone workstations as well as hardware-optimized scalable supercomputers such as the Cray XMT.

  3. Query | Department of Energy

    Office of Energy Efficiency and Renewable Energy (EERE) Indexed Site

    Query Query Query Query_1.pdf Query (104.29 KB) More Documents & Publications DOE Retrospective Review Plan and Burden Reduction Report July 29, 2013 DOE EO 13563 January 2014 Update Report and Burden Reduction Efforts DOE Retrospective Review Plan and Burden Reduction Report - December 18, 2012

  4. How can I query data on OpenEI and generate a map? | OpenEI Community

    Open Energy Info (EERE)

    How can I query data on OpenEI and generate a map? Home > Groups > Developer I'd like to have an Ask or SPARQL query display as a map in the OpenEI wiki section. What are the...

  5. Query

    Energy Savers [EERE]

    DOEIN CIA 12182006 12172008 Executive Office of the President, Council on Dixon, Robert K. DOEEE Environmental Quality (CEQ) 10232007 2232009 Dumaine, Carol CIA DOEIN ...

  6. compound queries | OpenEI Community

    Open Energy Info (EERE)

    queries developer Google maps maps multicolor result formats results Semantic Mediawiki Hi all, Recently, a couple of people on OpenEI have asked me how to do compound (or...

  7. ask queries | OpenEI Community

    Open Energy Info (EERE)

    queries developer Google maps maps multicolor result formats results Semantic Mediawiki Hi all, Recently, a couple of people on OpenEI have asked me how to do compound (or...

  8. Oregon Certified Water Right Examiners Query Webpage | Open Energy...

    Open Energy Info (EERE)

    Not Provided DOI Not Provided Check for DOI availability: http:crossref.org Online Internet link for Oregon Certified Water Right Examiners Query Webpage Citation State of...

  9. Increasing ask query limit | OpenEI Community

    Open Energy Info (EERE)

    via json. For example, this query only returns two entries: http:en.openei.orgservicesrestutilityrates?versionlatest&formatjsonplain&offset9998&limit30&detailb...

  10. Query-Driven Visualization and Analysis

    SciTech Connect (OSTI)

    Ruebel, Oliver; Bethel, E. Wes; Prabhat, Mr.; Wu, Kesheng

    2012-11-01

    This report focuses on an approach to high performance visualization and analysis, termed query-driven visualization and analysis (QDV). QDV aims to reduce the amount of data that needs to be processed by the visualization, analysis, and rendering pipelines. The goal of the data reduction process is to separate out data that is "scientifically interesting'' and to focus visualization, analysis, and rendering on that interesting subset. The premise is that for any given visualization or analysis task, the data subset of interest is much smaller than the larger, complete data set. This strategy---extracting smaller data subsets of interest and focusing of the visualization processing on these subsets---is complementary to the approach of increasing the capacity of the visualization, analysis, and rendering pipelines through parallelism. This report discusses the fundamental concepts in QDV, their relationship to different stages in the visualization and analysis pipelines, and presents QDV's application to problems in diverse areas, ranging from forensic cybersecurity to high energy physics.

  11. HDF5-FastQuery: An API for Simplifying Access to Data Storage,Retrieval, Indexing and Querying

    SciTech Connect (OSTI)

    Bethel, E. Wes; Gosink, Luke; Shalf, John; Stockinger, Kurt; Wu,Kesheng

    2006-06-15

    This work focuses on research and development activities that bridge a gap between fundamental data management technology index, query, storage and retrieval and use of such technology in computational and computer science algorithms and applications. The work has resulted in a streamlined applications programming interface (API) that simplifies data storage and retrieval using the HDF5 data I/O library, and eases use of the FastBit compressed bitmap indexing software for data indexing/querying. The API, which we call HDF5-FastQuery, will have broad applications in domain sciences as well as associated data analysis and visualization applications.

  12. Multicolor Maps from Compound Queries | OpenEI Community

    Open Energy Info (EERE)

    queries developer Google maps maps multicolor result formats results Semantic Mediawiki Hi all, Recently, a couple of people on OpenEI have asked me how to do compound (or...

  13. T-703: Cisco Unified Communications Manager Open Query Interface Lets

    Office of Energy Efficiency and Renewable Energy (EERE) Indexed Site

    Remote Users Obtain Database Contents | Department of Energy 703: Cisco Unified Communications Manager Open Query Interface Lets Remote Users Obtain Database Contents T-703: Cisco Unified Communications Manager Open Query Interface Lets Remote Users Obtain Database Contents August 26, 2011 - 3:45pm Addthis PROBLEM: A vulnerability was reported in Cisco Unified Communications Manager. A remote user can obtain database contents PLATFORM: Cisco Unified Communications Manager 6.x, 7.x, 8.0, 8.5

  14. Parallel Index and Query for Large Scale Data Analysis

    SciTech Connect (OSTI)

    Chou, Jerry; Wu, Kesheng; Ruebel, Oliver; Howison, Mark; Qiang, Ji; Prabhat,; Austin, Brian; Bethel, E. Wes; Ryne, Rob D.; Shoshani, Arie

    2011-07-18

    Modern scientific datasets present numerous data management and analysis challenges. State-of-the-art index and query technologies are critical for facilitating interactive exploration of large datasets, but numerous challenges remain in terms of designing a system for process- ing general scientific datasets. The system needs to be able to run on distributed multi-core platforms, efficiently utilize underlying I/O infrastructure, and scale to massive datasets. We present FastQuery, a novel software framework that address these challenges. FastQuery utilizes a state-of-the-art index and query technology (FastBit) and is designed to process mas- sive datasets on modern supercomputing platforms. We apply FastQuery to processing of a massive 50TB dataset generated by a large scale accelerator modeling code. We demonstrate the scalability of the tool to 11,520 cores. Motivated by the scientific need to search for inter- esting particles in this dataset, we use our framework to reduce search time from hours to tens of seconds.

  15. Secure quantum private information retrieval using phase-encoded queries

    SciTech Connect (OSTI)

    Olejnik, Lukasz

    2011-08-15

    We propose a quantum solution to the classical private information retrieval (PIR) problem, which allows one to query a database in a private manner. The protocol offers privacy thresholds and allows the user to obtain information from a database in a way that offers the potential adversary, in this model the database owner, no possibility of deterministically establishing the query contents. This protocol may also be viewed as a solution to the symmetrically private information retrieval problem in that it can offer database security (inability for a querying user to steal its contents). Compared to classical solutions, the protocol offers substantial improvement in terms of communication complexity. In comparison with the recent quantum private queries [Phys. Rev. Lett. 100, 230502 (2008)] protocol, it is more efficient in terms of communication complexity and the number of rounds, while offering a clear privacy parameter. We discuss the security of the protocol and analyze its strengths and conclude that using this technique makes it challenging to obtain the unconditional (in the information-theoretic sense) privacy degree; nevertheless, in addition to being simple, the protocol still offers a privacy level. The oracle used in the protocol is inspired both by the classical computational PIR solutions as well as the Deutsch-Jozsa oracle.

  16. Large-Scale Continuous Subgraph Queries on Streams

    SciTech Connect (OSTI)

    Choudhury, Sutanay; Holder, Larry; Chin, George; Feo, John T.

    2011-11-30

    Graph pattern matching involves finding exact or approximate matches for a query subgraph in a larger graph. It has been studied extensively and has strong applications in domains such as computer vision, computational biology, social networks, security and finance. The problem of exact graph pattern matching is often described in terms of subgraph isomorphism which is NP-complete. The exponential growth in streaming data from online social networks, news and video streams and the continual need for situational awareness motivates a solution for finding patterns in streaming updates. This is also the prime driver for the real-time analytics market. Development of incremental algorithms for graph pattern matching on streaming inputs to a continually evolving graph is a nascent area of research. Some of the challenges associated with this problem are the same as found in continuous query (CQ) evaluation on streaming databases. This paper reviews some of the representative work from the exhaustively researched field of CQ systems and identifies important semantics, constraints and architectural features that are also appropriate for HPC systems performing real-time graph analytics. For each of these features we present a brief discussion of the challenge encountered in the database realm, the approach to the solution and state their relevance in a high-performance, streaming graph processing framework.

  17. Towards Optimal Multi-Dimensional Query Processing with BitmapIndices

    SciTech Connect (OSTI)

    Rotem, Doron; Stockinger, Kurt; Wu, Kesheng

    2005-09-30

    Bitmap indices have been widely used in scientific applications and commercial systems for processing complex, multi-dimensional queries where traditional tree-based indices would not work efficiently. This paper studies strategies for minimizing the access costs for processing multi-dimensional queries using bitmap indices with binning. Innovative features of our algorithm include (a) optimally placing the bin boundaries and (b) dynamically reordering the evaluation of the query terms. In addition, we derive several analytical results concerning optimal bin allocation for a probabilistic query model. Our experimental evaluation with real life data shows an average I/O cost improvement of at least a factor of 10 for multi-dimensional queries on datasets from two different applications. Our experiments also indicate that the speedup increases with the number of query dimensions.

  18. V-204: A specially crafted query can cause BIND to terminate...

    Office of Energy Efficiency and Renewable Energy (EERE) Indexed Site

    affected source distributions may crash with assertion failures triggered in the same fashion. IMPACT: A specially crafted DNS query could cause the DNS service to terminate...

  19. Minimizing I/O Costs of Multi-Dimensional Queries with BitmapIndices

    SciTech Connect (OSTI)

    Rotem, Doron; Stockinger, Kurt; Wu, Kesheng

    2006-03-30

    Bitmap indices have been widely used in scientific applications and commercial systems for processing complex,multi-dimensional queries where traditional tree-based indices would not work efficiently. A common approach for reducing the size of a bitmap index for high cardinality attributes is to group ranges of values of an attribute into bins and then build a bitmap for each bin rather than a bitmap for each value of the attribute. Binning reduces storage costs,however, results of queries based on bins often require additional filtering for discarding it false positives, i.e., records in the result that do not satisfy the query constraints. This additional filtering,also known as ''candidate checking,'' requires access to the base data on disk and involves significant I/O costs. This paper studies strategies for minimizing the I/O costs for ''candidate checking'' for multi-dimensional queries. This is done by determining the number of bins allocated for each dimension and then placing bin boundaries in optimal locations. Our algorithms use knowledge of data distribution and query workload. We derive several analytical results concerning optimal bin allocation for a probabilistic query model. Our experimental evaluation with real life data shows an average I/O cost improvement of at least a factor of 10 for multi-dimensional queries on datasets from two different applications. Our experiments also indicate that the speedup increases with the number of query dimensions.

  20. Energy Information, Data, and other Resources | OpenEI

    Open Energy Info (EERE)

    SPARQL SPARQL OpenEI's SPARQL endpoint is accessible at sparql Sample SPARQL queries are available at resource...

  1. Cyber Graph Queries for Geographically Distributed Data Centers

    SciTech Connect (OSTI)

    Berry, Jonathan W.; Collins, Michael; Kearns, Aaron; Phillips, Cynthia A.; Saia, Jared

    2015-05-01

    We present new algorithms for a distributed model for graph computations motivated by limited information sharing we first discussed in [20]. Two or more independent entities have collected large social graphs. They wish to compute the result of running graph algorithms on the entire set of relationships. Because the information is sensitive or economically valuable, they do not wish to simply combine the information in a single location. We consider two models for computing the solution to graph algorithms in this setting: 1) limited-sharing: the two entities can share only a polylogarithmic size subgraph; 2) low-trust: the entities must not reveal any information beyond the query answer, assuming they are all honest but curious. We believe this model captures realistic constraints on cooperating autonomous data centers. We have algorithms in both setting for s - t connectivity in both models. We also give an algorithm in the low-communication model for finding a planted clique. This is an anomaly- detection problem, finding a subgraph that is larger and denser than expected. For both the low- communication algorithms, we exploit structural properties of social networks to prove perfor- mance bounds better than what is possible for general graphs. For s - t connectivity, we use known properties. For planted clique, we propose a new property: bounded number of triangles per node. This property is based upon evidence from the social science literature. We found that classic examples of social networks do not have the bounded-triangles property. This is because many social networks contain elements that are non-human, such as accounts for a business, or other automated accounts. We describe some initial attempts to distinguish human nodes from automated nodes in social networks based only on topological properties.

  2. Practical private database queries based on a quantum-key-distribution protocol

    SciTech Connect (OSTI)

    Jakobi, Markus; Simon, Christoph; Gisin, Nicolas; Bancal, Jean-Daniel; Branciard, Cyril; Walenta, Nino; Zbinden, Hugo

    2011-02-15

    Private queries allow a user, Alice, to learn an element of a database held by a provider, Bob, without revealing which element she is interested in, while limiting her information about the other elements. We propose to implement private queries based on a quantum-key-distribution protocol, with changes only in the classical postprocessing of the key. This approach makes our scheme both easy to implement and loss tolerant. While unconditionally secure private queries are known to be impossible, we argue that an interesting degree of security can be achieved by relying on fundamental physical principles instead of unverifiable security assumptions in order to protect both the user and the database. We think that the scope exists for such practical private queries to become another remarkable application of quantum information in the footsteps of quantum key distribution.

  3. U-038: BIND 9 Resolver crashes after logging an error in query.c

    Broader source: Energy.gov [DOE]

    A remote server can cause the target connected client to crash. Organizations across the Internet are reporting crashes interrupting service on BIND 9 nameservers performing recursive queries. Affected servers crash after logging an error in query.c with the following message: "INSIST(! dns_rdataset_isassociated(sigrdataset))" Multiple versions are reported as being affected, including all currently supported release versions of ISC BIND 9. ISC is actively investigating the root cause and working to produce patches which avoid the crash.

  4. U-039: ISC Update: BIND 9 Resolver crashes after logging an error in query.c

    Broader source: Energy.gov [DOE]

    A remote server can cause the target connected client to crash. Organizations across the Internet are reporting crashes interrupting service on BIND 9 nameservers performing recursive queries. Affected servers crash after logging an error in query.c with the following message: "INSIST(! dns_rdataset_isassociated(sigrdataset))" Multiple versions are reported as being affected, including all currently supported release versions of ISC BIND 9. ISC is actively investigating the root cause and working to produce patches which avoid the crash.

  5. Composing Data Parallel Code for a SPARQL Graph Engine

    SciTech Connect (OSTI)

    Castellana, Vito G.; Tumeo, Antonino; Villa, Oreste; Haglin, David J.; Feo, John

    2013-09-08

    Big data analytics process large amount of data to extract knowledge from them. Semantic databases are big data applications that adopt the Resource Description Framework (RDF) to structure metadata through a graph-based representation. The graph based representation provides several benefits, such as the possibility to perform in memory processing with large amounts of parallelism. SPARQL is a language used to perform queries on RDF-structured data through graph matching. In this paper we present a tool that automatically translates SPARQL queries to parallel graph crawling and graph matching operations. The tool also supports complex SPARQL constructs, which requires more than basic graph matching for their implementation. The tool generates parallel code annotated with OpenMP pragmas for x86 Shared-memory Multiprocessors (SMPs). With respect to commercial database systems such as Virtuoso, our approach reduces memory occupation due to join operations and provides higher performance. We show the scaling of the automatically generated graph-matching code on a 48-core SMP.

  6. An Application of Multivariate Statistical Analysis for Query-Driven Visualization

    SciTech Connect (OSTI)

    Gosink, Luke J.; Garth, Christoph; Anderson, John C.; Bethel, E. Wes; Joy, Kenneth I.

    2010-03-01

    Abstract?Driven by the ability to generate ever-larger, increasingly complex data, there is an urgent need in the scientific community for scalable analysis methods that can rapidly identify salient trends in scientific data. Query-Driven Visualization (QDV) strategies are among the small subset of techniques that can address both large and highly complex datasets. This paper extends the utility of QDV strategies with a statistics-based framework that integrates non-parametric distribution estimation techniques with a new segmentation strategy to visually identify statistically significant trends and features within the solution space of a query. In this framework, query distribution estimates help users to interactively explore their query's solution and visually identify the regions where the combined behavior of constrained variables is most important, statistically, to their inquiry. Our new segmentation strategy extends the distribution estimation analysis by visually conveying the individual importance of each variable to these regions of high statistical significance. We demonstrate the analysis benefits these two strategies provide and show how they may be used to facilitate the refinement of constraints over variables expressed in a user's query. We apply our method to datasets from two different scientific domains to demonstrate its broad applicability.

  7. Tool For Editing Structured Query Language Text Within ORACLE Forms Applications

    Energy Science and Technology Software Center (OSTI)

    1991-02-01

    SQTTEXT is an ORACLE SQL*Forms application that allows a programmer to view and edit all the Structured Query Language (SQL) text for a given application on one screen. This application is an outgrowth of the prototyping of an on-line system dictionary for the Worldwide Household Goods Information system for Transportation-Modernization decision support system being prototyped by the Oak Ridge National Laboratory, but it can be applied to all SQL*Forms software development, debugging, and maintenance.

  8. Computer systems and methods for the query and visualization of multidimensional databases

    DOE Patents [OSTI]

    Stolte, Chris; Tang, Diane L.; Hanrahan, Patrick

    2006-08-08

    A method and system for producing graphics. A hierarchical structure of a database is determined. A visual table, comprising a plurality of panes, is constructed by providing a specification that is in a language based on the hierarchical structure of the database. In some cases, this language can include fields that are in the database schema. The database is queried to retrieve a set of tuples in accordance with the specification. A subset of the set of tuples is associated with a pane in the plurality of panes.

  9. Computer systems and methods for the query and visualization of multidimensional database

    DOE Patents [OSTI]

    Stolte, Chris; Tang, Diane L.; Hanrahan, Patrick

    2010-05-11

    A method and system for producing graphics. A hierarchical structure of a database is determined. A visual table, comprising a plurality of panes, is constructed by providing a specification that is in a language based on the hierarchical structure of the database. In some cases, this language can include fields that are in the database schema. The database is queried to retrieve a set of tuples in accordance with the specification. A subset of the set of tuples is associated with a pane in the plurality of panes.

  10. A METHOD FOR ESTIMATING GAS PRESSURE IN 3013 CONTAINERS USING AN ISP DATABASE QUERY

    SciTech Connect (OSTI)

    Friday, G; L. G. Peppers, L; D. K. Veirs, D

    2008-07-31

    The U.S. Department of Energy's Integrated Surveillance Program (ISP) is responsible for the storage and surveillance of plutonium-bearing material. During storage, plutonium-bearing material has the potential to generate hydrogen gas from the radiolysis of adsorbed water. The generation of hydrogen gas is a safety concern, especially when a container is breached within a glove box during destructive evaluation. To address this issue, the DOE established a standard (DOE, 2004) that sets the criteria for the stabilization and packaging of material for up to 50 years. The DOE has now packaged most of its excess plutonium for long-term storage in compliance with this standard. As part of this process, it is desirable to know within reasonable certainty the total maximum pressure of hydrogen and other gases within the 3013 container if safety issues and compliance with the DOE standards are to be attained. The principal goal of this investigation is to document the method and query used to estimate total (i.e. hydrogen and other gases) gas pressure within a 3013 container based on the material properties and estimated moisture content contained in the ISP database. Initial attempts to estimate hydrogen gas pressure in 3013 containers was based on G-values (hydrogen gas generation per energy input) derived from small scale samples. These maximum G-values were used to calculate worst case pressures based on container material weight, assay, wattage, moisture content, container age, and container volume. This paper documents a revised hydrogen pressure calculation that incorporates new surveillance results and includes a component for gases other than hydrogen. The calculation is produced by executing a query of the ISP database. An example of manual mathematical computations from the pressure equation is compared and evaluated with results from the query. Based on the destructive evaluation of 17 containers, the estimated mean absolute pressure was significantly higher

  11. ISABELA-QA: query-driven analytics with ISABELA-compressed extreme-scale

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    scientific data | Argonne Leadership Computing Facility ISABELA-QA: query-driven analytics with ISABELA-compressed extreme-scale scientific data Authors: Lakshminarasimhan,S., Jenkins, J., Latham,R., Samatova, N.F., Arkatkar, I., Gong, Z., Kolla, H., Chen, J., Ku, S.H. Chang, C.S., Ethier, S., Klasky, S. Efficient analytics of scientific data from extreme-scale simulations is quickly becoming a top-notch priority. The increasing simulation output data sizes demand for a paradigm shift in how

  12. Query and Visualization of extremely large network datasets over the web using Quadtree based KML Regional Network Links

    SciTech Connect (OSTI)

    Dadi, Upendra; Liu, Cheng; Vatsavai, Raju

    2009-01-01

    Geographic data sets are often very large in size. Interactive visualization of such data at all scales is not easy because of the limited resolution of the monitors and inability of visualization applications to handle the volume of data. This is especially true for large vector datasets. The end user s experience is frequently unsatisfactory when exploring such data over the web using a naive application. Network bandwidth is another contributing factor to the low performance. In this paper, a Quadtree based technique to visualize extremely large spatial network datasets over the web is described. It involves using custom developed algorithms leveraging a PostGIS database as the data source and Google Earth as the visualization client. This methodology supports both point and range queries along with non-spatial queries. This methodology is demonstrated using a network dataset consisting of several million links. The methodology is based on using some of the powerful features of KML (Keyhole Markup Language). Keyhole Markup Language (KML) is an Open Geospatial Consortium (OGC) standard for displaying geospatial data on Earth browsers. One of the features of KML is the notion of Network Links. Using network links, a wide range of geospatial data sources such as geodatabases, static files and geospatial data services can be simultaneously accessed and visualized seamlessly. Using the network links combined with Level of Detail principle, view based rendering and intelligent server and client-side caching, scalability in visualizing extremely large spatial datasets can be achieved.

  13. HDF5-FastQuery: Accelerating Complex Queries

    Office of Scientific and Technical Information (OSTI)

    ... On the Performance of Bitmap Indices for High Cardinality Attributes. In International Conference on Very Large Data Bases (VLDB), Toronto, Canada, August 31 - September 3, 2004, ...

  14. HDF5-FastQuery: Accelerating Complex Queries

    Office of Scientific and Technical Information (OSTI)

    ... These groups act as a container of various metadata which in our approach is ... is based on the HDF5 indexing function H5INquery, and takes three arrays as function input. ...

  15. Complex Queries | Open Energy Information

    Open Energy Info (EERE)

    Electricity Markets Afghanistan-NREL Mission Afghanistan-NREL Resource Maps and Toolkits China-NREL Cooperation Dominica Island-NREL Cooperation Egypt-NREL Energy Activities...

  16. Example Queries | Open Energy Information

    Open Energy Info (EERE)

    Avoca, New York Avoca, Pennsylvania Avoca, Wisconsin Avocado Heights, California Avon Lake, Ohio Avon Park, Florida Avon, Alabama Avon, Colorado Avon, Connecticut Avon,...

  17. User:Woodjr/Sandbox/Sparql2 | Open Energy Information

    Open Energy Info (EERE)

    SandboxSparql2 < User:Woodjr | Sandbox Jump to: navigation, search States of the USA Retrieved from "http:en.openei.orgwindex.php?titleUser:WoodjrSandbox...

  18. User:Woodjr/Sandbox/Sparql3 | Open Energy Information

    Open Energy Info (EERE)

    Sparql3 < User:Woodjr | Sandbox Jump to: navigation, search States of the USA which have a geographic "point" defined in DBpedia Loading map... "minzoom":false,"mappingservice"...

  19. User:Woodjr/Sandbox/Sparql | Open Energy Information

    Open Energy Info (EERE)

    WoodjrSandboxSparql < User:Woodjr | Sandbox Jump to: navigation, search Software Developed by Companies Founded in California Retrieved from "http:en.openei.orgw...

  20. Table Name query? | OpenEI Community

    Open Energy Info (EERE)

    - 06:39 Groups Menu You must login in order to post into this group. Recent content Hello-Sorry for the delay in... Use of DynamicAggregationProcessor I submitted a pull...

  1. Natural Gas Annual Respondent Query System

    Gasoline and Diesel Fuel Update (EIA)

    (Volumes in Thousand Cubic Feet, Prices in Dollars per Thousand Cubic Feet) Form EIA-176 * User Guide * Definitions, Sources, & Notes Natural Gas Deliveries (2011 - 2014)...

  2. Querying Allocations Using cbank | Argonne Leadership Computing...

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    20130101 -b 20130131 charges on all projects for a particular user cbank charges -u username -a 20130401 get specifics on a single charge cbank detail charge NNNNN list ...

  3. Image subregion querying using color correlograms

    DOE Patents [OSTI]

    Huang, Jing; Kumar, Shanmugasundaram Ravi; Mitra, Mandar; Zhu, Wei-Jing

    2002-01-01

    A color correlogram (10) is a representation expressing the spatial correlation of color and distance between pixels in a stored image. The color correlogram (10) may be used to distinguish objects in an image as well as between images in a plurality of images. By intersecting a color correlogram of an image object with correlograms of images to be searched, those images which contain the objects are identified by the intersection correlogram.

  4. developer | OpenEI Community

    Open Energy Info (EERE)

    apps, lod, sparql and community will continue to function normally. Additionally, web services that rely on Ask queries (utility rate database API) may have some downtime...

  5. HDF5-FastQuery: Accelerating Complex Queries on HDF Datasets...

    Office of Scientific and Technical Information (OSTI)

    the scientific user community since they provide multi-dimensional storage and retrieval. ... Subject: 99 GENERAL AND MISCELLANEOUSMATHEMATICS, COMPUTING, AND INFORMATION SCIENCE; ...

  6. T-703: Cisco Unified Communications Manager Open Query Interface...

    Office of Energy Efficiency and Renewable Energy (EERE) Indexed Site

    A remote user can obtain database contents, including authentication credentials. ... This may include authentication credentials, configuration details, and other sensitive ...

  7. Improving Estimation Accuracy of Aggregate Queries on Data Cubes

    SciTech Connect (OSTI)

    Pourabbas, Elaheh; Shoshani, Arie

    2008-08-15

    In this paper, we investigate the problem of estimation of a target database from summary databases derived from a base data cube. We show that such estimates can be derived by choosing a primary database which uses a proxy database to estimate the results. This technique is common in statistics, but an important issue we are addressing is the accuracy of these estimates. Specifically, given multiple primary and multiple proxy databases, that share the same summary measure, the problem is how to select the primary and proxy databases that will generate the most accurate target database estimation possible. We propose an algorithmic approach for determining the steps to select or compute the source databases from multiple summary databases, which makes use of the principles of information entropy. We show that the source databases with the largest number of cells in common provide the more accurate estimates. We prove that this is consistent with maximizing the entropy. We provide some experimental results on the accuracy of the target database estimation in order to verify our results.

  8. Category:Query Results Templates | Open Energy Information

    Open Energy Info (EERE)

    The following 4 pages are in this category, out of 4 total. D Template:DefineVariables L Template:LabelActivities Template:LabelValuePair S Template:SubPageListHelper Retrieved...

  9. Graph Mining Meets the Semantic Web

    SciTech Connect (OSTI)

    Lee, Sangkeun; Sukumar, Sreenivas R; Lim, Seung-Hwan

    2015-01-01

    The Resource Description Framework (RDF) and SPARQL Protocol and RDF Query Language (SPARQL) were introduced about a decade ago to enable flexible schema-free data interchange on the Semantic Web. Today, data scientists use the framework as a scalable graph representation for integrating, querying, exploring and analyzing data sets hosted at different sources. With increasing adoption, the need for graph mining capabilities for the Semantic Web has emerged. We address that need through implementation of three popular iterative Graph Mining algorithms (Triangle count, Connected component analysis, and PageRank). We implement these algorithms as SPARQL queries, wrapped within Python scripts. We evaluate the performance of our implementation on 6 real world data sets and show graph mining algorithms (that have a linear-algebra formulation) can indeed be unleashed on data represented as RDF graphs using the SPARQL query interface.

  10. Template:LabelValuePair | Open Energy Information

    Open Energy Info (EERE)

    is typically used to display the results of an ask or sparql query in a simple label: value format. It is used by many pages, including the sub pages for country profiles, and is...

  11. Massive-scale RDF Processing Using Compressed Bitmap Indexes

    SciTech Connect (OSTI)

    Madduri, Kamesh; Wu, Kesheng

    2011-05-26

    The Resource Description Framework (RDF) is a popular data model for representing linked data sets arising from the web, as well as large scienti#12;c data repositories such as UniProt. RDF data intrinsically represents a labeled and directed multi-graph. SPARQL is a query language for RDF that expresses subgraph pattern-#12;nding queries on this implicit multigraph in a SQL- like syntax. SPARQL queries generate complex intermediate join queries; to compute these joins e#14;ciently, we propose a new strategy based on bitmap indexes. We store the RDF data in column-oriented structures as compressed bitmaps along with two dictionaries. This paper makes three new contributions. (i) We present an e#14;cient parallel strategy for parsing the raw RDF data, building dictionaries of unique entities, and creating compressed bitmap indexes of the data. (ii) We utilize the constructed bitmap indexes to e#14;ciently answer SPARQL queries, simplifying the join evaluations. (iii) To quantify the performance impact of using bitmap indexes, we compare our approach to the state-of-the-art triple-store RDF-3X. We #12;nd that our bitmap index-based approach to answering queries is up to an order of magnitude faster for a variety of SPARQL queries, on gigascale RDF data sets.

  12. Enabling Graph Mining in RDF Triplestores using SPARQL for Holistic In-situ Graph Analysis

    SciTech Connect (OSTI)

    Lee, Sangkeun; Sukumar, Sreenivas R; Hong, Seokyong; Lim, Seung-Hwan

    2016-01-01

    The graph analysis is now considered as a promising technique to discover useful knowledge in data with a new perspective. We envi- sion that there are two dimensions of graph analysis: OnLine Graph Analytic Processing (OLGAP) and Graph Mining (GM) where each respectively focuses on subgraph pattern matching and automatic knowledge discovery in graph. Moreover, as these two dimensions aim to complementarily solve complex problems, holistic in-situ graph analysis which covers both OLGAP and GM in a single system is critical for minimizing the burdens of operating multiple graph systems and transferring intermediate result-sets between those systems. Nevertheless, most existing graph analysis systems are only capable of one dimension of graph analysis. In this work, we take an approach to enabling GM capabilities (e.g., PageRank, connected-component analysis, node eccentricity, etc.) in RDF triplestores, which are originally developed to store RDF datasets and provide OLGAP capability. More specifically, to achieve our goal, we implemented six representative graph mining algorithms using SPARQL. The approach allows a wide range of available RDF data sets directly applicable for holistic graph analysis within a system. For validation of our approach, we evaluate performance of our implementations with nine real-world datasets and three different computing environments - a laptop computer, an Amazon EC2 instance, and a shared-memory Cray XMT2 URIKA-GD graph-processing appliance. The experimen- tal results show that our implementation can provide promising and scalable performance for real world graph analysis in all tested environments. The developed software is publicly available in an open-source project that we initiated.

  13. Enabling Graph Mining in RDF Triplestores using SPARQL for Holistic In-situ Graph Analysis

    DOE Public Access Gateway for Energy & Science Beta (PAGES Beta)

    Lee, Sangkeun; Sukumar, Sreenivas R; Hong, Seokyong; Lim, Seung-Hwan

    2016-01-01

    The graph analysis is now considered as a promising technique to discover useful knowledge in data with a new perspective. We envi- sion that there are two dimensions of graph analysis: OnLine Graph Analytic Processing (OLGAP) and Graph Mining (GM) where each respectively focuses on subgraph pattern matching and automatic knowledge discovery in graph. Moreover, as these two dimensions aim to complementarily solve complex problems, holistic in-situ graph analysis which covers both OLGAP and GM in a single system is critical for minimizing the burdens of operating multiple graph systems and transferring intermediate result-sets between those systems. Nevertheless, most existingmore » graph analysis systems are only capable of one dimension of graph analysis. In this work, we take an approach to enabling GM capabilities (e.g., PageRank, connected-component analysis, node eccentricity, etc.) in RDF triplestores, which are originally developed to store RDF datasets and provide OLGAP capability. More specifically, to achieve our goal, we implemented six representative graph mining algorithms using SPARQL. The approach allows a wide range of available RDF data sets directly applicable for holistic graph analysis within a system. For validation of our approach, we evaluate performance of our implementations with nine real-world datasets and three different computing environments - a laptop computer, an Amazon EC2 instance, and a shared-memory Cray XMT2 URIKA-GD graph-processing appliance. The experimen- tal results show that our implementation can provide promising and scalable performance for real world graph analysis in all tested environments. The developed software is publicly available in an open-source project that we initiated.« less

  14. EAGLE: 'EAGLE'Is an' Algorithmic Graph Library for Exploration

    SciTech Connect (OSTI)

    2015-01-16

    The Resource Description Framework (RDF) and SPARQL Protocol and RDF Query Language (SPARQL) were introduced about a decade ago to enable flexible schema-free data interchange on the Semantic Web. Today data scientists use the framework as a scalable graph representation for integrating, querying, exploring and analyzing data sets hosted at different sources. With increasing adoption, the need for graph mining capabilities for the Semantic Web has emerged. Today there is no tools to conduct "graph mining" on RDF standard data sets. We address that need through implementation of popular iterative Graph Mining algorithms (Triangle count, Connected component analysis, degree distribution, diversity degree, PageRank, etc.). We implement these algorithms as SPARQL queries, wrapped within Python scripts and call our software tool as EAGLE. In RDF style, EAGLE stands for "EAGLE 'Is an' algorithmic graph library for exploration. EAGLE is like 'MATLAB' for 'Linked Data.'

  15. EAGLE: 'EAGLE'Is an' Algorithmic Graph Library for Exploration

    Energy Science and Technology Software Center (OSTI)

    2015-01-16

    The Resource Description Framework (RDF) and SPARQL Protocol and RDF Query Language (SPARQL) were introduced about a decade ago to enable flexible schema-free data interchange on the Semantic Web. Today data scientists use the framework as a scalable graph representation for integrating, querying, exploring and analyzing data sets hosted at different sources. With increasing adoption, the need for graph mining capabilities for the Semantic Web has emerged. Today there is no tools to conduct "graphmore » mining" on RDF standard data sets. We address that need through implementation of popular iterative Graph Mining algorithms (Triangle count, Connected component analysis, degree distribution, diversity degree, PageRank, etc.). We implement these algorithms as SPARQL queries, wrapped within Python scripts and call our software tool as EAGLE. In RDF style, EAGLE stands for "EAGLE 'Is an' algorithmic graph library for exploration. EAGLE is like 'MATLAB' for 'Linked Data.'« less

  16. Geothermal Prospector: Supporting Geothermal Analysis Through Spatial Data Visualization and Querying Tools

    SciTech Connect (OSTI)

    Getman, Daniel; Anderson, Arlene; Augustine, Chad

    2015-09-02

    Determining opportunities for geothermal energy can involve a significant investment in data collection and analysis. Analysts within a variety of industry and research domains collect and use these data; however, determining the existence and availability of data needed for a specific analysis activity can be challenging and represents one of the initial barriers to geothermal development [2]. This paper describes the motivating factors involved in designing and building the Geothermal Prospector application, how it can be used to reduce risks and costs related to geothermal exploration, and where it fits within the larger collection of tools that is the National Geothermal Data System (NGDS) [5].

  17. Computer systems and methods for the query and visualization of multidimensional databases

    DOE Patents [OSTI]

    Stolte, Chris; Tang, Diane L.; Hanrahan, Patrick

    2012-03-20

    In response to a user request, a computer generates a graphical user interface on a computer display. A schema information region of the graphical user interface includes multiple operand names, each operand name associated with one or more fields of a multi-dimensional database. A data visualization region of the graphical user interface includes multiple shelves. Upon detecting a user selection of the operand names and a user request to associate each user-selected operand name with a respective shelf in the data visualization region, the computer generates a visual table in the data visualization region in accordance with the associations between the operand names and the corresponding shelves. The visual table includes a plurality of panes, each pane having at least one axis defined based on data for the fields associated with a respective operand name.

  18. Computer systems and methods for the query and visualization of multidimensional databases

    DOE Patents [OSTI]

    Stolte, Chris; Tang, Diane L; Hanrahan, Patrick

    2015-03-03

    A computer displays a graphical user interface on its display. The graphical user interface includes a schema information region and a data visualization region. The schema information region includes multiple operand names, each operand corresponding to one or more fields of a multi-dimensional database that includes at least one data hierarchy. The data visualization region includes a columns shelf and a rows shelf. The computer detects user actions to associate one or more first operands with the columns shelf and to associate one or more second operands with the rows shelf. The computer generates a visual table in the data visualization region in accordance with the user actions. The visual table includes one or more panes. Each pane has an x-axis defined based on data for the one or more first operands, and each pane has a y-axis defined based on data for the one or more second operands.

  19. Computer systems and methods for the query and visualization of multidimensional databases

    DOE Patents [OSTI]

    Stolte, Chris; Tang, Diane L.; Hanrahan, Patrick

    2015-11-10

    A computer displays a graphical user interface on its display. The graphical user interface includes a schema information region and a data visualization region. The schema information region includes a plurality of fields of a multi-dimensional database that includes at least one data hierarchy. The data visualization region includes a columns shelf and a rows shelf. The computer detects user actions to associate one or more first fields with the columns shelf and to associate one or more second fields with the rows shelf. The computer generates a visual table in the data visualization region in accordance with the user actions. The visual table includes one or more panes. Each pane has an x-axis defined based on data for the one or more first fields, and each pane has a y-axis defined based on data for the one or more second fields.

  20. Computer systems and methods for the query and visualization of multidimensional databases

    DOE Patents [OSTI]

    Stolte, Chris; Tang, Diane L; Hanrahan, Patrick

    2014-04-29

    In response to a user request, a computer generates a graphical user interface on a computer display. A schema information region of the graphical user interface includes multiple operand names, each operand name associated with one or more fields of a multi-dimensional database. A data visualization region of the graphical user interface includes multiple shelves. Upon detecting a user selection of the operand names and a user request to associate each user-selected operand name with a respective shelf in the data visualization region, the computer generates a visual table in the data visualization region in accordance with the associations between the operand names and the corresponding shelves. The visual table includes a plurality of panes, each pane having at least one axis defined based on data for the fields associated with a respective operand name.

  1. Computer systems and methods for the query and visualization of multidimensional databases

    DOE Patents [OSTI]

    Stolte, Chris; Tang, Diane L.; Hanrahan, Patrick

    2011-02-01

    In response to a user request, a computer generates a graphical user interface on a computer display. A schema information region of the graphical user interface includes multiple operand names, each operand name associated with one or more fields of a multi-dimensional database. A data visualization region of the graphical user interface includes multiple shelves. Upon detecting a user selection of the operand names and a user request to associate each user-selected operand name with a respective shelf in the data visualization region, the computer generates a visual table in the data visualization region in accordance with the associations between the operand names and the corresponding shelves. The visual table includes a plurality of panes, each pane having at least one axis defined based on data for the fields associated with a respective operand name.

  2. Accelerating semantic graph databases on commodity clusters

    SciTech Connect (OSTI)

    Morari, Alessandro; Castellana, Vito G.; Haglin, David J.; Feo, John T.; Weaver, Jesse R.; Tumeo, Antonino; Villa, Oreste

    2013-10-06

    We are developing a full software system for accelerating semantic graph databases on commodity cluster that scales to hundreds of nodes while maintaining constant query throughput. Our framework comprises a SPARQL to C++ compiler, a library of parallel graph methods and a custom multithreaded runtime layer, which provides a Partitioned Global Address Space (PGAS) programming model with fork/join parallelism and automatic load balancing over a commodity clusters. We present preliminary results for the compiler and for the runtime.

  3. Developer | OpenEI Community

    Open Energy Info (EERE)

    ask queries Type Term Title Author Replies Last Post sort icon Blog entry ask queries Multicolor Maps from Compound Queries Jweers 16 May 2013 - 14:22...

  4. Developer | OpenEI Community

    Open Energy Info (EERE)

    compound queries Type Term Title Author Replies Last Post sort icon Blog entry compound queries Multicolor Maps from Compound Queries Jweers 16 May 2013 - 14:22...

  5. FALDO: A semantic standard for describing the location of nucleotide and protein feature annotation

    DOE Public Access Gateway for Energy & Science Beta (PAGES Beta)

    Bolleman, Jerven T.; Mungall, Christopher J.; Strozzi, Francesco; Baran, Joachim; Dumontier, Michel; Bonnal, Raoul J. P.; Buels, Robert; Hoehndorf, Robert; Fujisawa, Takatomo; Katayama, Toshiaki; et al

    2016-06-13

    In this study, nucleotide and protein sequence feature annotations are essential to understand biology on the genomic, transcriptomic, and proteomic level. Using Semantic Web technologies to query biological annotations, there was no standard that described this potentially complex location information as subject-predicate-object triples. Here, we have developed an ontology, the Feature Annotation Location Description Ontology (FALDO), to describe the positions of annotated features on linear and circular sequences. FALDO can be used to describe nucleotide features in sequence records, protein annotations, and glycan binding sites, among other features in coordinate systems of the aforementioned “omics” areas. Using the same datamore » format to represent sequence positions that are independent of file formats allows us to integrate sequence data from multiple sources and data types. The genome browser JBrowse is used to demonstrate accessing multiple SPARQL endpoints to display genomic feature annotations, as well as protein annotations from UniProt mapped to genomic locations. In conclusion, our ontology allows users to uniformly describe – and potentially merge – sequence annotations from multiple sources. Data sources using FALDO can prospectively be retrieved using federalised SPARQL queries against public SPARQL endpoints and/or local private triple stores.« less

  6. FALDO: a semantic standard for describing the location of nucleotide and protein feature annotation

    DOE Public Access Gateway for Energy & Science Beta (PAGES Beta)

    Bolleman, Jerven T.; Mungall, Christopher J.; Strozzi, Francesco; Baran, Joachim; Dumontier, Michel; Bonnal, Raoul J. P.; Buels, Robert; Hoehndorf, Robert; Fujisawa, Takatomo; Katayama, Toshiaki; et al

    2016-06-13

    Nucleotide and protein sequence feature annotations are essential to understand biology on the genomic, transcriptomic, and proteomic level. Using Semantic Web technologies to query biological annotations, there was no standard that described this potentially complex location information as subject-predicate-object triples. In this paper, we have developed an ontology, the Feature Annotation Location Description Ontology (FALDO), to describe the positions of annotated features on linear and circular sequences. FALDO can be used to describe nucleotide features in sequence records, protein annotations, and glycan binding sites, among other features in coordinate systems of the aforementioned “omics” areas. Using the same data formatmore » to represent sequence positions that are independent of file formats allows us to integrate sequence data from multiple sources and data types. The genome browser JBrowse is used to demonstrate accessing multiple SPARQL endpoints to display genomic feature annotations, as well as protein annotations from UniProt mapped to genomic locations. Our ontology allows users to uniformly describe – and potentially merge – sequence annotations from multiple sources. Finally, data sources using FALDO can prospectively be retrieved using federalised SPARQL queries against public SPARQL endpoints and/or local private triple stores.« less

  7. Enabling Graph Appliance for Genome Assembly

    SciTech Connect (OSTI)

    Singh, Rina; Graves, Jeffrey A; Lee, Sangkeun; Sukumar, Sreenivas R; Shankar, Mallikarjun

    2015-01-01

    In recent years, there has been a huge growth in the amount of genomic data available as reads generated from various genome sequencers. The number of reads generated can be huge, ranging from hundreds to billions of nucleotide, each varying in size. Assembling such large amounts of data is one of the challenging computational problems for both biomedical and data scientists. Most of the genome assemblers developed have used de Bruijn graph techniques. A de Bruijn graph represents a collection of read sequences by billions of vertices and edges, which require large amounts of memory and computational power to store and process. This is the major drawback to de Bruijn graph assembly. Massively parallel, multi-threaded, shared memory systems can be leveraged to overcome some of these issues. The objective of our research is to investigate the feasibility and scalability issues of de Bruijn graph assembly on Cray s Urika-GD system; Urika-GD is a high performance graph appliance with a large shared memory and massively multithreaded custom processor designed for executing SPARQL queries over large-scale RDF data sets. However, to the best of our knowledge, there is no research on representing a de Bruijn graph as an RDF graph or finding Eulerian paths in RDF graphs using SPARQL for potential genome discovery. In this paper, we address the issues involved in representing a de Bruin graphs as RDF graphs and propose an iterative querying approach for finding Eulerian paths in large RDF graphs. We evaluate the performance of our implementation on real world ebola genome datasets and illustrate how genome assembly can be accomplished with Urika-GD using iterative SPARQL queries.

  8. Toward a Data Scalable Solution for Facilitating Discovery of Science Resources

    SciTech Connect (OSTI)

    Weaver, Jesse R.; Castellana, Vito G.; Morari, Alessandro; Tumeo, Antonino; Purohit, Sumit; Chappell, Alan R.; Haglin, David J.; Villa, Oreste; Choudhury, Sutanay; Schuchardt, Karen L.; Feo, John T.

    2014-12-31

    Science is increasingly motivated by the need to process larger quantities of data. It is facing severe challenges in data collection, management, and processing, so much so that the computational demands of data scaling are competing with, and in many fields surpassing, the traditional objective of decreasing processing time. Example domains with large datasets include astronomy, biology, genomics, climate/weather, and material sciences. This paper presents a real-world use case in which we wish to answer queries pro- vided by domain scientists in order to facilitate discovery of relevant science resources. The problem is that the metadata for these science resources is very large and is growing quickly, rapidly increasing the need for a data scaling solution. We propose a system SGEM designed for answering graph-based queries over large datasets on cluster architectures, and we re- port performance results for queries on the current RDESC dataset of nearly 1.4 billion triples, and on the well-known BSBM SPARQL query benchmark.

  9. In-Memory Graph Databases for Web-Scale Data

    SciTech Connect (OSTI)

    Castellana, Vito G.; Morari, Alessandro; Weaver, Jesse R.; Tumeo, Antonino; Haglin, David J.; Villa, Oreste; Feo, John

    2015-03-01

    RDF databases have emerged as one of the most relevant way for organizing, integrating, and managing expo- nentially growing, often heterogeneous, and not rigidly structured data for a variety of scientific and commercial fields. In this paper we discuss the solutions integrated in GEMS (Graph database Engine for Multithreaded Systems), a software framework for implementing RDF databases on commodity, distributed-memory high-performance clusters. Unlike the majority of current RDF databases, GEMS has been designed from the ground up to primarily employ graph-based methods. This is reflected in all the layers of its stack. The GEMS framework is composed of: a SPARQL-to-C++ compiler, a library of data structures and related methods to access and modify them, and a custom runtime providing lightweight software multithreading, network messages aggregation and a partitioned global address space. We provide an overview of the framework, detailing its component and how they have been closely designed and customized to address issues of graph methods applied to large-scale datasets on clusters. We discuss in details the principles that enable automatic translation of the queries (expressed in SPARQL, the query language of choice for RDF databases) to graph methods, and identify differences with respect to other RDF databases.

  10. Energy Information, Data, and other Resources | OpenEI

    Open Energy Info (EERE)

    provides inline queries in the form of an Ask query, which can also be modified into a web service (see OpenEI REST services documentation). Ask queries can be executed here and...

  11. All | OpenEI Community

    Open Energy Info (EERE)

    All Home > All By term Q & A Term: ask queries Type Term Title Author Replies Last Post sort icon Blog entry ask queries Multicolor Maps from Compound Queries Jweers 16 May 2013 -...

  12. Sounding Board V.1.0

    Energy Science and Technology Software Center (OSTI)

    2006-10-10

    Sounding Board allows users to query multiple models simultaneously, finding relevant experts, related terms, and historical text related to one's query.

  13. Help:External SPARQL integration | Open Energy Information

    Open Energy Info (EERE)

    navigation, search Integrating with Reegle logo.png OpenEI is engaged in an ongoing linked open data collaboration with Reegle1. This page serves to document a few of the...

  14. User:Woodjr/Sandbox/Sparql4 | Open Energy Information

    Open Energy Info (EERE)

    2005 1 4,506,411.00 MO 2005 1 8,248,149.00 ND 2005 1 2,760,136.00 NJ 2005 1 116,877.00 OK 2005 1 4,413,489.00 OR 2005 1 3,804,311.00 SC 2005 1 8,712,013.00 VT 2005 1 65,139.00 WA...

  15. Microsoft Word - NG_ResQrySys_UsersGuide_Sept2015-FINAL.docx

    Gasoline and Diesel Fuel Update (EIA)

    ... the Query System is a web-based system, no download or installation is necessary. All that is needed to run the Query System is a PC with up-to-date web-browsing software (such ...

  16. Geospatial | OpenEI Community

    Open Energy Info (EERE)

    Geospatial > Posts by term Content Group Activity By term Q & A Feeds ask queries (1) compound queries (1) data (1) developer (1) geospatial data (1) GIS (1) GIS data (1) Global...

  17. Sqlog

    Energy Science and Technology Software Center (OSTI)

    2007-08-22

    The sqlog software implements a system for creation, query, and maintenance of database for SLURM job history.

  18. Search for: All records | SciTech Connect

    Office of Scientific and Technical Information (OSTI)

    in ITER. less Full Text Available April 2013, Physical Review Letters (Novembe r2012) Prev Next SOLR Query Details...

  19. Estimating Missing Features to Improve Multimedia Information Retrieval

    SciTech Connect (OSTI)

    Bagherjeiran, A; Love, N S; Kamath, C

    2006-09-28

    Retrieval in a multimedia database usually involves combining information from different modalities of data, such as text and images. However, all modalities of the data may not be available to form the query. The retrieval results from such a partial query are often less than satisfactory. In this paper, we present an approach to complete a partial query by estimating the missing features in the query. Our experiments with a database of images and their associated captions show that, with an initial text-only query, our completion method has similar performance to a full query with both image and text features. In addition, when we use relevance feedback, our approach outperforms the results obtained using a full query.

  20. Unhappy with internal corporate search? : learn tips and tricks for building a controlled vocabulary ontology.

    SciTech Connect (OSTI)

    Arpin, Bettina Karin Schimanski; Jones, Brian S.; Bemesderfer, Joy; Ralph, Mark E.; Miller, Jennifer L

    2010-06-01

    Are your employees unhappy with internal corporate search? Frequent complaints include: too many results to sift through; results are unrelated/outdated; employees aren't sure which terms to search for. One way to improve intranet search is to implement a controlled vocabulary ontology. Employing this takes the guess work out of searching, makes search efficient and precise, educates employees about the lingo used within the corporation, and allows employees to contribute to the corpus of terms. It promotes internal corporate search to rival its superior sibling, internet search. We will cover our experiences, lessons learned, and conclusions from implementing a controlled vocabulary ontology at Sandia National Laboratories. The work focuses on construction of this ontology from the content perspective and the technical perspective. We'll discuss the following: (1) The tool we used to build a polyhierarchical taxonomy; (2) Examples of two methods of indexing the content: traditional 'back of the book' and folksonomy word-mapping; (3) Tips on how to build future search capabilities while building the basic controlled vocabulary; (4) How to implement the controlled vocabulary as an ontology that mimics Google's search suggestions; (5) Making the user experience more interactive and intuitive; and (6) Sorting suggestions based on preferred, alternate and related terms using SPARQL queries. In summary, future improvements will be presented, including permitting end-users to add, edit and remove terms, and filtering on different subject domains.

  1. SWNVF: database contents

    National Nuclear Security Administration (NNSA)

    Nevada Test Site (NTS). The database complies with protocols of Structured Query Language (SQL), allowing construction of relationships among these data, from...

  2. Electric Power Annual 2014 - U.S. Energy Information Administration

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Overview Data Electricity Data Browser (interactive query tool with charting & mapping) Summary Sales (consumption), ... to ultimate customers by end-use sector XLS Table 2.5. ...

  3. Electric Power Monthly - U.S. Energy Information Administration

    Annual Energy Outlook [U.S. Energy Information Administration (EIA)]

    Overview Data Electricity Data Browser (interactive query tool with charting & mapping) Summary Sales (consumption), ... Customers: Total by End-Use Sector 5.2 Revenue from ...

  4. Electricity Monthly Update - Energy Information Administration

    Annual Energy Outlook [U.S. Energy Information Administration (EIA)]

    Overview Data Electricity Data Browser (interactive query tool with charting & mapping) Summary Sales (consumption), ... demand levels that occurred towards the end of the month. ...

  5. Gateway:ECOWAS Clean Energy Gateway | Open Energy Information

    Open Energy Info (EERE)

    Policy Organizations (3) add West African Companies (4) add West African Programs (76The part "|Programs and Projects" of the query was not understood. Results might not...

  6. Search for: All records | SciTech Connect

    Office of Scientific and Technical Information (OSTI)

    ... factors as well as explain some mutational effects. less August 2012 , American Chemical Society (ACS) Prev Next Switch to Detail View for this search SOLR Query Details

  7. Search for: All records | SciTech Connect

    Office of Scientific and Technical Information (OSTI)

    Thomas C. ; Petsko, Gregory A. ; Ringe, Dagmar ; Brandeis) April 2016 , American Chemical Society (ACS) Prev Next Switch to Detail View for this search SOLR Query Details

  8. Prueba 3 | Open Energy Information

    Open Energy Info (EERE)

    de Redes (98) add Empresas de Energas Renovables (12) add Programas y Proyectos (1157The part "|Programs and Projects" of the query was not understood....

  9. OSTI, US Dept of Energy, Office of Scientific and Technical Informatio...

    Office of Scientific and Technical Information (OSTI)

    Using web-crawling technology, the search capability allows users to enter a single query for a technology transfer term; the search feature returns a consolidated, ...

  10. OSTI, US Dept of Energy Office of Scientific and Technical Information...

    Office of Scientific and Technical Information (OSTI)

    Web "mashups" combine multiple productsservices into a single application for the purpose ... one-stop simultaneous searching of multiple networked data resources via a single query. ...

  11. OSTI, US Dept of Energy Office of Scientific and Technical Information...

    Office of Scientific and Technical Information (OSTI)

    ... The NLEBeta employs two highly effective web tools - federated search, the simultaneous search of multiple web pages and databases in real time via a single query, and relevancy ...

  12. A Proactive Learning Framework for Non-Intrusive Load Monitoring...

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    NILM systems perform analysis on whole-building data taken at the main power panel to ... learning, an advanced machine learning method that interactively queries the user for ...

  13. Load flow analysis: Base cases, data, diagrams, and results ...

    Office of Scientific and Technical Information (OSTI)

    The report summarizes the load flow model construction, simulation, and validation and describes the general capabilities of an information query system designed to access load ...

  14. T-617: BIND RPZ Processing Flaw Lets Remote Users Deny Service

    Broader source: Energy.gov [DOE]

    When a name server is configured with a response policy zone (RPZ), queries for type RRSIG can trigger a server crash.

  15. T-559: Stack-based buffer overflow in oninit in IBM Informix...

    Office of Energy Efficiency and Renewable Energy (EERE) Indexed Site

    exploit this vulnerability. The specific flaw exists within the oninit process bound to TCP port 9088 when processing the arguments to the USELASTCOMMITTED option in a SQL query....

  16. Template:UnderDevelopment | Open Energy Information

    Open Energy Info (EERE)

    source History View New Pages Recent Changes All Special Pages Semantic SearchQuerying Get Involved Help Apps Datasets Community Login | Sign Up Search Template Edit History...

  17. Allocations | Argonne Leadership Computing Facility

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Allocation Management Determining Allocation Requirements Querying Allocations Using cbank MiraCetusVesta Cooley Policies Documentation Feedback Please provide feedback to help...

  18. Search for: All records | DOE Data Explorer

    Office of Scientific and Technical Information (OSTI)

    Save Results Excel (limit 2000) CSV (limit 5000) XML (limit 5000) Have feedback or suggestions for a way to improve these results? Prev Next SOLR Query Details Close...

  19. Search for: All records | SciTech Connect

    Office of Scientific and Technical Information (OSTI)

    Barbeau, P.S. ; Stanford U., Phys. Dept. ; Beauchamp, E. ; Laurentian U. ; Belov, V. ; Moscow, ITEP ; et al Full Text Available March 2013 Prev Next SOLR Query Details...

  20. Search for: All records | SciTech Connect

    Office of Scientific and Technical Information (OSTI)

    L. ; Winiarz, Jeffrey G. ; Kaiser, Helmut ; Taub, Haskell ; Veith, Gabriel M. December 2015 , Elsevier Prev Next Switch to Detail View for this search SOLR Query Details

  1. Search for: All records | SciTech Connect

    Office of Scientific and Technical Information (OSTI)

    and one where protein globules cause inter-aggregate clustering. less December 2015 , Elsevier Prev Next Switch to Detail View for this search SOLR Query Details

  2. Search for: All records | SciTech Connect

    Office of Scientific and Technical Information (OSTI)

    Xiao Mei ; Du, Zehui ; Schuh, Christopher A. ; Tamura, Nobumichi ; Gan, Chee Lip April 2016 , Elsevier Prev Next Switch to Detail View for this search SOLR Query Details

  3. Search for: All records | SciTech Connect

    Office of Scientific and Technical Information (OSTI)

    ; Liang, Shuquan ; Liu, Jun ; Wang, Wenjun ; Lei, Ming ; Tang, Shasha ; Yang, Qian March 2015 , Elsevier Prev Next Switch to Detail View for this search SOLR Query Details

  4. Search for: All records | SciTech Connect

    Office of Scientific and Technical Information (OSTI)

    ... retrieval algorithm for aerosol optical depth (AOD) in the presence of clouds using ... Richland, WA Prev Next Switch to Detail View for this search SOLR Query Details

  5. Template:TATNav | Open Energy Information

    Open Energy Info (EERE)

    source History View New Pages Recent Changes All Special Pages Semantic SearchQuerying Get Involved Help Apps Datasets Community Login | Sign Up Search Template Edit History...

  6. Template:ResourceLibraryTabs | Open Energy Information

    Open Energy Info (EERE)

    source History View New Pages Recent Changes All Special Pages Semantic SearchQuerying Get Involved Help Apps Datasets Community Login | Sign Up Search Template Edit History...

  7. Widget:CSC-CSS | Open Energy Information

    Open Energy Info (EERE)

    source History View New Pages Recent Changes All Special Pages Semantic SearchQuerying Get Involved Help Apps Datasets Community Login | Sign Up Search Widget Edit History...

  8. Template:LEDSLACNavs | Open Energy Information

    Open Energy Info (EERE)

    source History View New Pages Recent Changes All Special Pages Semantic SearchQuerying Get Involved Help Apps Datasets Community Login | Sign Up Search Template Edit History...

  9. Template:LEDSGPFooter | Open Energy Information

    Open Energy Info (EERE)

    source History View New Pages Recent Changes All Special Pages Semantic SearchQuerying Get Involved Help Apps Datasets Community Login | Sign Up Search Template Edit History...

  10. Template:WebServiceGraphic | Open Energy Information

    Open Energy Info (EERE)

    source History View New Pages Recent Changes All Special Pages Semantic SearchQuerying Get Involved Help Apps Datasets Community Login | Sign Up Search Template Edit History...

  11. Template:WFSPNav | Open Energy Information

    Open Energy Info (EERE)

    source History View New Pages Recent Changes All Special Pages Semantic SearchQuerying Get Involved Help Apps Datasets Community Login | Sign Up Search Template Edit History...

  12. Form:RAPID-BestPractices | Open Energy Information

    Open Energy Info (EERE)

    source History View New Pages Recent Changes All Special Pages Semantic SearchQuerying Get Involved Help Apps Datasets Community Login | Sign Up Search Form Edit History...

  13. Template:Organization | Open Energy Information

    Open Energy Info (EERE)

    source History View New Pages Recent Changes All Special Pages Semantic SearchQuerying Get Involved Help Apps Datasets Community Login | Sign Up Search Template Edit History...

  14. Widget:ContactFinder | Open Energy Information

    Open Energy Info (EERE)

    source History View New Pages Recent Changes All Special Pages Semantic SearchQuerying Get Involved Help Apps Datasets Community Login | Sign Up Search Widget Edit History...

  15. Widget:MailChimp | Open Energy Information

    Open Energy Info (EERE)

    source History View New Pages Recent Changes All Special Pages Semantic SearchQuerying Get Involved Help Apps Datasets Community Login | Sign Up Search Widget Edit History...

  16. Template:RAPID-Nav | Open Energy Information

    Open Energy Info (EERE)

    source History View New Pages Recent Changes All Special Pages Semantic SearchQuerying Get Involved Help Apps Datasets Community Login | Sign Up Search Template Edit History...

  17. Form:GeothermalResourceArea | Open Energy Information

    Open Energy Info (EERE)

    source History View New Pages Recent Changes All Special Pages Semantic SearchQuerying Get Involved Help Apps Datasets Community Login | Sign Up Search Form Edit History...

  18. Template:WindCover | Open Energy Information

    Open Energy Info (EERE)

    source History View New Pages Recent Changes All Special Pages Semantic SearchQuerying Get Involved Help Apps Datasets Community Login | Sign Up Search Template Edit History...

  19. Template:NEPA CX | Open Energy Information

    Open Energy Info (EERE)

    source History View New Pages Recent Changes All Special Pages Semantic SearchQuerying Get Involved Help Apps Datasets Community Login | Sign Up Search Template Edit History...

  20. Montana/Wind Resources | Open Energy Information

    Open Energy Info (EERE)

    source History View New Pages Recent Changes All Special Pages Semantic SearchQuerying Get Involved Help Apps Datasets Community Login | Sign Up Search Page Edit History Montana...

  1. Template:RegulatoryToolkitTabs | Open Energy Information

    Open Energy Info (EERE)

    source History View New Pages Recent Changes All Special Pages Semantic SearchQuerying Get Involved Help Apps Datasets Community Login | Sign Up Search Template Edit History...

  2. Template:Reflist | Open Energy Information

    Open Energy Info (EERE)

    source History View New Pages Recent Changes All Special Pages Semantic SearchQuerying Get Involved Help Apps Datasets Community Login | Sign Up Search Template Edit History...

  3. Cultural Resources | Open Energy Information

    Open Energy Info (EERE)

    source History View New Pages Recent Changes All Special Pages Semantic SearchQuerying Get Involved Help Apps Datasets Community Login | Sign Up Search Page Edit History Cultural...

  4. Widget:UtilityRateFinder | Open Energy Information

    Open Energy Info (EERE)

    source History View New Pages Recent Changes All Special Pages Semantic SearchQuerying Get Involved Help Apps Datasets Community Login | Sign Up Search Widget Edit History...

  5. Form:RRSection | Open Energy Information

    Open Energy Info (EERE)

    source History View New Pages Recent Changes All Special Pages Semantic SearchQuerying Get Involved Help Apps Datasets Community Login | Sign Up Search Form Edit History...

  6. Arizona/Wind Resources | Open Energy Information

    Open Energy Info (EERE)

    source History View New Pages Recent Changes All Special Pages Semantic SearchQuerying Get Involved Help Apps Datasets Community Login | Sign Up Search Page Edit History Arizona...

  7. OSTI, US Dept of Energy Office of Scientific and Technical Information...

    Office of Scientific and Technical Information (OSTI)

    Rather than relying on search algorithms that identify a specific query term, semantic search uses more complex contextual relationships among people, places and things. It is an ...

  8. OSTIblog Articles in the STI reports Topic | OSTI, US Dept of...

    Office of Scientific and Technical Information (OSTI)

    Rather than relying on search algorithms that identify a specific query term, semantic search uses more complex contextual relationships among people, places and things. It is an ...

  9. NREL: Energy Analysis - Nick Langle

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    HTML CSS Javascript jQuery UI UX Responsive Web Design Semantic MediaWiki SEO Primary research interests Web Design & Usability Data Visualization & Analytics Geospatial ...

  10. In-Situ Microphysics from the RACORO IOP (Dataset) | SciTech...

    Office of Scientific and Technical Information (OSTI)

    We highly recommend that anyone wishing to use these files do so in a collaborative endeavor and we welcome queries and opportunities for collaboration. There are caveats ...

  11. Search for: All records | SciTech Connect

    Office of Scientific and Technical Information (OSTI)

    is discussed. Full Text Available April 2014 , Institute of Electrical and Electronics Engineers (IEEE) Prev Next Switch to Detail View for this search SOLR Query Details

  12. OCIO Technology Summit: Data Analytics | Department of Energy

    Office of Energy Efficiency and Renewable Energy (EERE) Indexed Site

    at the Energy Information Administration demonstrated the Electricity Data Browser influence on creating the visualization of data, bringing together maps, query tools, and...

  13. File:02SiteConsiderations (1).pdf | Open Energy Information

    Open Energy Info (EERE)

    source History View New Pages Recent Changes All Special Pages Semantic SearchQuerying Get Involved Help Apps Datasets Community Login | Sign Up Search File Edit History...

  14. Energy Events | Open Energy Information

    Open Energy Info (EERE)

    Events Jump to: navigation, search Upcoming Events You need to have JavaScript enabled to view the interactive timeline. Further results for this query.WEEKMONTHYEAREvent:Preparing...

  15. Category:Tech Potential Properties | Open Energy Information

    Open Energy Info (EERE)

    source History View New Pages Recent Changes All Special Pages Semantic SearchQuerying Get Involved Help Apps Datasets Community Login | Sign Up Search Category Edit History...

  16. User:GregZiebold/Sector test | Open Energy Information

    Open Energy Info (EERE)

    search Query all sector types for Companies: Bioenergy Biofuels Biomass Buildings Carbon Efficiency Geothermal energy Hydro Hydrogen Marine and Hydrokinetic Ocean Renewable Energy...

  17. Search for: All records | SciTech Connect

    Office of Scientific and Technical Information (OSTI)

    potential applications in high density nonvolatile storage in the future. less October 2015 , Wiley Prev Next Switch to Detail View for this search SOLR Query Details

  18. Form:Marine and Hydrokinetic Technology Project Milestone | Open...

    Open Energy Info (EERE)

    source History View New Pages Recent Changes All Special Pages Semantic SearchQuerying Get Involved Help Apps Datasets Community Login | Sign Up Search Form Edit History...

  19. Form:Marine and Hydrokinetic Technology Project | Open Energy...

    Open Energy Info (EERE)

    source History View New Pages Recent Changes All Special Pages Semantic SearchQuerying Get Involved Help Apps Datasets Community Login | Sign Up Search Form Edit History...

  20. Contact DMSE | The Ames Laboratory

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Contact DMSE Division of Materials Sciences and Engineering Director Business Manager General Inquiries Web Queries Matthew Kramer Susan Elsner Julie Dredla Sarah Wiley 125 Metals...

  1. Properties | Open Energy Information

    Open Energy Info (EERE)

    View New Pages Recent Changes All Special Pages Semantic SearchQuerying Get Involved Help Apps Datasets Community Login | Sign Up Search Special page Properties Jump to:...

  2. Inline_System

    Energy Science and Technology Software Center (OSTI)

    2010-02-01

    Inline_System replaces a small subset of file query and manipulation commands, on computing platforms that do not offer a complete standard POSIX environment.

  3. Search for: All records | SciTech Connect

    Office of Scientific and Technical Information (OSTI)

    rhombohedral structure in TbMn2. Full Text Available January 2015 , Royal Society of Chemistry Prev Next Switch to Detail View for this search SOLR Query Details

  4. Coal Markets

    U.S. Energy Information Administration (EIA) Indexed Site

    Coal Glossary FAQS Overview Data Coal Data Browser (interactive query tool with charting and mapping) Summary Prices Reserves Consumption Production Stocks Imports, exports ...

  5. User:Woodjr/Sandbox/GoogleEarth | Open Energy Information

    Open Energy Info (EERE)

    < User:Woodjr | Sandbox Jump to: navigation, search Demonstration of an experimental "GoogleEarth" result format for ask queries. Based on the Thematic Mapping API....

  6. Collegiate Wind Competition | Open Energy Information

    Open Energy Info (EERE)

    Querying Get Involved Help Apps Datasets Community Login | Sign Up Search Page Edit History Collegiate Wind Competition Jump to: navigation, search Wind competition.jpg The...

  7. Efficient Data Management for Knowledge Discovery in Large-Scale Geospatial Imagery Collections

    SciTech Connect (OSTI)

    Baldwin, C; Abdulla, G

    2006-01-24

    We describe the results of our investigation on supporting ad-hoc and continuous queries over data streams. The major problem we address here is how to identify and utilize metadata for smart caching and to support queries over streaming and archived or historical data.

  8. Climatepipes: User-friendly data access, data manipulation, data analysis and visualization of community climate models Phase II

    SciTech Connect (OSTI)

    Chaudhary, Aashish

    2015-09-02

    In Phase I, we successfully developed a web-based tool that provides workflow and form-based interfaces for accessing, querying, and visualizing interesting datasets from one or more sources. For Phase II of the project, we have implemented mechanisms for supporting more elaborate and relevant queries.

  9. QQACCT

    Energy Science and Technology Software Center (OSTI)

    2015-01-01

    batchacct provides convenient library and command-line access to batch system accounting data for GridEngine and SLURM schedulers. It can be used to perform queries useful for data analysis of the accounting data alone or for integrative analysis in the context of a larger query.

  10. Multidimensional structured data visualization method and apparatus, text visualization method and apparatus, method and apparatus for visualizing and graphically navigating the world wide web, method and apparatus for visualizing hierarchies

    DOE Patents [OSTI]

    Risch, John S.; Dowson, Scott T.

    2012-03-06

    A method of displaying correlations among information objects includes receiving a query against a database; obtaining a query result set; and generating a visualization representing the components of the result set, the visualization including one of a plane and line to represent a data field, nodes representing data values, and links showing correlations among fields and values. Other visualization methods and apparatus are disclosed.

  11. Multidimensional structured data visualization method and apparatus, text visualization method and apparatus, method and apparatus for visualizing and graphically navigating the world wide web, method and apparatus for visualizing hierarchies

    DOE Patents [OSTI]

    Risch, John S.; Dowson, Scott T.; Hart, Michelle L.; Hatley, Wes L.

    2008-05-13

    A method of displaying correlations among information objects comprises receiving a query against a database; obtaining a query result set; and generating a visualization representing the components of the result set, the visualization including one of a plane and line to represent a data field, nodes representing data values, and links showing correlations among fields and values. Other visualization methods and apparatus are disclosed.

  12. SAPLE: Sandia Advanced Personnel Locator Engine.

    SciTech Connect (OSTI)

    Procopio, Michael J.

    2010-04-01

    We present the Sandia Advanced Personnel Locator Engine (SAPLE) web application, a directory search application for use by Sandia National Laboratories personnel. SAPLE's purpose is to return Sandia personnel 'results' as a function of user search queries, with its mission to make it easier and faster to find people at Sandia. To accomplish this, SAPLE breaks from more traditional directory application approaches by aiming to return the correct set of results while placing minimal constraints on the user's query. Two key features form the core of SAPLE: advanced search query interpretation and inexact string matching. SAPLE's query interpretation permits the user to perform compound queries when typing into a single search field; where able, SAPLE infers the type of field that the user intends to search on based on the value of the search term. SAPLE's inexact string matching feature yields a high-quality ranking of personnel search results even when there are no exact matches to the user's query. This paper explores these two key features, describing in detail the architecture and operation of SAPLE. Finally, an extensive analysis on logged search query data taken from an 11-week sample period is presented.

  13. EIA Open Data - Doc - U.S. Energy Information Administration (EIA)

    Gasoline and Diesel Fuel Update (EIA)

    API Commands EIA's API uses a modified RESTful architecture, where a separate URI is used for each query command with query string variables, both required and optional, providing input parameters. Two such query string input parameters apply to all commands: api_key: Required. A valid API key is required and may be obtained from Registration out: Optional. Valid values are "xml" or "json". If missing or any other value, the API call will return JSON formatted output. API

  14. Developer | OpenEI Community

    Open Energy Info (EERE)

    Q & A Feeds American Clean Skies Foundation (1) API (3) APIs (1) Apps (1) ask queries (1) Big Data (1) bug (2) challenge (1) citation (1) citing (1) clean energy (1) cleanweb (2)...

  15. Utility Rate API v2 | OpenEI Community

    Open Energy Info (EERE)

    Hi, I am running into one issue with the API and that is using an ask query to get the data in the embedded multiple instance template fields. This is the last big problem that...

  16. User:Jayhuggins/Test | Open Energy Information

    Open Energy Info (EERE)

    6 Compound Queries 7 Pipe Escape 8 Parser Functions 9 Maps 10 Math 11 Loops 12 External Data 13 Dynamic Functions 14 Category Test 15 Array 16 Number Format 17 UUID 18 InputBox...

  17. OpenEI Community Central | OpenEI Community

    Open Energy Info (EERE)

    (1) acres (1) adoption (1) American Clean Skies Foundation (1) Apps (1) ask queries (1) Big Data (1) biofuel art (1) building (1) building load (1) building load data (1) car (1)...

  18. Search for: All records | SciTech Connect

    Office of Scientific and Technical Information (OSTI)

    Switch to Detail View for this search SciTech Connect Search Results Page 1 of 1 Search ... October 1972 Prev Next Switch to Detail View for this search SOLR Query ...

  19. PowerPoint Presentation

    Office of Energy Efficiency and Renewable Energy (EERE) Indexed Site

    ...w.gsaelibrary.gsa.govElibMainsinDetails.do?executeQueryYES&scheduleNumber84&fl ag&filter&specialItemNumber246+53 Become a GSA Schedule 84, SIN 246-53 vendor (2 steps) ...

  20. OSTI, US Dept of Energy, Office of Scientific and Technical Informatio...

    Office of Scientific and Technical Information (OSTI)

    ... Is an offset needed? Does the site's query script need to incorporate parameters that will block old, pre-Harvesting records from coming into OSTI's system again? Who at the site ...

  1. Sulphur Hot Springs Geothermal Area | Open Energy Information

    Open Energy Info (EERE)

    form View source History View New Pages Recent Changes All Special Pages Semantic SearchQuerying Get Involved Help Apps Datasets Community Login | Sign Up Search Page Edit with...

  2. Lake City Hot Springs Geothermal Area | Open Energy Information

    Open Energy Info (EERE)

    form View source History View New Pages Recent Changes All Special Pages Semantic SearchQuerying Get Involved Help Apps Datasets Community Login | Sign Up Search Page Edit with...

  3. Maazama Well Geothermal Area | Open Energy Information

    Open Energy Info (EERE)

    form View source History View New Pages Recent Changes All Special Pages Semantic SearchQuerying Get Involved Help Apps Datasets Community Login | Sign Up Search Page Edit with...

  4. Virgin Islands Wtr&Pwr Auth | Open Energy Information

    Open Energy Info (EERE)

    form View source History View New Pages Recent Changes All Special Pages Semantic SearchQuerying Get Involved Help Apps Datasets Community Login | Sign Up Search Page Edit with...

  5. In-situ sampling of a large-scale particle simulation for interactive...

    Office of Scientific and Technical Information (OSTI)

    The limiting technology in this situation is analogous to the problem in many population surveys: there aren't enough human resources to query a large population. To cope with the ...

  6. Help:SubObjects | Open Energy Information

    Open Energy Info (EERE)

    except the subobject is listed too. If this is not included, it might actually be OK, but it's working with it in - so there you have it. You'll notice an ask query is in...

  7. OpenEI Community - Utility+Utility Access Map

    Open Energy Info (EERE)

    the Special Ask page, in the query box enter the following:

    &91;&91;Category:Utility...

  8. Finding Utility Companies Under a Given Utility ID | OpenEI Community

    Open Energy Info (EERE)

    utility company pages under a given utility id. From the Special Ask page, in the query box enter the following: Category:Utility CompaniesEiaUtilityId::15248 substituting...

  9. Utility+Utility Access Map | OpenEI Community

    Open Energy Info (EERE)

    utility company pages under a given utility id. From the Special Ask page, in the query box enter the following: Category:Utility CompaniesEiaUtilityId::15248 substituting...

  10. Solar Power In China | Open Energy Information

    Open Energy Info (EERE)

    Solar Power In China Jump to: navigation, search This article is a stub. You can help OpenEI by expanding it. Working on ask query to display all Chinese solar companies TODO:...

  11. Title 43 CFR 3201 Available Lands | Open Energy Information

    Open Energy Info (EERE)

    form View source History View New Pages Recent Changes All Special Pages Semantic SearchQuerying Get Involved Help Apps Datasets Community Login | Sign Up Search Page Edit with...

  12. Title 25 USC 323 Rights-of-way for all purposes across any Indian...

    Open Energy Info (EERE)

    form View source History View New Pages Recent Changes All Special Pages Semantic SearchQuerying Get Involved Help Apps Datasets Community Login | Sign Up Search Page Edit with...

  13. Alaska - AS 42.45.045 - Renewable Energy Grant Fund and Recommendation...

    Open Energy Info (EERE)

    form View source History View New Pages Recent Changes All Special Pages Semantic SearchQuerying Get Involved Help Apps Datasets Community Login | Sign Up Search Page Edit with...

  14. Title 10, Chapter 49 Protection of Navigable Waters and Shorelands...

    Open Energy Info (EERE)

    form View source History View New Pages Recent Changes All Special Pages Semantic SearchQuerying Get Involved Help Apps Datasets Community Login | Sign Up Search Page Edit with...

  15. How can an external application get access to OpenEI images and...

    Open Energy Info (EERE)

    How can an external application get access to OpenEI images and thumbnails? Home > Groups > Developer I'm building an external application in Simile Exhibit. Through an Ask query...

  16. Search for: All records | SciTech Connect

    Office of Scientific and Technical Information (OSTI)

    which are of great significance to the in-depth understanding of the phase transformation ... March 2016 , Elsevier Prev Next Switch to Detail View for this search SOLR Query ...

  17. Kanuti Geothermal Area | Open Energy Information

    Open Energy Info (EERE)

    form View source History View New Pages Recent Changes All Special Pages Semantic SearchQuerying Get Involved Help Apps Datasets Community Login | Sign Up Search Page Edit with...

  18. HCP Handbook | Open Energy Information

    Open Energy Info (EERE)

    form View source History View New Pages Recent Changes All Special Pages Semantic SearchQuerying Get Involved Help Apps Datasets Community Login | Sign Up Search Page Edit with...

  19. maps | OpenEI Community

    Open Energy Info (EERE)

    queries developer Google maps maps multicolor result formats results Semantic Mediawiki Hi all, Recently, a couple of people on OpenEI have asked me how to do compound (or...

  20. multicolor | OpenEI Community

    Open Energy Info (EERE)

    queries developer Google maps maps multicolor result formats results Semantic Mediawiki Hi all, Recently, a couple of people on OpenEI have asked me how to do compound (or...

  1. Google maps | OpenEI Community

    Open Energy Info (EERE)

    queries developer Google maps maps multicolor result formats results Semantic Mediawiki Hi all, Recently, a couple of people on OpenEI have asked me how to do compound (or...

  2. results | OpenEI Community

    Open Energy Info (EERE)

    queries developer Google maps maps multicolor result formats results Semantic Mediawiki Hi all, Recently, a couple of people on OpenEI have asked me how to do compound (or...

  3. Semantic Mediawiki | OpenEI Community

    Open Energy Info (EERE)

    queries developer Google maps maps multicolor result formats results Semantic Mediawiki Hi all, Recently, a couple of people on OpenEI have asked me how to do compound (or...

  4. result formats | OpenEI Community

    Open Energy Info (EERE)

    queries developer Google maps maps multicolor result formats results Semantic Mediawiki Hi all, Recently, a couple of people on OpenEI have asked me how to do compound (or...

  5. OSTI, US Dept of Energy Office of Scientific and Technical Information...

    Office of Scientific and Technical Information (OSTI)

    Federated search, for those of you new to the term, is the simultaneous search of multiple online databases or web resources from a single query. The Wikipedia article on federated ...

  6. OSTIblog Articles in the milestones Topic | OSTI, US Dept of...

    Office of Scientific and Technical Information (OSTI)

    Federated search, for those of you new to the term, is the simultaneous search of multiple online databases or web resources from a single query. The Wikipedia article on federated ...

  7. Microsoft Word - Plain Language Compliance Report 2016.doc

    Energy Savers [EERE]

    For example, visitors can enter a search term and find that information on Energy.gov; ... found most user search queries consist of terms that are simple and in Plain Language. ...

  8. DOE - NNSA/NFO -- Search

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    U.S. DOENNSA - Nevada Field Office Search this Website To enter a query, type in a few descriptive ... of the page Return results where my terms occur Sort Sort by relevance Sort by ...

  9. OSTI Employees Cited for Central Roles in Science.gov | OSTI...

    Office of Scientific and Technical Information (OSTI)

    in the creation and ongoing development of Science.gov, a federal science Web portal. ... for ranking science queries across 30 deep Web scientific databases and 1,800 Web sites. ...

  10. Microsoft Word - rhubc2_DOE-SC-ARM-0901_final_rev3.24.09.doc

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    ... This query led us to meet with Dr. Michel Cure at the University of Valparaiso. Dr. Cure is starting a project to forecast water vapor amounts over the Atacama Astronomical Park, ...

  11. Thermo Scientific Sulfur Dioxide Analyzer Instrument Handbook...

    Office of Scientific and Technical Information (OSTI)

    in National Instruments LabView that both controls the Model 43i-TLE Analyzer AND queries the unit for all measurement and housekeeping data. The LabView vi (the software program ...

  12. November-December | U.S. DOE Office of Science (SC)

    Office of Science (SC) Website

    ... To test the hypotheses of brain structure, dynamics and function, the team built a ... This is why being able to test the speed of the query execution on the Dash compute system ...

  13. Polls | OpenEI Community

    Open Energy Info (EERE)

    May 2012 - 13:48 by Rmckeel * The Utility Rate web service * The Incentive web service * Web services to query by geographic location or shape * Mediawiki Ask examples & tutorials...

  14. ScienceLab | OSTI, US Dept of Energy, Office of Scientific and...

    Office of Scientific and Technical Information (OSTI)

    For single-query access to an array of federal science education resources, intended for ... Innovative, state-of-the-art audio indexing and speech recognition technology is used to ...

  15. User:Nlangle/Timeline Test | Open Energy Information

    Open Energy Info (EERE)

    enabled to view the interactive timeline. Further results for this query.DECADEFederal Oil and Gas Royalty Simplification and Fairness Act of 19961996-01-010Year: 1996 Federal...

  16. User:Woodjr/Sandbox/GeoMap | Open Energy Information

    Open Energy Info (EERE)

    GeoMap < User:Woodjr | Sandbox Jump to: navigation, search Demonstration of an experimental "GeoMap" result format for ask queries. Based on Google's GeoMap Visualization API....

  17. Direct-Current Resistivity At Cove Fort Area - Liquid (Combs...

    Open Energy Info (EERE)

    form View source History View New Pages Recent Changes All Special Pages Semantic SearchQuerying Get Involved Help Apps Datasets Community Login | Sign Up Search Page Edit with...

  18. OpenEI Community Central | OpenEI Community

    Open Energy Info (EERE)

    OpenEI Community Central Home > OpenEI Community Central > Posts by term > OpenEI Community Central Content Group Activity By term Q & A Feeds Term: compound queries Type Term...

  19. NREL: Energy Analysis - Jon Weers

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Senior Web Applications Engineer On staff since March 2010 Phone number: 303-275-3701 E-mail: jon.weers@nrel.gov Areas of expertise Interactive web applications (PHP, jQuery, ...

  20. Publications about OSTI | OSTI, US Dept of Energy Office of Scientific...

    Office of Scientific and Technical Information (OSTI)

    and WorldWideScience.org, the patron gains access to multiple, geographically dispersed deep web databases and can search all of the constituent sources with a single query. ...

  1. OSTI, US Dept of Energy Office of Scientific and Technical Information...

    Office of Scientific and Technical Information (OSTI)

    In 2005, the idea of creating a global science gateway for the web was conceived at OSTI. ... a single query, and it would be available at no cost to anyone anywhere with web access. ...

  2. Science.gov 3.0 Launched | Department of Energy

    Energy Savers [EERE]

    early viewing of results while the database and Web site searches continue in real time. ... A single query searches across 30 databases and 1,800 Web sites. Science.gov allows users ...

  3. OSTI, US Dept of Energy Office of Scientific and Technical Information...

    Office of Scientific and Technical Information (OSTI)

    the simultaneous search of multiple online databases or web resources from a single query. ... be developed as part of many scientific research projects in today's Web 2.0 environment. ...

  4. Science.gov 3.0 Launched; Offers Increased Precision Searches...

    Office of Science (SC) Website

    early viewing of results while the database and Web site searches continue in real time. ... A single query searches across 30 databases and 1,800 Web sites. Science.gov allows users ...

  5. OSTIblog Articles in the milestones Topic | OSTI, US Dept of...

    Office of Scientific and Technical Information (OSTI)

    the simultaneous search of multiple online databases or web resources from a single query. ... Today it is the Internet transformation, especially the Web. As a leader in making the Web ...

  6. Web Support

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Web Support We want to be able to respond promptly to your queries. To expedite our response, please check the specific website or page in question for the name of the appropriate ...

  7. Porotomo Subtask 3.9 Build FEM Configuration

    DOE Data Explorer [Office of Scientific and Technical Information (OSTI)]

    Tabrez Ali

    mesh.vtk: Self contained VTK file that contains mesh information and can be directly visualized in Paraview/Visit mesh.png: Image of mesh as visualized in Paraview nodes.csv: Nodal coordinates of the mesh in UTM coordinates (m). nodes_rotated.csv: Nodal coordinates of the mesh in rotated (X/Y/Z) coordinates (m). cells.csv: Connectivity data query_points.csv: List of points (centroid of cells) that will be used to query the geologic database

  8. LANL Contacts

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Lab Contacts Lab Contacts Addresses and phone numbers for general information, journalistic queries, business needs, community programs and employee resources. General Employee directory Emergency communication Communications Office (505) 667-7000 Ethics & Audits Internal Audit: (505) 665-3104 Ethics Office: (505) 667-7506 Fax: (505) 665-3664 ethics@lanl.gov Journalist queries Communications Office (505) 667-7000 Media contacts Lab mailing address Los Alamos National Laboratory P.O. Box 1663

  9. Buildings Energy Data Book

    Buildings Energy Data Book [EERE]

    Explore Survey Data from the Energy Information Administration Follow the links below to two easy-to-use query tools, developed exclusively for this website. With these tools you can explore results from the Commercial Buildings Energy Consumption Survey (CBECS) and the Residential Energy Consumption Survey (RECS). Commercial Buildings Energy Index Use this custom query tool to analyze micro data from CBECS 2003. Residential Buildings Energy Index Use this custom Microsoft Excel pivot table to

  10. Geometric Algorithms for Modeling, Motion, and Animation (GAMMA): Collision Detection Videos from the University of North Carolina GAMMA Research Group

    DOE Data Explorer [Office of Scientific and Technical Information (OSTI)]

    Physically based modeling simulations depend highly on the physical interaction between objects in a scene. Complex physics engines require fast, accurate, and robust proximity queries to maintain a realistic simulation at interactive rates. We couple our proximity query research with physically based modeling to ensure that our packages provide the capabilities of today's physics engines.[Copied from http://www.cs.unc.edu/~geom/collide/index.shtml

  11. Department of Energy (DOE) OpenNet documents

    Office of Scientific and Technical Information (OSTI)

    Help The Search Help section provides assistance in the following areas: Search features Field definitions Query examples Adjacency This database allows you to search for exact character strings or phrases. This is known as a literal search and provides the same functionality as an adjacency operator. To search, use double quotes around the phrase entered in the query box. For example, using the Basic Search on the Home Page, or the Full-Text and field on the Advanced Search page, enter:

  12. Porotomo Subtask 3.9 Build FEM Configuration

    SciTech Connect (OSTI)

    Tabrez Ali

    2015-06-30

    mesh.vtk: Self contained VTK file that contains mesh information and can be directly visualized in Paraview/Visit mesh.png: Image of mesh as visualized in Paraview nodes.csv: Nodal coordinates of the mesh in UTM coordinates (m). nodes_rotated.csv: Nodal coordinates of the mesh in rotated (X/Y/Z) coordinates (m). cells.csv: Connectivity data query_points.csv: List of points (centroid of cells) that will be used to query the geologic database

  13. Efficient binning for bitmap indices on high-cardinality attributes

    SciTech Connect (OSTI)

    Rotem, Doron; Stockinger, Kurt; Wu, Kesheng

    2004-11-17

    Bitmap indexing is a common technique for indexing high-dimensional data in data warehouses and scientific applications. Though efficient for low-cardinality attributes, query processing can be rather costly for high-cardinality attributes due to the large storage requirements for the bitmap indices. Binning is a common technique for reducing storage costs of bitmap indices. This technique partitions the attribute values into a number of ranges, called bins, and uses bitmap vectors to represent bins (attribute ranges) rather than distinct values. Although binning may reduce storage costs, it may increase the access costs of queries that do not fall on exact bin boundaries (edge bins). For this kind of queries the original data values associated with edge bins must be accessed, in order to check them against the query constraints.In this paper we study the problem of finding optimal locations for the bin boundaries in order to minimize these access costs subject to storage constraints. We propose a dynamic programming algorithm for optimal partitioning of attribute values into bins that takes into account query access patterns as well as data distribution statistics. Mathematical analysis and experiments on real life data sets show that the optimal partitioning achieved by this algorithm can lead to a significant improvement in the access costs of bitmap indexing systems for high-cardinality attributes.

  14. Method for localizing and isolating an errant process step

    DOE Patents [OSTI]

    Tobin, Jr., Kenneth W.; Karnowski, Thomas P.; Ferrell, Regina K.

    2003-01-01

    A method for localizing and isolating an errant process includes the steps of retrieving from a defect image database a selection of images each image having image content similar to image content extracted from a query image depicting a defect, each image in the selection having corresponding defect characterization data. A conditional probability distribution of the defect having occurred in a particular process step is derived from the defect characterization data. A process step as a highest probable source of the defect according to the derived conditional probability distribution is then identified. A method for process step defect identification includes the steps of characterizing anomalies in a product, the anomalies detected by an imaging system. A query image of a product defect is then acquired. A particular characterized anomaly is then correlated with the query image. An errant process step is then associated with the correlated image.

  15. Fast Search for Dynamic Multi-Relational Graphs

    SciTech Connect (OSTI)

    Choudhury, Sutanay; Holder, Larry; Chin, George; Feo, John T.

    2013-06-23

    Acting on time-critical events by processing ever growing social media or news streams is a major technical challenge. Many of these data sources can be modeled as multi-relational graphs. Continuous queries or techniques to search for rare events that typically arise in monitoring applications have been studied extensively for relational databases. This work is dedicated to answer the question that emerges naturally: how can we efficiently execute a continuous query on a dynamic graph? This paper presents an exact subgraph search algorithm that exploits the temporal characteristics of representative queries for online news or social media monitoring. The algorithm is based on a novel data structure called the that leverages the structural and semantic characteristics of the underlying multi-relational graph. The paper concludes with extensive experimentation on several real-world datasets that demonstrates the validity of this approach.

  16. Toward a Data Scalable Solution for Facilitating Discovery of Scientific Data Resources

    SciTech Connect (OSTI)

    Chappell, Alan R.; Choudhury, Sutanay; Feo, John T.; Haglin, David J.; Morari, Alessandro; Purohit, Sumit; Schuchardt, Karen L.; Tumeo, Antonino; Weaver, Jesse R.; Villa, Oreste

    2013-11-18

    Science is increasingly motivated by the need to process larger quantities of data. It is facing severe challenges in data collection, management, and processing, so much so that the computational demands of "data scaling" are competing with, and in many fields surpassing, the traditional objective of decreasing processing time. Example domains with large datasets include astronomy, biology, genomic, climate and weather, and material sciences. This paper presents a real-world use case in which we wish to answer queries provided by domain scientists in order to facilitate discovery of relevant science resources. The problem is that the metadata for these science resources is very large and is growing quickly, rapidly increasing the need for a data scaling solution. We propose the use of our SGEM stack -- a system designed for answering graph-based queries over large datasets on cluster architectures -- for answering complex queries over the metadata, and we report early results for our current capability.

  17. PCard Data Analysis Tool

    Energy Science and Technology Software Center (OSTI)

    2005-04-01

    The Procurement Card data analysis and monitoring tool enables due-diligence review using predefined user-created queries and reports. The system tracks individual compliance emails. More specifically, the tool: - Helps identify exceptions or questionable and non-compliant purchases, - Creates audit random sample on request, - Allows users to create and run new or ad-hoc queries and reports, - Monitors disputed charges, - Creates predefined Emails to Cardholders requesting documentation and/or clarification, - Tracks audit status, notes,more » Email status (date sent, response), audit resolution.« less

  18. PCard Data Analysis Tool

    SciTech Connect (OSTI)

    Hilts, Jim

    2005-04-01

    The Procurement Card data analysis and monitoring tool enables due-diligence review using predefined user-created queries and reports. The system tracks individual compliance emails. More specifically, the tool: - Helps identify exceptions or questionable and non-compliant purchases, - Creates audit random sample on request, - Allows users to create and run new or ad-hoc queries and reports, - Monitors disputed charges, - Creates predefined Emails to Cardholders requesting documentation and/or clarification, - Tracks audit status, notes, Email status (date sent, response), audit resolution.

  19. Preliminary Results on Uncertainty Quantification for Pattern Analytics

    SciTech Connect (OSTI)

    Stracuzzi, David John; Brost, Randolph; Chen, Maximillian Gene; Malinas, Rebecca; Peterson, Matthew Gregor; Phillips, Cynthia A.; Robinson, David G.; Woodbridge, Diane

    2015-09-01

    This report summarizes preliminary research into uncertainty quantification for pattern ana- lytics within the context of the Pattern Analytics to Support High-Performance Exploitation and Reasoning (PANTHER) project. The primary focus of PANTHER was to make large quantities of remote sensing data searchable by analysts. The work described in this re- port adds nuance to both the initial data preparation steps and the search process. Search queries are transformed from does the specified pattern exist in the data? to how certain is the system that the returned results match the query? We show example results for both data processing and search, and discuss a number of possible improvements for each.

  20. Enterprise Middleware for Scientific Data

    SciTech Connect (OSTI)

    Thomson, Judi; Chappell, Alan R.; Almquist, Justin P.

    2003-02-27

    We describe an enterprise middleware system that integrates, from a user’s perspective, data located on disparate data storage devices without imposing additional requirements upon those storage mechanisms. The system provides advanced search capabilities by exploiting a repository of metadata that describes the integrated data. This search mechanism integrates information from a collection of XML metadata documents with diverse schema. Users construct queries using familiar search terms, and the enterprise system uses domain representations and vocabulary mappings to translate the user’s query, expanding the search to include other potentially relevant data. The enterprise architecture allows flexibility with respect to domain dependent processing of user data and metadata

  1. Method and system for efficiently searching an encoded vector index

    DOE Patents [OSTI]

    Bui, Thuan Quang; Egan, Randy Lynn; Kathmann, Kevin James

    2001-09-04

    Method and system aspects for efficiently searching an encoded vector index are provided. The aspects include the translation of a search query into a candidate bitmap, and the mapping of data from the candidate bitmap into a search result bitmap according to entry values in the encoded vector index. Further, the translation includes the setting of a bit in the candidate bitmap for each entry in a symbol table that corresponds to candidate of the search query. Also included in the mapping is the identification of a bit value in the candidate bitmap pointed to by an entry in an encoded vector.

  2. Search tool plug-in: imploements latent topic feedback

    Energy Science and Technology Software Center (OSTI)

    2011-09-23

    IRIS is a search tool plug-in that is used to implement latent topic feedback for enhancing text navigation. It accepts a list of returned documents from an information retrieval wywtem that is generated from keyword search queries. Data is pulled directly from a topic information database and processed by IRIS to determine the most prominent and relevant topics, along with topic-ngrams, associated with the list of returned documents. User selected topics are then used tomore » expand the query and presumabley refine the search results.« less

  3. Annotated Bibliography for the DEWPOINT project

    SciTech Connect (OSTI)

    Oehmen, Christopher S.

    2009-04-21

    This bibliography covers aspects of the Detection and Early Warning of Proliferation from Online INdicators of Threat (DEWPOINT) project including 1) data management and querying, 2) baseline and advanced methods for classifying free text, and 3) algorithms to achieve the ultimate goal of inferring intent from free text sources. Metrics for assessing the quality and correctness of classification are addressed in the second group. Data management and querying include methods for efficiently storing, indexing, searching, and organizing the data we expect to operate on within the DEWPOINT project.

  4. Ensemble Data Analysis ENvironment (EDEN)

    SciTech Connect (OSTI)

    Steed, Chad Allen

    2012-08-01

    The EDEN toolkit facilitates exploratory data analysis and visualization of global climate model simulation datasets. EDEN provides an interactive graphical user interface (GUI) that helps the user visually construct dynamic queries of the characteristically large climate datasets using temporal ranges, variable selections, and geographic areas of interest. EDEN reads the selected data into a multivariate visualization panel which features an extended implementation of parallel coordinates plots as well as interactive scatterplots. The user can query data in the visualization panel using mouse gestures to analyze different ranges of data. The visualization panel provides coordinated multiple views whereby selections made in one plot are propagated to the other plots.

  5. An efficient compression scheme for bitmap indices

    SciTech Connect (OSTI)

    Wu, Kesheng; Otoo, Ekow J.; Shoshani, Arie

    2004-04-13

    When using an out-of-core indexing method to answer a query, it is generally assumed that the I/O cost dominates the overall query response time. Because of this, most research on indexing methods concentrate on reducing the sizes of indices. For bitmap indices, compression has been used for this purpose. However, in most cases, operations on these compressed bitmaps, mostly bitwise logical operations such as AND, OR, and NOT, spend more time in CPU than in I/O. To speedup these operations, a number of specialized bitmap compression schemes have been developed; the best known of which is the byte-aligned bitmap code (BBC). They are usually faster in performing logical operations than the general purpose compression schemes, but, the time spent in CPU still dominates the total query response time. To reduce the query response time, we designed a CPU-friendly scheme named the word-aligned hybrid (WAH) code. In this paper, we prove that the sizes of WAH compressed bitmap indices are about two words per row for large range of attributes. This size is smaller than typical sizes of commonly used indices, such as a B-tree. Therefore, WAH compressed indices are not only appropriate for low cardinality attributes but also for high cardinality attributes.In the worst case, the time to operate on compressed bitmaps is proportional to the total size of the bitmaps involved. The total size of the bitmaps required to answer a query on one attribute is proportional to the number of hits. These indicate that WAH compressed bitmap indices are optimal. To verify their effectiveness, we generated bitmap indices for four different datasets and measured the response time of many range queries. Tests confirm that sizes of compressed bitmap indices are indeed smaller than B-tree indices, and query processing with WAH compressed indices is much faster than with BBC compressed indices, projection indices and B-tree indices. In addition, we also verified that the average query response time

  6. Sifting Through a Trillion Electrons

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Sifting Through a Trillion Electrons Sifting Through a Trillion Electrons Berkeley researchers design strategies for extracting interesting data from massive scientific datasets June 26, 2012 Linda Vu, lvu@lbl.gov, +1 510 495 2402 VPIC1.jpg After querying a dataset of approximately 114,875,956,837 particles for those with Energy values less than 1.5, FastQuery identifies 57,740,614 particles, which are mapped on this plot. Image by Oliver Rubel, Berkeley Lab. Modern research tools like

  7. Monitoring jobs with qs

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Jobs » Monitoring jobs with qs Monitoring jobs with qs qs is an alternative tool to the SGE-provided qstat for querying the queue status developed at NERSC. qs provides an enhanced user interface designed to make it easier to see resource requests, utilization, and job position in the queue. qs provides a centralized web-service that can be queried using either the provided "qs" client, or by HTTP connection to the qs server. qs reports data from a cached copy of the genepool UGE

  8. GenoGraphics for OpenWindows trademark

    SciTech Connect (OSTI)

    Hagstrom, R.; Overbeek, R.; Price, M.; Zawada, D. ); Michaels, G.S.; Taylor, R. . Div. of Computer Research and Technology); Yoshida, Kaoru )

    1992-04-01

    GenoGraphics is a generic utility for constructing and querying one-dimensional linear plots. The outgrowth of a request from Dr. Cassandra Smith for a tool to facilitate her genome mapping research. GenoGraphics development has benefited from a continued collaboration with her. Written in Sun Microsystem's OpenWindows environment and the BTOL toolkit developed at Argonne National Laboratory. GenoGraphics provides an interactive, intuitive, graphical interface. Its features include: viewing multiple maps simultaneously, zooming, and querying by mouse clicking. By expediting plot generation, GenoGraphics gives the scientist more time to analyze data and a novel means for deducing conclusions.

  9. GenoGraphics for OpenWindows{trademark}

    SciTech Connect (OSTI)

    Hagstrom, R.; Overbeek, R.; Price, M.; Zawada, D.; Michaels, G.S.; Taylor, R.; Yoshida, Kaoru

    1992-04-01

    GenoGraphics is a generic utility for constructing and querying one-dimensional linear plots. The outgrowth of a request from Dr. Cassandra Smith for a tool to facilitate her genome mapping research. GenoGraphics development has benefited from a continued collaboration with her. Written in Sun Microsystem`s OpenWindows environment and the BTOL toolkit developed at Argonne National Laboratory. GenoGraphics provides an interactive, intuitive, graphical interface. Its features include: viewing multiple maps simultaneously, zooming, and querying by mouse clicking. By expediting plot generation, GenoGraphics gives the scientist more time to analyze data and a novel means for deducing conclusions.

  10. InterMine Webservices for Phytozome

    SciTech Connect (OSTI)

    Carlson, Joseph; Hayes, David; Goodstein, David; Rokhsar, Daniel

    2014-01-10

    A data warehousing framework for biological information provides a useful infrastructure for providers and users of genomic data. For providers, the infrastructure give them a consistent mechanism for extracting raw data. While for the users, the web services supported by the software allows them to make either simple and common, or complex and unique, queries of the data

  11. Vista Version 0.7

    Energy Science and Technology Software Center (OSTI)

    2005-05-05

    Vista is a database management system tailored to the needs of scientific computing. It provides data storage for Index Sets, topological relationships, parameters, and fields. It provides scoping capabilities for data along with a nice way of managing attribute queries. It is an in-core database that is intended to replace the majority of data structures used in scientific software.

  12. Linked-View Parallel Coordinate Plot Renderer

    Energy Science and Technology Software Center (OSTI)

    2011-06-28

    This software allows multiple linked views for interactive querying via map-based data selection, bar chart analytic overlays, and high dynamic range (HDR) line renderings. The major component of the visualization package is a parallel coordinate renderer with binning, curved layouts, shader-based rendering, and other techniques to allow interactive visualization of multidimensional data.

  13. RTOSPlanner v 0.9

    Energy Science and Technology Software Center (OSTI)

    2012-01-05

    RTOSPianner provides a generic robot motion planning capability that interfaces directly with the SMART kinematics and dynamics engine. It provides rapid setup, syncronization and query routines for driving a robot modelled within SMART and kinApps. It requires the following packages to run: core SMART, core Umbra, Esmart, and kinApps.

  14. List of utility company aliases | OpenEI Community

    Open Energy Info (EERE)

    ntro%5D&p%5Boutro%5D&p%5Bdefault%5D&eqyes to query various fields from the aliases. Hope that helps - feel free to edit your question if that didn't answer everything Rmckeel...

  15. Optimal Chunking of Large Multidimensional Arrays for Data Warehousing

    SciTech Connect (OSTI)

    Otoo, Ekow J; Otoo, Ekow J.; Rotem, Doron; Seshadri, Sridhar

    2008-02-15

    Very large multidimensional arrays are commonly used in data intensive scientific computations as well as on-line analytical processingapplications referred to as MOLAP. The storage organization of such arrays on disks is done by partitioning the large global array into fixed size sub-arrays called chunks or tiles that form the units of data transfer between disk and memory. Typical queries involve the retrieval of sub-arrays in a manner that access all chunks that overlap the query results. An important metric of the storage efficiency is the expected number of chunks retrieved over all such queries. The question that immediately arises is"what shapes of array chunks give the minimum expected number of chunks over a query workload?" The problem of optimal chunking was first introduced by Sarawagi and Stonebraker who gave an approximate solution. In this paper we develop exact mathematical models of the problem and provide exact solutions using steepest descent and geometric programming methods. Experimental results, using synthetic and real life workloads, show that our solutions are consistently within than 2.0percent of the true number of chunks retrieved for any number of dimensions. In contrast, the approximate solution of Sarawagi and Stonebraker can deviate considerably from the true result with increasing number of dimensions and also may lead to suboptimal chunk shapes.

  16. Visualization and Analysis in Support of Fusion Science

    SciTech Connect (OSTI)

    Sanderson, Allen R.

    2012-10-01

    This report summarizes the results of the award for “Visualization and Analysis in Support of Fusion Science.” With this award our main efforts have been to develop and deploy visualization and analysis tools in three areas 1) magnetic field line analysis 2) query based visualization and 3) comparative visualization.

  17. AISL Development Toolkit

    Energy Science and Technology Software Center (OSTI)

    2012-09-13

    AISLDT is a library of utility functions supporting other AISL software. Code provides various utility functions for Common Lisp, including an object-oriented database, distributed objects, logic query engine, web content management, chart drawing, packet sniffing, text processing, and various data structures.

  18. GASIS demonstration

    SciTech Connect (OSTI)

    Vidas, E.H.

    1995-04-01

    A prototype of the GASIS database and retrieval software has been developed and is the subject of this poster session and computer demonstration. The prototype consists of test or preliminary versions of the GASIS Reservoir Data System and Source Directory datasets and the software for query and retrieval. The prototype reservoir database covers the Rocky Mountain region and contains the full GASIS data matrix (all GASIS data elements) that will eventually be included on the CD-ROM. It is populated for development purposes primarily by the information included in the Rocky Mountain Gas Atlas. The software has been developed specifically for GASIS using Foxpro for Windows. The application is an executable file that does not require Foxpro to run. The reservoir database software includes query and retrieval, screen display, report generation, and data export functions. Basic queries by state, basin, or field name will be assisted by scrolling selection lists. A detailed query screen will allow record selection on the basis of any data field, such as depth, cumulative production, or geological age. Logical operators can be applied to any-numeric data element or combination of elements. Screen display includes a {open_quotes}browse{close_quotes} display with one record per row and a detailed single record display. Datasets can be exported in standard formats for manipulation with other software packages. The Source Directory software will allow record retrieval by database type or subject area.

  19. Semantic Space Analyst

    Energy Science and Technology Software Center (OSTI)

    2004-04-15

    The Semantic Space Analyst (SSA) is software for analyzing a text corpus, discovering relationships among terms, and allowing the user to explore that information in different ways. It includes features for displaying and laying out terms and relationships visually, for generating such maps from manual queries, for discovering differences between corpora. Data can also be exported to Microsoft Excel.

  20. A Selectivity based approach to Continuous Pattern Detection in Streaming Graphs

    SciTech Connect (OSTI)

    Choudhury, Sutanay; Holder, Larry; Chin, George; Agarwal, Khushbu; Feo, John T.

    2015-02-02

    Cyber security is one of the most significant technical challenges in current times. Detecting adversarial activities, prevention of theft of intellectual properties and customer data is a high priority for corporations and government agencies around the world. Cyber defenders need to analyze massive-scale, high-resolution network flows to identify, categorize, and mitigate attacks involving net- works spanning institutional and national boundaries. Many of the cyber attacks can be described as subgraph patterns, with promi- nent examples being insider infiltrations (path queries), denial of service (parallel paths) and malicious spreads (tree queries). This motivates us to explore subgraph matching on streaming graphs in a continuous setting. The novelty of our work lies in using the subgraph distributional statistics collected from the streaming graph to determine the query processing strategy. We introduce a Lazy Search" algorithm where the search strategy is decided on a vertex-to-vertex basis depending on the likelihood of a match in the vertex neighborhood. We also propose a metric named Relative Selectivity" that is used to se- lect between different query processing strategies. Our experiments performed on real online news, network traffic stream and a syn- thetic social network benchmark demonstrate 10-100x speedups over selectivity agnostic approaches.

  1. MATRIX AND VECTOR SERVICES

    Energy Science and Technology Software Center (OSTI)

    2001-10-18

    PETRA V2 provides matrix and vector services and the ability construct, query, and use matrix and vector objects that are used and computed by TRILINOS solvers. It provides all basic matr5ix and vector operations for solvers in TRILINOS.

  2. Computing quality scores and uncertainty for approximate pattern matching in geospatial semantic graphs

    SciTech Connect (OSTI)

    Stracuzzi, David John; Brost, Randolph C.; Phillips, Cynthia A.; Robinson, David G.; Wilson, Alyson G.; Woodbridge, Diane M. -K.

    2015-09-26

    Geospatial semantic graphs provide a robust foundation for representing and analyzing remote sensor data. In particular, they support a variety of pattern search operations that capture the spatial and temporal relationships among the objects and events in the data. However, in the presence of large data corpora, even a carefully constructed search query may return a large number of unintended matches. This work considers the problem of calculating a quality score for each match to the query, given that the underlying data are uncertain. As a result, we present a preliminary evaluation of three methods for determining both match quality scores and associated uncertainty bounds, illustrated in the context of an example based on overhead imagery data.

  3. Computing quality scores and uncertainty for approximate pattern matching in geospatial semantic graphs

    DOE Public Access Gateway for Energy & Science Beta (PAGES Beta)

    Stracuzzi, David John; Brost, Randolph C.; Phillips, Cynthia A.; Robinson, David G.; Wilson, Alyson G.; Woodbridge, Diane M. -K.

    2015-09-26

    Geospatial semantic graphs provide a robust foundation for representing and analyzing remote sensor data. In particular, they support a variety of pattern search operations that capture the spatial and temporal relationships among the objects and events in the data. However, in the presence of large data corpora, even a carefully constructed search query may return a large number of unintended matches. This work considers the problem of calculating a quality score for each match to the query, given that the underlying data are uncertain. As a result, we present a preliminary evaluation of three methods for determining both match qualitymore » scores and associated uncertainty bounds, illustrated in the context of an example based on overhead imagery data.« less

  4. Method and system for the diagnosis of disease using retinal image content and an archive of diagnosed human patient data

    DOE Patents [OSTI]

    Tobin, Kenneth W; Karnowski, Thomas P; Chaum, Edward

    2013-08-06

    A method for diagnosing diseases having retinal manifestations including retinal pathologies includes the steps of providing a CBIR system including an archive of stored digital retinal photography images and diagnosed patient data corresponding to the retinal photography images, the stored images each indexed in a CBIR database using a plurality of feature vectors, the feature vectors corresponding to distinct descriptive characteristics of the stored images. A query image of the retina of a patient is obtained. Using image processing, regions or structures in the query image are identified. The regions or structures are then described using the plurality of feature vectors. At least one relevant stored image from the archive based on similarity to the regions or structures is retrieved, and an eye disease or a disease having retinal manifestations in the patient is diagnosed based on the diagnosed patient data associated with the relevant stored image(s).

  5. Shards v 1.0

    Energy Science and Technology Software Center (OSTI)

    2009-07-28

    Shards is a library of Shared Discretization Tools intended to support development of computer codes for the numerical solution of Partial Differential Equations (PDEs). The library comprises of two categories of tools: methods to manage and access information about cell topologies used in mesh-based methods for PDEs, and methods to work with multi-dimensional arrays used to store numerical data in corresponding computer codes. The basic cell topology functionality of Shards includes methods to query adjacenciesmore » of subcells, find subcell permutation with respect to a global cell and create user-defined custom cell topologies. Multi-dimensional array part of the package provides specialized compile-time dimension tags, multi-index access methods, rank and dimension queries.« less

  6. Final Report: Efficient Databases for MPC Microdata

    SciTech Connect (OSTI)

    Michael A. Bender; Martin Farach-Colton; Bradley C. Kuszmaul

    2011-08-31

    The purpose of this grant was to develop the theory and practice of high-performance databases for massive streamed datasets. Over the last three years, we have developed fast indexing technology, that is, technology for rapidly ingesting data and storing that data so that it can be efficiently queried and analyzed. During this project we developed the technology so that high-bandwidth data streams can be indexed and queried efficiently. Our technology has been proven to work data sets composed of tens of billions of rows when the data streams arrives at over 40,000 rows per second. We achieved these numbers even on a single disk driven by two cores. Our work comprised (1) new write-optimized data structures with better asymptotic complexity than traditional structures, (2) implementation, and (3) benchmarking. We furthermore developed a prototype of TokuFS, a middleware layer that can handle microdata I/O packaged up in an MPI-IO abstraction.

  7. The PANTHER User Experience

    SciTech Connect (OSTI)

    Coram, Jamie L.; Morrow, James D.; Perkins, David Nikolaus

    2015-09-01

    This document describes the PANTHER R&D Application, a proof-of-concept user interface application developed under the PANTHER Grand Challenge LDRD. The purpose of the application is to explore interaction models for graph analytics, drive algorithmic improvements from an end-user point of view, and support demonstration of PANTHER technologies to potential customers. The R&D Application implements a graph-centric interaction model that exposes analysts to the algorithms contained within the GeoGraphy graph analytics library. Users define geospatial-temporal semantic graph queries by constructing search templates based on nodes, edges, and the constraints among them. Users then analyze the results of the queries using both geo-spatial and temporal visualizations. Development of this application has made user experience an explicit driver for project and algorithmic level decisions that will affect how analysts one day make use of PANTHER technologies.

  8. AmeriFlux Network Data from the ORNL AmeriFlux Website

    DOE Data Explorer [Office of Scientific and Technical Information (OSTI)]

    The AmeriFlux network was established in 1996 to provide continuous observations of ecosystem level exchanges of CO2, water, energy and momentum spanning diurnal, synoptic, seasonal, and interannual time scales. It is fed by sites from North America, Central America, and South America. DOE's CDIAC stores and maintains AmeriFlux data, and this web site explains the different levels of data available there, with links to the CDIAC ftp site. A separate web-based data interface is also provided; it allows users to graph, query, and download Level 2 data for up to four sites at a time. Data may be queried by site, measurement period, or parameter. More than 550 site-years of level 2 data are available from AmeriFlux sites through the interface.

  9. Security of statistical data bases: invasion of privacy through attribute correlational modeling

    SciTech Connect (OSTI)

    Palley, M.A.

    1985-01-01

    This study develops, defines, and applies a statistical technique for the compromise of confidential information in a statistical data base. Attribute Correlational Modeling (ACM) recognizes that the information contained in a statistical data base represents real world statistical phenomena. As such, ACM assumes correlational behavior among the database attributes. ACM proceeds to compromise confidential information through creation of a regression model, where the confidential attribute is treated as the dependent variable. The typical statistical data base may preclude the direct application of regression. In this scenario, the research introduces the notion of a synthetic data base, created through legitimate queries of the actual data base, and through proportional random variation of responses to these queries. The synthetic data base is constructed to resemble the actual data base as closely as possible in a statistical sense. ACM then applies regression analysis to the synthetic data base, and utilizes the derived model to estimate confidential information in the actual database.

  10. Chunking of Large Multidimensional Arrays

    SciTech Connect (OSTI)

    Rotem, Doron; Otoo, Ekow J.; Seshadri, Sridhar

    2007-02-28

    Data intensive scientific computations as well on-lineanalytical processing applications as are done on very large datasetsthat are modeled as k-dimensional arrays. The storage organization ofsuch arrays on disks is done by partitioning the large global array intofixed size hyper-rectangular sub-arrays called chunks or tiles that formthe units of data transfer between disk and memory. Typical queriesinvolve the retrieval of sub-arrays in a manner that accesses all chunksthat overlap the query results. An important metric of the storageefficiency is the expected number of chunks retrieved over all suchqueries. The question that immediately arises is "what shapes of arraychunks give the minimum expected number of chunks over a query workload?"In this paper we develop two probabilistic mathematical models of theproblem and provide exact solutions using steepest descent and geometricprogramming methods. Experimental results, using synthetic workloads onreal life data sets, show that our chunking is much more efficient thanthe existing approximate solutions.

  11. Provenance management in Swift with implementation details.

    SciTech Connect (OSTI)

    Gadelha, L. M. R; Clifford, B.; Mattoso, M.; Wilde, M.; Foster, I.

    2011-04-01

    The Swift parallel scripting language allows for the specification, execution and analysis of large-scale computations in parallel and distributed environments. It incorporates a data model for recording and querying provenance information. In this article we describe these capabilities and evaluate interoperability with other systems through the use of the Open Provenance Model. We describe Swift's provenance data model and compare it to the Open Provenance Model. We also describe and evaluate activities performed within the Third Provenance Challenge, which consisted of implementing a specific scientific workflow, capturing and recording provenance information of its execution, performing provenance queries, and exchanging provenance information with other systems. Finally, we propose improvements to both the Open Provenance Model and Swift's provenance system.

  12. U.S. Coal Reserves

    U.S. Energy Information Administration (EIA) Indexed Site

    Coal Glossary › FAQS › Overview Data Coal Data Browser (interactive query tool with charting and mapping) Summary Prices Reserves Consumption Production Stocks Imports, exports & distribution Coal-fired electric power plants Transportation costs to electric power sector International All coal data reports Analysis & Projections Major Topics Most popular Consumption Environment Imports & exports Industry characteristics Prices Production Projections Recurring Reserves Stocks All

  13. Electricity Transmission Success Stories - Energy Innovation Portal

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Electricity Glossary › FAQS › Overview Data Electricity Data Browser (interactive query tool with charting & mapping) Summary Sales (consumption), revenue, prices & customers Generation and thermal output Capacity of electric power plants Consumption of fuels used to generate electricity Receipts of fossil-fuels for electricity generation Average cost of fossil-fuels for electricity generation Fossil-fuel stocks for electricity generation Cost, revenue and expense statistics for...

  14. Coal - U.S. Energy Information Administration (EIA)

    U.S. Energy Information Administration (EIA) Indexed Site

    Coal Glossary › FAQS › Overview Data Coal Data Browser (interactive query tool with charting and mapping) Summary Prices Reserves Consumption Production Stocks Imports, exports & distribution Coal-fired electric power plants Transportation costs to electric power sector International All coal data reports Analysis & Projections Major Topics Most popular Consumption Environment Imports & exports Industry characteristics Prices Production Projections Recurring Reserves Stocks All

  15. Quarterly Coal Report - Energy Information Administration

    U.S. Energy Information Administration (EIA) Indexed Site

    Coal Glossary › FAQS › Overview Data Coal Data Browser (interactive query tool with charting and mapping) Summary Prices Reserves Consumption Production Stocks Imports, exports & distribution Coal-fired electric power plants Transportation costs to electric power sector International All coal data reports Analysis & Projections Major Topics Most popular Consumption Environment Imports & exports Industry characteristics Prices Production Projections Recurring Reserves Stocks All

  16. UNITED STATES ATOMIC ENERGY COMMISSION SAC200063~~0oooo Frank K. Pittman, Director, /Division of Waste Management and Trans-

    Office of Legacy Management (LM)

    SAC200063~~0oooo .- Frank K. Pittman, Director, /Division of Waste Management and Trans- portation, Headquarters CONTAMIWATRD EE-AEC-OWNED OR IEASED FACILITIES This memorandum responds to your TWX dated October 30, 1973, requesting certain information on the above subject. Unfortunately, some of the documentation necessary to answer your queries is no Longer available due to the records disposal program or the agreements pre- vailing at the time of release or transfer of the facilities. From

  17. Quick Facts

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Coal Glossary › FAQS › Overview Data Coal Data Browser (interactive query tool with charting and mapping) Summary Prices Reserves Consumption Production Stocks Imports, exports & distribution Coal-fired electric power plants Transportation costs to electric power sector International All coal data reports Analysis & Projections Major Topics Most popular Consumption Environment Imports & exports Industry characteristics Prices Production Projections Recurring Reserves Stocks All

  18. Sandia National Laboratories: Business Opportunities Website

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Prospective Suppliers What Sandia Looks For In Our Suppliers What Does Sandia Buy? Business Opportunities Website Small Business Working with Sandia Business Opportunities Website The Business Opportunities Website (BOW) is located at the following URL: https://supplierportal.sandia.gov/OA_HTML/snl/AbstractQuery.jsp NOTE: Internet Explore and Firefox are preferred browsers. Users may encounter an error when using other browsers to view the above link. Twenty-four hours a day, 7 days a week, and

  19. Test program element II blanket and shield thermal-hydraulic and thermomechanical testing, experimental facility survey

    SciTech Connect (OSTI)

    Ware, A.G.; Longhurst, G.R.

    1981-12-01

    This report presents results of a survey conducted by EG and G Idaho to determine facilities available to conduct thermal-hydraulic and thermomechanical testing for the Department of Energy Office of Fusion Energy First Wall/Blanket/Shield Engineering Test Program. In response to EG and G queries, twelve organizations (in addition to EG and G and General Atomic) expressed interest in providing experimental facilities. A variety of methods of supplying heat is available.

  20. AmiGO: online access to ontology and annotation data

    SciTech Connect (OSTI)

    Carbon, Seth; Ireland, Amelia; Mungall, Christopher J.; Shu, ShengQiang; Marshall, Brad; Lewis, Suzanna

    2009-01-15

    AmiGO is a web application that allows users to query, browse, and visualize ontologies and related gene product annotation (association) data. AmiGO can be used online at the Gene Ontology (GO) website to access the data provided by the GO Consortium; it can also be downloaded and installed to browse local ontologies and annotations. AmiGO is free open source software developed and maintained by the GO Consortium.

  1. Decontamination and Decommisioning Equipment Tracking System

    Energy Science and Technology Software Center (OSTI)

    1994-08-26

    DDETS is Relational Data Base Management System (RDBMS) which incorporates 1-D (code 39) and 2-D (PDF417) bar codes into its equipment tracking capabilities. DDETS is compatible with the Reportable Excess Automated Property System (REAPS), and has add, edit, delete and query capabilities for tracking equipment being decontaminated and decommissioned. In addition, bar code technology is utilized in the inventory tracking and shipping of equipment.

  2. Method for indexing and retrieving manufacturing-specific digital imagery based on image content

    DOE Patents [OSTI]

    Ferrell, Regina K.; Karnowski, Thomas P.; Tobin, Jr., Kenneth W.

    2004-06-15

    A method for indexing and retrieving manufacturing-specific digital images based on image content comprises three steps. First, at least one feature vector can be extracted from a manufacturing-specific digital image stored in an image database. In particular, each extracted feature vector corresponds to a particular characteristic of the manufacturing-specific digital image, for instance, a digital image modality and overall characteristic, a substrate/background characteristic, and an anomaly/defect characteristic. Notably, the extracting step includes generating a defect mask using a detection process. Second, using an unsupervised clustering method, each extracted feature vector can be indexed in a hierarchical search tree. Third, a manufacturing-specific digital image associated with a feature vector stored in the hierarchicial search tree can be retrieved, wherein the manufacturing-specific digital image has image content comparably related to the image content of the query image. More particularly, can include two data reductions, the first performed based upon a query vector extracted from a query image. Subsequently, a user can select relevant images resulting from the first data reduction. From the selection, a prototype vector can be calculated, from which a second-level data reduction can be performed. The second-level data reduction can result in a subset of feature vectors comparable to the prototype vector, and further comparable to the query vector. An additional fourth step can include managing the hierarchical search tree by substituting a vector average for several redundant feature vectors encapsulated by nodes in the hierarchical search tree.

  3. Accounting - What happened with that job?

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Accounting - What happened with that job? Accounting - What happened with that job? On genepool there are three options for accessing information on your past jobs: Genepool completed jobs webpage (genepool only) The UGE provided tool: qacct (genepool or phoebe) The NERSC provided tool: qqacct - Query Queue Accounting data (genepool or phoebe) Everytime a job is completed - either failed or successful, the UGE batch system writes an entry into its accounting logs. These accounting logs contain a

  4. Allocation Management | Argonne Leadership Computing Facility

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Allocation Management Determining Allocation Requirements Querying Allocations Using cbank Mira/Cetus/Vesta Cooley Policies Documentation Feedback Please provide feedback to help guide us as we continue to build documentation for our new computing resource. [Feedback Form] Allocation Management Allocations require management - balance checks, resource allocation, requesting more time, etc. Checking for an active allocation To determine if there is an active allocation, check Running Jobs. For

  5. getnim - NIM's Command Line Interface

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    getnim command getnim - NIM's Command Line Interface This page describes the inquiry-only command called getnim that users can use interactively and in scripts to get their account balances. GETNIM(l) NERSC GETNIM(l) NAME getnim - query the NERSC banking database for remaining allocation, resources and repository information SYNOPSIS getnim [ options ] -Rrname or getnim [ options ] -Rrname { -uuid | -Uuname } or getnim [ options ][ -D ] { -uuid | -Uuname } or getnim [ options ] -Rrname { -l | -L

  6. BioMon: A Google Earth Based Continuous Biomass Monitoring System (Demo Paper)

    SciTech Connect (OSTI)

    Vatsavai, Raju

    2009-01-01

    We demonstrate a Google Earth based novel visualization system for continuous monitoring of biomass at regional and global scales. This system is integrated with a back-end spatiotemporal data mining system that continuously detects changes using high temporal resolution MODIS images. In addition to the visualization, we demonstrate novel query features of the system that provides insights into the current conditions of the landscape.

  7. Automated Nuclear Data Test Suite

    Energy Science and Technology Software Center (OSTI)

    2013-01-09

    Provides python routines to create a database of test problems in a user-defined directory tree, to query the database using user-defined parameters, to generate a list of test urns, to automatically run with user-defined particle transport codes. Includes natural isotope abundance data, and a table of benchmark effective for fast critical assemblies. Does not include input decks, cross-section libraries, or particle transport codes.

  8. Compact Mesh Generator

    Energy Science and Technology Software Center (OSTI)

    2007-02-02

    The CMG is a small, lightweight, structured mesh generation code. It features a simple text input parser that allows setup of various meshes via a small set of text commands. Mesh generation data can be output to text, the silo file format, or the API can be directly queried by applications. It can run serially or in parallel via MPI. The CMG includes the ability to specify varius initial conditions on a mesh via meshmore » tags.« less

  9. System and method for generating a relationship network

    DOE Patents [OSTI]

    Franks, Kasian; Myers, Cornelia A.; Podowski, Raf M.

    2011-07-26

    A computer-implemented system and process for generating a relationship network is disclosed. The system provides a set of data items to be related and generates variable length data vectors to represent the relationships between the terms within each data item. The system can be used to generate a relationship network for documents, images, or any other type of file. This relationship network can then be queried to discover the relationships between terms within the set of data items.

  10. System and method for generating a relationship network

    SciTech Connect (OSTI)

    Franks, Kasian; Myers, Cornelia A; Podowski, Raf M

    2015-05-05

    A computer-implemented system and process for generating a relationship network is disclosed. The system provides a set of data items to be related and generates variable length data vectors to represent the relationships between the terms within each data item. The system can be used to generate a relationship network for documents, images, or any other type of file. This relationship network can then be queried to discover the relationships between terms within the set of data items.

  11. Large-Scale Geospatial Indexing for Image-Based Retrieval and Analysis

    SciTech Connect (OSTI)

    Tobin Jr, Kenneth William; Bhaduri, Budhendra L; Bright, Eddie A; Cheriydat, Anil; Karnowski, Thomas Paul; Palathingal, Paul J; Potok, Thomas E; Price, Jeffery R

    2005-12-01

    We describe a method for indexing and retrieving high-resolution image regions in large geospatial data libraries. An automated feature extraction method is used that generates a unique and specific structural description of each segment of a tessellated input image file. These tessellated regions are then merged into similar groups and indexed to provide flexible and varied retrieval in a query-by-example environment.

  12. FE0005961_UIllinois | netl.doe.gov

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    RVA: 3-D Visualization and Analysis Software to Support Management of Unconventional Oil and Gas Resources Last Reviewed 12/2/2015 DE-FE0005961 Goal This project will produce a state-of-the-art 3-D visualization and analysis software package targeted for improving development of oil and gas resources. The software [RVA (Reservoir Visualization and Analysis)] will display data, models, and reservoir simulation results and have the ability to jointly visualize and query data from geologic models

  13. Cobalt Scheduler | Argonne Leadership Computing Facility

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Coal Glossary › FAQS › Overview Data Coal Data Browser (interactive query tool with charting and mapping) Summary Prices Reserves Consumption Production Stocks Imports, exports & distribution Coal-fired electric power plants Transportation costs to electric power sector International All coal data reports Analysis & Projections Major Topics Most popular Consumption Environment Imports & exports Industry characteristics Prices Production Projections Recurring Reserves Stocks All

  14. Electrical Energy Storage A DOE ENERGY FRONTIER RESEARCH CENTER

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Electricity Glossary › FAQS › Overview Data Electricity Data Browser (interactive query tool with charting & mapping) Summary Sales (consumption), revenue, prices & customers Generation and thermal output Capacity of electric power plants Consumption of fuels used to generate electricity Receipts of fossil-fuels for electricity generation Average cost of fossil-fuels for electricity generation Fossil-fuel stocks for electricity generation Cost, revenue and expense statistics for...

  15. Electricity Transmission and Distribution Technologies Available for

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Electricity Glossary › FAQS › Overview Data Electricity Data Browser (interactive query tool with charting & mapping) Summary Sales (consumption), revenue, prices & customers Generation and thermal output Capacity of electric power plants Consumption of fuels used to generate electricity Receipts of fossil-fuels for electricity generation Average cost of fossil-fuels for electricity generation Fossil-fuel stocks for electricity generation Cost, revenue and expense statistics for...

  16. Buildings Energy Data Book

    Buildings Energy Data Book [EERE]

    Current and Past EditionsGlossaryPopular TablesQuery Tools Contact Us Search What Is the Buildings Energy Data Book? The Data Book includes statistics on residential and commercial building energy consumption. Data tables contain statistics related to construction, building technologies, energy consumption, and building characteristics. The Building Technologies Program within the U.S. Department of Energy's Office of Energy Efficiency and Renewable Energy developed this resource to provide a

  17. 2015 GTP GTO Peer Review

    Office of Energy Efficiency and Renewable Energy (EERE) Indexed Site

    Prospector Project Officer: Arlene Anderson Total Project Funding: $157,000 May 11, 2015 This presentation does not contain any proprietary confidential, or otherwise restricted information. PI: Dan Getman National Renewable Energy Laboratory Systems Analysis, Resources Assessment, Data System Development & Population, Education 2 | US DOE Geothermal Office eere.energy.gov Relevance/Impact of Research Objective * The Geothermal Prospector provides access to explore, query, visualize, and

  18. Temporal Representation in Semantic Graphs

    SciTech Connect (OSTI)

    Levandoski, J J; Abdulla, G M

    2007-08-07

    A wide range of knowledge discovery and analysis applications, ranging from business to biological, make use of semantic graphs when modeling relationships and concepts. Most of the semantic graphs used in these applications are assumed to be static pieces of information, meaning temporal evolution of concepts and relationships are not taken into account. Guided by the need for more advanced semantic graph queries involving temporal concepts, this paper surveys the existing work involving temporal representations in semantic graphs.

  19. Electric Sales, Revenue, and Average Price 2011 - Energy Information

    U.S. Energy Information Administration (EIA) Indexed Site

    Administration Electricity Glossary › FAQS › Overview Data Electricity Data Browser (interactive query tool with charting & mapping) Summary Sales (consumption), revenue, prices & customers Generation and thermal output Capacity of electric power plants Consumption of fuels used to generate electricity Receipts of fossil-fuels for electricity generation Average cost of fossil-fuels for electricity generation Fossil-fuel stocks for electricity generation Cost, revenue and expense

  20. DOE - Office of Legacy Management -- New Canaan Site - CT 08

    Office of Legacy Management (LM)

    Canaan Site - CT 08 FUSRAP Considered Sites Site: NEW CANAAN SITE (CT.08) Eliminated from consideration under FUSRAP Designated Name: Not Designated Alternate Name: None Location: New Canaan , Connecticut CT.08-1 Evaluation Year: 1985 CT.08-2 Site Operations: None; Investigation of area prompted by public query; no site found in New Canaan. CT.08-1 Site Disposition: Eliminated - No AEC site located in this city CT.08-2 Radioactive Materials Handled: No Primary Radioactive Materials Handled: None

  1. At the intersection of past and future-The Lab's archives

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    The Lab's Archives Community Connections: Your link to news and opportunities from Los Alamos National Laboratory Latest Issue: September 1, 2016 all issues All Issues » submit At the intersection of past and future-The Lab's archives The Archives staff typically handle about 60 requests a month for everything from Freedom of Information Act queries to calls from journalists and television producers. January 1, 2013 dummy image Read our archives Contacts Editor Linda Anderman Email Community

  2. Nuclear Fuel Cycle Reasoner: PNNL FY13 Report

    SciTech Connect (OSTI)

    Hohimer, Ryan E.; Strasburg, Jana D.

    2013-09-30

    In Fiscal Year 2012 (FY12) PNNL implemented a formal reasoning framework and applied it to a specific challenge in nuclear nonproliferation. The Semantic Nonproliferation Analysis Platform (SNAP) was developed as a preliminary graphical user interface to demonstrate the potential power of the underlying semantic technologies to analyze and explore facts and relationships relating to the nuclear fuel cycle (NFC). In Fiscal Year 2013 (FY13) the SNAP demonstration was enhanced with respect to query and navigation usability issues.

  3. Worldwide report: Arms control, [July 26, 1986

    SciTech Connect (OSTI)

    1986-07-26

    This report contains translations/transcriptions of articles and/or broadcasts on arms control. Titles include; Soviet Spokesman Explains Far East Arms Cut; Delegation attends Society Naval Exercise; Defense Minister Queried on Military Reductions; Further on Society Force Withdrawals from Poland; Criteria of Military-Strategic Parity, Sufficiency; Further on Allegations of CW Materiel Sale to Iran; Reports on Nuclear, Chemical Warheads Denied; and others.

  4. JPRS report: Arms control, [July 11, 1989

    SciTech Connect (OSTI)

    1989-07-11

    This report contains translations/transcriptions of articles and/or broadcasts on arms control. Titles include: Soviet Spokesman Explains Far East Arms Cut; Delegation Attends Soviet Naval Exercise; Defense Minister Queried on Military Reductions; Further on Soviet Force Withdrawals from Poland; Criteria of Military-Strategic Parity, Sufficiency; Further on Allegations of CW Materiel Sale to Iran; Reports on Nuclear, Chemical Warheads Denied; and others.

  5. Worldwide report: Arms control, [19 July 1985

    SciTech Connect (OSTI)

    1985-07-19

    This report contains translations/transcriptions of articles and/or broadcasts on arms control. Titles include: Soviet Spokesman Explains Far East Arms Cut; Delegation attends Soviet Naval Exercise; Defense Minister Queried on Military Reductions; Further on Soviet Force Withdrawals from Poland; Criteria of Military-Strategic Parity, Sufficiency; Further on Allegations of CW Materiel Sale to Iran; Reports on Nuclear, Chemical Warheads Denied; and others.

  6. U.S. Federal Agencies | OSTI, US Dept of Energy Office of Scientific and

    Office of Scientific and Technical Information (OSTI)

    Technical Information U.S. Federal Agencies Science Search Tools Home DOE Collections Journal Sources Library and University Tools U.S. Federal Agencies Global Databases Customized Resources This interagency initiative of 19 U.S. government science organizations within 15 Federal Agencies provides a search of over 60 scientific databases and 200 million pages of science information with just one query. Science.gov is a gateway to over 2,200 scientific websites. For more information about

  7. National Library of Energy (BETA): the Department of Energy's National

    Office of Scientific and Technical Information (OSTI)

    Resource for Energy Literacy, Innovation and Security - Help Help Simple Search Advanced Search Search Results Search Tools Selecting, Downloading and Printing Results Emailing Results Search Tips Simple Search A simple search from the homepage will search all of the collections in the application, merge the results and rank them according to how relevant they are to your query. To conduct a Simple Search: Type in your keyword(s), like "deep web technologies" and select Search.

  8. DOE Science Showcase - Biofuels in the databases | OSTI, US Dept of

    Office of Scientific and Technical Information (OSTI)

    Energy Office of Scientific and Technical Information DOE Science Showcase - Biofuels in the databases The new ScienceCinema provides access points where the term biofuels is spoken in DOE multimedia presentations. DOE Green Energy renewable energy portal offers biofuels related research. Science Accelerator returns results for biofuels from DOE resources with just one query: DOE Data Explorer DOE Information Bridge Energy Citations Database Federal R&D Project Summaries Biofuels in the

  9. OSTI, US Dept of Energy Office of Scientific and Technical Information |

    Office of Scientific and Technical Information (OSTI)

    Speeding access to science information from DOE and Beyond Enormous STI Content Made Easily Searchable by OSTI by Dr. Walt Warnick on Thu, July 01, 2010 1591 mwws.jpg Enormous STI Content Made Easily Searchable by OSTI Read more about 1591 We have integrated about ten OSTI products dealing with technical reports, e-prints, patents, conference proceedings, project summaries, etc., so that they are all searchable via s single query. The integrated product allows users to search without first

  10. Identification of candidate genes in Populus cell wall biosynthesis using text-mining, co-expression network and comparative genomics

    SciTech Connect (OSTI)

    Yang, Xiaohan; Ye, Chuyu; Bisaria, Anjali; Tuskan, Gerald A; Kalluri, Udaya C

    2011-01-01

    Populus is an important bioenergy crop for bioethanol production. A greater understanding of cell wall biosynthesis processes is critical in reducing biomass recalcitrance, a major hindrance in efficient generation of ethanol from lignocellulosic biomass. Here, we report the identification of candidate cell wall biosynthesis genes through the development and application of a novel bioinformatics pipeline. As a first step, via text-mining of PubMed publications, we obtained 121 Arabidopsis genes that had the experimental evidences supporting their involvement in cell wall biosynthesis or remodeling. The 121 genes were then used as bait genes to query an Arabidopsis co-expression database and additional genes were identified as neighbors of the bait genes in the network, increasing the number of genes to 548. The 548 Arabidopsis genes were then used to re-query the Arabidopsis co-expression database and re-construct a network that captured additional network neighbors, expanding to a total of 694 genes. The 694 Arabidopsis genes were computationally divided into 22 clusters. Queries of the Populus genome using the Arabidopsis genes revealed 817 Populus orthologs. Functional analysis of gene ontology and tissue-specific gene expression indicated that these Arabidopsis and Populus genes are high likelihood candidates for functional genomics in relation to cell wall biosynthesis.

  11. Reliability Availability Serviceability

    Energy Science and Technology Software Center (OSTI)

    2006-09-18

    Our work is aimed at providing a data store for system-level events and presenting a flexible query interface to those events. The work extends the functinality provided by the open source Request Tracker (RT) (http://www.bestpractical.com/rt) project witht the Asset Tracker (AT) addon (http://atwiki.chaka.net). We have developed an Event Tracker add-on to RT and an interface for gathering, dispatching, and inserting system events into Event Tracker. Data sources include data from all components of the system.more » Data is initially sent to a defined set of data filters. The data filters are capable of discarding specified data, throttling input, handling context-sensitive input, passing data through an external shell pipe command, and compressing multiple data enteries into a single event. The filters then pass the data on to an event dispatch engine. The dispatcher can print events to the screen as they happen, track them in the database, forward them on, or pass them on to an external command. By collecting all of the data into a single database, we are able to leverage the Query Builder interface supplied by RT to create, save, and restore almost any kind of query imaginable.« less

  12. Expediting Scientific Data Analysis with Reorganization of Data

    SciTech Connect (OSTI)

    Byna, Surendra; Wu, Kesheng

    2013-08-19

    Data producers typically optimize the layout of data files to minimize the write time. In most cases, data analysis tasks read these files in access patterns different from the write patterns causing poor read performance. In this paper, we introduce Scientific Data Services (SDS), a framework for bridging the performance gap between writing and reading scientific data. SDS reorganizes data to match the read patterns of analysis tasks and enables transparent data reads from the reorganized data. We implemented a HDF5 Virtual Object Layer (VOL) plugin to redirect the HDF5 dataset read calls to the reorganized data. To demonstrate the effectiveness of SDS, we applied two parallel data organization techniques: a sort-based organization on a plasma physics data and a transpose-based organization on mass spectrometry imaging data. We also extended the HDF5 data access API to allow selection of data based on their values through a query interface, called SDS Query. We evaluated the execution time in accessing various subsets of data through existing HDF5 Read API and SDS Query. We showed that reading the reorganized data using SDS is up to 55X faster than reading the original data.

  13. Sandia Cognitive Aide V2.0

    Energy Science and Technology Software Center (OSTI)

    2004-04-15

    The Sandia Cognitive Aide (SCA) collects data from personal computer use and uses this information to make suggestions to the user. It records interactions with MS Outlook, MS Word, MS PowerPoint, and the Internet Explorer, indexing email messages, documents, presentation, and web pages accessed, The user can then query the indexed documents from any Windows application. The system also suggests what it believes to be relevant terms to a given query. The software provides facilitiesmore » for constructing and submitting queries to WWW search engines. This version of the software also enables the user to define different "task contexts" within the user works. The contexts are defined in terms of related terms. The user can associate documents with these contexts. The contexts can be searched as well as the documents. This software is designed to access and utilize the cognitive model being build by Sandia National Laboratories, org. 15311 and uses the STANLEY text analysis library.« less

  14. A Metadata-Rich File System

    SciTech Connect (OSTI)

    Ames, S; Gokhale, M B; Maltzahn, C

    2009-01-07

    Despite continual improvements in the performance and reliability of large scale file systems, the management of file system metadata has changed little in the past decade. The mismatch between the size and complexity of large scale data stores and their ability to organize and query their metadata has led to a de facto standard in which raw data is stored in traditional file systems, while related, application-specific metadata is stored in relational databases. This separation of data and metadata requires considerable effort to maintain consistency and can result in complex, slow, and inflexible system operation. To address these problems, we have developed the Quasar File System (QFS), a metadata-rich file system in which files, metadata, and file relationships are all first class objects. In contrast to hierarchical file systems and relational databases, QFS defines a graph data model composed of files and their relationships. QFS includes Quasar, an XPATH-extended query language for searching the file system. Results from our QFS prototype show the effectiveness of this approach. Compared to the defacto standard, the QFS prototype shows superior ingest performance and comparable query performance on user metadata-intensive operations and superior performance on normal file metadata operations.

  15. Data Management Architectures

    SciTech Connect (OSTI)

    Critchlow, Terence J.; Abdulla, Ghaleb; Becla, Jacek; Kleese van Dam, Kerstin; Lang, Sam; McGuinness, Deborah L.

    2012-10-31

    Data management is the organization of information to support efficient access and analysis. For data intensive computing applications, the speed at which relevant data can be accessed is a limiting factor in terms of the size and complexity of computation that can be performed. Data access speed is impacted by the size of the relevant subset of the data, the complexity of the query used to define it, and the layout of the data relative to the query. As the underlying data sets become increasingly complex, the questions asked of it become more involved as well. For example, geospatial data associated with a city is no longer limited to the map data representing its streets, but now also includes layers identifying utility lines, key points, locations and types of businesses within the city limits, tax information for each land parcel, satellite imagery, and possibly even street-level views. As a result, queries have gone from simple questions, such as "how long is Main Street?", to much more complex questions such as "taking all other factors into consideration, are the property values of houses near parks higher than those under power lines, and if so, by what percentage". Answering these questions requires a coherent infrastructure, integrating the relevant data into a format optimized for the questions being asked.

  16. Efficient Analysis of Live and Historical Streaming Data and itsApplication to Cybersecurity

    SciTech Connect (OSTI)

    Reiss, Frederick; Stockinger, Kurt; Wu, Kesheng; Shoshani, Arie; Hellerstein, Joseph M.

    2007-04-06

    Applications that query data streams in order to identifytrends, patterns, or anomalies can often benefit from comparing the livestream data with archived historical stream data. However, searching thishistorical data in real time has been considered so far to beprohibitively expensive. One of the main bottlenecks is the update costsof the indices over the archived data. In this paper, we address thisproblem by using our highly-efficient bitmap indexing technology (calledFastBit) and demonstrate that the index update operations aresufficiently efficient for this bottleneck to be removed. We describe ourprototype system based on the TelegraphCQ streaming query processor andthe FastBit bitmap index. We present a detailed performance evaluation ofour system using a complex query workload for analyzing real networktraffic data. The combined system uses TelegraphCQ to analyze streams oftraffic information and FastBit to correlate current behaviors withhistorical trends. We demonstrate that our system can simultaneouslyanalyze (1) live streams with high data rates and (2) a large repositoryof historical stream data.

  17. Cosmetic Outcomes and Complications Reported by Patients Having Undergone Breast-Conserving Treatment

    SciTech Connect (OSTI)

    Hill-Kayser, Christine E.; Vachani, Carolyn; Hampshire, Margaret K.; Di Lullo, Gloria A.; Metz, James M.

    2012-07-01

    Purpose: Over the past 30 years, much work in treatment of breast cancer has contributed to improvement of cosmetic and functional outcomes. The goal of breast-conservation treatment (BCT) is avoidance of mastectomy through use of lumpectomy and adjuvant radiation. Modern data demonstrate 'excellent' or 'good' cosmesis in >90% of patients treated with BCT. Methods and Materials: Patient-reported data were gathered via a convenience sample frame from breast cancer survivors using a publically available, free, Internet-based tool for creation of survivorship care plans. During use of the tool, breast cancer survivors are queried as to the cosmetic appearance of the treated breast, as well as perceived late effects. All data have been maintained anonymously with internal review board approval. Results: Three hundred fifty-four breast cancer survivors having undergone BCT and voluntarily using this tool were queried with regard to breast cosmesis and perceived late effects. Median diagnosis age was 48 years, and median current age 52 years. 'Excellent' cosmesis was reported by 27% (n = 88), 'Good' by 44% (n = 144), 'Fair' by 24% (n = 81), and 'Poor' by 5% (n = 18). Of the queries posted to survivors after BCT, late effects most commonly reported were cognitive changes (62%); sexual concerns (52%); changes in texture and color of irradiated skin (48%); chronic pain, numbness, or tingling (35%); and loss of flexibility in the irradiated area (30%). Survivors also described osteopenia/osteoporosis (35%), cardiopulmonary problems (12%), and lymphedema (19%). Conclusions: This anonymous tool uses a convenience sample frame to gather patient reported assessments of cosmesis and complications after breast cancer. Among the BCT population, cosmetic assessment by survivors appears less likely to be 'excellent' or 'good' than would be expected, with 30% of BCT survivors reporting 'fair' or 'poor' cosmesis. Patient reported incidence of chronic pain, as well as cognitive and

  18. Interoperable PKI Data Distribution in Computational Grids

    SciTech Connect (OSTI)

    Pala, Massimiliano; Cholia, Shreyas; Rea, Scott A.; Smith, Sean W.

    2008-07-25

    One of the most successful working examples of virtual organizations, computational grids need authentication mechanisms that inter-operate across domain boundaries. Public Key Infrastructures(PKIs) provide sufficient flexibility to allow resource managers to securely grant access to their systems in such distributed environments. However, as PKIs grow and services are added to enhance both security and usability, users and applications must struggle to discover available resources-particularly when the Certification Authority (CA) is alien to the relying party. This article presents how to overcome these limitations of the current grid authentication model by integrating the PKI Resource Query Protocol (PRQP) into the Grid Security Infrastructure (GSI).

  19. REMS Webinar Survey

    Office of Energy Efficiency and Renewable Energy (EERE) Indexed Site

    REMS Webinar Survey Hosted by the Office of Environment, Health, Safety and Security Thank you for participating in the inaugural DOE REMS Webinar that was held on Tuesday, March 8 at 1:00 pm EST. Please take a moment to respond to this survey. Excellent Good Fair Poor n/a Webinar access and login process Topics covered: ● PII ● Reporting requirements ● Site descriptions ● ALARA success ● Query Tool ● Visualization tools Relevance to your work Presentation and materials Length (time)

  20. USGS Annual Water Data Reports

    SciTech Connect (OSTI)

    2012-04-01

    Water resources data are published annually for use by engineers, scientists, managers, educators, and the general public. These archival products supplement direct access to current and historical water data provided by the National Water Information System (NWIS). Beginning with Water Year 2006, annual water data reports are available as individual electronic Site Data Sheets for the entire Nation for retrieval, download, and localized printing on demand. National distribution includes tabular and map interfaces for search, query, display and download of data. Data provided include extreme and mean discharge rates.

  1. UNITED STATES ATOMIC ENERGY COMMISSION

    Office of Legacy Management (LM)

    lLB"O"L"P"E OPC"AT10*s OCFlCC ..a .0x s.00 ALSUOULIQUL. "6" YLXICO "98s Nov 28 1973, Frank K. Pittmsn, Director, 'Division of Waste Management and Trans- portation, Headquarters CONTAMINATED KK-AEC-OWNED OR LEASED FACILITIES This memorandum responds to your TWK dat.ed October 30, 1973, requesting certain information on the above subject. Unfortunately, same of the documentation necessary to answer your queries is no longer available due to the records

  2. Feature Based Tolerancing Product Modeling V4.1

    Energy Science and Technology Software Center (OSTI)

    2001-11-30

    FBTol is a component technology in the form of software linkable library. The purpose of FBToI is to augment the shape of a nominal solid model with an explicit representation of a product’s tolerances and other non-shape attributes. This representation enforces a complete and unambiguous definition of non-shape information, permits an open architecture to dynamically create, modify, delete, and query tolerance information, and incorporates verify and checking algorithms to assure the quality of the tolerancemore » design.« less

  3. OSTI, US Dept of Energy Office of Scientific and Technical Information |

    Office of Scientific and Technical Information (OSTI)

    Speeding access to science information from DOE and Beyond Science.gov's Unique Collaboration by Valerie Allen on Mon, September 14, 2009 Science.gov is a one-stop portal for federal government science information. Over 200 million pages of science information from 14 federal agencies may be searched through a single query. How far we have come in the past decade! You may not be aware that Science.gov was developed and is governed by the Science.gov Alliance, a group of science information

  4. Yushu

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Yushu SciDB @ NERSC --- 1 --- Array Like Science Data " - More common than you think --- 2 --- SciDB, parallel processing without parallel programming Everything i n A rrays - Locate a n e lement a t O(constant) - Can b e v ery s parse - Best f or m achine/simula9on generated s tructure d ata - Good f or m etadata t oo * Query---like l anguage, a uto--- paralleliza:on * Do C alcula:ons i nside t he DB --- 3 --- Yushu Y ao NERSC SciDB Testbed * Partner u p w ith S cience T eams - Hold t heir

  5. Adding Data Management Services to Parallel File Systems

    SciTech Connect (OSTI)

    Brandt, Scott

    2015-03-04

    The objective of this project, called DAMASC for “Data Management in Scientific Computing”, is to coalesce data management with parallel file system management to present a declarative interface to scientists for managing, querying, and analyzing extremely large data sets efficiently and predictably. Managing extremely large data sets is a key challenge of exascale computing. The overhead, energy, and cost of moving massive volumes of data demand designs where computation is close to storage. In current architectures, compute/analysis clusters access data in a physically separate parallel file system and largely leave it scientist to reduce data movement. Over the past decades the high-end computing community has adopted middleware with multiple layers of abstractions and specialized file formats such as NetCDF-4 and HDF5. These abstractions provide a limited set of high-level data processing functions, but have inherent functionality and performance limitations: middleware that provides access to the highly structured contents of scientific data files stored in the (unstructured) file systems can only optimize to the extent that file system interfaces permit; the highly structured formats of these files often impedes native file system performance optimizations. We are developing Damasc, an enhanced high-performance file system with native rich data management services. Damasc will enable efficient queries and updates over files stored in their native byte-stream format while retaining the inherent performance of file system data storage via declarative queries and updates over views of underlying files. Damasc has four key benefits for the development of data-intensive scientific code: (1) applications can use important data-management services, such as declarative queries, views, and provenance tracking, that are currently available only within database systems; (2) the use of these services becomes easier, as they are provided within a familiar file

  6. Towards a Relation Extraction Framework for Cyber-Security Concepts

    SciTech Connect (OSTI)

    Jones, Corinne L; Bridges, Robert A; Huffer, Kelly M; Goodall, John R

    2015-01-01

    In order to assist security analysts in obtaining information pertaining to their network, such as novel vulnerabilities, exploits, or patches, information retrieval methods tailored to the security domain are needed. As labeled text data is scarce and expensive, we follow developments in semi-supervised NLP and implement a bootstrapping algorithm for extracting security entities and their relationships from text. The algorithm requires little input data, specifically, a few relations or patterns (heuristics for identifying relations), and incorporates an active learning component which queries the user on the most important decisions to prevent drifting the desired relations. Preliminary testing on a small corpus shows promising results, obtaining precision of .82.

  7. Geospatial Toolkits and Resource Maps for Selected Countries from the National Renewable Energy Laboratory (NREL)

    DOE Data Explorer [Office of Scientific and Technical Information (OSTI)]

    NREL developed the Geospatial Toolkit (GsT), a map-based software application that integrates resource data and geographic information systems (GIS) for integrated resource assessment. A variety of agencies within countries, along with global datasets, provided country-specific data. Originally developed in 2005, the Geospatial Toolkit was completely redesigned and re-released in November 2010 to provide a more modern, easier-to-use interface with considerably faster analytical querying capabilities. Toolkits are available for 21 countries and each one can be downloaded separately. The source code for the toolkit is also available. [Taken and edited from http://www.nrel.gov/international/geospatial_toolkits.html

  8. Method for gathering and summarizing internet information

    DOE Patents [OSTI]

    Potok, Thomas E.; Elmore, Mark Thomas; Reed, Joel Wesley; Treadwell, Jim N.; Samatova, Nagiza Faridovna

    2008-01-01

    A computer method of gathering and summarizing large amounts of information comprises collecting information from a plurality of information sources (14, 51) according to respective maps (52) of the information sources (14), converting the collected information from a storage format to XML-language documents (26, 53) and storing the XML-language documents in a storage medium, searching for documents (55) according to a search query (13) having at least one term and identifying the documents (26) found in the search, and displaying the documents as nodes (33) of a tree structure (32) having links (34) and nodes (33) so as to indicate similarity of the documents to each other.

  9. System for gathering and summarizing internet information

    DOE Patents [OSTI]

    Potok, Thomas E.; Elmore, Mark Thomas; Reed, Joel Wesley; Treadwell, Jim N.; Samatova, Nagiza Faridovna

    2006-07-04

    A computer method of gathering and summarizing large amounts of information comprises collecting information from a plurality of information sources (14, 51) according to respective maps (52) of the information sources (14), converting the collected information from a storage format to XML-language documents (26, 53) and storing the XML-language documents in a storage medium, searching for documents (55) according to a search query (13) having at least one term and identifying the documents (26) found in the search, and displaying the documents as nodes (33) of a tree structure (32) having links (34) and nodes (33) so as to indicate similarity of the documents to each other.

  10. Method for gathering and summarizing internet information

    DOE Patents [OSTI]

    Potok, Thomas E.; Elmore, Mark Thomas; Reed, Joel Wesley; Treadwell, Jim N.; Samatova, Nagiza Faridovna

    2010-04-06

    A computer method of gathering and summarizing large amounts of information comprises collecting information from a plurality of information sources (14, 51) according to respective maps (52) of the information sources (14), converting the collected information from a storage format to XML-language documents (26, 53) and storing the XML-language documents in a storage medium, searching for documents (55) according to a search query (13) having at least one term and identifying the documents (26) found in the search, and displaying the documents as nodes (33) of a tree structure (32) having links (34) and nodes (33) so as to indicate similarity of the documents to each other.

  11. Co-op

    Energy Science and Technology Software Center (OSTI)

    2007-05-25

    Co-op is primarily middleware software, a runtime system for the support of the Cooperative Parallel Programming model. This model is based on using whole SPMD applications as components in a scalable programming, and having them treat one another as single objects and communicate via remote method invocation. Also included is some application level software: (1) a metric space database library for managing data items located in an arbitrary metric space and retrieving based on nearestmore » neighbor queries; and (2) a Krieging extrapolation library for use in implementing adaptive sampling for generic multiscale simulations.« less

  12. Second Line of Defense Master Spares Catalog

    SciTech Connect (OSTI)

    Henderson, Dale L.; Muller, George; Mercier, Theresa M.; Brigantic, Robert T.; Perkins, Casey J.; Cooley, Scott K.

    2012-11-20

    This catalog is intended to be a comprehensive listing of repair parts, components, kits, and consumable items used on the equipment deployed at SLD sites worldwide. The catalog covers detection, CAS, network, ancillary equipment, and tools. The catalog is backed by a Master Parts Database which is used to generate the standard report views of the catalog. The master parts database is a relational database containing a record for every part in the master parts catalog along with supporting tables for normalizing fields in the records. The database also includes supporting queries, database maintenance forms, and reports.

  13. Construction of file database management

    SciTech Connect (OSTI)

    MERRILL,KYLE J.

    2000-03-01

    This work created a database for tracking data analysis files from multiple lab techniques and equipment stored on a central file server. Experimental details appropriate for each file type are pulled from the file header and stored in a searchable database. The database also stores specific location and self-directory structure for each data file. Queries can be run on the database according to file type, sample type or other experimental parameters. The database was constructed in Microsoft Access and Visual Basic was used for extraction of information from the file header.

  14. Using Web and Social Media for Influenza Surveillance

    SciTech Connect (OSTI)

    Corley, Courtney D.; Cook, Diane; Mikler, Armin R.; Singh, Karan P.

    2010-01-04

    Analysis of Google influenza-like-illness (ILI) search queries has shown a strongly correlated pattern with Centers for Disease Control (CDC) and Prevention seasonal ILI reporting data.Web and social media provide another resource to detect increases in ILI. This paper evaluates trends in blog posts that discuss influenza. Our key finding is that from 5-October 2008 to 31-January 2009 a high correlation exists between the frequency of posts, containing influenza keywords, per week and CDC influenza-like-illness surveillance data.

  15. StreamWorks - A system for Dynamic Graph Search

    SciTech Connect (OSTI)

    Choudhury, Sutanay; Holder, Larry; Chin, George; Ray, Abhik; Beus, Sherman J.; Feo, John T.

    2013-06-11

    Acting on time-critical events by processing ever growing social media, news or cyber data streams is a major technical challenge. Many of these data sources can be modeled as multi-relational graphs. Mining and searching for subgraph patterns in a continuous setting requires an efficient approach to incremental graph search. The goal of our work is to enable real-time search capabilities for graph databases. This demonstration will present a dynamic graph query system that leverages the structural and semantic characteristics of the underlying multi-relational graph.

  16. Automated Feature Generation in Large-Scale Geospatial Libraries for Content-Based Indexing.

    SciTech Connect (OSTI)

    Tobin Jr, Kenneth William; Bhaduri, Budhendra L; Bright, Eddie A; Cheriydat, Anil; Karnowski, Thomas Paul; Palathingal, Paul J; Potok, Thomas E; Price, Jeffery R

    2006-05-01

    We describe a method for indexing and retrieving high-resolution image regions in large geospatial data libraries. An automated feature extraction method is used that generates a unique and specific structural description of each segment of a tessellated input image file. These tessellated regions are then merged into similar groups, or sub-regions, and indexed to provide flexible and varied retrieval in a query-by-example environment. The methods of tessellation, feature extraction, sub-region clustering, indexing, and retrieval are described and demonstrated using a geospatial library representing a 153 km2 region of land in East Tennessee at 0.5 m per pixel resolution.

  17. DOE Research and Development Accomplishments XML Service

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    XML Service This XML service is a mechanism for searching the DOE R&D Accomplishments Database, full-text documents, and Web pages, either through a query string in a browser or via a computer application. It is based upon open standards Web protocols and facilitates communications and collaborations of applications and people. Search results are returned in an XML format. This format can be easily parsed, making it simple to add to a federated search. Specifics about the DOE R&D

  18. Notices

    National Nuclear Security Administration (NNSA)

    4908 Federal Register / Vol. 77, No. 173 / Thursday, September 6, 2012 / Notices and 214 of the Commission's Regulations (18 CFR 385.211 and 385.214) on or before 5 p.m. Eastern time on the specified comment date. Protests may be considered, but intervention is necessary to become a party to the proceeding. The filings are accessible in the Commission's eLibrary system by clicking on the links or querying the docket number. eFiling is encouraged. More detailed information relating to filing

  19. Sandia Equation of State Model Library

    Energy Science and Technology Software Center (OSTI)

    2013-08-29

    The software provides a general interface for querying thermodynamic states of material models along with implementation of both general and specific equation of state models. In particular, models are provided for the IAPWS-IF97 and IAPWS95 water standards as well as the associated water standards for viscosity, thermal conductivity, and surface tension. The interface supports implementation of models in a variety of independent variable spaces. Also, model support routines are included that allow for coupling ofmore » models and determination and representation of phase boundaries.« less

  20. Existing generating assets squeezed as new project starts slow

    SciTech Connect (OSTI)

    Jones, R.B.; Tiffany, E.D.

    2009-01-15

    Most forecasting reports concentrate on political or regulatory events to predict future industry trends. Frequently overlooked are the more empirical performance trends of the principal power generation technologies. Solomon and Associates queried its many power plant performance databases and crunched some numbers to identify those trends. Areas of investigation included reliability, utilization (net output factor and net capacity factor) and cost (operating costs). An in-depth analysis for North America and Europe is presented in this article, by region and by regeneration technology. 4 figs., 2 tabs.

  1. Navigating nuclear science: Enhancing analysis through visualization

    SciTech Connect (OSTI)

    Irwin, N.H.; Berkel, J. van; Johnson, D.K.; Wylie, B.N.

    1997-09-01

    Data visualization is an emerging technology with high potential for addressing the information overload problem. This project extends the data visualization work of the Navigating Science project by coupling it with more traditional information retrieval methods. A citation-derived landscape was augmented with documents using a text-based similarity measure to show viability of extension into datasets where citation lists do not exist. Landscapes, showing hills where clusters of similar documents occur, can be navigated, manipulated and queried in this environment. The capabilities of this tool provide users with an intuitive explore-by-navigation method not currently available in today`s retrieval systems.

  2. Relational Blackboard

    Energy Science and Technology Software Center (OSTI)

    2012-09-11

    The Relational Blackboard (RBB) is an extension of the H2 Relational Database to support discrete events and timeseries data. The original motivation for RBB is as a knowledge base for cognitive systems and simulations. It is useful wherever there is a need for persistent storage of timeseries (i.e. samples of a continuous process generating numerical data) and semantic labels for the data. The RBB is an extension to the H2 Relational Database, which is open-source.more » RBB is a set of stored procedures for H2 allowing data to be labeled, queried, and resampled.« less

  3. Streamnet; Northwest Aquatic Information Network, 2002 Annual Report.

    SciTech Connect (OSTI)

    Schmidt, Bruce

    2003-02-07

    A primary focus of the StreamNet project in FY-02 was maintenance and update of ongoing data types. Significant progress was made toward updating data for the primary data categories in the StreamNet regional database. Data updates had been slowed in previous years due to the time required for conversion of georeferencing for most data types from the 1:250,000 scale River Reach Number (RRN) system to the 1:100,000 Longitude-Latitude Identifier (LLID) system. In addition, data relating to Protected Areas and Smolt Density Model results, the last data sets in the StreamNet database still in the 1:250,000 RRN format, were converted this year to the LLID system, making them available through the on-line Web Query System. The Protected Areas data were also made available through an on-line interactive mapping application. All routine project activities continued, including project administration at the full project and cooperating project levels, project management through the StreamNet Steering Committee, maintenance of databases and Internet data delivery systems, and providing data related services to the Northwest Power Planning Council's (NWPPC) Fish and Wildlife Program. As part of system management, a new web server was put in operation, significantly improving speed and reliability of Internet data delivery. The web based data query system was modified to utilize ColdFusion, in preparation for a full conversion to ColdFusion from the custom programming in Delphi. This greatly increased flexibility and the ability to modify query system function, correct errors, and develop new query capabilities. All project participants responded to numerous requests for information (data, maps, technical assistance, etc.) throughout the year. A significant accomplishment this year was resolution of long standing differences in how fish distribution is defined and presented. By focusing strictly on definitions related to current distribution (ignoring potential and historic

  4. Rapid Exploitation and Analysis of Documents

    SciTech Connect (OSTI)

    Buttler, D J; Andrzejewski, D; Stevens, K D; Anastasiu, D; Gao, B

    2011-11-28

    Analysts are overwhelmed with information. They have large archives of historical data, both structured and unstructured, and continuous streams of relevant messages and documents that they need to match to current tasks, digest, and incorporate into their analysis. The purpose of the READ project is to develop technologies to make it easier to catalog, classify, and locate relevant information. We approached this task from multiple angles. First, we tackle the issue of processing large quantities of information in reasonable time. Second, we provide mechanisms that allow users to customize their queries based on latent topics exposed from corpus statistics. Third, we assist users in organizing query results, adding localized expert structure over results. Forth, we use word sense disambiguation techniques to increase the precision of matching user generated keyword lists with terms and concepts in the corpus. Fifth, we enhance co-occurrence statistics with latent topic attribution, to aid entity relationship discovery. Finally we quantitatively analyze the quality of three popular latent modeling techniques to examine under which circumstances each is useful.

  5. POSet Ontology Categorizer

    SciTech Connect (OSTI)

    Miniszewski, Sue M.

    2005-03-01

    POSet Ontology Categorizer (POSOC) V1.0 The POSet Ontology Categorizer (POSOC) software package provides tools for creating and mining of poset-structured ontologies, such as the Gene Ontology (GO). Given a list of weighted query items (ex.genes,proteins, and/or phrases) and one or more focus nodes, POSOC determines the ordered set of GO nodes that summarize the query, based on selections of a scoring function, pseudo-distance measure, specificity level, and cluster determination. Pseudo-distance measures provided are minimum chain length, maximum chain length, average of extreme chain lengths, and average of all chain lengths. A low specificity level, such as -1 or 0, results in a general set of clusters. Increasing the specificity results in more specific results in more specific and lighter clusters. POSOC cluster results can be compared agaist known results by calculations of precision, recall, and f-score for graph neighborhood relationships. This tool has been used in understanding the function of a set of genes, finding similar genes, and annotating new proteins. The POSOC software consists of a set of Java interfaces, classes, and programs that run on Linux or Windows platforms. It incorporates graph classes from OpenJGraph (openjgraph.sourceforge.net).

  6. The Configuration Space Toolkit (C-Space Toolkit or CSTK) Ver. 2.5 beta

    Energy Science and Technology Software Center (OSTI)

    2010-02-24

    The C-Space Toolkit provides a software library that makes it easier to program motion planning, simulation, robotics, and virtual reality codes using the Configuration Space abstraction. Key functionality (1) enables the user to special create representations of movable and stationary rigid geometric objects, and (2) perform fast distance, interference (clash) detection, collision detection, closest-feature pairs, and contact queries in terms of object configuration. Not only can queries be computed at any given point in configurationmore » space, but they can be done exactly over linear-translational path segments and approximately for rotational path segments. Interference detection and distance computations can be done with respect to the Minkowski sum of the original geometry and a piece of convex geometry. The Toolkit takes as raw model input (1) collections of convex polygons that form the boundaries of models and (2) convex polyhedra, cones, cylinders, and discs that are models and model components. Configurations are given in terms of homogeneous transforms. A simple OpenGL-based system for displaying and animating the geometric objects is included in the implementation. This version, 2.5 Beta, incorporates feature additions and enhancements, improvements in algorithms, improved robustness, bug fixes and cleaned-up source code, better compliance with standards and recent programming convention, changes to the build process for the software, support for more recent hardware and software platforms, and improvements to documentation and source-code comments.« less

  7. POSet Ontology Categorizer

    Energy Science and Technology Software Center (OSTI)

    2005-03-01

    POSet Ontology Categorizer (POSOC) V1.0 The POSet Ontology Categorizer (POSOC) software package provides tools for creating and mining of poset-structured ontologies, such as the Gene Ontology (GO). Given a list of weighted query items (ex.genes,proteins, and/or phrases) and one or more focus nodes, POSOC determines the ordered set of GO nodes that summarize the query, based on selections of a scoring function, pseudo-distance measure, specificity level, and cluster determination. Pseudo-distance measures provided are minimum chainmore » length, maximum chain length, average of extreme chain lengths, and average of all chain lengths. A low specificity level, such as -1 or 0, results in a general set of clusters. Increasing the specificity results in more specific results in more specific and lighter clusters. POSOC cluster results can be compared agaist known results by calculations of precision, recall, and f-score for graph neighborhood relationships. This tool has been used in understanding the function of a set of genes, finding similar genes, and annotating new proteins. The POSOC software consists of a set of Java interfaces, classes, and programs that run on Linux or Windows platforms. It incorporates graph classes from OpenJGraph (openjgraph.sourceforge.net).« less

  8. Geometric Algorithms for Modeling, Motion, and Animation (GAMMA): Collision Detection Videos from the University of North Carolina GAMMA Research Group

    DOE Data Explorer [Office of Scientific and Technical Information (OSTI)]

    Collision detection has been a fundamental problem in computer animation, physically-based modeling, geometric modeling, and robotics. In these applications, interactions between moving objects are modeled by dynamic constraints and contact analysis. The objects' motions are constrained by various interactions, including collisions. A virtual environment, like a walkthrough, creates a computer-generated world, filled with virtual objects. Such an environment should give the user a feeling of presence, which includes making the images of both the user and the surrounding objects feel solid. For example, the objects should not pass through each other, and things should move as expected when pushed, pulled or grasped. Such actions require accurate collision detection, if they are to achieve any degree of realism. However, there may be hundreds, even thousands of objects in the virtual world, so a naive algorithm could take a long time just to check for possible collisions as the user moves. This is not acceptable for virtual environments, where the issues of interactivity impose fundamental constraints on the system. A fast and interactive collision detection algorithm is a fundamental component of a complex virtual environment. Physically based modeling simulations depend highly on the physical interaction between objects in a scene. Complex physics engines require fast, accurate, and robust proximity queries to maintain a realistic simulation at interactive rates. We couple our proximity query research with physically based modeling to ensure that our packages provide the capabilities of today's physics engines.[Copied from http://www.cs.unc.edu/~geom/collide/index.shtml

  9. Hanford Environmental Information System (HEIS). Volume 1, User`s guide

    SciTech Connect (OSTI)

    Not Available

    1994-01-14

    The Hanford Environmental Information System (HEIS) is a consolidated set of automated resources that effectively manage the data gathered during environmental monitoring and restoration of the Hanford Site. HEIS includes an integrated database that provides consistent and current data to all users and promotes sharing of data by the entire user community. HEIS is an information system with an inclusive database. Although the database is the nucleus of the system, HEIS also provides user access software: query-by-form data entry, extraction, and browsing facilities; menu-driven reporting facilities; an ad hoc query facility; and a geographic information system (GIS). These features, with the exception of the GIS, are described in this manual set. Because HEIS contains data from the entire Hanford Site, many varieties of data are included and have.been divided into subject areas. Related subject areas comprise several volumes of the manual set. The manual set includes a data dictionary that lists all of the fields in the HEIS database, with their definitions and a cross reference of their locations in the database; definitions of data qualifiers for analytical results; and a mapping between the HEIS software functions and the keyboard keys for each of the supported terminals or terminal emulators.

  10. United States Transuranium and Uranium Registries. Annual report, February 1, 2003 - January 31, 2004

    SciTech Connect (OSTI)

    Alldredge, J. R.; Brumbaugh, T. L.; Ehrhart, Susan M.; Elliston, J. T.; Filipy, R. E.; James, A. C.; Pham, M. V.; Wood, T. G.; Sasser, L. B.

    2004-01-31

    This year was my fourteenth year with the U. S. Transuranium and Uranium Registries (USTUR). How time flies! Since I became the director of the program five years ago, one of my primary goals was to increase the usefulness of the large USTUR database that consists of six tables containing personal information, medical histories, radiation exposure histories, causes of death, and the results of radiochemical analysis of organ samples collected at autopsy. It is essential that a query of one or more of these tables by USTUR researchers or by collaborating researchers provides complete and reliable information. Also, some of the tables (those without personal identifiers) are destined to appear on the USTUR website for the use of the scientific community. I am pleased to report that most of the data in the database have now been verified and formatted for easy query. It is important to note that no data were discarded; copies of the original tables were retained and the original paper documents are still available for further verification of values as needed.

  11. Compressing bitmap indexes for faster search operations

    SciTech Connect (OSTI)

    Wu, Kesheng; Otoo, Ekow J.; Shoshani, Arie

    2002-04-25

    In this paper, we study the effects of compression on bitmap indexes. The main operations on the bitmaps during query processing are bitwise logical operations such as AND, OR, NOT, etc. Using the general purpose compression schemes, such as gzip, the logical operations on the compressed bitmaps are much slower than on the uncompressed bitmaps. Specialized compression schemes, like the byte-aligned bitmap code(BBC), are usually faster in performing logical operations than the general purpose schemes, but in many cases they are still orders of magnitude slower than the uncompressed scheme. To make the compressed bitmap indexes operate more efficiently, we designed a CPU-friendly scheme which we refer to as the word-aligned hybrid code (WAH). Tests on both synthetic and real application data show that the new scheme significantly outperforms well-known compression schemes at a modest increase in storage space. Compared to BBC, a scheme well-known for its operational efficiency, WAH performs logical operations about 12 times faster and uses only 60 percent more space. Compared to the uncompressed scheme, in most test cases WAH is faster while still using less space. We further verified with additional tests that the improvement in logical operation speed translates to similar improvement in query processing speed.

  12. Multi-Level Bitmap Indexes for Flash Memory Storage

    SciTech Connect (OSTI)

    Wu, Kesheng; Madduri, Kamesh; Canon, Shane

    2010-07-23

    Due to their low access latency, high read speed, and power-efficient operation, flash memory storage devices are rapidly emerging as an attractive alternative to traditional magnetic storage devices. However, tests show that the most efficient indexing methods are not able to take advantage of the flash memory storage devices. In this paper, we present a set of multi-level bitmap indexes that can effectively take advantage of flash storage devices. These indexing methods use coarsely binned indexes to answer queries approximately, and then use finely binned indexes to refine the answers. Our new methods read significantly lower volumes of data at the expense of an increased disk access count, thus taking full advantage of the improved read speed and low access latency of flash devices. To demonstrate the advantage of these new indexes, we measure their performance on a number of storage systems using a standard data warehousing benchmark called the Set Query Benchmark. We observe that multi-level strategies on flash drives are up to 3 times faster than traditional indexing strategies on magnetic disk drives.

  13. Coherent Image Layout using an Adaptive Visual Vocabulary

    SciTech Connect (OSTI)

    Dillard, Scott E.; Henry, Michael J.; Bohn, Shawn J.; Gosink, Luke J.

    2013-03-06

    When querying a huge image database containing millions of images, the result of the query may still contain many thousands of images that need to be presented to the user. We consider the problem of arranging such a large set of images into a visually coherent layout, one that places similar images next to each other. Image similarity is determined using a bag-of-features model, and the layout is constructed from a hierarchical clustering of the image set by mapping an in-order traversal of the hierarchy tree into a space-filling curve. This layout method provides strong locality guarantees so we are able to quantitatively evaluate performance using standard image retrieval benchmarks. Performance of the bag-of-features method is best when the vocabulary is learned on the image set being clustered. Because learning a large, discriminative vocabulary is a computationally demanding task, we present a novel method for efficiently adapting a generic visual vocabulary to a particular dataset. We evaluate our clustering and vocabulary adaptation methods on a variety of image datasets and show that adapting a generic vocabulary to a particular set of images improves performance on both hierarchical clustering and image retrieval tasks.

  14. 'Big Data' Collaboration: Exploring, Recording and Sharing Enterprise Knowledge

    SciTech Connect (OSTI)

    Sukumar, Sreenivas R; Ferrell, Regina Kay

    2013-01-01

    As data sources and data size proliferate, knowledge discovery from "Big Data" is starting to pose several challenges. In this paper, we address a specific challenge in the practice of enterprise knowledge management while extracting actionable nuggets from diverse data sources of seemingly-related information. In particular, we address the challenge of archiving knowledge gained through collaboration, dissemination and visualization as part of the data analysis, inference and decision-making lifecycle. We motivate the implementation of an enterprise data-discovery and knowledge recorder tool, called SEEKER based on real world case-study. We demonstrate SEEKER capturing schema and data-element relationships, tracking the data elements of value based on the queries and the analytical artifacts that are being created by analysts as they use the data. We show how the tool serves as digital record of institutional domain knowledge and a documentation for the evolution of data elements, queries and schemas over time. As a knowledge management service, a tool like SEEKER saves enterprise resources and time by avoiding analytic silos, expediting the process of multi-source data integration and intelligently documenting discoveries from fellow analysts.

  15. Security Profile Inspector for UNIX Systems

    Energy Science and Technology Software Center (OSTI)

    1995-04-01

    SPI/U3.2 consists of five tools used to assess and report the security posture of computers running the UNIX operating system. The tools are: Access Control Test: A rule-based system which identifies sequential dependencies in UNIX access controls. Binary Authentication Tool: Evaluates the release status of system binaries by comparing a crypto-checksum to provide table entries. Change Detection Tool: Maintains and applies a snapshot of critical system files and attributes for purposes of change detection. Configurationmore » Query Language: Accepts CQL-based scripts (provided) to evaluate queries over the status of system files, configuration of services and many other elements of UNIX system security. Password Security Inspector: Tests for weak or aged passwords. The tools are packaged with a forms-based user interface providing on-line context-sensistive help, job scheduling, parameter management and output report management utilities. Tools may be run independent of the UI.« less

  16. Predicting and Detecting Emerging Cyberattack Patterns Using StreamWorks

    SciTech Connect (OSTI)

    Chin, George; Choudhury, Sutanay; Feo, John T.; Holder, Larry

    2014-06-30

    The number and sophistication of cyberattacks on industries and governments have dramatically grown in recent years. To counter this movement, new advanced tools and techniques are needed to detect cyberattacks in their early stages such that defensive actions may be taken to avert or mitigate potential damage. From a cybersecurity analysis perspective, detecting cyberattacks may be cast as a problem of identifying patterns in computer network traffic. Logically and intuitively, these patterns may take on the form of a directed graph that conveys how an attack or intrusion propagates through the computers of a network. Such cyberattack graphs could provide cybersecurity analysts with powerful conceptual representations that are natural to express and analyze. We have been researching and developing graph-centric approaches and algorithms for dynamic cyberattack detection. The advanced dynamic graph algorithms we are developing will be packaged into a streaming network analysis framework known as StreamWorks. With StreamWorks, a scientist or analyst may detect and identify precursor events and patterns as they emerge in complex networks. This analysis framework is intended to be used in a dynamic environment where network data is streamed in and is appended to a large-scale dynamic graph. Specific graphical query patterns are decomposed and collected into a graph query library. The individual decomposed subpatterns in the library are continuously and efficiently matched against the dynamic graph as it evolves to identify and detect early, partial subgraph patterns. The scalable emerging subgraph pattern algorithms will match on both structural and semantic network properties.

  17. Storing files in a parallel computing system using list-based index to identify replica files

    DOE Patents [OSTI]

    Faibish, Sorin; Bent, John M.; Tzelnic, Percy; Zhang, Zhenhua; Grider, Gary

    2015-07-21

    Improved techniques are provided for storing files in a parallel computing system using a list-based index to identify file replicas. A file and at least one replica of the file are stored in one or more storage nodes of the parallel computing system. An index for the file comprises at least one list comprising a pointer to a storage location of the file and a storage location of the at least one replica of the file. The file comprises one or more of a complete file and one or more sub-files. The index may also comprise a checksum value for one or more of the file and the replica(s) of the file. The checksum value can be evaluated to validate the file and/or the file replica(s). A query can be processed using the list.

  18. Methods for modeling impact-induced reactivity changes in small reactors.

    SciTech Connect (OSTI)

    Tallman, Tyler N.; Radel, Tracy E.; Smith, Jeffrey A.; Villa, Daniel L.; Smith, Brandon M.; Radel, Ross F.; Lipinski, Ronald J.; Wilson, Paul Philip Hood

    2010-10-01

    This paper describes techniques for determining impact deformation and the subsequent reactivity change for a space reactor impacting the ground following a potential launch accident or for large fuel bundles in a shipping container following an accident. This technique could be used to determine the margin of subcriticality for such potential accidents. Specifically, the approach couples a finite element continuum mechanics model (Pronto3D or Presto) with a neutronics code (MCNP). DAGMC, developed at the University of Wisconsin-Madison, is used to enable MCNP geometric queries to be performed using Pronto3D output. This paper summarizes what has been done historically for reactor launch analysis, describes the impact criticality analysis methodology, and presents preliminary results using representative reactor designs.

  19. Materials Databases Infrastructure Constructed by First Principles Calculations: A Review

    SciTech Connect (OSTI)

    Lin, Lianshan

    2015-10-13

    The First Principles calculations, especially the calculation based on High-Throughput Density Functional Theory, have been widely accepted as the major tools in atom scale materials design. The emerging super computers, along with the powerful First Principles calculations, have accumulated hundreds of thousands of crystal and compound records. The exponential growing of computational materials information urges the development of the materials databases, which not only provide unlimited storage for the daily increasing data, but still keep the efficiency in data storage, management, query, presentation and manipulation. This review covers the most cutting edge materials databases in materials design, and their hot applications such as in fuel cells. By comparing the advantages and drawbacks of these high-throughput First Principles materials databases, the optimized computational framework can be identified to fit the needs of fuel cell applications. The further development of high-throughput DFT materials database, which in essence accelerates the materials innovation, is discussed in the summary as well.

  20. LDRD final report : first application of geospatial semantic graphs to SAR image data.

    SciTech Connect (OSTI)

    Brost, Randolph C.; McLendon, William Clarence,

    2013-01-01

    Modeling geospatial information with semantic graphs enables search for sites of interest based on relationships between features, without requiring strong a priori models of feature shape or other intrinsic properties. Geospatial semantic graphs can be constructed from raw sensor data with suitable preprocessing to obtain a discretized representation. This report describes initial work toward extending geospatial semantic graphs to include temporal information, and initial results applying semantic graph techniques to SAR image data. We describe an efficient graph structure that includes geospatial and temporal information, which is designed to support simultaneous spatial and temporal search queries. We also report a preliminary implementation of feature recognition, semantic graph modeling, and graph search based on input SAR data. The report concludes with lessons learned and suggestions for future improvements.

  1. CPTAC Assay Portal: a repository of targeted proteomic assays

    SciTech Connect (OSTI)

    Whiteaker, Jeffrey R.; Halusa, Goran; Hoofnagle, Andrew N.; Sharma, Vagisha; MacLean, Brendan; Yan, Ping; Wrobel, John; Kennedy, Jacob; Mani, DR; Zimmerman, Lisa J.; Meyer, Matthew R.; Mesri, Mehdi; Rodriguez, Henry; Abbateillo, Susan E.; Boja, Emily; Carr, Steven A.; Chan, Daniel W.; Chen, Xian; Chen, Jing; Davies, Sherri; Ellis, Matthew; Fenyo, David; Hiltket, Tara; Ketchum, Karen; Kinsinger, Christopher; Kuhn, Eric; Liebler, Daniel; Lin, De; Liu, Tao; Loss, Michael; MacCoss, Michael; Qian, Weijun; Rivers, Robert; Rodland, Karin D.; Ruggles, Kelly; Scott, Mitchell; Smith, Richard D.; Thomas, Stefani N.; Townsend, Reid; Whiteley, Gordon; Wu, Chaochao; Zhang, Hui; Zhang, Zhen; Paulovich, Amanda G.

    2014-06-27

    To address these issues, the Clinical Proteomic Tumor Analysis Consortium (CPTAC) of the National Cancer Institute (NCI) has launched an Assay Portal (http://assays.cancer.gov) to serve as a public repository of well-characterized quantitative, MS-based, targeted proteomic assays. The purpose of the CPTAC Assay Portal is to facilitate widespread adoption of targeted MS assays by disseminating SOPs, reagents, and assay characterization data for highly characterized assays. A primary aim of the NCI-supported portal is to bring together clinicians or biologists and analytical chemists to answer hypothesis-driven questions using targeted, MS-based assays. Assay content is easily accessed through queries and filters, enabling investigators to find assays to proteins relevant to their areas of interest. Detailed characterization data are available for each assay, enabling researchers to evaluate assay performance prior to launching the assay in their own laboratory.

  2. Unified Parallel Software

    Energy Science and Technology Software Center (OSTI)

    2003-12-01

    UPS (Unified Paralled Software is a collection of software tools libraries, scripts, executables) that assist in parallel programming. This consists of: o libups.a C/Fortran callable routines for message passing (utilities written on top of MPI) and file IO (utilities written on top of HDF). o libuserd-HDF.so EnSight user-defined reader for visualizing data files written with UPS File IO. o ups_libuserd_query, ups_libuserd_prep.pl, ups_libuserd_script.pl Executables/scripts to get information from data files and to simplify the use ofmore » EnSight on those data files. o ups_io_rm/ups_io_cp Manipulate data files written with UPS File IO These tools are portable to a wide variety of Unix platforms.« less

  3. High-performance Computing Applied to Semantic Databases

    SciTech Connect (OSTI)

    Goodman, Eric L.; Jimenez, Edward; Mizell, David W.; al-Saffar, Sinan; Adolf, Robert D.; Haglin, David J.

    2011-06-02

    To-date, the application of high-performance computing resources to Semantic Web data has largely focused on commodity hardware and distributed memory platforms. In this paper we make the case that more specialized hardware can offer superior scaling and close to an order of magnitude improvement in performance. In particular we examine the Cray XMT. Its key characteristics, a large, global shared-memory, and processors with a memory-latency tolerant design, offer an environment conducive to programming for the Semantic Web and have engendered results that far surpass current state of the art. We examine three fundamental pieces requisite for a fully functioning semantic database: dictionary encoding, RDFS inference, and query processing. We show scaling up to 512 processors (the largest configuration we had available), and the ability to process 20 billion triples completely in-memory.

  4. Frontier: High Performance Database Access Using Standard Web Components in a Scalable Multi-Tier Architecture

    SciTech Connect (OSTI)

    Kosyakov, S.; Kowalkowski, J.; Litvintsev, D.; Lueking, L.; Paterno, M.; White, S.P.; Autio, Lauri; Blumenfeld, B.; Maksimovic, P.; Mathis, M.; /Johns Hopkins U.

    2004-09-01

    A high performance system has been assembled using standard web components to deliver database information to a large number of broadly distributed clients. The CDF Experiment at Fermilab is establishing processing centers around the world imposing a high demand on their database repository. For delivering read-only data, such as calibrations, trigger information, and run conditions data, we have abstracted the interface that clients use to retrieve data objects. A middle tier is deployed that translates client requests into database specific queries and returns the data to the client as XML datagrams. The database connection management, request translation, and data encoding are accomplished in servlets running under Tomcat. Squid Proxy caching layers are deployed near the Tomcat servers, as well as close to the clients, to significantly reduce the load on the database and provide a scalable deployment model. Details the system's construction and use are presented, including its architecture, design, interfaces, administration, performance measurements, and deployment plan.

  5. High-performance computing applied to semantic databases.

    SciTech Connect (OSTI)

    al-Saffar, Sinan; Jimenez, Edward Steven, Jr.; Adolf, Robert; Haglin, David; Goodman, Eric L.; Mizell, David

    2010-12-01

    To-date, the application of high-performance computing resources to Semantic Web data has largely focused on commodity hardware and distributed memory platforms. In this paper we make the case that more specialized hardware can offer superior scaling and close to an order of magnitude improvement in performance. In particular we examine the Cray XMT. Its key characteristics, a large, global shared-memory, and processors with a memory-latency tolerant design, offer an environment conducive to programming for the Semantic Web and have engendered results that far surpass current state of the art. We examine three fundamental pieces requisite for a fully functioning semantic database: dictionary encoding, RDFS inference, and query processing. We show scaling up to 512 processors (the largest configuration we had available), and the ability to process 20 billion triples completely in-memory.

  6. Tensor Algebra Library for NVidia Graphics Processing Units

    Energy Science and Technology Software Center (OSTI)

    2015-03-16

    This is a general purpose math library implementing basic tensor algebra operations on NVidia GPU accelerators. This software is a tensor algebra library that can perform basic tensor algebra operations, including tensor contractions, tensor products, tensor additions, etc., on NVidia GPU accelerators, asynchronously with respect to the CPU host. It supports a simultaneous use of multiple NVidia GPUs. Each asynchronous API function returns a handle which can later be used for querying the completion ofmore » the corresponding tensor algebra operation on a specific GPU. The tensors participating in a particular tensor operation are assumed to be stored in local RAM of a node or GPU RAM. The main research area where this library can be utilized is the quantum many-body theory (e.g., in electronic structure theory).« less

  7. Simrank: Rapid and sensitive general-purpose k-mer search tool

    SciTech Connect (OSTI)

    DeSantis, T.Z.; Keller, K.; Karaoz, U.; Alekseyenko, A.V; Singh, N.N.S.; Brodie, E.L; Pei, Z.; Andersen, G.L; Larsen, N.

    2011-04-01

    Terabyte-scale collections of string-encoded data are expected from consortia efforts such as the Human Microbiome Project (http://nihroadmap.nih.gov/hmp). Intra- and inter-project data similarity searches are enabled by rapid k-mer matching strategies. Software applications for sequence database partitioning, guide tree estimation, molecular classification and alignment acceleration have benefited from embedded k-mer searches as sub-routines. However, a rapid, general-purpose, open-source, flexible, stand-alone k-mer tool has not been available. Here we present a stand-alone utility, Simrank, which allows users to rapidly identify database strings the most similar to query strings. Performance testing of Simrank and related tools against DNA, RNA, protein and human-languages found Simrank 10X to 928X faster depending on the dataset. Simrank provides molecular ecologists with a high-throughput, open source choice for comparing large sequence sets to find similarity.

  8. NetState

    Energy Science and Technology Software Center (OSTI)

    2005-09-01

    NetState is a distributed network monitoring system. It uses passive sensors to develop status information on a target network. Two major features provided by NetState are version and port tracking. Version tracking maintains information about software and operating systems versions. Port tracking identifies information about active TOP and UDP ports. Multiple NetState sniffers can be deployed, one at each entry point of the target network. The sniffers monitor network traffic, then send the information tomore » the NetState server. The information is stored in centralized database which can then be accessed via standard SQL database queries or this web-based GUI, for further analysis and display.« less

  9. Scenario driven data modelling: a method for integrating diverse sources of data and data streams

    SciTech Connect (OSTI)

    Brettin, Thomas S.; Cottingham, Robert W.; Griffith, Shelton D.; Quest, Daniel J.

    2015-09-08

    A system and method of integrating diverse sources of data and data streams is presented. The method can include selecting a scenario based on a topic, creating a multi-relational directed graph based on the scenario, identifying and converting resources in accordance with the scenario and updating the multi-directed graph based on the resources, identifying data feeds in accordance with the scenario and updating the multi-directed graph based on the data feeds, identifying analytical routines in accordance with the scenario and updating the multi-directed graph using the analytical routines and identifying data outputs in accordance with the scenario and defining queries to produce the data outputs from the multi-directed graph.

  10. Geospatial Toolkit

    SciTech Connect (OSTI)

    2010-10-14

    The Geospatial Toolkit is an NREL-developed map-based software application that integrates resource data and other geographic information systems (GIS) data for integrated resource assessment. The non-resource, country-specific data for each toolkit comes from a variety of agencies within each country as well as from global datasets. Originally developed in 2005, the Geospatial Toolkit was completely redesigned and re-released in November 2010 to provide a more modern, easier-to-use interface with considerably faster analytical querying capabilities. The revised version of the Geospatial Toolkit has been released for all original toolkit countries/regions and each software package is made available on NREL's website,

  11. Geospatial Toolkit

    Energy Science and Technology Software Center (OSTI)

    2010-10-14

    The Geospatial Toolkit is an NREL-developed map-based software application that integrates resource data and other geographic information systems (GIS) data for integrated resource assessment. The non-resource, country-specific data for each toolkit comes from a variety of agencies within each country as well as from global datasets. Originally developed in 2005, the Geospatial Toolkit was completely redesigned and re-released in November 2010 to provide a more modern, easier-to-use interface with considerably faster analytical querying capabilities. Themore » revised version of the Geospatial Toolkit has been released for all original toolkit countries/regions and each software package is made available on NREL's website,« less

  12. Battery Life Estimator (BLE) Data Analysis Software v. 1.2

    SciTech Connect (OSTI)

    Thomas, Edward; Bloom, Ira; Battaglia, Vincent; & Christopherson, Jon

    2010-02-24

    The purpose of this software is estimate the useable life of rechargeable batteries (e.g., lithium-ion). The software employs a generalized statistical approach to model cell data in the context of accelerated aging experiments. The cell performance is modeled in two parts. The first part consists of a deterministic degradation model which models the average cell behavior. The second part relates to the statistical variation in performance of the cells (error model). Experimental data from an accelerated aging experiment will be input from an Excel worksheet. The software will then query the user for a specific model form (within the generalized model framework). Model parameters will be estimated by the software using various statistical methodologies. Average cell life will be predicted using the estimated model parameters. The uncertainty in the estimated cell life will also be computed using bootstrap simulations. This software can be used in several modes: 1) fit only, 2) fit and simulation, and 3) simulation only

  13. geryon v. 0.1

    Energy Science and Technology Software Center (OSTI)

    2010-04-28

    Geryon is intended to be a simple library for managing the CUDA Runtime, CUDA Driver, and OpenCL APIs with a consistent interface * Change from one API to another by simply changing the namespace * Use multiple APIs in the same code * Lightweight (only include files no build required) * Manage device query and selection * Simple vector and matrix containers * Simple routines for data copy and type casting * Simple routines formore » data I/O * Simple classes for managing device timing * Simple classes for managing kernel compilation and execution The primary application is to facilitate writing a single code that can be compiled using the CUDA Runtime API, the CUDA Driver API, or OpenCL.« less

  14. An organizational survey of the Pittsburgh Energy Technology Center

    SciTech Connect (OSTI)

    Stock, D.A.; Shurberg, D.A.; Haber, S.B.

    1991-09-01

    An Organizational Survey (OS) was administrated at the Pittsburgh Energy Technology Center (PETC) that queried employees on the subjects of organizational culture, various aspects of communications, employee commitment, work group cohesion, coordination of work, environmental, safety, and health concerns, hazardous nature of work, safety and overall job satisfaction. The purpose of the OS is to measure in a quantitative and objective way the notion of culture''; that is, the values attitudes, and beliefs of the individuals working within the organization. In addition, through the OS, a broad sample of individuals can be reached that would probably not be interviewed or observed during the course of a typical assessment. The OS also provides a descriptive profile of the organization at one point in time that can then be compared to a profile taken at a different point in time to assess changes in the culture of the organization.

  15. An organizational survey of the Pittsburgh Energy Technology Center

    SciTech Connect (OSTI)

    Stock, D.A.; Shurberg, D.A.; Haber, S.B.

    1991-09-01

    An Organizational Survey (OS) was administrated at the Pittsburgh Energy Technology Center (PETC) that queried employees on the subjects of organizational culture, various aspects of communications, employee commitment, work group cohesion, coordination of work, environmental, safety, and health concerns, hazardous nature of work, safety and overall job satisfaction. The purpose of the OS is to measure in a quantitative and objective way the notion of ``culture``; that is, the values attitudes, and beliefs of the individuals working within the organization. In addition, through the OS, a broad sample of individuals can be reached that would probably not be interviewed or observed during the course of a typical assessment. The OS also provides a descriptive profile of the organization at one point in time that can then be compared to a profile taken at a different point in time to assess changes in the culture of the organization.

  16. Needle Federated Search Engine

    Energy Science and Technology Software Center (OSTI)

    2009-12-01

    The Idaho National Laboratory (INL) has combined a number of technologies, tools, and resources to accomplish a new means of federating search results. The resulting product is a search engine called Needle, an open-source-based tool that the INL uses internally for researching across a wide variety of information repositories. Needle has a flexible search interface that allows end users to point at any available data source. A user can select multiple sources such as commercialmore » databases (Web of Science, Engineering Index), external resources (WorldCat, Google Scholar), and internal corporate resources (email, document management system, library collections) in a single interface with one search query. In the future, INL hopes to offer this open-source engine to the public. This session will outline the development processes for making Needle™s search interface and simplifying the federation of internal and external data sources.« less

  17. RTDB: A memory resident real-time object database

    SciTech Connect (OSTI)

    Jerzy M. Nogiec; Eugene Desavouret

    2003-06-04

    RTDB is a fast, memory-resident object database with built-in support for distribution. It constitutes an attractive alternative for architecting real-time solutions with multiple, possibly distributed, processes or agents sharing data. RTDB offers both direct and navigational access to stored objects, with local and remote random access by object identifiers, and immediate direct access via object indices. The database supports transparent access to objects stored in multiple collaborating dispersed databases and includes a built-in cache mechanism that allows for keeping local copies of remote objects, with specifiable invalidation deadlines. Additional features of RTDB include a trigger mechanism on objects that allows for issuing events or activating handlers when objects are accessed or modified and a very fast, attribute based search/query mechanism. The overall architecture and application of RTDB in a control and monitoring system is presented.

  18. BitPredator: A Discovery Algorithm for BitTorrent Initial Seeders and Peers

    SciTech Connect (OSTI)

    Borges, Raymond; Patton, Robert M; Kettani, Houssain; Masalmah, Yahya

    2011-01-01

    There is a large amount of illegal content being replicated through peer-to-peer (P2P) networks where BitTorrent is dominant; therefore, a framework to profile and police it is needed. The goal of this work is to explore the behavior of initial seeds and highly active peers to develop techniques to correctly identify them. We intend to establish a new methodology and software framework for profiling BitTorrent peers. This involves three steps: crawling torrent indexers for keywords in recently added torrents using Really Simple Syndication protocol (RSS), querying torrent trackers for peer list data and verifying Internet Protocol (IP) addresses from peer lists. We verify IPs using active monitoring methods. Peer behavior is evaluated and modeled using bitfield message responses. We also design a tool to profile worldwide file distribution by mapping IP-to-geolocation and linking to WHOIS server information in Google Earth.

  19. A semantic characterization of an algorithm for estimating others` beliefs from observation

    SciTech Connect (OSTI)

    Isozaki, Hideki; Katsuno, Hirofumi

    1996-12-31

    Human beings often estimate others beliefs and intentions when they interact with others. Estimation of others beliefs will be useful also in controlling the behavior and utterances of artificial agents, especially when lines of communication are unstable or slow. But, devising such estimation algorithms and background theories for the algorithms is difficult, because of many factors affecting one`s belief. We have proposed an algorithm that estimates others beliefs from observation in the changing world. Experimental results show that this algorithm returns natural answers to various queries. However, the algorithm is only heuristic, and how the algorithm deals with beliefs and their changes is not entirely clear. We propose certain semantics based on a nonstandard structure for modal logic. By using these semantics, we shed light on a logical meaning of the belief estimation that the algorithm deals with. We also discuss how the semantics and the algorithm can be generalized.

  20. Supporting Mutual Understanding in a Visual Dialogue Between Analyst and Computer

    SciTech Connect (OSTI)

    Chappell, Alan R.; Cowell, Andrew J.; Thurman, David A.; Thomson, Judi R.

    2004-09-20

    The Knowledge Associates for Novel Intelligence (KANI) project is developing a system of automated associates to actively support and participate in the information analysis task. The primary goal of KANI is to use automatically extracted information in a reasoning system that draws on the strengths of both a human analyst and automated reasoning. The interface between the two agents is a key element in achieving this goal. The KANI interface seeks to support a visual dialogue with mixed-initiative manipulation of information and reasoning components. To be successful, the interface must achieve mutual understanding between the analyst and KANI of the others actions. Toward this mutual understanding, KANI allows the analyst to work at multiple levels of abstraction over the reasoning process, links the information presented across these levels to make use of interaction context, and provides querying facilities to allow exploration and explanation.

  1. An integrated computer modeling environment for regional land use, air quality, and transportation planning

    SciTech Connect (OSTI)

    Hanley, C.J.; Marshall, N.L.

    1997-04-01

    The Land Use, Air Quality, and Transportation Integrated Modeling Environment (LATIME) represents an integrated approach to computer modeling and simulation of land use allocation, travel demand, and mobile source emissions for the Albuquerque, New Mexico, area. This environment provides predictive capability combined with a graphical and geographical interface. The graphical interface shows the causal relationships between data and policy scenarios and supports alternative model formulations. Scenarios are launched from within a Geographic Information System (GIS), and data produced by each model component at each time step within a simulation is stored in the GIS. A menu-driven query system is utilized to review link-based results and regional and area-wide results. These results can also be compared across time or between alternative land use scenarios. Using this environment, policies can be developed and implemented based on comparative analysis, rather than on single-step future projections. 16 refs., 3 figs., 2 tabs.

  2. Tensor Algebra Library for NVidia Graphics Processing Units

    SciTech Connect (OSTI)

    Liakh, Dmitry

    2015-03-16

    This is a general purpose math library implementing basic tensor algebra operations on NVidia GPU accelerators. This software is a tensor algebra library that can perform basic tensor algebra operations, including tensor contractions, tensor products, tensor additions, etc., on NVidia GPU accelerators, asynchronously with respect to the CPU host. It supports a simultaneous use of multiple NVidia GPUs. Each asynchronous API function returns a handle which can later be used for querying the completion of the corresponding tensor algebra operation on a specific GPU. The tensors participating in a particular tensor operation are assumed to be stored in local RAM of a node or GPU RAM. The main research area where this library can be utilized is the quantum many-body theory (e.g., in electronic structure theory).

  3. MeSh ToolKit v1.2

    Energy Science and Technology Software Center (OSTI)

    2004-05-15

    MSTK or Mesh Toolkit is a mesh framework that allows users to represent, manipulate and query unstructured 3D arbitrary topology meshes in a general manner without the need to code their own data structures. MSTK is a flexible framework in that is allows (or will eventually allow) a wide variety of underlying representations for the mesh while maintaining a common interface. It will allow users to choose from different mesh representations either at initialization ormore » during the program execution so that the optimal data structures are used for the particular algorithm. The interaction of users and applications with MSTK is through a functional interface that acts as through the mesh always contains vertices, edges, faces and regions and maintains connectivity between all these entities.« less

  4. FastBit: Interactively Searching Massive Data

    SciTech Connect (OSTI)

    Wu, Kesheng; Ahern, Sean; Bethel, E. Wes; Chen, Jacqueline; Childs, Hank; Cormier-Michel, Estelle; Geddes, Cameron; Gu, Junmin; Hagen, Hans; Hamann, Bernd; Koegler, Wendy; Lauret, Jerome; Meredith, Jeremy; Messmer, Peter; Otoo, Ekow; Perevoztchikov, Victor; Poskanzer, Arthur; Prabhat,; Rubel, Oliver; Shoshani, Arie; Sim, Alexander; Stockinger, Kurt; Weber, Gunther; Zhang, Wei-Ming

    2009-06-23

    As scientific instruments and computer simulations produce more and more data, the task of locating the essential information to gain insight becomes increasingly difficult. FastBit is an efficient software tool to address this challenge. In this article, we present a summary of the key underlying technologies, namely bitmap compression, encoding, and binning. Together these techniques enable FastBit to answer structured (SQL) queries orders of magnitude faster than popular database systems. To illustrate how FastBit is used in applications, we present three examples involving a high-energy physics experiment, a combustion simulation, and an accelerator simulation. In each case, FastBit significantly reduces the response time and enables interactive exploration on terabytes of data.

  5. High Performance Multivariate Visual Data Exploration for Extremely Large Data

    SciTech Connect (OSTI)

    Rubel, Oliver; Wu, Kesheng; Childs, Hank; Meredith, Jeremy; Geddes, Cameron G.R.; Cormier-Michel, Estelle; Ahern, Sean; Weber, Gunther H.; Messmer, Peter; Hagen, Hans; Hamann, Bernd; Bethel, E. Wes; Prabhat,

    2008-08-22

    One of the central challenges in modern science is the need to quickly derive knowledge and understanding from large, complex collections of data. We present a new approach that deals with this challenge by combining and extending techniques from high performance visual data analysis and scientific data management. This approach is demonstrated within the context of gaining insight from complex, time-varying datasets produced by a laser wakefield accelerator simulation. Our approach leverages histogram-based parallel coordinates for both visual information display as well as a vehicle for guiding a data mining operation. Data extraction and subsetting are implemented with state-of-the-art index/query technology. This approach, while applied here to accelerator science, is generally applicable to a broad set of science applications, and is implemented in a production-quality visual data analysis infrastructure. We conduct a detailed performance analysis and demonstrate good scalability on a distributed memory Cray XT4 system.

  6. System and method for anomaly detection

    DOE Patents [OSTI]

    Scherrer, Chad

    2010-06-15

    A system and method for detecting one or more anomalies in a plurality of observations is provided. In one illustrative embodiment, the observations are real-time network observations collected from a stream of network traffic. The method includes performing a discrete decomposition of the observations, and introducing derived variables to increase storage and query efficiencies. A mathematical model, such as a conditional independence model, is then generated from the formatted data. The formatted data is also used to construct frequency tables which maintain an accurate count of specific variable occurrence as indicated by the model generation process. The formatted data is then applied to the mathematical model to generate scored data. The scored data is then analyzed to detect anomalies.

  7. Byna-NERSC-ASCR-2017.pptx

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Requirements f or S cien0fic D ata M anagement Suren B yna Scien,fic D ata M anagement G roup Computa,onal R esearch D ivision Lawrence B erkeley L ab NERSC ASCR Requirements for 2017 January 15, 2014 LBNL Projects * m1248 repo * Arie S hoshani, S uren B yna, A lex S im, J ohn W u * Searching s cien,fic d ata * FastBit a nd F astQuery * Scien,fic D ata S ervices ( SDS) f ramework * Transparent d ata r eorganiza,on f or b eQer d ata a ccess * Redirec,on o f d ata r ead c alls f or r eorganized d

  8. Solution NMR Structure of Hypothetical Protein CV_2116 Encoded by a Viral Prophage Element in Chromobacterium violaceum

    SciTech Connect (OSTI)

    Yang, Yunhuang; Ramelot, Theresa A.; Cort, John R.; Garcia, Maite; Yee, Adelinda; Arrowsmith, Cheryl H.; Kennedy, Michael A.

    2012-06-14

    CV{_}2116 from Chromobacterium violaceum is a small hypothetical protein of 82 amino acids. A PSI-BLAST search using the CV{_}2116 protein sequence as a query identified only two hits, both with amino acid sequence identities of less than 40%. After the CV{_}2116 gene was cloned into the p15TvLic expression plasmid and transformed into E. coli, the desired CV{_}2116 protein was expressed and purified. A high quality solution structure of CV{_}2116 was determined by NMR spectroscopy. The tertiary structure of CV{_}2116 adopts a novel alpha + beta fold containing two anti-parallel beta sheets and one alpha helix in the C-terminal end. CV{_}2116 does not belong to any known protein sequence families and no similar structures exist in the protein data bank. To date, no function of CV{_}2116 can be derived from either sequence or structural similarity searches.

  9. System of and method for transparent management of data objects in containers across distributed heterogenous resources

    DOE Patents [OSTI]

    Moore, Reagan W.; Rajasekar, Arcot; Wan, Michael Y.

    2007-09-11

    A system of and method for maintaining data objects in containers across a network of distributed heterogeneous resources in a manner which is transparent to a client. A client request pertaining to containers is resolved by querying meta data for the container, processing the request through one or more copies of the container maintained on the system, updating the meta data for the container to reflect any changes made to the container as a result processing the re quest, and, if a copy of the container has changed, changing the status of the copy to indicate dirty status or synchronizing the copy to one or more other copies that may be present on the system.

  10. System of and method for transparent management of data objects in containers across distributed heterogenous resources

    DOE Patents [OSTI]

    Moore, Reagan W.; Rajasekar, Arcot; Wan, Michael Y.

    2010-09-21

    A system of and method for maintaining data objects in containers across a network of distributed heterogeneous resources in a manner which is transparent to a client. A client request pertaining to containers is resolved by querying meta data for the container, processing the request through one or more copies of the container maintained on the system, updating the meta data for the container to reflect any changes made to the container as a result processing the request, and, if a copy of the container has changed, changing the status of the copy to indicate dirty status or synchronizing the copy to one or more other copies that may be present on the system.

  11. System of and method for transparent management of data objects in containers across distributed heterogenous resources

    DOE Patents [OSTI]

    Moore, Reagan W.; Rajasekar, Arcot; Wan, Michael Y.

    2004-01-13

    A system of and method for maintaining data objects in containers across a network of distributed heterogeneous resources in a manner which is transparent to a client. A client request pertaining to containers is resolved by querying meta data for the container, processing the request through one or more copies of the container maintained on the system, updating the meta data for the container to reflect any changes made to the container as a result processing the request, and, if a copy of the container has changed, changing the status of the copy to indicate dirty status or synchronizing the copy to one or more other copies that may be present on the system.

  12. Webinar: Demonstration of NREL’s BioEnergy Atlas Tools

    Broader source: Energy.gov [DOE]

    The National Renewable Energy Laboratory (NREL) will host a free webinar on December 16 demonstrating how to use the BioEnergy Atlas tools. The U.S. Department of Energy’s Bioenergy Technologies Office funded the BioEnergy Atlas tools, which include the BioFuels and BioPower Atlases. These tools are designed as first-pass visualization tools that allow users to view many bioenergy and related datasets in Google Maps. Users can query and download map data and view incentives and state energy data, as well as select an area on the map for estimated biofuels or biopower production potential. The webinar will review the data source and date of bioenergy data layers. The NREL team will show users how to view and download data behind the map, how to view state energy data and incentives, and how to view and edit potential biofuel or biopower production in a geographical location.

  13. Computer systems and methods for visualizStolte; Chris ing data

    DOE Patents [OSTI]

    Stolte, Chris; Hanrahan, Patrick

    2013-01-29

    A method for forming a visual plot using a hierarchical structure of a dataset. The dataset comprises a measure and a dimension. The dimension consists of a plurality of levels. The plurality of levels form a dimension hierarchy. The visual plot is constructed based on a specification. A first level from the plurality of levels is represented by a first component of the visual plot. A second level from the plurality of levels is represented by a second component of the visual plot. The dataset is queried to retrieve data in accordance with the specification. The data includes all or a portion of the dimension and all or a portion of the measure. The visual plot is populated with the retrieved data in accordance with the specification.

  14. Computer systems and methods for visualizing data

    DOE Patents [OSTI]

    Stolte, Chris; Hanrahan, Patrick

    2010-07-13

    A method for forming a visual plot using a hierarchical structure of a dataset. The dataset comprises a measure and a dimension. The dimension consists of a plurality of levels. The plurality of levels form a dimension hierarchy. The visual plot is constructed based on a specification. A first level from the plurality of levels is represented by a first component of the visual plot. A second level from the plurality of levels is represented by a second component of the visual plot. The dataset is queried to retrieve data in accordance with the specification. The data includes all or a portion of the dimension and all or a portion of the measure. The visual plot is populated with the retrieved data in accordance with the specification.

  15. Model Components of the Certification Framework for Geologic Carbon Sequestration Risk Assessment

    SciTech Connect (OSTI)

    Oldenburg, Curtis M.; Bryant, Steven L.; Nicot, Jean-Philippe; Kumar, Navanit; Zhang, Yingqi; Jordan, Preston; Pan, Lehua; Granvold, Patrick; Chow, Fotini K.

    2009-06-01

    We have developed a framework for assessing the leakage risk of geologic carbon sequestration sites. This framework, known as the Certification Framework (CF), emphasizes wells and faults as the primary potential leakage conduits. Vulnerable resources are grouped into compartments, and impacts due to leakage are quantified by the leakage flux or concentrations that could potentially occur in compartments under various scenarios. The CF utilizes several model components to simulate leakage scenarios. One model component is a catalog of results of reservoir simulations that can be queried to estimate plume travel distances and times, rather than requiring CF users to run new reservoir simulations for each case. Other model components developed for the CF and described here include fault characterization using fault-population statistics; fault connection probability using fuzzy rules; well-flow modeling with a drift-flux model implemented in TOUGH2; and atmospheric dense-gas dispersion using a mesoscale weather prediction code.

  16. NGNP Risk Management Database: A Model for Managing Risk

    SciTech Connect (OSTI)

    John Collins

    2009-09-01

    To facilitate the implementation of the Risk Management Plan, the Next Generation Nuclear Plant (NGNP) Project has developed and employed an analytical software tool called the NGNP Risk Management System (RMS). A relational database developed in Microsoft Access, the RMS provides conventional database utility including data maintenance, archiving, configuration control, and query ability. Additionally, the tools design provides a number of unique capabilities specifically designed to facilitate the development and execution of activities outlined in the Risk Management Plan. Specifically, the RMS provides the capability to establish the risk baseline, document and analyze the risk reduction plan, track the current risk reduction status, organize risks by reference configuration system, subsystem, and component (SSC) and Area, and increase the level of NGNP decision making.

  17. Event heap: a coordination infrastructure for dynamic heterogeneous application interactions in ubiquitous computing environments

    DOE Patents [OSTI]

    Johanson, Bradley E.; Fox, Armando; Winograd, Terry A.; Hanrahan, Patrick M.

    2010-04-20

    An efficient and adaptive middleware infrastructure called the Event Heap system dynamically coordinates application interactions and communications in a ubiquitous computing environment, e.g., an interactive workspace, having heterogeneous software applications running on various machines and devices across different platforms. Applications exchange events via the Event Heap. Each event is characterized by a set of unordered, named fields. Events are routed by matching certain attributes in the fields. The source and target versions of each field are automatically set when an event is posted or used as a template. The Event Heap system implements a unique combination of features, both intrinsic to tuplespaces and specific to the Event Heap, including content based addressing, support for routing patterns, standard routing fields, limited data persistence, query persistence/registration, transparent communication, self-description, flexible typing, logical/physical centralization, portable client API, at most once per source first-in-first-out ordering, and modular restartability.

  18. Thematic World Wide Web Visualization System

    Energy Science and Technology Software Center (OSTI)

    1996-10-10

    WebTheme is a system designed to facilitate world wide web information access and retrieval through visualization. It consists of two principal pieces, a WebTheme Server which allows users to enter in a query and automatocally harvest and process information of interest, and a WebTheme browser, which allows users to work with both Galaxies and Themescape visualizations of their data within a JAVA capable world wide web browser. WebTheme is an Internet solution, meaning that accessmore » to the server and the resulting visualizations can all be performed through the use of a WWW browser. This allows users to access and interact with SPIRE (Spatial Paradigm for Information Retrieval and Exploration) based visualizations through a web browser regardless of what computer platforms they are running on. WebTheme is specifically designed to create databases by harvesting and processing WWW home pages available on the Internet.« less

  19. Generative inspection process planner for integrated production

    SciTech Connect (OSTI)

    Brown, C.W. . Kansas City Div.); Gyorog, D.A. . Dept. of Mechanical Engineering)

    1990-04-01

    This work describes the design prototype development of a generative process planning system for dimensional inspection. The system, IPPEX (Inspection Process Planning EXpert), is a rule-based expert system for integrated production. Using as advanced product modeler, relational databases, and artificial intelligence techniques, IPPEX generates the process plan and part program for the dimensional inspection of products using CMMs. Through an application interface, the IPPEX system software accesses product definition from the product modeler. The modeler is a solid geometric modeler coupled with a dimension and tolerance modeler. Resource data regarding the machines, probes, and fixtures are queried from databases. IPPEX represents inspection process knowledge as production rules and incorporates an embedded inference engine to perform decision making. The IPPEX system, its functional architecture, system architecture, system approach, product modeling environment, inspection features, inspection knowledge, hierarchical planning strategy, user interface formats, and other fundamental issues related to inspection planning and part programming for CMMs are described. 27 refs., 16 figs., 4 tabs.

  20. Ensemble Data Analysis ENvironment (EDEN)

    Energy Science and Technology Software Center (OSTI)

    2012-08-01

    The EDEN toolkit facilitates exploratory data analysis and visualization of global climate model simulation datasets. EDEN provides an interactive graphical user interface (GUI) that helps the user visually construct dynamic queries of the characteristically large climate datasets using temporal ranges, variable selections, and geographic areas of interest. EDEN reads the selected data into a multivariate visualization panel which features an extended implementation of parallel coordinates plots as well as interactive scatterplots. The user can querymore » data in the visualization panel using mouse gestures to analyze different ranges of data. The visualization panel provides coordinated multiple views whereby selections made in one plot are propagated to the other plots.« less

  1. Multivariate Data EXplorer (MDX)

    Energy Science and Technology Software Center (OSTI)

    2012-08-01

    The MDX toolkit facilitates exploratory data analysis and visualization of multivariate datasets. MDX provides and interactive graphical user interface to load, explore, and modify multivariate datasets stored in tabular forms. MDX uses an extended version of the parallel coordinates plot and scatterplots to represent the data. The user can perform rapid visual queries using mouse gestures in the visualization panels to select rows or columns of interest. The visualization panel provides coordinated multiple views wherebymore » selections made in one plot are propagated to the other plots. Users can also export selected data or reconfigure the visualization panel to explore relationships between columns and rows in the data.« less

  2. Evaluation of Potential LSST Spatial Indexing Strategies

    SciTech Connect (OSTI)

    Nikolaev, S; Abdulla, G; Matzke, R

    2006-10-13

    The LSST requirement for producing alerts in near real-time, and the fact that generating an alert depends on knowing the history of light variations for a given sky position, both imply that the clustering information for all detections is available at any time during the survey. Therefore, any data structure describing clustering of detections in LSST needs to be continuously updated, even as new detections are arriving from the pipeline. We call this use case ''incremental clustering'', to reflect this continuous updating of clustering information. This document describes the evaluation results for several potential LSST incremental clustering strategies, using: (1) Neighbors table and zone optimization to store spatial clusters (a.k.a. Jim Grey's, or SDSS algorithm); (2) MySQL built-in R-tree implementation; (3) an external spatial index library which supports a query interface.

  3. Semantic Features for Classifying Referring Search Terms

    SciTech Connect (OSTI)

    May, Chandler J.; Henry, Michael J.; McGrath, Liam R.; Bell, Eric B.; Marshall, Eric J.; Gregory, Michelle L.

    2012-05-11

    When an internet user clicks on a result in a search engine, a request is submitted to the destination web server that includes a referrer field containing the search terms given by the user. Using this information, website owners can analyze the search terms leading to their websites to better understand their visitors needs. This work explores some of the features that can be used for classification-based analysis of such referring search terms. We present initial results for the example task of classifying HTTP requests countries of origin. A system that can accurately predict the country of origin from query text may be a valuable complement to IP lookup methods which are susceptible to the obfuscation of dereferrers or proxies. We suggest that the addition of semantic features improves classifier performance in this example application. We begin by looking at related work and presenting our approach. After describing initial experiments and results, we discuss paths forward for this work.

  4. Midcontinent Interactive Digital Carbon Atlas and Relational Database (MIDCARB)

    SciTech Connect (OSTI)

    Timothy R. Carr; Scott W. White

    2002-06-01

    This annual report describes progress of the project entitled ''Midcontinent Interactive Digital Carbon Atlas and Relational Database (MIDCARB)''. This project, funded by the Department of Energy, is a cooperative project that assembles a consortium of five states (Indiana, Illinois, Kansas, Kentucky and Ohio) to construct an online distributed Relational Database Management System (RDBMS) and Geographic Information System (GIS) covering aspects of carbon dioxide geologic sequestration (http://www.midcarb.org). The system links the five states in the consortium into a coordinated regional database system consisting of datasets useful to industry, regulators and the public. The project is working to provide advanced distributed computing solutions to link database servers across the five states into a single system where data is maintained at the local level but is accessed through a single Web portal and can be queried, assembled, analyzed and displayed. Each individual state has strengths in data gathering, data manipulation and data display, including GIS mapping, custom application development, web development, and database design. Sharing of expertise provides the critical mass of technical expertise to improve CO{sub 2} databases and data access in all states. This project improves the flow of data across servers in the five states and increases the amount and quality of available digital data. The MIDCARB project is developing improved online tools to provide real-time display and analyze CO{sub 2} sequestration data. The system links together data from sources, sinks and transportation within a spatial database that can be queried online. Visualization of high quality and current data can assist decision makers by providing access to common sets of high quality data in a consistent manner.

  5. An Ontology Design Pattern for Surface Water Features

    SciTech Connect (OSTI)

    Sinha, Gaurav; Mark, David; Kolas, Dave; Varanka, Dalia; Romero, Boleslo E; Feng, Chen-Chieh; Usery, Lynn; Liebermann, Joshua; Sorokine, Alexandre

    2014-01-01

    Surface water is a primary concept of human experience but concepts are captured in cultures and languages in many different ways. Still, many commonalities can be found due to the physical basis of many of the properties and categories. An abstract ontology of surface water features based only on those physical properties of landscape features has the best potential for serving as a foundational domain ontology. It can then be used to systematically incor-porate concepts that are specific to a culture, language, or scientific domain. The Surface Water ontology design pattern was developed both for domain knowledge distillation and to serve as a conceptual building-block for more complex surface water ontologies. A fundamental distinction is made in this on-tology between landscape features that act as containers (e.g., stream channels, basins) and the bodies of water (e.g., rivers, lakes) that occupy those containers. Concave (container) landforms semantics are specified in a Dry module and the semantics of contained bodies of water in a Wet module. The pattern is imple-mented in OWL, but Description Logic axioms and a detailed explanation is provided. The OWL ontology will be an important contribution to Semantic Web vocabulary for annotating surface water feature datasets. A discussion about why there is a need to complement the pattern with other ontologies, es-pecially the previously developed Surface Network pattern is also provided. Fi-nally, the practical value of the pattern in semantic querying of surface water datasets is illustrated through a few queries and annotated geospatial datasets.

  6. ACTIVE LEARNING TO OVERCOME SAMPLE SELECTION BIAS: APPLICATION TO PHOTOMETRIC VARIABLE STAR CLASSIFICATION

    SciTech Connect (OSTI)

    Richards, Joseph W.; Starr, Dan L.; Miller, Adam A.; Bloom, Joshua S.; Butler, Nathaniel R.; Berian James, J.; Brink, Henrik; Long, James P.; Rice, John

    2012-01-10

    Despite the great promise of machine-learning algorithms to classify and predict astrophysical parameters for the vast numbers of astrophysical sources and transients observed in large-scale surveys, the peculiarities of the training data often manifest as strongly biased predictions on the data of interest. Typically, training sets are derived from historical surveys of brighter, more nearby objects than those from more extensive, deeper surveys (testing data). This sample selection bias can cause catastrophic errors in predictions on the testing data because (1) standard assumptions for machine-learned model selection procedures break down and (2) dense regions of testing space might be completely devoid of training data. We explore possible remedies to sample selection bias, including importance weighting, co-training, and active learning (AL). We argue that AL-where the data whose inclusion in the training set would most improve predictions on the testing set are queried for manual follow-up-is an effective approach and is appropriate for many astronomical applications. For a variable star classification problem on a well-studied set of stars from Hipparcos and Optical Gravitational Lensing Experiment, AL is the optimal method in terms of error rate on the testing data, beating the off-the-shelf classifier by 3.4% and the other proposed methods by at least 3.0%. To aid with manual labeling of variable stars, we developed a Web interface which allows for easy light curve visualization and querying of external databases. Finally, we apply AL to classify variable stars in the All Sky Automated Survey, finding dramatic improvement in our agreement with the ASAS Catalog of Variable Stars, from 65.5% to 79.5%, and a significant increase in the classifier's average confidence for the testing set, from 14.6% to 42.9%, after a few AL iterations.

  7. Muscle Logic: New Knowledge Resource for Anatomy Enables Comprehensive Searches of the Literature on the Feeding Muscles of Mammals

    DOE Public Access Gateway for Energy & Science Beta (PAGES Beta)

    Druzinsky, Robert E.; Balhoff, James P.; Crompton, Alfred W.; Done, James; German, Rebecca Z.; Haendel, Melissa A.; Herrel, Anthony; Herring, Susan W.; Lapp, Hilmar; Mabee, Paula M.; et al

    2016-02-12

    Here we present the Mammalian Feeding Muscle Ontology (MFMO), a multi-species ontology focused on anatomical structures that participate in feeding and other oral/pharyngeal behaviors. A unique feature of the MFMO is that a simple, computable, definition of each muscle, which includes its attachments and innervation, is true across mammals. This construction mirrors the logical foundation of comparative anatomy and permits searches using language familiar to biologists. Further, it provides a template for muscles that will be useful in extending any anatomy ontology. The MFMO is developed to support the Feeding Experiments End-User Database Project (FEED, https://feedexp.org/), a publicly-available, online repositorymore » for physiological data collected from in vivo studies of feeding (e.g., mastication, biting, swallowing) in mammals. Currently the MFMO is integrated into FEED and also into two literature-specific implementations of Textpresso, a text-mining system that facilitates powerful searches of a corpus of scientific publications. We evaluate the MFMO by asking questions that test the ability of the ontology to return appropriate answers (competency questions). Lastly, we compare the results of queries of the MFMO to results from similar searches in PubMed and Google Scholar. Our tests demonstrate that the MFMO is competent to answer queries formed in the common language of comparative anatomy, but PubMed and Google Scholar are not. Overall, our results show that by incorporating anatomical ontologies into searches, an expanded and anatomically comprehensive set of results can be obtained. The broader scientific and publishing communities should consider taking up the challenge of semantically enabled search capabilities.« less

  8. SU-E-T-544: A Radiation Oncology-Specific Multi-Institutional Federated Database: Initial Implementation

    SciTech Connect (OSTI)

    Hendrickson, K; Phillips, M; Fishburn, M; Evans, K; Banerian, S; Mayr, N; Wong, J; McNutt, T; Moore, J; Robertson, S

    2014-06-01

    Purpose: To implement a common database structure and user-friendly web-browser based data collection tools across several medical institutions to better support evidence-based clinical decision making and comparative effectiveness research through shared outcomes data. Methods: A consortium of four academic medical centers agreed to implement a federated database, known as Oncospace. Initial implementation has addressed issues of differences between institutions in workflow and types and breadth of structured information captured. This requires coordination of data collection from departmental oncology information systems (OIS), treatment planning systems, and hospital electronic medical records in order to include as much as possible the multi-disciplinary clinical data associated with a patients care. Results: The original database schema was well-designed and required only minor changes to meet institution-specific data requirements. Mobile browser interfaces for data entry and review for both the OIS and the Oncospace database were tailored for the workflow of individual institutions. Federation of database queries--the ultimate goal of the project--was tested using artificial patient data. The tests serve as proof-of-principle that the system as a whole--from data collection and entry to providing responses to research queries of the federated database--was viable. The resolution of inter-institutional use of patient data for research is still not completed. Conclusions: The migration from unstructured data mainly in the form of notes and documents to searchable, structured data is difficult. Making the transition requires cooperation of many groups within the department and can be greatly facilitated by using the structured data to improve clinical processes and workflow. The original database schema design is critical to providing enough flexibility for multi-institutional use to improve each institution s ability to study outcomes, determine best practices

  9. SU-E-T-357: Semi-Automated Knowledge-Based Radiation Therapy (KBRT) Planning for Head-And-Neck Cancer (HNC): Can KBRT Plans Achieve Better Results Than Manual Planning?

    SciTech Connect (OSTI)

    Lutzky, C; Grzetic, S; Lo, J; Das, S

    2014-06-01

    Purpose: Knowledge Based Radiation Therapy Treatment (KBRT) planning can be used to semi-automatically generate IMRT plans for new patients using constraints derived from previously manually-planned, geometrically similar patients. We investigate whether KBRT plans can achieve greater dose sparing than manual plans using optimized, organspecific constraint weighting factors. Methods: KBRT planning of HNC radiotherapy cases geometrically matched each new (query) case to one of the 105 clinically approved plans in our database. The dose distribution of the planned match was morphed to fit the querys geometry. Dose-volume constraints extracted from the morphed dose distribution were used to run the IMRT optimization with no user input. In the first version, all constraints were multiplied by a weighting factor of 0.7. The weighting factors were then systematically optimized (in order of OARs with increasing separation from the target) to maximize sparing to each OAR without compromising other OARs. The optimized, second version plans were compared against the first version plans and the clinically approved plans for 45 unilateral/bilateral target cases using the dose metrics: mean, median and maximum (brainstem and cord) doses. Results: Compared to the first version, the second version significantly reduced mean/median contralateral parotid doses (>2Gy) for bilateral cases. Other changes between the two versions were not clinically meaningful. Compared to the original clinical plans, both bilateral and unilateral plans in the second version had lower average dose metrics for 5 of the 6 OARs. Compared to the original plans, the second version achieved dose sparing that was at least as good for all OARs and better for the ipsilateral parotid (bilateral) and oral cavity (bilateral/unilateral). Differences in planning target volume coverage metrics were not clinically significant. Conclusion: HNC-KBRT planning generated IMRT plans with at least equivalent dose sparing to

  10. Sequence modelling and an extensible data model for genomic database

    SciTech Connect (OSTI)

    Li, Peter Wei-Der Lawrence Berkeley Lab., CA )

    1992-01-01

    The Human Genome Project (HGP) plans to sequence the human genome by the beginning of the next century. It will generate DNA sequences of more than 10 billion bases and complex marker sequences (maps) of more than 100 million markers. All of these information will be stored in database management systems (DBMSs). However, existing data models do not have the abstraction mechanism for modelling sequences and existing DBMS's do not have operations for complex sequences. This work addresses the problem of sequence modelling in the context of the HGP and the more general problem of an extensible object data model that can incorporate the sequence model as well as existing and future data constructs and operators. First, we proposed a general sequence model that is application and implementation independent. This model is used to capture the sequence information found in the HGP at the conceptual level. In addition, abstract and biological sequence operators are defined for manipulating the modelled sequences. Second, we combined many features of semantic and object oriented data models into an extensible framework, which we called the Extensible Object Model'', to address the need of a modelling framework for incorporating the sequence data model with other types of data constructs and operators. This framework is based on the conceptual separation between constructors and constraints. We then used this modelling framework to integrate the constructs for the conceptual sequence model. The Extensible Object Model is also defined with a graphical representation, which is useful as a tool for database designers. Finally, we defined a query language to support this model and implement the query processor to demonstrate the feasibility of the extensible framework and the usefulness of the conceptual sequence model.

  11. Sequence modelling and an extensible data model for genomic database

    SciTech Connect (OSTI)

    Li, Peter Wei-Der |

    1992-01-01

    The Human Genome Project (HGP) plans to sequence the human genome by the beginning of the next century. It will generate DNA sequences of more than 10 billion bases and complex marker sequences (maps) of more than 100 million markers. All of these information will be stored in database management systems (DBMSs). However, existing data models do not have the abstraction mechanism for modelling sequences and existing DBMS`s do not have operations for complex sequences. This work addresses the problem of sequence modelling in the context of the HGP and the more general problem of an extensible object data model that can incorporate the sequence model as well as existing and future data constructs and operators. First, we proposed a general sequence model that is application and implementation independent. This model is used to capture the sequence information found in the HGP at the conceptual level. In addition, abstract and biological sequence operators are defined for manipulating the modelled sequences. Second, we combined many features of semantic and object oriented data models into an extensible framework, which we called the ``Extensible Object Model``, to address the need of a modelling framework for incorporating the sequence data model with other types of data constructs and operators. This framework is based on the conceptual separation between constructors and constraints. We then used this modelling framework to integrate the constructs for the conceptual sequence model. The Extensible Object Model is also defined with a graphical representation, which is useful as a tool for database designers. Finally, we defined a query language to support this model and implement the query processor to demonstrate the feasibility of the extensible framework and the usefulness of the conceptual sequence model.

  12. Retrieval Using Texture Features in High Resolution Multi-spectral Satellite Imagery

    SciTech Connect (OSTI)

    Newsam, S D; Kamath, C

    2004-01-22

    Texture features have long been used in remote sensing applications to represent and retrieve image regions similar to a query region. Various representations of texture have been proposed based on the Fourier power spectrum, spatial co-occurrence, wavelets, Gabor filters, etc. These representations vary in their computational complexity and their suitability for representing different region types. Much of the work done thus far has focused on panchromatic imagery at low to moderate spatial resolutions, such as images from Landsat 1-7 which have a resolution of 15-30 m/pixel, and from SPOT 1-5 which have a resolution of 2.5-20 m/pixel. However, it is not clear which texture representation works best for the new classes of high resolution panchromatic (60-100 cm/pixel) and multi-spectral (4 bands for red, green, blue, and near infra-red at 2.4-4 m/pixel) imagery. It is also not clear how the different spectral bands should be combined. In this paper, we investigate the retrieval performance of several different texture representations using multi-spectral satellite images from IKONOS. A query-by-example framework, along with a manually chosen ground truth dataset, allows different combinations of texture representations and spectral bands to be compared. We focus on the specific problem of retrieving inhabited regions from images of urban and rural scenes. Preliminary results show that (1) the use of all spectral bands improves the retrieval performance, and (2) co-occurrence, wavelet and Gabor texture features perform comparably.

  13. HPC Analytics Support. Requirements for Uncertainty Quantification Benchmarks

    SciTech Connect (OSTI)

    Paulson, Patrick R.; Purohit, Sumit; Rodriguez, Luke R.

    2015-05-01

    This report outlines techniques for extending benchmark generation products so they support uncertainty quantification by benchmarked systems. We describe how uncertainty quantification requirements can be presented to candidate analytical tools supporting SPARQL. We describe benchmark data sets for evaluating uncertainty quantification, as well as an approach for using our benchmark generator to produce data sets for generating benchmark data sets.

  14. National Carbon Sequestration Database and Geographic Information System (NatCarb)

    SciTech Connect (OSTI)

    Kenneth Nelson; Timothy Carr

    2009-03-31

    This annual and final report describes the results of the multi-year project entitled 'NATional CARBon Sequestration Database and Geographic Information System (NatCarb)' (http://www.natcarb.org). The original project assembled a consortium of five states (Indiana, Illinois, Kansas, Kentucky and Ohio) in the midcontinent of the United States (MIDCARB) to construct an online distributed Relational Database Management System (RDBMS) and Geographic Information System (GIS) covering aspects of carbon dioxide (CO{sub 2}) geologic sequestration. The NatCarb system built on the technology developed in the initial MIDCARB effort. The NatCarb project linked the GIS information of the Regional Carbon Sequestration Partnerships (RCSPs) into a coordinated regional database system consisting of datasets useful to industry, regulators and the public. The project includes access to national databases and GIS layers maintained by the NatCarb group (e.g., brine geochemistry) and publicly accessible servers (e.g., USGS, and Geography Network) into a single system where data are maintained and enhanced at the local level, but are accessed and assembled through a single Web portal to facilitate query, assembly, analysis and display. This project improves the flow of data across servers and increases the amount and quality of available digital data. The purpose of NatCarb is to provide a national view of the carbon capture and storage potential in the U.S. and Canada. The digital spatial database allows users to estimate the amount of CO{sub 2} emitted by sources (such as power plants, refineries and other fossil-fuel-consuming industries) in relation to geologic formations that can provide safe, secure storage sites over long periods of time. The NatCarb project worked to provide all stakeholders with improved online tools for the display and analysis of CO{sub 2} carbon capture and storage data through a single website portal (http://www.natcarb.org/). While the external project is

  15. Improving the Availability and Delivery of Critical Information for Tight Gas Resource Development in the Appalachian Basin

    SciTech Connect (OSTI)

    Mary Behling; Susan Pool; Douglas Patchen; John Harper

    2008-12-31

    in 1-foot intervals from 11 cores, and approximately 260 references for these plays. A primary objective of the research was to make data and information available free to producers through an on-line data delivery model designed for public access on the Internet. The web-based application that was developed utilizes ESRI's ArcIMS GIS software to deliver both well-based and play-based data that are searchable through user-originated queries, and allows interactive regional geographic and geologic mapping that is play-based. System tools help users develop their customized spatial queries. A link also has been provided to the West Virginia Geological Survey's 'pipeline' system for accessing all available well-specific data for more than 140,000 wells in West Virginia. However, only well-specific queries by API number are permitted at this time. The comprehensive project web site (http://www.wvgs.wvnet.edu/atg) resides on West Virginia Geological Survey's servers and links are provided from the Pennsylvania Geological Survey and Appalachian Oil and Natural Gas Research Consortium web sites.

  16. DOE SBIR Phase II Final Report: Distributed Relevance Ranking in Heterogeneous Document Collections

    SciTech Connect (OSTI)

    Abe Lederman

    2007-01-08

    the user to select from a number of collections grouped by category and enter a query expression (See Illustration 1 - Science.gov 3.0 Search Page). After the user clicks search a results page is displayed that provides a list of results from the selected collections ordered by relevance based on the query expression the user provided. Our grid based solution to deep web search and document ranking has already gained attention within DOE, other Government Agencies and a fortune 50 company. We are committed to the continued development of grid based solutions to large scale data access, filtering, and presentation problems within the domain of Information Retrieval and the more general categories of content management, data mining and data analysis.

  17. National Computational Infrastructure for LatticeGauge Theory SciDAC-2 Closeout Report

    SciTech Connect (OSTI)

    Bapty, Theodore; Dubey, Abhishek

    2013-07-18

    As part of the reliability project work, researchers from Vanderbilt University, Fermi National Laboratory and Illinois Institute of technology developed a real-time cluster fault-tolerant cluster monitoring framework. The goal for the scientific workflow project is to investigate and develop domain-specific workflow tools for LQCD to help effectively orchestrate, in parallel, computational campaigns consisting of many loosely-coupled batch processing jobs. Major requirements for an LQCD workflow system include: a system to manage input metadata, e.g. physics parameters such as masses, a system to manage and permit the reuse of templates describing workflows, a system to capture data provenance information, a systems to manage produced data, a means of monitoring workflow progress and status, a means of resuming or extending a stopped workflow, fault tolerance features to enhance the reliability of running workflows. In summary, these achievements are reported: • Implemented a software system to manage parameters. This includes a parameter set language based on a superset of the JSON data-interchange format, parsers in multiple languages (C++, Python, Ruby), and a web-based interface tool. It also includes a templating system that can produce input text for LQCD applications like MILC. • Implemented a monitoring sensor framework in software that is in production on the Fermilab USQCD facility. This includes equipment health, process accounting, MPI/QMP process tracking, and batch system (Torque) job monitoring. All sensor data are available from databases, and various query tools can be used to extract common data patterns and perform ad hoc searches. Common batch system queries such as job status are available in command line tools and are used in actual workflow-based production by a subset of Fermilab users. • Developed a formal state machine model for scientific workflow and reliability systems. This includes the use of Vanderbilt’s Generic Modeling

  18. Measuring the Interestingness of Articles in a Limited User Environment

    SciTech Connect (OSTI)

    Pon, R K

    2008-10-06

    Search engines, such as Google, assign scores to news articles based on their relevancy to a query. However, not all relevant articles for the query may be interesting to a user. For example, if the article is old or yields little new information, the article would be uninteresting. Relevancy scores do not take into account what makes an article interesting, which varies from user to user. Although methods such as collaborative filtering have been shown to be effective in recommendation systems, in a limited user environment, there are not enough users that would make collaborative filtering effective. A general framework, called iScore, is presented for defining and measuring the 'interestingness' of articles, incorporating user-feedback. iScore addresses various aspects of what makes an article interesting, such as topic relevancy, uniqueness, freshness, source reputation, and writing style. It employs various methods to measure these features and uses a classifier operating on these features to recommend articles. The basic iScore configuration is shown to improve recommendation results by as much as 20%. In addition to the basic iScore features, additional features are presented to address the deficiencies of existing feature extractors, such as one that tracks multiple topics, called MTT, and a version of the Rocchio algorithm that learns its parameters online as it processes documents, called eRocchio. The inclusion of both MTT and eRocchio into iScore is shown to improve iScore recommendation results by as much as 3.1% and 5.6%, respectively. Additionally, in TREC11 Adaptive Filter Task, eRocchio is shown to be 10% better than the best filter in the last run of the task. In addition to these two major topic relevancy measures, other features are also introduced that employ language models, phrases, clustering, and changes in topics to improve recommendation results. These additional features are shown to improve recommendation results by iScore by up to 14

  19. Geospatial Analysis and Technical Assistance for Power Plant Siting Interagency

    SciTech Connect (OSTI)

    Neher, L A

    2002-03-07

    The focus of this contract (in the summer and fall of 2001) was originally to help the California Energy Commission (CEC) locate and evaluate potential sites for electric power generation facilities and to assist the CEC in addressing areas of congestion on transmission lines and natural gas supply line corridors. Subsequent events have reduced the immediate urgency, although not the ultimate need for such analyses. Software technology for deploying interactive geographic information systems (GIS) accessible over the Internet have developed to the point that it is now practical to develop and publish GIS web sites that have substantial viewing, movement, query, and even map-making capabilities. As part of a separate project not funded by the CEC, the GIS Center at LLNL, on an experimental basis, has developed a web site to explore the technical difficulties as well as the interest in such a web site by agencies and others concerned with energy research. This exploratory effort offers the potential or developing an interactive GIS web site for use by the CEC for energy research, policy analysis, site evaluation, and permit and regulatory matters. To help ground the geospatial capabilities in the realistic requirements and needs of the CEC staff, the CEC requested that the GIS Center conduct interviews of several CEC staff persons to establish their current and envisioned use of spatial data and requirements for geospatial analyses. This survey will help define a web-accessible central GIS database for the CEC, which will augment the well-received work of the CEC Cartography Unit. Individuals within each siting discipline have been contacted and their responses to three question areas have been summarized. The web-based geospatial data and analytical tools developed within this project will be available to CEC staff for initial area studies, queries, and informal, small-format maps. It is not designed for fine cartography or for large-format posters such as the

  20. Genomic insights into the evolution of hybrid isoprenoid biosynthetic gene clusters in the MAR4 marine streptomycete clade

    DOE Public Access Gateway for Energy & Science Beta (PAGES Beta)

    Gallagher, Kelley A.; Jensen, Paul R.

    2015-11-17

    Background: Considerable advances have been made in our understanding of the molecular genetics of secondary metabolite biosynthesis. Coupled with increased access to genome sequence data, new insight can be gained into the diversity and distributions of secondary metabolite biosynthetic gene clusters and the evolutionary processes that generate them. Here we examine the distribution of gene clusters predicted to encode the biosynthesis of a structurally diverse class of molecules called hybrid isoprenoids (HIs) in the genus Streptomyces. These compounds are derived from a mixed biosynthetic origin that is characterized by the incorporation of a terpene moiety onto a variety of chemicalmore » scaffolds and include many potent antibiotic and cytotoxic agents. Results: One hundred and twenty Streptomyces genomes were searched for HI biosynthetic gene clusters using ABBA prenyltransferases (PTases) as queries. These enzymes are responsible for a key step in HI biosynthesis. The strains included 12 that belong to the ‘MAR4’ clade, a largely marine-derived lineage linked to the production of diverse HI secondary metabolites. We found ABBA PTase homologs in all of the MAR4 genomes, which averaged five copies per strain, compared with 21 % of the non-MAR4 genomes, which averaged one copy per strain. Phylogenetic analyses suggest that MAR4 PTase diversity has arisen by a combination of horizontal gene transfer and gene duplication. Furthermore, there is evidence that HI gene cluster diversity is generated by the horizontal exchange of orthologous PTases among clusters. Many putative HI gene clusters have not been linked to their secondary metabolic products, suggesting that MAR4 strains will yield additional new compounds in this structure class. Finally, we confirm that the mevalonate pathway is not always present in genomes that contain HI gene clusters and thus is not a reliable query for identifying strains with the potential to produce HI secondary metabolites

  1. NATIONAL CARBON SEQUESTRATION DATABASE AND GEOGRAPHIC INFORMATION SYSTEM (NATCARB) FORMER TITLE-MIDCONTINENT INTERACTIVE DIGITAL CARBON ATLAS AND RELATIONAL DATABASE (MIDCARB)

    SciTech Connect (OSTI)

    Timothy R. Carr

    2004-07-16

    This annual report describes progress in the third year of the three-year project entitled ''Midcontinent Interactive Digital Carbon Atlas and Relational Database (MIDCARB)''. The project assembled a consortium of five states (Indiana, Illinois, Kansas, Kentucky and Ohio) to construct an online distributed Relational Database Management System (RDBMS) and Geographic Information System (GIS) covering aspects of carbon dioxide (CO{sub 2}) geologic sequestration (http://www.midcarb.org). The system links the five states in the consortium into a coordinated regional database system consisting of datasets useful to industry, regulators and the public. The project has been extended and expanded as a ''NATional CARBon Sequestration Database and Geographic Information System (NATCARB)'' to provide national coverage across the Regional CO{sub 2} Partnerships, which currently cover 40 states (http://www.natcarb.org). Advanced distributed computing solutions link database servers across the five states and other publicly accessible servers (e.g., USGS) into a single system where data is maintained and enhanced at the local level but is accessed and assembled through a single Web portal and can be queried, assembled, analyzed and displayed. This project has improved the flow of data across servers and increased the amount and quality of available digital data. The online tools used in the project have improved in stability and speed in order to provide real-time display and analysis of CO{sub 2} sequestration data. The move away from direct database access to web access through eXtensible Markup Language (XML) has increased stability and security while decreasing management overhead. The MIDCARB viewer has been simplified to provide improved display and organization of the more than 125 layers and data tables that have been generated as part of the project. The MIDCARB project is a functional demonstration of distributed management of data systems that cross the boundaries

  2. Applied Parallel Metadata Indexing

    SciTech Connect (OSTI)

    Jacobi, Michael R

    2012-08-01

    The GPFS Archive is parallel archive is a parallel archive used by hundreds of users in the Turquoise collaboration network. It houses 4+ petabytes of data in more than 170 million files. Currently, users must navigate the file system to retrieve their data, requiring them to remember file paths and names. A better solution might allow users to tag data with meaningful labels and searach the archive using standard and user-defined metadata, while maintaining security. last summer, I developed the backend to a tool that adheres to these design goals. The backend works by importing GPFS metadata into a MongoDB cluster, which is then indexed on each attribute. This summer, the author implemented security and developed the user interfae for the search tool. To meet security requirements, each database table is associated with a single user, which only stores records that the user may read, and requires a set of credentials to access. The interface to the search tool is implemented using FUSE (Filesystem in USErspace). FUSE is an intermediate layer that intercepts file system calls and allows the developer to redefine how those calls behave. In the case of this tool, FUSE interfaces with MongoDB to issue queries and populate output. A FUSE implementation is desirable because it allows users to interact with the search tool using commands they are already familiar with. These security and interface additions are essential for a usable product.

  3. A Run-Time Verification Framework for Smart Grid Applications Implemented on Simulation Frameworks

    SciTech Connect (OSTI)

    Ciraci, Selim; Sozer, Hasan; Tekinerdogan, Bedir

    2013-05-18

    Smart grid applications are implemented and tested with simulation frameworks as the developers usually do not have access to large sensor networks to be used as a test bed. The developers are forced to map the implementation onto these frameworks which results in a deviation between the architecture and the code. On its turn this deviation makes it hard to verify behavioral constraints that are de- scribed at the architectural level. We have developed the ConArch toolset to support the automated verification of architecture-level behavioral constraints. A key feature of ConArch is programmable mapping for architecture to the implementation. Here, developers implement queries to identify the points in the target program that correspond to architectural interactions. ConArch generates run- time observers that monitor the flow of execution between these points and verifies whether this flow conforms to the behavioral constraints. We illustrate how the programmable mappings can be exploited for verifying behavioral constraints of a smart grid appli- cation that is implemented with two simulation frameworks.

  4. Integration of remote sensing and geographic information systems for Great Lakes water quality monitoring

    SciTech Connect (OSTI)

    Lathrop, R.G. Jr.

    1988-01-01

    The utility of three operational satellite remote sensing systems, namely, the Landsat Thematic Mapper (TM), the SPOT High Resolution Visible (HRV) sensors and the NOAA Advanced Very High Resolution Radiometer (AVHRR), were evaluated as a means of estimating water quality and surface temperature. Empirical calibration through linear regression techniques was used to relate near-simultaneously acquired satellite radiance/reflectance data and water quality observations obtained in Green Bay and the nearshore waters of Lake Michigan. Four dates of TM and one date each of SPOT and AVHRR imagery/surface reference data were acquired and analyzed. Highly significant relationships were identified between the TM and SPOT data and secchi disk depth, nephelometric turbidity, chlorophyll a, total suspended solids (TSS), absorbance, and surface temperature (TM only). The AVHRR data were not analyzed independently but were used for comparison with the TM data. Calibrated water quality image maps were input to a PC-based raster GIS package, EPPL7. Pattern interpretation and spatial analysis techniques were used to document the circulation dynamics and model mixing processes in Green Bay. A GIS facilitates the retrieval, query and spatial analysis of mapped information and provides the framework for an integrated operational monitoring system for the Great Lakes.

  5. Visualization Gallery from the Computational Research Division at Lawrence Berkeley National Laboratory

    DOE Data Explorer [Office of Scientific and Technical Information (OSTI)]

    This excellent collection of visualization vignettes highlights research work done by the LBNL/NERSC Visualization Group and its collaborators from 1993 to the present. Images lead to technical explanations and project details, helping users to branch out to other related sources. Titles of the projects provide clues both to the imaging focus of the research and the scientific discipline for which the visualizations are intended. Only a few of the many titles/images/projects are listed here: 1) Hybrid Parallelism for Volume Rendering at Large Scale Analysis of Laser Wakefield Particle Acceleration Data; 2) Visualization of Microearthquake Data from Enhanced Geothermal Systems; 3) PointCloudXplore: Visualization and Analysis of 3D Gene Expression Data; 4) Visualization of Quantum Monte-Carlo simulations; 5) Global Cloud Resolving Models; 6) Visualization of large-scale GFDL/NOAA climate simulations; 7) Direct Numerical Simulation of Turbulent Flame Quenching by Fine Water Droplets; 8) Visualization of Magneto-rotational instability and turbulent angular momentum transport; 9) Sunfall: Visual Analytics for Astrophysics; 10) Fast Contour Descriptor Algorithm for Supernova Image Classification; 11) Supernova Recognition Using Support Vector Machines; 12) High Performance Visualization - Query-Driven Network Traffic Analysis; 13) Visualization of Magneto-rotational instability and turbulent angular momentum transport; 14) Life Sciences: Cell Division of Caulobacter Crescentus; 15) Electron Cloud Simulations.

  6. Orchestrating Distributed Resource Ensembles for Petascale Science

    SciTech Connect (OSTI)

    Baldin, Ilya; Mandal, Anirban; Ruth, Paul; Yufeng, Xin

    2014-04-24

    Distributed, data-intensive computational science applications of interest to DOE scientific com- munities move large amounts of data for experiment data management, distributed analysis steps, remote visualization, and accessing scientific instruments. These applications need to orchestrate ensembles of resources from multiple resource pools and interconnect them with high-capacity multi- layered networks across multiple domains. It is highly desirable that mechanisms are designed that provide this type of resource provisioning capability to a broad class of applications. It is also important to have coherent monitoring capabilities for such complex distributed environments. In this project, we addressed these problems by designing an abstract API, enabled by novel semantic resource descriptions, for provisioning complex and heterogeneous resources from multiple providers using their native provisioning mechanisms and control planes: computational, storage, and multi-layered high-speed network domains. We used an extensible resource representation based on semantic web technologies to afford maximum flexibility to applications in specifying their needs. We evaluated the effectiveness of provisioning using representative data-intensive ap- plications. We also developed mechanisms for providing feedback about resource performance to the application, to enable closed-loop feedback control and dynamic adjustments to resource allo- cations (elasticity). This was enabled through development of a novel persistent query framework that consumes disparate sources of monitoring data, including perfSONAR, and provides scalable distribution of asynchronous notifications.

  7. Secure Information Sharing

    Energy Science and Technology Software Center (OSTI)

    2005-09-09

    We are develoing a peer-to-peer system to support secure, location independent information sharing in the scientific community. Once complete, this system will allow seamless and secure sharing of information between multiple collaborators. The owners of information will be able to control how the information is stored, managed. ano shared. In addition, users will have faster access to information updates within a collaboration. Groups collaborating on scientific experiments have a need to share information and data.more » This information and data is often represented in the form of files and database entries. In a typical scientific collaboration, there are many different locations where data would naturally be stored. This makes It difficult for collaborators to find and access the information they need. Our goal is to create a lightweight file-sharing system that makes it’easy for collaborators to find and use the data they need. This system must be easy-to-use, easy-to-administer, and secure. Our information-sharing tool uses group communication, in particular the InterGroup protocols, to reliably deliver each query to all of the current participants in a scalable manner, without having to discover all of their identities. We will use the Secure Group Layer (SGL) and Akenti to provide security to the participants of our environment, SGL will provide confldentiality, integrity, authenticity, and authorization enforcement for the InterGroup protocols and Akenti will provide access control to other resources.« less

  8. Arctic & Offshore Technical Data System

    Energy Science and Technology Software Center (OSTI)

    1990-07-01

    AORIS is a computerized information system to assist the technology and planning community in the development of Arctic oil and gas resources. In general, AORIS is geographically dependent and, where possible, site specific. The main topics are sea ice, geotechnology, oceanography, meteorology, and Arctic engineering, as they relate to such offshore oil and gas activities as exploration, production, storage, and transportation. AORIS consists of a directory component that identifies 85 Arctic energy-related databases and tellsmore » how to access them; a bibliographic/management information system or bibliographic component containing over 8,000 references and abstracts on Arctic energy-related research; and a scientific and engineering information system, or data component, containing over 800 data sets, in both tabular and graphical formats, on sea ice characteristics taken from the bibliographic citations. AORIS also contains much of the so-called grey literature, i.e., data and/or locations of Arctic data collected, but never published. The three components are linked so the user may easily move from one component to another. A generic information system is provided to allow users to create their own information systems. The generic programs have the same query and updating features as AORIS, except that there is no directory component.« less

  9. Parallel Environment for the Creation of Stochastics 1.0

    Energy Science and Technology Software Center (OSTI)

    2011-01-06

    PECOS is a computational library for creating and manipulating realizations of stochastic quantities, including scalar uncertain variables, random fields, and stochastic processes. It offers a unified interface to univariate and multivariate polynomial approximations using either orthogonal or interpolation polynomials; numerical integration drivers for Latin hypercube sampling, quadrature, cubature, and sparse grids; and fast Fourier transforms using third party libraries. The PECOS core also offers statistical utilities and transformations between various representations of stochastic uncertainty. PECOSmore » provides a C++ API through which users can generate and transform realizations of stochastic quantities. It is currently used by Sandia’s DAKOTA, Stokhos, and Encore software packages for uncertainty quantification and verification. PECOS generates random sample sets and multi-dimensional integration grids, typically used in forward propagation of scalar uncertainty in computational models (uncertainty quantification (UQ)). PECOS also generates samples of random fields (RFs) and stochastic processes (SPs) from a set of user-defined power spectral densities (PSDs). The RF/SP may be either Gaussian or non-Gaussian and either stationary or nonstationary, and the resulting sample is intended for run-time query by parallel finite element simulation codes. Finally, PECOS supports nonlinear transformations of random variables via the Nataf transformation and extensions.« less

  10. Mesh infrastructure for coupled multiprocess geophysical simulations

    DOE Public Access Gateway for Energy & Science Beta (PAGES Beta)

    Garimella, Rao V.; Perkins, William A.; Buksas, Mike W.; Berndt, Markus; Lipnikov, Konstantin; Coon, Ethan; Moulton, John D.; Painter, Scott L.

    2014-01-01

    We have developed a sophisticated mesh infrastructure capability to support large scale multiphysics simulations such as subsurface flow and reactive contaminant transport at storage sites as well as the analysis of the effects of a warming climate on the terrestrial arctic. These simulations involve a wide range of coupled processes including overland flow, subsurface flow, freezing and thawing of ice rich soil, accumulation, redistribution and melting of snow, biogeochemical processes involving plant matter and finally, microtopography evolution due to melting and degradation of ice wedges below the surface. In addition to supporting the usual topological and geometric queries about themore » mesh, the mesh infrastructure adds capabilities such as identifying columnar structures in the mesh, enabling deforming of the mesh subject to constraints and enabling the simultaneous use of meshes of different dimensionality for subsurface and surface processes. The generic mesh interface is capable of using three different open source mesh frameworks (MSTK, MOAB and STKmesh) under the hood allowing the developers to directly compare them and choose one that is best suited for the application's needs. We demonstrate the results of some simulations using these capabilities as well as present a comparison of the performance of the different mesh frameworks.« less

  11. Medical and Transmission Vector Vocabulary Alignment with Schema.org

    SciTech Connect (OSTI)

    Smith, William P.; Chappell, Alan R.; Corley, Courtney D.

    2015-04-21

    Available biomedical ontologies and knowledge bases currently lack formal and standards-based interconnections between disease, disease vector, and drug treatment vocabularies. The PNNL Medical Linked Dataset (PNNL-MLD) addresses this gap. This paper describes the PNNL-MLD, which provides a unified vocabulary and dataset of drug, disease, side effect, and vector transmission background information. Currently, the PNNL-MLD combines and curates data from the following research projects: DrugBank, DailyMed, Diseasome, DisGeNet, Wikipedia Infobox, Sider, and PharmGKB. The main outcomes of this effort are a dataset aligned to Schema.org, including a parsing framework, and extensible hooks ready for integration with selected medical ontologies. The PNNL-MLD enables researchers more quickly and easily to query distinct datasets. Future extensions to the PNNL-MLD will include Traditional Chinese Medicine, broader interlinks across genetic structures, a larger thesaurus of synonyms and hypernyms, explicit coding of diseases and drugs across research systems, and incorporating vector-borne transmission vocabularies.

  12. The risk assessment information system

    SciTech Connect (OSTI)

    Kerr, S.B.; Bonczek, R.R.; McGinn, C.W.; Land, M.L.; Bloom, L.D.; Sample, B.E.; Dolislager, F.G.

    1998-06-01

    In an effort to provide service-oriented environmental risk assessment expertise, the Department of Energy (DOE) Center for Risk Excellence (CRE) and DOE Oak Ridge Operations Office (ORO) are sponsoring Oak Ridge National Laboratory (ORNL) to develop a web-based system for disseminating risk tools and information to its users. This system, the Risk Assessment Information System (RAIS), was initially developed to support the site-specific needs of the DOE-ORO Environmental Restoration Risk Assessment Program. With support from the CRE, the system is currently being expanded to benefit all DOE risk information users and can be tailored to meet site-specific needs. Taking advantage of searchable and executable databases, menu-driven queries, and data downloads, using the latest World Wide Web technologies, the RAIS offers essential tools that are used in the risk assessment process or anywhere from project scoping to implementation. The RAIS tools can be located directly at http://risk.lsd.ornl.gov/homepage/rap{_}tool.htm or through the CRE`s homepage at http://www.doe.gov/riskcenter/home.html.

  13. Using Arc/Info GIS to help implement the National Pollutant Discharge Elimination System (NPDES) stormwater permit for Los Angeles County

    SciTech Connect (OSTI)

    Levine, D.A.; Pace, P.J.; Woods, J.A.; DePoto, W.

    1997-06-01

    One of Los Angeles County Department of Public Works` many responsibilities is to manage non-point pollution that enters the storm drain network within Los Angeles County. The management of this non-point source pollution is mandated by the NPDES guidelines under the Federal Clean Water Act. These guidelines require the County to monitor the drainage network and the storm water and urban runoff flowing through it. The County covers over 3,117 square miles, with the NPDES Permit covering over 3,100 square miles and over 2500 miles of storm drains. A proposed solution to monitor and manage this vast geographic area is centered upon an Arc/Info GIS. Some of the many concerns which need to be addressed include the administration and evaluation of Best Management Practices (BMP`s), storm drain inspection for illegal connections and illicit discharges, and pollutant load assessment and modeling. The storm drain network and other coverages will be related to external data bases currently used for facility management and planning. This system would be used for query purposes to perform spatial modeling and {open_quotes}what if{close_quotes} scenarios needed to create maps and reports required by the permit and to evaluate various BMP implementation strategies.

  14. TrustBuilder2

    Energy Science and Technology Software Center (OSTI)

    2007-07-20

    TrustBuilder2 is a flexible framework for supporting research in the area trust negotiation protocols, designed to allow researchers to quickly prototype and experiment with various approaches to trust negotiation. In Trustbuilder2, the primary components of a trust negotiation system are represented using abstract interfaces. Any or all of these components can be implemented or extended by users of the TrustBuilder2 system, thereby making the system's functionality easily extensible. The TrustBuilder2 configuration files can be modifiedmore » to load these custom components in place of the default system components; this facilitates the use of new features without modifications to the underlying runtime system. In our implementation, we provide support for one negotiation strategy, a policy compliance checker based on Jess (the Java Expert System Shell), query interfaces enabling access to disk-based credential and policy repositories, a credential chain construction algorithm, two credential chain verification routines, and both graphical and text-based logging facilities. Trustbuilder2 also supports the interposition of user-defined plug-ins at communication points between system components to allow for easy monitoring of system activity or the modification of messages passed between components.« less

  15. Compact Graph Representations and Parallel Connectivity Algorithms for Massive Dynamic Network Analysis

    SciTech Connect (OSTI)

    Madduri, Kamesh; Bader, David A.

    2009-02-15

    Graph-theoretic abstractions are extensively used to analyze massive data sets. Temporal data streams from socioeconomic interactions, social networking web sites, communication traffic, and scientific computing can be intuitively modeled as graphs. We present the first study of novel high-performance combinatorial techniques for analyzing large-scale information networks, encapsulating dynamic interaction data in the order of billions of entities. We present new data structures to represent dynamic interaction networks, and discuss algorithms for processing parallel insertions and deletions of edges in small-world networks. With these new approaches, we achieve an average performance rate of 25 million structural updates per second and a parallel speedup of nearly28 on a 64-way Sun UltraSPARC T2 multicore processor, for insertions and deletions to a small-world network of 33.5 million vertices and 268 million edges. We also design parallel implementations of fundamental dynamic graph kernels related to connectivity and centrality queries. Our implementations are freely distributed as part of the open-source SNAP (Small-world Network Analysis and Partitioning) complex network analysis framework.

  16. Streaming data analytics via message passing with application to graph algorithms

    SciTech Connect (OSTI)

    Plimpton, Steven J.; Shead, Tim

    2014-05-06

    The need to process streaming data, which arrives continuously at high-volume in real-time, arises in a variety of contexts including data produced by experiments, collections of environmental or network sensors, and running simulations. Streaming data can also be formulated as queries or transactions which operate on a large dynamic data store, e.g. a distributed database. We describe a lightweight, portable framework named PHISH which enables a set of independent processes to compute on a stream of data in a distributed-memory parallel manner. Datums are routed between processes in patterns defined by the application. PHISH can run on top of either message-passing via MPI or sockets via ZMQ. The former means streaming computations can be run on any parallel machine which supports MPI; the latter allows them to run on a heterogeneous, geographically dispersed network of machines. We illustrate how PHISH can support streaming MapReduce operations, and describe streaming versions of three algorithms for large, sparse graph analytics: triangle enumeration, subgraph isomorphism matching, and connected component finding. Lastly, we also provide benchmark timings for MPI versus socket performance of several kernel operations useful in streaming algorithms.

  17. Streaming data analytics via message passing with application to graph algorithms

    DOE Public Access Gateway for Energy & Science Beta (PAGES Beta)

    Plimpton, Steven J.; Shead, Tim

    2014-05-06

    The need to process streaming data, which arrives continuously at high-volume in real-time, arises in a variety of contexts including data produced by experiments, collections of environmental or network sensors, and running simulations. Streaming data can also be formulated as queries or transactions which operate on a large dynamic data store, e.g. a distributed database. We describe a lightweight, portable framework named PHISH which enables a set of independent processes to compute on a stream of data in a distributed-memory parallel manner. Datums are routed between processes in patterns defined by the application. PHISH can run on top of eithermore » message-passing via MPI or sockets via ZMQ. The former means streaming computations can be run on any parallel machine which supports MPI; the latter allows them to run on a heterogeneous, geographically dispersed network of machines. We illustrate how PHISH can support streaming MapReduce operations, and describe streaming versions of three algorithms for large, sparse graph analytics: triangle enumeration, subgraph isomorphism matching, and connected component finding. Lastly, we also provide benchmark timings for MPI versus socket performance of several kernel operations useful in streaming algorithms.« less

  18. Global prevalence and distribution of genes and microorganisms involved in mercury methylation

    SciTech Connect (OSTI)

    Podar, Mircea; Gilmour, C C; Brandt, Craig C; Bullock, Allyson L; Brown, Steven D; Crable, Bryan R; Palumbo, Anthony Vito; Somenahally, Anil C; Elias, Dwayne A

    2015-01-01

    Mercury methylation produces the neurotoxic, highly bioaccumulative methylmercury (MeHg). Recent identification of the methylation genes (hgcAB) provides the foundation for broadly evaluating microbial Hg-methylation potential in nature without making explicit rate measurements. We queried hgcAB diversity and distribution in all available microbial metagenomes, encompassing most environments. The genes were found in nearly all anaerobic, but not in aerobic, environments including oxygenated layers of the open ocean. Critically, hgcAB was effectively absent in ~1500 human microbiomes, suggesting a low risk of endogenous MeHg production. New potential methylation habitats were identified, including invertebrate guts, thawing permafrost, coastal dead zones , soils, sediments, and extreme environments, suggesting multiple routes for MeHg entry into food webs. Several new taxonomic groups potentially capable of Hg-methylation emerged, including lineages having no cultured representatives. We begin to address long-standing evolutionary questions about Hg-methylation and ancient carbon fixation mechanisms while generating a new global view of Hg-methylation potential.

  19. Mesh infrastructure for coupled multiprocess geophysical simulations

    SciTech Connect (OSTI)

    Garimella, Rao V.; Perkins, William A.; Buksas, Mike W.; Berndt, Markus; Lipnikov, Konstantin; Coon, Ethan; Moulton, John D.; Painter, Scott L.

    2014-01-01

    We have developed a sophisticated mesh infrastructure capability to support large scale multiphysics simulations such as subsurface flow and reactive contaminant transport at storage sites as well as the analysis of the effects of a warming climate on the terrestrial arctic. These simulations involve a wide range of coupled processes including overland flow, subsurface flow, freezing and thawing of ice rich soil, accumulation, redistribution and melting of snow, biogeochemical processes involving plant matter and finally, microtopography evolution due to melting and degradation of ice wedges below the surface. In addition to supporting the usual topological and geometric queries about the mesh, the mesh infrastructure adds capabilities such as identifying columnar structures in the mesh, enabling deforming of the mesh subject to constraints and enabling the simultaneous use of meshes of different dimensionality for subsurface and surface processes. The generic mesh interface is capable of using three different open source mesh frameworks (MSTK, MOAB and STKmesh) under the hood allowing the developers to directly compare them and choose one that is best suited for the application's needs. We demonstrate the results of some simulations using these capabilities as well as present a comparison of the performance of the different mesh frameworks.

  20. Arctic & Offshore Technical Data System

    Energy Science and Technology Software Center (OSTI)

    1990-07-01

    AORIS is a computerized information system to assist the technology and planning community in the development of Arctic oil and gas resources. In general, AORIS is geographically dependent and, where possible, site specific. The main topics are sea ice, geotechnology, oceanography, meteorology, and Arctic engineering, as they relate to such offshore oil and gas activities as exploration, production, storage, and transportation. AORIS consists of a directory component that identifies 85 Arctic energy-related databases and tellsmorehow to access them; a bibliographic/management information system or bibliographic component containing over 8,000 references and abstracts on Arctic energy-related research; and a scientific and engineering information system, or data component, containing over 800 data sets, in both tabular and graphical formats, on sea ice characteristics taken from the bibliographic citations. AORIS also contains much of the so-called grey literature, i.e., data and/or locations of Arctic data collected, but never published. The three components are linked so the user may easily move from one component to another. A generic information system is provided to allow users to create their own information systems. The generic programs have the same query and updating features as AORIS, except that there is no directory component.less

  1. Limited-memory adaptive snapshot selection for proper orthogonal decomposition

    SciTech Connect (OSTI)

    Oxberry, Geoffrey M.; Kostova-Vassilevska, Tanya; Arrighi, Bill; Chand, Kyle

    2015-04-02

    Reduced order models are useful for accelerating simulations in many-query contexts, such as optimization, uncertainty quantification, and sensitivity analysis. However, offline training of reduced order models can have prohibitively expensive memory and floating-point operation costs in high-performance computing applications, where memory per core is limited. To overcome this limitation for proper orthogonal decomposition, we propose a novel adaptive selection method for snapshots in time that limits offline training costs by selecting snapshots according an error control mechanism similar to that found in adaptive time-stepping ordinary differential equation solvers. The error estimator used in this work is related to theory bounding the approximation error in time of proper orthogonal decomposition-based reduced order models, and memory usage is minimized by computing the singular value decomposition using a single-pass incremental algorithm. Results for a viscous Burgers’ test problem demonstrate convergence in the limit as the algorithm error tolerances go to zero; in this limit, the full order model is recovered to within discretization error. The resulting method can be used on supercomputers to generate proper orthogonal decomposition-based reduced order models, or as a subroutine within hyperreduction algorithms that require taking snapshots in time, or within greedy algorithms for sampling parameter space.

  2. The Integrated Cloud-based Environmental Data Management System at Los Alamos National Laboratory - 13391

    SciTech Connect (OSTI)

    Schultz Paige, Karen; Gomez, Penny; Patel, Nita P.; EchoHawk, Chris; Dorries, Alison M.

    2013-07-01

    In today's world, instant access to information is taken for granted. The national labs are no exception; our data users expect immediate access to their data. Los Alamos National Laboratory (LANL) has collected over ten million records, and the data needs to be accessible to scientists as well as the public. The data span a wide range of media, analytes, time periods, formats, and quality and have traditionally existed in scattered databases, making comprehensive work with the data impossible. Recently, LANL has successfully integrated all their environmental data into a single, cloud-based, web-accessible data management system. The system combines data transparency to the public with immediate access required by the technical staff. The use of automatic electronic data validation has been critical to immediate data access while saving millions of dollars and increasing data consistency and quality. The system includes a Google Maps based GIS tool that is simple enough for people to locate potentially contaminated sites near their home or workplace, and complex enough to allow scientists to plot and trend their data at the surface and at depth as well as over time. A variety of formatted reports can be run at any desired frequency to report the most current data available in the data base. The advanced user can also run free form queries of the data base. This data management system has saved LANL time and money, an increasingly important accomplishment during periods of budget cuts with increasing demand for immediate electronic services. (authors)

  3. Intelligent Object-Oriented GIS Engine W/dynamic Coupling to Modeled Objects

    Energy Science and Technology Software Center (OSTI)

    1997-02-12

    The GEOVIEWER is an intelligent object-oriented Geographic Information System (GIS) engine that provides not only a spatially-optimized object representation, but also direct linkage to the underlying object, its data and behaviors. Tools are incorporated to perform tasks involving typical GIS functionality, data ingestion, linkage to external models, and integration with other application frameworks. The GOEVIEWER module was designed to provide GIS functionality to create, query, view, and manipulate software objects within a selected area undermore » investigation in a simulation system. Many of these objects are not stored in a format conductive to efficient GIS usage. Their dynamic nature, complexity, and the sheer number of possible entity classes preclude effective integration with traditional GIS technologies due to the loosely coupled nature of their data representations. The primary difference between GEOVIEWER and standard GIS packages is that standard GIS packages offer static views of geospatial data while GEOVIEWER can be dynamically coupled to models and/or applications producing data and, therefore, display changes in geometry, attributes or behavior as they occur in the simulation.« less

  4. Web-Based Geographic Information System Tool for Accessing Hanford Site Environmental Data

    SciTech Connect (OSTI)

    Triplett, Mark B.; Seiple, Timothy E.; Watson, David J.; Charboneau, Briant L.; Morse, John G.

    2014-11-15

    Data volume, complexity, and access issues pose severe challenges for analysts, regulators and stakeholders attempting to efficiently use legacy data to support decision making at the U.S. Department of Energy’s (DOE) Hanford Site. DOE has partnered with the Pacific Northwest National Laboratory (PNNL) on the PHOENIX (PNNL-Hanford Online Environmental Information System) project, which seeks to address data access, transparency, and integration challenges at Hanford to provide effective decision support. PHOENIX is a family of spatially-enabled web applications providing quick access to decades of valuable scientific data and insight through intuitive query, visualization, and analysis tools. PHOENIX realizes broad, public accessibility by relying only on ubiquitous web-browsers, eliminating the need for specialized software. It accommodates a wide range of users with intuitive user interfaces that require little or no training to quickly obtain and visualize data. Currently, PHOENIX is actively hosting three applications focused on groundwater monitoring, groundwater clean-up performance reporting, and in-tank monitoring. PHOENIX-based applications are being used to streamline investigative and analytical processes at Hanford, saving time and money. But more importantly, by integrating previously isolated datasets and developing relevant visualization and analysis tools, PHOENIX applications are enabling DOE to discover new correlations hidden in legacy data, allowing them to more effectively address complex issues at Hanford.

  5. Quick start user%3CU%2B2019%3Es guide for the PATH/AWARE decision support system.

    SciTech Connect (OSTI)

    Knowlton, Robert G.; Melton, Brad Joseph; Einfeld, Wayne; Tucker, Mark D; Franco, David Oliver; Yang, Lynn I.

    2013-06-01

    The Prioritization Analysis Tool for All-Hazards/Analyzer for Wide Area Restoration Effectiveness (PATH/AWARE) software system, developed by Sandia National Laboratories, is a comprehensive decision support tool designed to analyze situational awareness, as well as response and recovery actions, following a wide-area release of chemical, biological or radiological materials. The system provides capability to prioritize critical infrastructure assets and services for restoration. It also provides a capability to assess resource needs (e.g., number of sampling teams, laboratory capacity, decontamination units, etc.), timelines for consequence management activities, and costs. PATH/AWARE is a very comprehensive tool set with a considerable amount of database information managed through a Microsoft SQL (Structured Query Language) database engine, a Geographical Information System (GIS) engine that provides comprehensive mapping capabilities, as well as comprehensive decision logic to carry out the functional aspects of the tool set. This document covers the basic installation and operation of the PATH/AWARE tool in order to give the user enough information to start using the tool. A companion users manual is under development with greater specificity of the PATH/AWARE functionality.

  6. Open-Source GIS

    SciTech Connect (OSTI)

    Vatsavai, Raju; Burk, Thomas E; Lime, Steve

    2012-01-01

    The components making up an Open Source GIS are explained in this chapter. A map server (Sect. 30.1) can broadly be defined as a software platform for dynamically generating spatially referenced digital map products. The University of Minnesota MapServer (UMN Map Server) is one such system. Its basic features are visualization, overlay, and query. Section 30.2 names and explains many of the geospatial open source libraries, such as GDAL and OGR. The other libraries are FDO, JTS, GEOS, JCS, MetaCRS, and GPSBabel. The application examples include derived GIS-software and data format conversions. Quantum GIS, its origin and its applications explained in detail in Sect. 30.3. The features include a rich GUI, attribute tables, vector symbols, labeling, editing functions, projections, georeferencing, GPS support, analysis, and Web Map Server functionality. Future developments will address mobile applications, 3-D, and multithreading. The origins of PostgreSQL are outlined and PostGIS discussed in detail in Sect. 30.4. It extends PostgreSQL by implementing the Simple Feature standard. Section 30.5 details the most important open source licenses such as the GPL, the LGPL, the MIT License, and the BSD License, as well as the role of the Creative Commons.

  7. Low-Level Waste Forum notes and summary reports for 1994. Volume 9, Number 4, July 1994

    SciTech Connect (OSTI)

    1994-07-01

    This issue includes the following articles: Federal Facility Compliance Act Task Force forms mixed waste workgroup; Illinois Department of Nuclear Safety considers construction of centralized storage facility; Midwest Commission agrees on capacity limit, advisory committee; EPA responds to California site developer`s queries regarding application of air pollutant standards; county-level disqualification site screening of Pennsylvania complete; Texas Compact legislation introduced in US Senate; Generators ask court to rule in their favor on surcharge rebates lawsuit; Vermont authority and Battelle settle wetlands dispute; Eighth Circuit affirms decision in Nebraska community consent lawsuit; Nebraska court dismisses action filed by Boyd County local monitoring committee; NC authority, Chem-Nuclear, and Stowe exonerated; Senator Johnson introduces legislation to transfer Ward Valley site; Representative Dingell writes to Clinton regarding disposal of low-level radioactive waste; NAS committee on California site convenes; NRC to improve public petition process; NRC releases draft proposed rule on criteria for decontamination and closure of NRC-licensed facilities; and EPA names first environmental justice federal advisory council.

  8. Computational Cell Environment: A Problem Solving Environment for integrating diverse biological data

    SciTech Connect (OSTI)

    Klicker, Kyle R.; Singhal, Mudita; Stephan, Eric G.; Trease, Lynn L.; Gracio, Deborah K.

    2004-06-22

    Biologists and bioinformaticists face the ever-increasing challenge of managing large datasets queried from diverse data sources. Genomics and proteomics databases such as the National Center for Biotechnology (NCBI), Kyoto Encyclopedia of Genes and Genomes (KEGG), and the European Molecular Biology Laboratory (EMBL) are becoming the standard biological data department stores that biologists visit on a regular basis to obtain the supplies necessary for conducting their research. However, much of the data that biologists retrieve from these databases needs to be further managed and organized in a meaningful way so that the researcher can focus on the problem that they are trying to investigate and share their data and findings with other researchers. We are working towards developing a problem-solving environment called the Computational Cell Environment (CCE) that provides connectivity to these diverse data stores and provides data retrieval, management, and analysis through all aspects of biological study. In this paper we discuss the system and database design of CCE. We also outline a few problems encountered at various stages of its development and the design decisions taken to resolve them.

  9. Distributed Merge Trees

    SciTech Connect (OSTI)

    Morozov, Dmitriy; Weber, Gunther

    2013-01-08

    Improved simulations and sensors are producing datasets whose increasing complexity exhausts our ability to visualize and comprehend them directly. To cope with this problem, we can detect and extract significant features in the data and use them as the basis for subsequent analysis. Topological methods are valuable in this context because they provide robust and general feature definitions. As the growth of serial computational power has stalled, data analysis is becoming increasingly dependent on massively parallel machines. To satisfy the computational demand created by complex datasets, algorithms need to effectively utilize these computer architectures. The main strength of topological methods, their emphasis on global information, turns into an obstacle during parallelization. We present two approaches to alleviate this problem. We develop a distributed representation of the merge tree that avoids computing the global tree on a single processor and lets us parallelize subsequent queries. To account for the increasing number of cores per processor, we develop a new data structure that lets us take advantage of multiple shared-memory cores to parallelize the work on a single node. Finally, we present experiments that illustrate the strengths of our approach as well as help identify future challenges.

  10. A representation for efficient temporal reasoning

    SciTech Connect (OSTI)

    Delgrande, J.P.; Gupta, A.

    1996-12-31

    It has been observed that the temporal reasoning component in a knowledge-based system is frequently a bottleneck. We investigate here a class of graphs appropriate for an interesting class of temporal domains and for which very efficient algorithms for reasoning are obtained, that of series-parallel graphs. These graphs can be used for example to model process execution, as well as various planning or scheduling activities. Events are represented by nodes of a graph and relationships are represented by edges labeled by {le} or <. Graphs are composed using a sequence of series and parallel steps (recursively) on series-parallel graphs. We show that there is an O(n) time preprocessing algorithm that allows us to answer queries about the events in O(l) time. Our results make use of a novel embedding of the graphs on the plane that is of independent interest. Finally we argue that these results may be incorporated in general graphs representing temporal events by extending the approach of Gerevini and Schubert.

  11. Geometric reasoning about assembly tools

    SciTech Connect (OSTI)

    Wilson, R.H.

    1997-01-01

    Planning for assembly requires reasoning about various tools used by humans, robots, or other automation to manipulate, attach, and test parts and subassemblies. This paper presents a general framework to represent and reason about geometric accessibility issues for a wide variety of such assembly tools. Central to the framework is a use volume encoding a minimum space that must be free in an assembly state to apply a given tool, and placement constraints on where that volume must be placed relative to the parts on which the tool acts. Determining whether a tool can be applied in a given assembly state is then reduced to an instance of the FINDPLACE problem. In addition, the author presents more efficient methods to integrate the framework into assembly planning. For tools that are applied either before or after their target parts are mated, one method pre-processes a single tool application for all possible states of assembly of a product in polynomial time, reducing all later state-tool queries to evaluations of a simple expression. For tools applied after their target parts are mated, a complementary method guarantees polynomial-time assembly planning. The author presents a wide variety of tools that can be described adequately using the approach, and surveys tool catalogs to determine coverage of standard tools. Finally, the author describes an implementation of the approach in an assembly planning system and experiments with a library of over one hundred manual and robotic tools and several complex assemblies.

  12. Global prevalence and distribution of genes and microorganisms involved in mercury methylation

    DOE Public Access Gateway for Energy & Science Beta (PAGES Beta)

    Podar, Mircea; Gilmour, C. C.; Brandt, Craig C.; Soren, Allyson; Brown, Steven D.; Crable, Bryan R.; Palumbo, Anthony Vito; Somenahally, Anil C.; Elias, Dwayne A.

    2015-01-01

    Mercury methylation produces the neurotoxic, highly bioaccumulative methylmercury (MeHg). Recent identification of the methylation genes (hgcAB) provides the foundation for broadly evaluating microbial Hg-methylation potential in nature without making explicit rate measurements. We first queried hgcAB diversity and distribution in all available microbial metagenomes, encompassing most environments. The genes were found in nearly all anaerobic, but not in aerobic, environments including oxygenated layers of the open ocean. Critically, hgcAB was effectively absent in ~1500 human microbiomes, suggesting a low risk of endogenous MeHg production. New potential methylation habitats were identified, including invertebrate guts, thawing permafrost, coastal dead zones, soils, sediments,more » and extreme environments, suggesting multiple routes for MeHg entry into food webs. Several new taxonomic groups potentially capable of Hg-methylation emerged, including lineages having no cultured representatives. We then begin to address long-standing evolutionary questions about Hg-methylation and ancient carbon fixation mechanisms while generating a new global view of Hg-methylation potential.« less

  13. CERCLA {section}103 and EPCRA {section}304 Release Notification Requirements update

    SciTech Connect (OSTI)

    Not Available

    1995-01-01

    This guidance document updates and clarifies information provided in an earlier guidance document published by the US Environmental Protection Agency (EPA) entitled Guidance for Federal Facilities on Release Notification Requirements under CERCLA and SARA Title III (EPA 9360.7-06; November 1990). Since publication of that earlier guidance document, several significant events have occurred that affect the reporting obligations of facilities owned or operated by the Department of Energy (DOE), including the publication of Executive Order 12856--Federal Compliance with Right-to-Know Laws and Pollution Prevention Requirements--and a rejection by the US Court of Appeals of EPA`s interpretation of the term release into the environment. In preparing this guidance document, the Office of Environmental Policy and Assistance, RCRA/CERCLA Division (EH-413), has documented responses to queries from DOE field elements on CERCLA and EPCRA release reporting requirements, as well as incorporating those Questions and Answers from the previous document that remain germane to DOE`s reporting obligations under CERCLA and EPCRA.

  14. Understanding the Complexities of Subnational Incentives in Supporting a National Market for Distributed Photovoltaics

    SciTech Connect (OSTI)

    Bush, B.; Doris, E.; Getman, D.

    2014-09-01

    Subnational policies pertaining to photovoltaic (PV) systems have increased in volume in recent years and federal incentives are set to be phased out over the next few. Understanding how subnational policies function within and across jurisdictions, thereby impacting PV market development, informs policy decision making. This report was developed for subnational policy-makers and researchers in order to aid the analysis on the function of PV system incentives within the emerging PV deployment market. The analysis presented is based on a 'logic engine,' a database tool using existing state, utility, and local incentives allowing users to see the interrelationships between PV system incentives and parameters, such as geographic location, technology specifications, and financial factors. Depending on how it is queried, the database can yield insights into which combinations of incentives are available and most advantageous to the PV system owner or developer under particular circumstances. This is useful both for individual system developers to identify the most advantageous incentive packages that they qualify for as well as for researchers and policymakers to better understand the patch work of incentives nationwide as well as how they drive the market.

  15. Towards the Integration of APECS and VE-Suite for Virtual Power Plant Co-Simulation

    SciTech Connect (OSTI)

    Zitney, S.E.; McCorkle, D.; Yang, C.; Jordan, T.; Swensen, D.; Bryden, M.

    2007-05-01

    Process modeling and simulation tools are widely used for the design and operation of advanced power generation systems. These tools enable engineers to solve the critical process systems engineering problems that arise throughout the lifecycle of a power plant, such as designing a new process, troubleshooting a process unit or optimizing operations of the full process. To analyze the impact of complex thermal and fluid flow phenomena on overall power plant performance, the Department of Energy’s (DOE) National Energy Technology Laboratory (NETL) has developed the Advanced Process Engineering Co-Simulator (APECS). The APECS system is an integrated software suite that combines process simulation (e.g., Aspen Plus) and high-fidelity equipment simulations such as those based on computational fluid dynamics (CFD), together with advanced analysis capabilities including case studies, sensitivity analysis, stochastic simulation for risk/uncertainty analysis, and multi-objective optimization. In this paper we discuss the initial phases of the integration of the APECS system with the immersive and interactive virtual engineering software, VE-Suite, developed at Iowa State University and Ames Laboratory. VE-Suite uses the ActiveX (OLE Automation) controls in the Aspen Plus process simulator wrapped by the CASI library developed by Reaction Engineering International to run process/CFD co-simulations and query for results. This integration represents a necessary step in the development of virtual power plant co-simulations that will ultimately reduce the time, cost, and technical risk of developing advanced power generation systems.

  16. NATIONAL GEODATABASE OF TIDAL STREAM POWER RESOURCE IN USA

    SciTech Connect (OSTI)

    Smith, Brennan T; Neary, Vincent S; Stewart, Kevin M

    2012-01-01

    A geodatabase of tidal constituents is developed to present the regional assessment of tidal stream power resource in the USA. Tidal currents are numerically modeled with the Regional Ocean Modeling System (ROMS) and calibrated with the available measurements of tidal current speeds and water level surfaces. The performance of the numerical model in predicting the tidal currents and water levels is assessed by an independent validation. The geodatabase is published on a public domain via a spatial database engine with interactive tools to select, query and download the data. Regions with the maximum average kinetic power density exceeding 500 W/m2 (corresponding to a current speed of ~1 m/s), total surface area larger than 0.5 km2 and depth greater than 5 m are defined as hotspots and documented. The regional assessment indicates that the state of Alaska (AK) has the largest number of locations with considerably high kinetic power density, followed by, Maine (ME), Washington (WA), Oregon (OR), California (CA), New Hampshire (NH), Massachusetts (MA), New York (NY), New Jersey (NJ), North and South Carolina (NC, SC), Georgia (GA), and Florida (FL).

  17. Materials Databases Infrastructure Constructed by First Principles Calculations: A Review

    DOE Public Access Gateway for Energy & Science Beta (PAGES Beta)

    Lin, Lianshan

    2015-10-13

    The First Principles calculations, especially the calculation based on High-Throughput Density Functional Theory, have been widely accepted as the major tools in atom scale materials design. The emerging super computers, along with the powerful First Principles calculations, have accumulated hundreds of thousands of crystal and compound records. The exponential growing of computational materials information urges the development of the materials databases, which not only provide unlimited storage for the daily increasing data, but still keep the efficiency in data storage, management, query, presentation and manipulation. This review covers the most cutting edge materials databases in materials design, and their hotmore » applications such as in fuel cells. By comparing the advantages and drawbacks of these high-throughput First Principles materials databases, the optimized computational framework can be identified to fit the needs of fuel cell applications. The further development of high-throughput DFT materials database, which in essence accelerates the materials innovation, is discussed in the summary as well.« less

  18. A Scalable Monitoring for the CMS Filter Farm Based on Elasticsearch

    SciTech Connect (OSTI)

    Andre, J.M.; et al.

    2015-12-23

    A flexible monitoring system has been designed for the CMS File-based Filter Farm making use of modern data mining and analytics components. All the metadata and monitoring information concerning data flow and execution of the HLT are generated locally in the form of small documents using the JSON encoding. These documents are indexed into a hierarchy of elasticsearch (es) clusters along with process and system log information. Elasticsearch is a search server based on Apache Lucene. It provides a distributed, multitenant-capable search and aggregation engine. Since es is schema-free, any new information can be added seamlessly and the unstructured information can be queried in non-predetermined ways. The leaf es clusters consist of the very same nodes that form the Filter Farm thus providing natural horizontal scaling. A separate central” es cluster is used to collect and index aggregated information. The fine-grained information, all the way to individual processes, remains available in the leaf clusters. The central es cluster provides quasi-real-time high-level monitoring information to any kind of client. Historical data can be retrieved to analyse past problems or correlate them with external information. We discuss the design and performance of this system in the context of the CMS DAQ commissioning for LHC Run 2.

  19. Feature-based Analysis of Plasma-based Particle Acceleration Data

    SciTech Connect (OSTI)

    Ruebel, Oliver; Geddes, Cameron G.R.; Chen, Min; Cormier-Michel, Estelle; Bethel, E. Wes

    2013-07-05

    Plasma-based particle accelerators can produce and sustain thousands of times stronger acceleration fields than conventional particle accelerators, providing a potential solution to the problem of the growing size and cost of conventional particle accelerators. To facilitate scientific knowledge discovery from the ever growing collections of accelerator simulation data generated by accelerator physicists to investigate next-generation plasma-based particle accelerator designs, we describe a novel approach for automatic detection and classification of particle beams and beam substructures due to temporal differences in the acceleration process, here called acceleration features. The automatic feature detection in combination with a novel visualization tool for fast, intuitive, query-based exploration of acceleration features enables an effective top-down data exploration process, starting from a high-level, feature-based view down to the level of individual particles. We describe the application of our analysis in practice to analyze simulations of single pulse and dual and triple colliding pulse accelerator designs, and to study the formation and evolution of particle beams, to compare substructures of a beam and to investigate transverse particle loss.

  20. Finding Text Information in the Ocean of Electronic Documents

    SciTech Connect (OSTI)

    Medvick, Patricia A.; Calapristi, Augustin J.

    2003-02-05

    Information management in natural resources has become an overwhelming task. A massive amount of electronic documents and data is now available for creating informed decisions. The problem is finding the relevant information to support the decision-making process. Determining gaps in knowledge in order to propose new studies or to determine which proposals to fund for maximum potential is a time-consuming and difficult task. Additionally, available data stores are increasing in complexity; they now may include not only text and numerical data, but also images, sounds, and video recordings. Information visualization specialists at Pacific Northwest National Laboratory (PNNL) have software tools for exploring electronic data stores and for discovering and exploiting relationships within data sets. These provide capabilities for unstructured text explorations, the use of data signatures (a compact format for the essence of a set of scientific data) for visualization (Wong et al 2000), visualizations for multiple query results (Havre et al. 2001), and others (http://www.pnl.gov/infoviz ). We will focus on IN-SPIRE, a MS Windows vision of PNNL’s SPIRE (Spatial Paradigm for Information Retrieval and Exploration). IN-SPIRE was developed to assist information analysts find and discover information in huge masses of text documents.

  1. GROK

    Energy Science and Technology Software Center (OSTI)

    2006-02-24

    GROK is web based Internet Protocol (IP) search tool designed to help the user find and analyze network sessions in close to real time (5 minute). It reliew on the output generated by a packet capture and session summary tool called BAG. The bag program runs on a linux system, and continuously generates 5 minute full packet capture ILIBPCAP files, Internet session summary files, and interface statistic files, round-robin, over a period limited to themore » amount of disc storage available to the system. In the LANL case, an 8 terabyte file system accomodates seven days of data (most of the time). Summary information, such as top 20 outgoing and incoming network services (such as www/tcp or 161/udp) along with network interface statistics which indicate the health of the capture system are plotted every 5 minutes for display by the GROK web server. The grok home page presents the analyst with a set of search criteia used to query the information being collected by the bag program. Since the information ultimately resides in "pcap" files, other pcap aware programs such as bro ethereal, nosehair, smacqq, snort, and tcpdump have been incorporated into groks web interface. Clickable documentation is available for each search criteria« less

  2. CORE (Common Operating Response Environment) Software Technology Suite

    SciTech Connect (OSTI)

    Gelston, Gariann; Rohlfing, Kerrie

    2015-05-26

    Agencies that oversee complex, multi-stakeholder programs need efficient, secure ways to link people and knowledge within and across organizations. The Common Operating Response Environment (CORE), a software suite developed by PNNL researchers does just that. The CORE tool—which is customizable for a multitude of uses—facilitates situational awareness by integrating diverse data streams without the need to reformat them, summarizing that information, and providing users with the information they need to rapidly understand and appropriately respond to situations. It is mobile device-ready, has a straightforward interface for ease of use across organizations and skill sets, and is incredibly configurable to the needs of each specific user, whether they require data summaries for high-level decision makers or tactical maps, operational data, or weather information for responders in the field. Information can be input into CORE and queried in a variety of ways—using customized forms, reports, visuals, or other organizational templates—according to the needs of each user’s organization, teams, and business processes. CORE data forms, for instance, could be accessed and used in real-time to capture information about vessels being inspected for nuclear material.

  3. Final project report

    SciTech Connect (OSTI)

    Nitin S. Baliga and Leroy Hood

    2008-11-12

    The proposed overarching goal for this project was the following: Data integration, simulation and visualization will facilitate metabolic and regulatory network prediction, exploration, and formulation of hypotheses. We stated three specific aims to achieve the overarching goal of this project: (1) Integration of multiple levels of information such as mRNA and protein levels, predicted protein-protein interactions/associations and gene function will enable construction of models describing environmental response and dynamic behavior. (2) Flexible tools for network inference will accelerate our understanding of biological systems. (3) Flexible exploration and queries of model hypotheses will provide focus and reveal novel dependencies. The underlying philosophy of these proposed aims is that an iterative cycle of experiments, experimental design, and verification will lead to a comprehensive and predictive model that will shed light on systems level mechanisms involved in responses elicited by living systems upon sensing a change in their environment. In the previous years report we demonstrated considerable progress in development of data standards, regulatory network inference and data visualization and exploration. We are pleased to report that several manuscripts describing these procedures have been published in top international peer reviewed journals including Genome Biology, PNAS, and Cell. The abstracts of these manuscripts are given and they summarize our accomplishments in this project.

  4. Method and apparatus for biological sequence comparison

    DOE Patents [OSTI]

    Marr, Thomas G.; Chang, William I-Wei

    1997-01-01

    A method and apparatus for comparing biological sequences from a known source of sequences, with a subject (query) sequence. The apparatus takes as input a set of target similarity levels (such as evolutionary distances in units of PAM), and finds all fragments of known sequences that are similar to the subject sequence at each target similarity level, and are long enough to be statistically significant. The invention device filters out fragments from the known sequences that are too short, or have a lower average similarity to the subject sequence than is required by each target similarity level. The subject sequence is then compared only to the remaining known sequences to find the best matches. The filtering member divides the subject sequence into overlapping blocks, each block being sufficiently large to contain a minimum-length alignment from a known sequence. For each block, the filter member compares the block with every possible short fragment in the known sequences and determines a best match for each comparison. The determined set of short fragment best matches for the block provide an upper threshold on alignment values. Regions of a certain length from the known sequences that have a mean alignment value upper threshold greater than a target unit score are concatenated to form a union. The current block is compared to the union and provides an indication of best local alignment with the subject sequence.

  5. The Wyodak-Anderson coal assessment, Powder River Basin, Wyoming and Montana -- An ArcView project

    SciTech Connect (OSTI)

    Flores, R.M.; Gunther, G.; Ochs, A.; Ellis, M.E.; Stricker, G.D.; Bader, L.R.

    1998-12-31

    In 1997, more than 305 million short tons of clean and compliant coal were produced from the Wyodak-Anderson and associated coal beds and zones of the Paleocene Fort Union Formation in the Powder River Basin, Wyoming and Montana. To date, all coal produced from the Wyodak-Anderson, which averages 0.47 percent sulfur and 6.44 percent ash, has met regulatory compliance standards. Twenty-eight percent of the total US coal production in 1997 was from the Wyodak-Anderson coal. Based on the current consumption rates and forecast by the Energy Information Administration (1996), the Wyodak-Anderson coal is projected to produce 413 million short tons by the year 2016. In addition, this coal deposit as well as other Fort Union coals have recently been targeted for exploration and development of methane gas. New US Geological Survey (USGS) digital products could provide valuable assistance in future mining and gas development in the Powder River Basin. An interactive format, with querying tools, using ArcView software will display the digital products of the resource assessment of Wyodak-Anderson coal, a part of the USGS National Coal Resource Assessment of the Powder River Basin. This ArcView project includes coverages of the data point distribution; land use; surface and subsurface ownerships; coal geology, stratigraphy, quality and geochemistry; and preliminary coal resource calculations. These coverages are displayed as map views, cross sections, tables, and charts.

  6. Data Intensive Architecture for Scalable Cyber Analytics

    SciTech Connect (OSTI)

    Olsen, Bryan K.; Johnson, John R.; Critchlow, Terence J.

    2011-12-19

    Cyber analysts are tasked with the identification and mitigation of network exploits and threats. These compromises are difficult to identify due to the characteristics of cyber communication, the volume of traffic, and the duration of possible attack. In this paper, we describe a prototype implementation designed to provide cyber analysts an environment where they can interactively explore a months worth of cyber security data. This prototype utilized On-Line Analytical Processing (OLAP) techniques to present a data cube to the analysts. The cube provides a summary of the data, allowing trends to be easily identified as well as the ability to easily pull up the original records comprising an event of interest. The cube was built using SQL Server Analysis Services (SSAS), with the interface to the cube provided by Tableau. This software infrastructure was supported by a novel hardware architecture comprising a Netezza TwinFin for the underlying data warehouse and a cube server with a FusionIO drive hosting the data cube. We evaluated this environment on a months worth of artificial, but realistic, data using multiple queries provided by our cyber analysts. As our results indicate, OLAP technology has progressed to the point where it is in a unique position to provide novel insights to cyber analysts, as long as it is supported by an appropriate data intensive architecture.

  7. Review and evaluation of paleohydrologic methodologies

    SciTech Connect (OSTI)

    Foley, M.G.; Zimmerman, D.A.; Doesburg, J.M.; Thorne, P.D.

    1982-12-01

    A literature review was conducted to identify methodologies that could be used to interpret paleohydrologic environments. Paleohydrology is the study of past hydrologic systems or of the past behavior of an existing hydrologic system. The purpose of the review was to evaluate how well these methodologies could be applied to the siting of low-level radioactive waste facilities. The computer literature search queried five bibliographical data bases containing over five million citations of technical journals, books, conference papers, and reports. Two data-base searches (United States Geological Survey - USGS) and a manual search were also conducted. The methodologies were examined for data requirements and sensitivity limits. Paleohydrologic interpretations are uncertain because of the effects of time on hydrologic and geologic systems and because of the complexity of fluvial systems. Paleoflow determinations appear in many cases to be order-of-magnitude estimates. However, the methodologies identified in this report mitigate this uncertainty when used collectively as well as independently. That is, the data from individual methodologies can be compared or combined to corroborate hydrologic predictions. In this manner, paleohydrologic methodologies are viable tools to assist in evaluating the likely future hydrology of low-level radioactive waste sites.

  8. GENOME-ENABLED DISCOVERY OF CARBON SEQUESTRATION GENES IN POPLAR

    SciTech Connect (OSTI)

    DAVIS J M

    2007-10-11

    Plants utilize carbon by partitioning the reduced carbon obtained through photosynthesis into different compartments and into different chemistries within a cell and subsequently allocating such carbon to sink tissues throughout the plant. Since the phytohormones auxin and cytokinin are known to influence sink strength in tissues such as roots (Skoog & Miller 1957, Nordstrom et al. 2004), we hypothesized that altering the expression of genes that regulate auxin-mediated (e.g., AUX/IAA or ARF transcription factors) or cytokinin-mediated (e.g., RR transcription factors) control of root growth and development would impact carbon allocation and partitioning belowground (Fig. 1 - Renewal Proposal). Specifically, the ARF, AUX/IAA and RR transcription factor gene families mediate the effects of the growth regulators auxin and cytokinin on cell expansion, cell division and differentiation into root primordia. Invertases (IVR), whose transcript abundance is enhanced by both auxin and cytokinin, are critical components of carbon movement and therefore of carbon allocation. Thus, we initiated comparative genomic studies to identify the AUX/IAA, ARF, RR and IVR gene families in the Populus genome that could impact carbon allocation and partitioning. Bioinformatics searches using Arabidopsis gene sequences as queries identified regions with high degrees of sequence similarities in the Populus genome. These Populus sequences formed the basis of our transgenic experiments. Transgenic modification of gene expression involving members of these gene families was hypothesized to have profound effects on carbon allocation and partitioning.

  9. Open Research Challenges with Big Data - A Data-Scientist s Perspective

    SciTech Connect (OSTI)

    Sukumar, Sreenivas R

    2015-01-01

    In this paper, we discuss data-driven discovery challenges of the Big Data era. We observe that recent innovations in being able to collect, access, organize, integrate, and query massive amounts of data from a wide variety of data sources have brought statistical data mining and machine learning under more scrutiny and evaluation for gleaning insights from the data than ever before. In that context, we pose and debate the question - Are data mining algorithms scaling with the ability to store and compute? If yes, how? If not, why not? We survey recent developments in the state-of-the-art to discuss emerging and outstanding challenges in the design and implementation of machine learning algorithms at scale. We leverage experience from real-world Big Data knowledge discovery projects across domains of national security, healthcare and manufacturing to suggest our efforts be focused along the following axes: (i) the data science challenge - designing scalable and flexible computational architectures for machine learning (beyond just data-retrieval); (ii) the science of data challenge the ability to understand characteristics of data before applying machine learning algorithms and tools; and (iii) the scalable predictive functions challenge the ability to construct, learn and infer with increasing sample size, dimensionality, and categories of labels. We conclude with a discussion of opportunities and directions for future research.

  10. Concept of Operations for Collaboration and Discovery from Big Data Across Enterprise Data Warehouses

    SciTech Connect (OSTI)

    Olama, Mohammed M; Nutaro, James J; Sukumar, Sreenivas R; McNair, Wade

    2013-01-01

    The success of data-driven business in government, science, and private industry is driving the need for seamless integration of intra and inter-enterprise data sources to extract knowledge nuggets in the form of correlations, trends, patterns and behaviors previously not discovered due to physical and logical separation of datasets. Today, as volume, velocity, variety and complexity of enterprise data keeps increasing, the next generation analysts are facing several challenges in the knowledge extraction process. Towards addressing these challenges, data-driven organizations that rely on the success of their analysts have to make investment decisions for sustainable data/information systems and knowledge discovery. Options that organizations are considering are newer storage/analysis architectures, better analysis machines, redesigned analysis algorithms, collaborative knowledge management tools, and query builders amongst many others. In this paper, we present a concept of operations for enabling knowledge discovery that data-driven organizations can leverage towards making their investment decisions. We base our recommendations on the experience gained from integrating multi-agency enterprise data warehouses at the Oak Ridge National Laboratory to design the foundation of future knowledge nurturing data-system architectures.

  11. RAG-3D: A search tool for RNA 3D substructures

    DOE Public Access Gateway for Energy & Science Beta (PAGES Beta)

    Zahran, Mai; Sevim Bayrak, Cigdem; Elmetwaly, Shereef; Schlick, Tamar

    2015-08-24

    In this study, to address many challenges in RNA structure/function prediction, the characterization of RNA's modular architectural units is required. Using the RNA-As-Graphs (RAG) database, we have previously explored the existence of secondary structure (2D) submotifs within larger RNA structures. Here we present RAG-3D—a dataset of RNA tertiary (3D) structures and substructures plus a web-based search tool—designed to exploit graph representations of RNAs for the goal of searching for similar 3D structural fragments. The objects in RAG-3D consist of 3D structures translated into 3D graphs, cataloged based on the connectivity between their secondary structure elements. Each graph is additionally describedmore » in terms of its subgraph building blocks. The RAG-3D search tool then compares a query RNA 3D structure to those in the database to obtain structurally similar structures and substructures. This comparison reveals conserved 3D RNA features and thus may suggest functional connections. Though RNA search programs based on similarity in sequence, 2D, and/or 3D structural elements are available, our graph-based search tool may be advantageous for illuminating similarities that are not obvious; using motifs rather than sequence space also reduces search times considerably. Ultimately, such substructuring could be useful for RNA 3D structure prediction, structure/function inference and inverse folding.« less

  12. SIFTER search: a web server for accurate phylogeny-based protein function prediction

    DOE Public Access Gateway for Energy & Science Beta (PAGES Beta)

    Sahraeian, Sayed M.; Luo, Kevin R.; Brenner, Steven E.

    2015-05-15

    We are awash in proteins discovered through high-throughput sequencing projects. As only a minuscule fraction of these have been experimentally characterized, computational methods are widely used for automated annotation. Here, we introduce a user-friendly web interface for accurate protein function prediction using the SIFTER algorithm. SIFTER is a state-of-the-art sequence-based gene molecular function prediction algorithm that uses a statistical model of function evolution to incorporate annotations throughout the phylogenetic tree. Due to the resources needed by the SIFTER algorithm, running SIFTER locally is not trivial for most users, especially for large-scale problems. The SIFTER web server thus provides access tomore » precomputed predictions on 16 863 537 proteins from 232 403 species. Users can explore SIFTER predictions with queries for proteins, species, functions, and homologs of sequences not in the precomputed prediction set. Lastly, the SIFTER web server is accessible at http://sifter.berkeley.edu/ and the source code can be downloaded.« less

  13. SIFTER search: a web server for accurate phylogeny-based protein function prediction

    SciTech Connect (OSTI)

    Sahraeian, Sayed M.; Luo, Kevin R.; Brenner, Steven E.

    2015-05-15

    We are awash in proteins discovered through high-throughput sequencing projects. As only a minuscule fraction of these have been experimentally characterized, computational methods are widely used for automated annotation. Here, we introduce a user-friendly web interface for accurate protein function prediction using the SIFTER algorithm. SIFTER is a state-of-the-art sequence-based gene molecular function prediction algorithm that uses a statistical model of function evolution to incorporate annotations throughout the phylogenetic tree. Due to the resources needed by the SIFTER algorithm, running SIFTER locally is not trivial for most users, especially for large-scale problems. The SIFTER web server thus provides access to precomputed predictions on 16 863 537 proteins from 232 403 species. Users can explore SIFTER predictions with queries for proteins, species, functions, and homologs of sequences not in the precomputed prediction set. Lastly, the SIFTER web server is accessible at http://sifter.berkeley.edu/ and the source code can be downloaded.

  14. RAG-3D: A search tool for RNA 3D substructures

    SciTech Connect (OSTI)

    Zahran, Mai; Sevim Bayrak, Cigdem; Elmetwaly, Shereef; Schlick, Tamar

    2015-08-24

    In this study, to address many challenges in RNA structure/function prediction, the characterization of RNA's modular architectural units is required. Using the RNA-As-Graphs (RAG) database, we have previously explored the existence of secondary structure (2D) submotifs within larger RNA structures. Here we present RAG-3D—a dataset of RNA tertiary (3D) structures and substructures plus a web-based search tool—designed to exploit graph representations of RNAs for the goal of searching for similar 3D structural fragments. The objects in RAG-3D consist of 3D structures translated into 3D graphs, cataloged based on the connectivity between their secondary structure elements. Each graph is additionally described in terms of its subgraph building blocks. The RAG-3D search tool then compares a query RNA 3D structure to those in the database to obtain structurally similar structures and substructures. This comparison reveals conserved 3D RNA features and thus may suggest functional connections. Though RNA search programs based on similarity in sequence, 2D, and/or 3D structural elements are available, our graph-based search tool may be advantageous for illuminating similarities that are not obvious; using motifs rather than sequence space also reduces search times considerably. Ultimately, such substructuring could be useful for RNA 3D structure prediction, structure/function inference and inverse folding.

  15. Assessment of US shipbuilding current capability to build a commercial OTEC platform and a cold water pipe

    SciTech Connect (OSTI)

    Komelasky, M. C.

    1980-03-01

    Lowry and Hoffman Associates Inc. (LHA) performed for ORI an analysis of the shipbuilding requirements for constructing an OTEC plant, and the available shipyard assets which could fulfill these requirements. In addition, several shipyards were queried concerning their attitudes towards OTEC. In assessing the shipbuilding requirements for an OTEC plant, four different platform configurations were studied and four different designs of the cold water pipe (CWP) were examined. The platforms were: a concrete ship design proposed by Lockheed; concrete spar designs with internal heat exchangers (IHE) (Rosenblatt) and external heat exchangers (XHE) (Lockheed); and a steel ship design proposed by Gibbs and Cox. The types of materials examined for CWP construction were: steel, fiber reinforced plastic (FPR), elastomer, and concrete. The report is organized io three major discussion areas. All the construction requirements are synthesized for the four platforms and CWPs, and general comments are made concerning their availability in the US. Specific shipbuilders facilities are reviewed for their applicability to building an OTEC plant, an assessment of the shipyards general interest in the OTEC program is presented providing an insight into their nearterm commercial outlook. The method of determining this interest will depend largely on a risk analysis of the OTEC system. Also included are factors which may comprise this analysis, and a methodology to ascertain the risk. In the appendices, various shipyard specifications are presented, shipyard assessment matrices are given, graphs of various shipyard economic outlooks are provided, and definitions of the risk factors are listed. (WHK)

  16. Agent-based method for distributed clustering of textual information

    DOE Patents [OSTI]

    Potok, Thomas E [Oak Ridge, TN; Reed, Joel W [Knoxville, TN; Elmore, Mark T [Oak Ridge, TN; Treadwell, Jim N [Louisville, TN

    2010-09-28

    A computer method and system for storing, retrieving and displaying information has a multiplexing agent (20) that calculates a new document vector (25) for a new document (21) to be added to the system and transmits the new document vector (25) to master cluster agents (22) and cluster agents (23) for evaluation. These agents (22, 23) perform the evaluation and return values upstream to the multiplexing agent (20) based on the similarity of the document to documents stored under their control. The multiplexing agent (20) then sends the document (21) and the document vector (25) to the master cluster agent (22), which then forwards it to a cluster agent (23) or creates a new cluster agent (23) to manage the document (21). The system also searches for stored documents according to a search query having at least one term and identifying the documents found in the search, and displays the documents in a clustering display (80) of similarity so as to indicate similarity of the documents to each other.

  17. The Environmental Assessment Management modification of CADET

    Energy Science and Technology Software Center (OSTI)

    1996-05-01

    The original CADET system (finalized in September 1995 as version 1.3) is a data collection and transfer system developed for the Headquarters Air Force Space Command (HQAFSPC) Environmental Compliance Assessment and Management Program (ECAMP). The system was designed as a tool for ECAMP evaluators to use to enter compliance related data while in the field and to subsequently store, modify, sort, query, and print the data and to electronically transfer the data into the Airmore » Force''s Work Information Management System Environmental Subsystem (WIMSES). The original CADET system was designed to match the database structure of the WIMSES ECAMP module that came on-line in 1992. In June 1995, the Department of Defense issued The Environmental Assessment Management (TEAM) Guide and ECAMP Supplement to the TEAM Guide. These included changes to the type and amount of data collected during an ECAMP assessment. The WIMSES database structure was not modified to match the TEAM Guide; however, the need for collecting and storing the ECAMP data remained. The HQAFSC decided to modify the CADET system to incorporate the changes specified in the ECAMP Supplement and to convert the system from simply a data entry and transfer tool to a data entry and storage system to manage ECAMP findings in lieu of the WIMSES ECAMP module. The revised software is designated as version 2.0 and nicknamed TEAM CADET to distinguish it from the original CADET system.« less

  18. Analyzing product test data in a relational database using SAS software

    SciTech Connect (OSTI)

    Orman, J.L.

    1991-01-01

    SAS software is being used to analyze product test data stored in an INGRES relational database. The database has been implemented at Allied-Signal in Kansas City on a Digital Equipment Corporation (DEC) VAX computer. The INGRES application development has been a joint project between Sandia National Laboratories and Allied-Signal. Application screens have been developed so that the user can query the database for selected data. Fourth generation language procedures are used to retrieve all data requested. FORTRAN and VAX/VMS DCL (DIGITAL Control Language) procedures are invoked from the application to create SAS data sets and dynamically build SAS programs that are executed to build custom reports or graphically display the retrieved test data along with control and specification limits. A retrieval screen has also been developed which invokes SAS software to calculate the mean and standard deviation of the retrieved data. These parameters are passed back into the application for display and may then be used as an aid in setting new control limits for future test runs. Screens have been developed to provide an interface for the user to select from a library of SAS programs, edit the selected program, and run the program with a user-defined SAS data set as input. This paper will give a brief description of the application screens and provide details of how information is passed between the application and SAS programs.

  19. Global prevalence and distribution of genes and microorganisms involved in mercury methylation

    SciTech Connect (OSTI)

    Podar, Mircea; Gilmour, C. C.; Brandt, Craig C.; Soren, Allyson; Brown, Steven D.; Crable, Bryan R.; Palumbo, Anthony Vito; Somenahally, Anil C.; Elias, Dwayne A.

    2015-01-01

    Mercury methylation produces the neurotoxic, highly bioaccumulative methylmercury (MeHg). Recent identification of the methylation genes (hgcAB) provides the foundation for broadly evaluating microbial Hg-methylation potential in nature without making explicit rate measurements. We first queried hgcAB diversity and distribution in all available microbial metagenomes, encompassing most environments. The genes were found in nearly all anaerobic, but not in aerobic, environments including oxygenated layers of the open ocean. Critically, hgcAB was effectively absent in ~1500 human microbiomes, suggesting a low risk of endogenous MeHg production. New potential methylation habitats were identified, including invertebrate guts, thawing permafrost, coastal dead zones, soils, sediments, and extreme environments, suggesting multiple routes for MeHg entry into food webs. Several new taxonomic groups potentially capable of Hg-methylation emerged, including lineages having no cultured representatives. We then begin to address long-standing evolutionary questions about Hg-methylation and ancient carbon fixation mechanisms while generating a new global view of Hg-methylation potential.

  20. Method and apparatus for biological sequence comparison

    DOE Patents [OSTI]

    Marr, T.G.; Chang, W.I.

    1997-12-23

    A method and apparatus are disclosed for comparing biological sequences from a known source of sequences, with a subject (query) sequence. The apparatus takes as input a set of target similarity levels (such as evolutionary distances in units of PAM), and finds all fragments of known sequences that are similar to the subject sequence at each target similarity level, and are long enough to be statistically significant. The invention device filters out fragments from the known sequences that are too short, or have a lower average similarity to the subject sequence than is required by each target similarity level. The subject sequence is then compared only to the remaining known sequences to find the best matches. The filtering member divides the subject sequence into overlapping blocks, each block being sufficiently large to contain a minimum-length alignment from a known sequence. For each block, the filter member compares the block with every possible short fragment in the known sequences and determines a best match for each comparison. The determined set of short fragment best matches for the block provide an upper threshold on alignment values. Regions of a certain length from the known sequences that have a mean alignment value upper threshold greater than a target unit score are concatenated to form a union. The current block is compared to the union and provides an indication of best local alignment with the subject sequence. 5 figs.

  1. Xgrid admin guide

    SciTech Connect (OSTI)

    Strauss, Charlie E M

    2010-01-01

    Xgrid, with a capital-X is the name for Apple's grid computing system. With a lower case x, xgrid is the name of the command line utility that clients can use, among other ways, to submit jobs to a controller. An Xgrid divides into three logical components: Agent, Controller and Client. Client computers submit jobs (a set of tasks) they want run to a Controller computer. The Controller queues the Client jobs and distributes tasks to Agent computers. Agent computers run the tasks and report their output and status back to the controller where it is stored until deleted by the Client. The Clients can asynchronously query the controller about the status of a job and the results. Any OSX computer can be any of these. A single mac can be more than one: it's possible to be Agent, Controller and Client at the same time. There is one Controller per Grid. Clients can submit jobs to Controllers of different grids. Agents can work for more than one grid. Xgrid's setup has a pleasantly small palette of choices. The first two decisions to make are the kind of authentication & authorization to use and if a shared file system is needed. A shared file system that all the agents can access can be very beneficial for many computing problems, but it is not appropriate for every network.

  2. Implementation of a laboratory information management system for environmental regulatory analyses

    SciTech Connect (OSTI)

    Spencer, W.A.; Aiken, H.B.; Spatz, T.L.; Miles, W.F.; Griffin, J.C.

    1993-09-07

    The Savannah River Technology Center created a second instance of its ORACLE based PEN LIMS to support site Environmental Restoration projects. The first instance of the database had been optimized for R&D support and did not implement rigorous sample tracking, verification, and holding times needed to support regulatory commitments. Much of the R&D instance was transferable such as the work control functions for backlog reports, work assignment sheets, and hazard communication support. A major enhancement of the regulatory LIMS was the addition of features to support a {open_quotes}standardized{close_quotes} electronic data format for environmental data reporting. The electronic format, called {open_quotes}AN92{close_quotes}, was developed by the site environmental monitoring organization and applies to both onsite and offsite environmental analytical contracts. This format incorporates EPA CLP data validation codes as well as details holding time and analytical result reporting requirements. The authors support this format by using special SQL queries to the database. The data is then automatically transferred to the environmental databases for trending and geological mapping.

  3. In-Situ Microphysics from the RACORO IOP

    DOE Data Explorer [Office of Scientific and Technical Information (OSTI)]

    McFarquhar, Greg

    These files were generated by Greg McFarquhar and Robert Jackson at the University of Illinois. Please contact mcfarq@atmos.uiuc.edu or rjackso2@atmos.uiuc.edu for more information or for assistance in interpreting the content of these files. We highly recommend that anyone wishing to use these files do so in a collaborative endeavor and we welcome queries and opportunities for collaboration. There are caveats associated with the use of the data which are difficult to thoroughly document and not all products for all time periods have been thoroughly examined. This is a value added data set of the best estimate of cloud microphysical parameters derived using data collected by the cloud microphysical probes installed on the Center for Interdisciplinary Remotely-Piloted Aircraft Studies (CIRPAS) Twin Otter during RACORO. These files contain best estimates of liquid size distributions N(D) in terms of droplet diameter D, liquid water content LWC, extinction of liquid drops beta, effective radius of cloud drops (re), total number concentration of droplets NT, and radar reflectivity factor Z at 1 second resolution.

  4. System description for DART (Decision Analysis for Remediation Technologies)

    SciTech Connect (OSTI)

    Nonte, J.; Bolander, T.; Nickelson, D.; Nielson, R.; Richardson, J.; Sebo, D.

    1997-09-01

    DART is a computer aided system populated with influence models to determine quantitative benefits derived by matching requirements and technologies. The DART database is populated with data from over 900 DOE sites from 10 Field Offices. These sites are either source terms, such as buried waste pits, or soil or groundwater contaminated plumes. The data, traceable to published documents, consists of site-specific data (contaminants, area, volume, depth, size, remedial action dates, site preferred remedial option), problems (e.g., offsite contaminant plume), and Site Technology Coordinating Group (STCG) need statements (also contained in the Ten-Year Plan). DART uses this data to calculate and derive site priorities, risk rankings, and site specific technology requirements. DART is also populated with over 900 industry and DOE SCFA technologies. Technology capabilities can be used to match technologies to waste sites based on the technology`s capability to meet site requirements and constraints. Queries may be used to access, sort, roll-up, and rank site data. Data roll-ups may be graphically displayed.

  5. Global disease monitoring and forecasting with Wikipedia

    SciTech Connect (OSTI)

    Generous, Nicholas; Fairchild, Geoffrey; Deshpande, Alina; Del Valle, Sara Y.; Priedhorsky, Reid; Salathé, Marcel

    2014-11-13

    Infectious disease is a leading threat to public health, economic stability, and other key social structures. Efforts to mitigate these impacts depend on accurate and timely monitoring to measure the risk and progress of disease. Traditional, biologically-focused monitoring techniques are accurate but costly and slow; in response, new techniques based on social internet data, such as social media and search queries, are emerging. These efforts are promising, but important challenges in the areas of scientific peer review, breadth of diseases and countries, and forecasting hamper their operational usefulness. We examine a freely available, open data source for this use: access logs from the online encyclopedia Wikipedia. Using linear models, language as a proxy for location, and a systematic yet simple article selection procedure, we tested 14 location-disease combinations and demonstrate that these data feasibly support an approach that overcomes these challenges. Specifically, our proof-of-concept yields models with up to 0.92, forecasting value up to the 28 days tested, and several pairs of models similar enough to suggest that transferring models from one location to another without re-training is feasible. Based on these preliminary results, we close with a research agenda designed to overcome these challenges and produce a disease monitoring and forecasting system that is significantly more effective, robust, and globally comprehensive than the current state of the art.

  6. Mercury-metadata data management system

    Energy Science and Technology Software Center (OSTI)

    2008-01-03

    Mercury is a federated metadata harvesting, search and retrieval tool based on both open source software and software developed at Oak Ridge National Laboratory. It was originally developed for NASA, USGS, and DOE. A major new version of Mercury (version 3.0) was developed during 2007 and released in early 2008. This Mercury 3.0 version provides orders of magnitude improvements in search speed, support for additional metadata formats, integration with Google Maps for spatial queries, facettedmore » type search, support for RSS delivery of search results, and ready customization to meet the needs of the multiple projects which use Mercury. For the end users, Mercury provides a single portal to very quickly search for data and information contained in disparate data management systems. It collects metadata and key data from contributing project servers distributed around the world and builds a centralized index. The Mercury search interfaces then allow the users to perform simple, fielded, spatial, and temporal searches across these metadata sources. This centralized repository of metadata with distributed data sources provides extremely fast search results to the user, while allowing data providers to advertise the availability of their data and maintain complete control and ownership of that data.« less

  7. Mercury Metadata Toolset

    Energy Science and Technology Software Center (OSTI)

    2009-09-08

    Mercury is a federated metadata harvesting, search and retrieval tool based on both open source software and software developed at Oak Ridge National Laboratory. It was originally developed for NASA, and the Mercury development consortium now includes funding from NASA, USGS, and DOE. A major new version of Mercury (version 3.0) was developed during 2007 and released in early 2008. This Mercury 3.0 version provides orders of magnitude improvements in search speed, support for additionalmore » metadata formats, integration with Google Maps for spatial queries, facetted type search, support for RSS delivery of search results, and ready customization to meet the needs of the multiple projects which use Mercury. For the end users, Mercury provides a single portal to very quickly search for data and information contained in disparate data management systems. It collects metadata and key data from contributing project servers distributed around the world and builds a centralized index. The Mercury search interfaces then allow the users to perform simple, fielded, spatial, and temporal searches across these metadata sources. This centralized repository of metadata with distributed data sources provides extremely fast search results to the user, while allowing data providers to advertise the availability of their data and maintain complete control and ownership of that data.« less

  8. System/subsystem specifications for the Worldwide Port System (WPS) Regional Integrated Cargo Database (ICDB)

    SciTech Connect (OSTI)

    Rollow, J.P.; Shipe, P.C.; Truett, L.F.; Faby, E.Z.; Fluker, J.; Grubb, J.; Hancock, B.R.; Ferguson, R.A.

    1995-11-20

    A system is being developed by the Military Traffic Management Command (MTMC) to provide data integration and worldwide management and tracking of surface cargo movements. The Integrated Cargo Database (ICDB) will be a data repository for the WPS terminal-level system, will be a primary source of queries and cargo traffic reports, will receive data from and provide data to other MTMC and non-MTMC systems, will provide capabilities for processing Advance Transportation Control and Movement Documents (ATCMDs), and will process and distribute manifests. This System/Subsystem Specifications for the Worldwide Port System Regional ICDB documents the system/subsystem functions, provides details of the system/subsystem analysis in order to provide a communication link between developers and operational personnel, and identifies interfaces with other systems and subsystems. It must be noted that this report is being produced near the end of the initial development phase of ICDB, while formal software testing is being done. Following the initial implementation of the ICDB system, maintenance contractors will be in charge of making changes and enhancing software modules. Formal testing and user reviews may indicate the need for additional software units or changes to existing ones. This report describes the software units that are components of this ICDB system as of August 1995.

  9. Global disease monitoring and forecasting with Wikipedia

    DOE Public Access Gateway for Energy & Science Beta (PAGES Beta)

    Generous, Nicholas; Fairchild, Geoffrey; Deshpande, Alina; Del Valle, Sara Y.; Priedhorsky, Reid; Salathé, Marcel

    2014-11-13

    Infectious disease is a leading threat to public health, economic stability, and other key social structures. Efforts to mitigate these impacts depend on accurate and timely monitoring to measure the risk and progress of disease. Traditional, biologically-focused monitoring techniques are accurate but costly and slow; in response, new techniques based on social internet data, such as social media and search queries, are emerging. These efforts are promising, but important challenges in the areas of scientific peer review, breadth of diseases and countries, and forecasting hamper their operational usefulness. We examine a freely available, open data source for this use: accessmore » logs from the online encyclopedia Wikipedia. Using linear models, language as a proxy for location, and a systematic yet simple article selection procedure, we tested 14 location-disease combinations and demonstrate that these data feasibly support an approach that overcomes these challenges. Specifically, our proof-of-concept yields models with up to 0.92, forecasting value up to the 28 days tested, and several pairs of models similar enough to suggest that transferring models from one location to another without re-training is feasible. Based on these preliminary results, we close with a research agenda designed to overcome these challenges and produce a disease monitoring and forecasting system that is significantly more effective, robust, and globally comprehensive than the current state of the art.« less

  10. In-Situ Microphysics from the RACORO IOP

    DOE Data Explorer [Office of Scientific and Technical Information (OSTI)]

    McFarquhar, Greg

    2013-11-08

    These files were generated by Greg McFarquhar and Robert Jackson at the University of Illinois. Please contact mcfarq@atmos.uiuc.edu or rjackso2@atmos.uiuc.edu for more information or for assistance in interpreting the content of these files. We highly recommend that anyone wishing to use these files do so in a collaborative endeavor and we welcome queries and opportunities for collaboration. There are caveats associated with the use of the data which are difficult to thoroughly document and not all products for all time periods have been thoroughly examined. This is a value added data set of the best estimate of cloud microphysical parameters derived using data collected by the cloud microphysical probes installed on the Center for Interdisciplinary Remotely-Piloted Aircraft Studies (CIRPAS) Twin Otter during RACORO. These files contain best estimates of liquid size distributions N(D) in terms of droplet diameter D, liquid water content LWC, extinction of liquid drops beta, effective radius of cloud drops (re), total number concentration of droplets NT, and radar reflectivity factor Z at 1 second resolution.

  11. Model Investigation of Temperature and Concentration Dependent Luminescence of Erbium-doped Tellurite Glasses

    SciTech Connect (OSTI)

    Ghoshal, S. K.; Sahar, M. R.; Rohani, M. S.; Tewari, H. S.

    2011-11-22

    Improving the up-conversion efficiency is the key issue in tellurite glasses. The quantum efficiency, radiative transition rate and lifetimes of excited states are greatly influenced by the optical properties of the host material, ligand field, multiphonon relaxation processes, impurities, temperature and concentration of erbium ions. We develop a comprehensive 4-level model to examine the radiative and nonradiative (NR) decay processes for the green ({sup 4}S{sub 3/2}{yields}{sup 4}I{sub 15/2}) and red ({sup 4}F{sub 9/2}{yields}{sup 4}I{sub 15/2}) emission over a temperature range of (10-340 K) and concentration range of (0.1-4.5 mol.%). Concentration dependent enhancement and thermal quenching of efficiency for up-conversion is investigated using the derived rate equations. These features are attributed to the NR energy transfer processes, trapped impurity effects, and thermal assisted hopping. The unusual nature of temperature and concentration dependent quenching effects for green and red emission is queries for further investigations. It is further suggested that to achieve higher infrared to visible up-converted efficiency in tellurite glasses the NR channels for energy and charge transfer by phonon and impurity mediated process has to be minimized. Our results on pump power dependent emission intensity, quantum efficiency, luminescence intensity, radiative lifetimes, and transition probabilities are in conformity with other experimental findings.

  12. Waste Isolation Pilot Plant (WIPP) Waste Information System (Public Access)

    DOE Data Explorer [Office of Scientific and Technical Information (OSTI)]

    The Waste Isolation Pilot Plant (WIPP) is a DOE facility located in the desert outside Carlsbad, New Mexico. Its mission is to safely dispose of defense-related transuranic radioactive waste. Disposal ôroomsö are carved out of the Permian Salt Formation deep below the desertÆs surface. The WIPP Waste Information Service (WWIS) was established in accordance with an Agreement between the United States Department of Energy and the New Mexico Environment Department, dated February 11, 2005, Docket Number HWB 04-07 (CO). The service provides information the containers emplaced at WIPP and the waste products they hold. The public may query by shipment number, location of waste stream or location of the container after it is placed at WIPP, date placed, and Haz Codes or other information about the waste stream profiles. For example, choosing the waste stream identified as ID-SDA-SLUDGE reveals that it may contain more than 20 chemical waste products, including arsenic, spent halogenated solvents, potassium cyanide, and chloroform. The system then tells you each numbered container that has this kind of sludge. Container data is available within 14 days after the containerÆs emplacement in the WIPP Repository.

  13. Field Trial of a Low-Cost, Distributed Plug Load Monitoring System

    SciTech Connect (OSTI)

    Auchter, B.; Cautley, D.; Ahl, D.; Earle, L.; Jin, X.

    2014-03-01

    Researchers have struggled to inventory and characterize the energy use profiles of the ever-growing category of so-called miscellaneous electric loads (MELs) because plug-load monitoring is cost-prohibitive to the researcher and intrusive to the homeowner. However, these data represent a crucial missing link to understanding how homes use energy. Detailed energy use profiles would enable the nascent automated home energy management (AHEM) industry to develop effective control algorithms that target consumer electronics and other plug loads. If utility and other efficiency programs are to incent AHEM devices, they need large-scale datasets that provide statistically meaningful justification of their investments by quantifying the aggregate energy savings achievable. To address this need, NREL researchers investigated a variety of plug-load measuring devices available commercially and tested them in the laboratory to identify the most promising candidates for field applications. This report centers around the lessons learned from a field validation of one proof-of-concept system, called Smartenit (formerly SimpleHomeNet). The system was evaluated based on the rate of successful data queries, reliability over a period of days to weeks, and accuracy. This system offers good overall performance when deployed with up to 10 end nodes in a residential environment, although deployment with more nodes and in a commercial environment is much less robust. NREL concludes that the current system is useful in selected field research projects, with the recommendation that system behavior is observed over time.

  14. Advanced cryogenics for cutting tools. Final report

    SciTech Connect (OSTI)

    Lazarus, L.J.

    1996-10-01

    The purpose of the investigation was to determine if cryogenic treatment improved the life and cost effectiveness of perishable cutting tools over other treatments or coatings. Test results showed that in five of seven of the perishable cutting tools tested there was no improvement in tool life. The other two tools showed a small gain in tool life, but not as much as when switching manufacturers of the cutting tool. The following conclusions were drawn from this study: (1) titanium nitride coatings are more effective than cryogenic treatment in increasing the life of perishable cutting tools made from all cutting tool materials, (2) cryogenic treatment may increase tool life if the cutting tool is improperly heat treated during its origination, and (3) cryogenic treatment was only effective on those tools made from less sophisticated high speed tool steels. As a part of a recent detailed investigation, four cutting tool manufacturers and two cutting tool laboratories were queried and none could supply any data to substantiate cryogenic treatment of perishable cutting tools.

  15. From Question Answering to Visual Exploration

    SciTech Connect (OSTI)

    McColgin, Dave W.; Gregory, Michelle L.; Hetzler, Elizabeth G.; Turner, Alan E.

    2006-08-11

    Research in Question Answering has focused on the quality of information retrieval or extraction using the metrics of precision and recall to judge success; these metrics drive toward finding the specific best answer(s) and are best supportive of a lookup type of search. These do not address the opportunity that users? natural language questions present for exploratory interactions. In this paper, we present an integrated Question Answering environment that combines a visual analytics tool for unstructured text and a state-of-the-art query expansion tool designed to compliment the cognitive processes associated with an information analysts work flow. Analysts are seldom looking for factoid answers to simple questions; their information needs are much more complex in that they may be interested in patterns of answers over time, conflicting information, and even related non-answer data may be critical to learning about a problem or reaching prudent conclusions. In our visual analytics tool, questions result in a comprehensive answer space that allows users to explore the variety within the answers and spot related information in the rest of the data. The exploratory nature of the dialog between the user and this system requires tailored evaluation methods that better address the evolving user goals and counter cognitive biases inherent to exploratory search tasks.

  16. Protein Structure Determination Using Protein Threading and Sparse NMR Data

    SciTech Connect (OSTI)

    Crawford, O.H.; Einstein, J.R.; Xu, D.; Xu, Y.

    1999-11-14

    It is well known that the NMR method for protein structure determination applies to small proteins and that its effectiveness decreases very rapidly as the molecular weight increases beyond about 30 kD. We have recently developed a method for protein structure determination that can fully utilize partial NMR data as calculation constraints. The core of the method is a threading algorithm that guarantees to find a globally optimal alignment between a query sequence and a template structure, under distance constraints specified by NMR/NOE data. Our preliminary tests have demonstrated that a small number of NMR/NOE distance restraints can significantly improve threading performance in both fold recognition and threading-alignment accuracy, and can possibly extend threading's scope of applicability from structural homologs to structural analogs. An accurate backbone structure generated by NMR-constrained threading can then provide a significant amount of structural information, equivalent to that provided by the NMR method with many NMR/NOE restraints; and hence can greatly reduce the amount of NMR data typically required for accurate structure determination. Our preliminary study suggests that a small number of NMR/NOE restraints may suffice to determine adequately the all-atom structure when those restraints are incorporated in a procedure combining threading, modeling of loops and sidechains, and molecular dynamics simulation. Potentially, this new technique can expand NMR's capability to larger proteins.

  17. Master equipment list -- Phase 1. Revision 1

    SciTech Connect (OSTI)

    Jech, J.B.

    1995-04-28

    The purpose of this document is to define the system requirements for the Master Equipment List (MEL) Phase 1 project. The intended audience for this document includes Data Automation Engineering (DAE), Configuration Management Improvement and Control Engineering (CMI and CE), Data Administration Council (DAC), and Tank Waste Remedial System (TWRS) personnel. The intent of Phase 1 is to develop a user-friendly system to support the immediate needs of the TWRS labeling program. Phase 1 will provide CMI and CE the ability to administrate, distribute, and maintain key information generated by the labeling program. CMI and CE is assigning new Equipment Identification Numbers (EINs) to selected equipment in Tank Farms per the TWRS Data Standard ``Tank Farm Equipment Identification Number``. The MEL Phase 1 system will be a multi-user system available through the HLAN network. It will provide basic functions such as view, query, and report, edit, data entry, password access control, administration and change control. The scope of Phase 1 data will encompass all Tank Farm Equipment identified by the labeling program. The data will consist of fields from the labeling program`s working database, relational key references and pointers, safety class information, and field verification data.

  18. Common Geometry Module

    Energy Science and Technology Software Center (OSTI)

    2005-01-01

    The Common Geometry Module (CGM) is a code library which provides geometry functionality used for mesh generation and other applications. This functionality includes that commonly found in solid modeling engines, like geometry creation, query and modification; CGM also includes capabilities not commonly found in solid modeling engines, like geometry decomposition tools and support for shared material interfaces. CGM is built upon the ACIS solid modeling engine, but also includes geometry capability developed beside and onmore » top of ACIS. CGM can be used as-is to provide geometry functionality for codes needing this capability. However, CGM can also be extended using derived classes in C++, allowing the geometric model to serve as the basis for other applications, for example mesh generation. CGM is supported on Sun Solaris, SGI, HP, IBM, DEC, Linux and Windows NT platforms. CGM also indudes support for loading ACIS models on parallel computers, using MPI-based communication. Future plans for CGM are to port it to different solid modeling engines, including Pro/Engineer or SolidWorks. CGM is being released into the public domain under an LGPL license; the ACIS-based engine is available to ACIS licensees on request.« less

  19. Battery Life Estimator (BLE) Data Analysis Software v. 1.2

    Energy Science and Technology Software Center (OSTI)

    2010-02-24

    The purpose of this software is estimate the useable life of rechargeable batteries (e.g., lithium-ion). The software employs a generalized statistical approach to model cell data in the context of accelerated aging experiments. The cell performance is modeled in two parts. The first part consists of a deterministic degradation model which models the average cell behavior. The second part relates to the statistical variation in performance of the cells (error model). Experimental data from anmore » accelerated aging experiment will be input from an Excel worksheet. The software will then query the user for a specific model form (within the generalized model framework). Model parameters will be estimated by the software using various statistical methodologies. Average cell life will be predicted using the estimated model parameters. The uncertainty in the estimated cell life will also be computed using bootstrap simulations. This software can be used in several modes: 1) fit only, 2) fit and simulation, and 3) simulation only« less

  20. Automatic Fault Characterization via Abnormality-Enhanced Classification

    SciTech Connect (OSTI)

    Bronevetsky, G; Laguna, I; de Supinski, B R

    2010-12-20

    Enterprise and high-performance computing systems are growing extremely large and complex, employing hundreds to hundreds of thousands of processors and software/hardware stacks built by many people across many organizations. As the growing scale of these machines increases the frequency of faults, system complexity makes these faults difficult to detect and to diagnose. Current system management techniques, which focus primarily on efficient data access and query mechanisms, require system administrators to examine the behavior of various system services manually. Growing system complexity is making this manual process unmanageable: administrators require more effective management tools that can detect faults and help to identify their root causes. System administrators need timely notification when a fault is manifested that includes the type of fault, the time period in which it occurred and the processor on which it originated. Statistical modeling approaches can accurately characterize system behavior. However, the complex effects of system faults make these tools difficult to apply effectively. This paper investigates the application of classification and clustering algorithms to fault detection and characterization. We show experimentally that naively applying these methods achieves poor accuracy. Further, we design novel techniques that combine classification algorithms with information on the abnormality of application behavior to improve detection and characterization accuracy. Our experiments demonstrate that these techniques can detect and characterize faults with 65% accuracy, compared to just 5% accuracy for naive approaches.

  1. SU-E-J-16: A Review of the Magnitude of Patient Imaging Shifts in Relation to Departmental Policy Changes

    SciTech Connect (OSTI)

    O'Connor, M; Sansourekidou, P

    2014-06-01

    Purpose: To evaluate how changes in imaging policy affect the magnitude of shifts applied to patients. Methods: In June 2012, the department's imaging policy was altered to require that any shifts derived from imaging throughout the course of treatment shall be considered systematic only after they were validated with two data points that are consistent in the same direction. Multiple additions and clarifications to the imaging policy were implemented throughout the course of the data collection, but they were mostly of administrative nature. Entered shifts were documented in MOSAIQ (Elekta AB) through the localization offset. The MOSAIQ database was queried to identify a possible trend. A total of 25,670 entries were analyzed, including four linear accelerators with a combination of MV planar, kV planar and kV three dimensional imaging. The monthly average of the magnitude of the vector was used. Plan relative offsets were excluded. During the evaluated period of time, one of the satellite facilities acquired and implemented Vision RT (AlignRT Inc). Results: After the new policy was implemented the shifts variance and standard deviation decreased. The decrease is linear with time elapsed. Vision RT implementation at one satellite facility reduced the number of overall shifts, specifically for breast patients. Conclusion: Changes in imaging policy have a significant effect on the magnitude of shifts applied to patients. Using two statistical points before applying a shift as persistent decreased the overall magnitude of the shifts applied to patients.

  2. A SOAP Web Service for accessing MODIS land product subsets

    SciTech Connect (OSTI)

    SanthanaVannan, Suresh K; Cook, Robert B; Pan, Jerry Yun; Wilson, Bruce E

    2011-01-01

    Remote sensing data from satellites have provided valuable information on the state of the earth for several decades. Since March 2000, the Moderate Resolution Imaging Spectroradiometer (MODIS) sensor on board NASA s Terra and Aqua satellites have been providing estimates of several land parameters useful in understanding earth system processes at global, continental, and regional scales. However, the HDF-EOS file format, specialized software needed to process the HDF-EOS files, data volume, and the high spatial and temporal resolution of MODIS data make it difficult for users wanting to extract small but valuable amounts of information from the MODIS record. To overcome this usability issue, the NASA-funded Distributed Active Archive Center (DAAC) for Biogeochemical Dynamics at Oak Ridge National Laboratory (ORNL) developed a Web service that provides subsets of MODIS land products using Simple Object Access Protocol (SOAP). The ORNL DAAC MODIS subsetting Web service is a unique way of serving satellite data that exploits a fairly established and popular Internet protocol to allow users access to massive amounts of remote sensing data. The Web service provides MODIS land product subsets up to 201 x 201 km in a non-proprietary comma delimited text file format. Users can programmatically query the Web service to extract MODIS land parameters for real time data integration into models, decision support tools or connect to workflow software. Information regarding the MODIS SOAP subsetting Web service is available on the World Wide Web (WWW) at http://daac.ornl.gov/modiswebservice.

  3. Reducing Information Overload in Large Seismic Data Sets

    SciTech Connect (OSTI)

    HAMPTON,JEFFERY W.; YOUNG,CHRISTOPHER J.; MERCHANT,BION J.; CARR,DORTHE B.; AGUILAR-CHANG,JULIO

    2000-08-02

    Event catalogs for seismic data can become very large. Furthermore, as researchers collect multiple catalogs and reconcile them into a single catalog that is stored in a relational database, the reconciled set becomes even larger. The sheer number of these events makes searching for relevant events to compare with events of interest problematic. Information overload in this form can lead to the data sets being under-utilized and/or used incorrectly or inconsistently. Thus, efforts have been initiated to research techniques and strategies for helping researchers to make better use of large data sets. In this paper, the authors present their efforts to do so in two ways: (1) the Event Search Engine, which is a waveform correlation tool and (2) some content analysis tools, which area combination of custom-built and commercial off-the-shelf tools for accessing, managing, and querying seismic data stored in a relational database. The current Event Search Engine is based on a hierarchical clustering tool known as the dendrogram tool, which is written as a MatSeis graphical user interface. The dendrogram tool allows the user to build dendrogram diagrams for a set of waveforms by controlling phase windowing, down-sampling, filtering, enveloping, and the clustering method (e.g. single linkage, complete linkage, flexible method). It also allows the clustering to be based on two or more stations simultaneously, which is important to bridge gaps in the sparsely recorded event sets anticipated in such a large reconciled event set. Current efforts are focusing on tools to help the researcher winnow the clusters defined using the dendrogram tool down to the minimum optimal identification set. This will become critical as the number of reference events in the reconciled event set continually grows. The dendrogram tool is part of the MatSeis analysis package, which is available on the Nuclear Explosion Monitoring Research and Engineering Program Web Site. As part of the research

  4. Threatened and Endangered Species Evaluation for Operating Commercial Nuclear Power Generating Plants

    SciTech Connect (OSTI)

    Sackschewsky, Michael R.

    2004-01-15

    &E species take, consultations, and evaluations of potential effects of operation on T&E species. This search recovered a total of approximately 100 documents from 13 sites. Sites that were in the relicensing or decommissioning processes were excluded from the ADAMS search. In general the ADAMS search did not reveal any serious deficiencies or compliance problems. The most notable finds were reports of takes of green sea turtles at Diablo Canyon. While these events were reported to both the NRC and to NOAA Fisheries, no record of interaction between the two federal agencies was found. Species potentially present at each site were determined via querying the Geographical, Environmental, and Siting Information System (GEn&SIS) database developed for the NRC by Lawrence Livermore National Laboratory. The results of these queries were compared against the 1997 review, and in the cases of sites that were in the relicensing process, with the results of those site specific evaluations. A total of 452 T&E species were identified as potentially occurring near one or more of the operating commercial nuclear power generating plants. Information about each of these species was gathered to support an assessment of the probability of occurrence at each of the reactor sites. Based on the assessments of which species are potentially affected at each site, and the information gathered through the ADAMS search, each site was assigned a priority value for follow-up evaluations. The priority listing did not include any sites that had entered the relicensing process, those where the licensee has indicated that they intend to enter the relicensing process before the end of 2005, or those that have entered the decommissioning process. Of the 39 remaining sites, those that were identified as the highest priority for follow-on evaluations are: Diablo Canyon, San Onofre, Crystal River, Harris, and Vogtle, followed by South Texas, Palo Verde, Salem, and Cooper.

  5. Genetic and Pharmacological Inhibition of PDK1 in Cancer Cells: Characterization of a Selective Allosteric Kinase Inhibitor

    SciTech Connect (OSTI)

    Nagashima, Kumiko; Shumway, Stuart D.; Sathyanarayanan, Sriram; Chen, Albert H.; Dolinski, Brian; Xu, Youyuan; Keilhack, Heike; Nguyen, Thi; Wiznerowicz, Maciej; Li, Lixia; Lutterbach, Bart A.; Chi, An; Paweletz, Cloud; Allison, Timothy; Yan, Youwei; Munshi, Sanjeev K.; Klippel, Anke; Kraus, Manfred; Bobkova, Ekaterina V.; Deshmukh, Sujal; Xu, Zangwei; Mueller, Uwe; Szewczak, Alexander A.; Pan, Bo-Sheng; Richon, Victoria; Pollock, Roy; Blume-Jensen, Peter; Northrup, Alan; Andersen, Jannik N.

    2013-11-20

    Phosphoinositide-dependent kinase 1 (PDK1) is a critical activator of multiple prosurvival and oncogenic protein kinases and has garnered considerable interest as an oncology drug target. Despite progress characterizing PDK1 as a therapeutic target, pharmacological support is lacking due to the prevalence of nonspecific inhibitors. Here, we benchmark literature and newly developed inhibitors and conduct parallel genetic and pharmacological queries into PDK1 function in cancer cells. Through kinase selectivity profiling and x-ray crystallographic studies, we identify an exquisitely selective PDK1 inhibitor (compound 7) that uniquely binds to the inactive kinase conformation (DFG-out). In contrast to compounds 1-5, which are classical ATP-competitive kinase inhibitors (DFG-in), compound 7 specifically inhibits cellular PDK1 T-loop phosphorylation (Ser-241), supporting its unique binding mode. Interfering with PDK1 activity has minimal antiproliferative effect on cells growing as plastic-attached monolayer cultures (i.e. standard tissue culture conditions) despite reduced phosphorylation of AKT, RSK, and S6RP. However, selective PDK1 inhibition impairs anchorage-independent growth, invasion, and cancer cell migration. Compound 7 inhibits colony formation in a subset of cancer cell lines (four of 10) and primary xenograft tumor lines (nine of 57). RNAi-mediated knockdown corroborates the PDK1 dependence in cell lines and identifies candidate biomarkers of drug response. In summary, our profiling studies define a uniquely selective and cell-potent PDK1 inhibitor, and the convergence of genetic and pharmacological phenotypes supports a role of PDK1 in tumorigenesis in the context of three-dimensional in vitro culture systems.

  6. InterMine Webservices for Phytozome (Rev2)

    SciTech Connect (OSTI)

    Carlson, Joseph; Goodstein, David; Rokhsar, Dan

    2014-07-10

    A datawarehousing framework for information provides a useful infrastructure for providers and users of genomic data. For providers, the infrastructure give them a consistent mechanism for extracting raw data. While for the users, the web services supported by the software allows them to make complex, and often unique, queries of the data. Previously, phytozome.net used BioMart to provide the infrastructure. As the complexity, scale and diversity of the dataset as grown, we decided to implement an InterMine web service on our servers. This change was largely motivated by the ability to have a more complex table structure and richer web reporting mechanism than BioMart. For InterMine to achieve its more complex database schema it requires an XML description of the data and an appropriate loader. Unlimited one-to-many and many-to-many relationship between the tables can be enabled in the schema. We have implemented support for:1.) Genomes and annotations for the data in Phytozome. This set is the 48 organisms currently stored in a back end CHADO datastore. The data loaders are modified versions of the CHADO data adapters from FlyMine. 2.) Interproscan results from all proteins in the Phytozome database. 3.) Clusters of proteins into a grouped heirarchically by similarity. 4.) Cufflinks results from tissue-specific RNA-Seq data of Phytozome organisms. 5.) Diversity data (GATK and SnpEFF results) from a set of individual organism. The last two datatypes are new in this implementation of our web services. We anticipate that the scale of these data will increase considerably in the near future.

  7. An affinity-structure database of helix-turn-helix: DNA complexes with a universal coordinate system

    DOE Public Access Gateway for Energy & Science Beta (PAGES Beta)

    AlQuraishi, Mohammed; Tang, Shengdong; Xia, Xide

    2015-11-19

    Molecular interactions between proteins and DNA molecules underlie many cellular processes, including transcriptional regulation, chromosome replication, and nucleosome positioning. Computational analyses of protein-DNA interactions rely on experimental data characterizing known protein-DNA interactions structurally and biochemically. While many databases exist that contain either structural or biochemical data, few integrate these two data sources in a unified fashion. Such integration is becoming increasingly critical with the rapid growth of structural and biochemical data, and the emergence of algorithms that rely on the synthesis of multiple data types to derive computational models of molecular interactions. We have developed an integrated affinity-structure database inmore » which the experimental and quantitative DNA binding affinities of helix-turn-helix proteins are mapped onto the crystal structures of the corresponding protein-DNA complexes. This database provides access to: (i) protein-DNA structures, (ii) quantitative summaries of protein-DNA binding affinities using position weight matrices, and (iii) raw experimental data of protein-DNA binding instances. Critically, this database establishes a correspondence between experimental structural data and quantitative binding affinity data at the single basepair level. Furthermore, we present a novel alignment algorithm that structurally aligns the protein-DNA complexes in the database and creates a unified residue-level coordinate system for comparing the physico-chemical environments at the interface between complexes. Using this unified coordinate system, we compute the statistics of atomic interactions at the protein-DNA interface of helix-turn-helix proteins. We provide an interactive website for visualization, querying, and analyzing this database, and a downloadable version to facilitate programmatic analysis. Lastly, this database will facilitate the analysis of protein-DNA interactions and the

  8. Adaptable Computing Environment/Self-Assembling Software

    Energy Science and Technology Software Center (OSTI)

    2007-09-25

    Complex software applications are difficult to learn to use and to remember how to use. Further, the user has no control over the functionality available in a given application. The software we use can be created and modified only by a relatively small group of elite, highly skilled artisans known as programmers. "Normal users" are powerless to create and modify software themselves, because the tools for software development, designed by and for programmers, are amore » barrier to entry. This software, when completed, will be a user-adaptable computing environment in which the user is really in control of his/her own software, able to adapt the system, make new parts of the system interactive, and even modify the behavior of the system itself. Som key features of the basic environment that have been implemented are (a) books in bookcases, where all data is stored, (b) context-sensitive compass menus (compass, because the buttons are located in compass directions relative to the mouose cursor position), (c) importing tabular data and displaying it in a book, (d) light-weight table querying/sorting, (e) a Reach&Get capability (sort of a "smart" copy/paste that prevents the user from copying invalid data), and (f) a LogBook that automatically logs all user actions that change data or the system itself. To bootstrap toward full end-user adaptability, we implemented a set of development tools. With the development tools, compass menus can be made and customized.« less

  9. Towards the Integration of APECS with VE-Suite to Create a Comprehensive Virtual Engineering Environment

    SciTech Connect (OSTI)

    McCorkle, D.; Yang, C.; Jordan, T.; Swensen, D.; Zitney, S.E.; Bryden, M.

    2007-06-01

    Modeling and simulation tools are becoming pervasive in the process engineering practice of designing advanced power generation facilities. These tools enable engineers to explore many what-if scenarios before cutting metal or constructing a pilot scale facility. While such tools enable investigation of crucial plant design aspects, typical commercial process simulation tools such as Aspen Plus®, gPROMS®, and HYSYS® still do not explore some plant design information, including computational fluid dynamics (CFD) models for complex thermal and fluid flow phenomena, economics models for policy decisions, operational data after the plant is constructed, and as-built information for use in as-designed models. Software tools must be created that allow disparate sources of information to be integrated if environments are to be constructed where process simulation information can be accessed. At the Department of Energy’s (DOE) National Energy Technology Laboratory (NETL), the Advanced Process Engineering Co-Simulator (APECS) has been developed as an integrated software suite that combines process simulation (e.g., Aspen Plus) and high-fidelity equipment simulation (e.g., Fluent® CFD), together with advanced analysis capabilities including case studies, sensitivity analysis, stochastic simulation for risk/uncertainty analysis, and multi-objective optimization. In this paper, we discuss the initial phases of integrating APECS with the immersive and interactive virtual engineering software, VE-Suite, developed at Iowa State University and Ames Laboratory. VE-Suite utilizes the ActiveX (OLE Automation) controls in Aspen Plus wrapped by the CASI library developed by Reaction Engineering International to run the process simulation and query for unit operation results. This integration permits any application that uses the VE-Open interface to integrate with APECS co-simulations, enabling construction of the comprehensive virtual engineering environment needed for the

  10. GenomeVista

    Energy Science and Technology Software Center (OSTI)

    2002-11-04

    Aligning large vertebrate genomes that are structurally complex poses a variety of problems not encountered on smaller scales. Such genomes are rich in repetitive elements and contain multiple segmental duplications, which increases the difficulty of identifying true orthologous SNA segments in alignments. The sizes of the sequences make many alignment algorithms designed for comparing single proteins extremely inefficient when processing large genomic intervals. We integrated both local and global alignment tools and developed a suitemore » of programs for automatically aligning large vertebrate genomes and identifying conserved non-coding regions in the alignments. Our method uses the BLAT local alignment program to find anchors on the base genome to identify regions of possible homology for a query sequence. These regions are postprocessed to find the best candidates which are then globally aligned using the AVID global alignment program. In the last step conserved non-coding segments are identified using VISTA. Our methods are fast and the resulting alignments exhibit a high degree of sensitivity, covering more than 90% of known coding exons in the human genome. The GenomeVISTA software is a suite of Perl programs that is built on a MySQL database platform. The scheduler gets control data from the database, builds a queve of jobs, and dispatches them to a PC cluster for execution. The main program, running on each node of the cluster, processes individual sequences. A Perl library acts as an interface between the database and the above programs. The use of a separate library allows the programs to function independently of the database schema. The library also improves on the standard Perl MySQL database interfere package by providing auto-reconnect functionality and improved error handling.« less

  11. Evaluating the Potential of Commercial GIS for Accelerator Configuration Management

    SciTech Connect (OSTI)

    T.L. Larrieu; Y.R. Roblin; K. White; R. Slominski

    2005-10-10

    The Geographic Information System (GIS) is a tool used by industries needing to track information about spatially distributed assets. A water utility, for example, must know not only the precise location of each pipe and pump, but also the respective pressure rating and flow rate of each. In many ways, an accelerator such as CEBAF (Continuous Electron Beam Accelerator Facility) can be viewed as an ''electron utility''. Whereas the water utility uses pipes and pumps, the ''electron utility'' uses magnets and RF cavities. At Jefferson lab we are exploring the possibility of implementing ESRI's ArcGIS as the framework for building an all-encompassing accelerator configuration database that integrates location, configuration, maintenance, and connectivity details of all hardware and software. The possibilities of doing so are intriguing. From the GIS, software such as the model server could always extract the most-up-to-date layout information maintained by the Survey & Alignment for lattice modeling. The Mechanical Engineering department could use ArcGIS tools to generate CAD drawings of machine segments from the same database. Ultimately, the greatest benefit of the GIS implementation could be to liberate operators and engineers from the limitations of the current system-by-system view of machine configuration and allow a more integrated regional approach. The commercial GIS package provides a rich set of tools for database-connectivity, versioning, distributed editing, importing and exporting, and graphical analysis and querying, and therefore obviates the need for much custom development. However, formidable challenges to implementation exist and these challenges are not only technical and manpower issues, but also organizational ones. The GIS approach would crosscut organizational boundaries and require departments, which heretofore have had free reign to manage their own data, to cede some control and agree to a centralized framework.

  12. SU-D-BRD-07: Automatic Patient Data Audit and Plan Quality Check to Support ARIA and Eclipse

    SciTech Connect (OSTI)

    Li, X; Li, H; Wu, Y; Mutic, S; Yang, D

    2014-06-01

    Purpose: To ensure patient safety and treatment quality in RT departments that use Varian ARIA and Eclipse, we developed a computer software system and interface functions that allow previously developed electron chart checking (EcCk) methodologies to support these Varian systems. Methods: ARIA and Eclipse store most patient information in its MSSQL database. We studied the contents in the hundreds database tables and identified the data elements used for patient treatment management and treatment planning. Interface functions were developed in both c-sharp and MATLAB to support data access from ARIA and Eclipse servers using SQL queries. These functions and additional data processing functions allowed the existing rules and logics from EcCk to support ARIA and Eclipse. Dose and structure information are important for plan quality check, however they are not stored in the MSSQL database but as files in Varian private formats, and cannot be processed by external programs. We have therefore implemented a service program, which uses the DB Daemon and File Daemon services on ARIA server to automatically and seamlessly retrieve dose and structure data as DICOM files. This service was designed to 1) consistently monitor the data access requests from EcCk programs, 2) translate the requests for ARIA daemon services to obtain dose and structure DICOM files, and 3) monitor the process and return the obtained DICOM files back to EcCk programs for plan quality check purposes. Results: EcCk, which was previously designed to only support MOSAIQ TMS and Pinnacle TPS, can now support Varian ARIA and Eclipse. The new EcCk software has been tested and worked well in physics new start plan check, IMRT plan integrity and plan quality checks. Conclusion: Methods and computer programs have been implemented to allow EcCk to support Varian ARIA and Eclipse systems. This project was supported by a research grant from Varian Medical System.

  13. AGR-2 Data Qualification Report for ATR Cycles 149B, 150A, 150B, 151A, and 151B

    SciTech Connect (OSTI)

    Michael L. Abbott; Binh T. Pham

    2012-06-01

    This report provides the data qualification status of AGR-2 fuel irradiation experimental data from Advanced Test Reactor (ATR) cycles 149B, 150A, 150B, 151A, and 151B), as recorded in the Nuclear Data Management and Analysis System (NDMAS). The AGR-2 data streams addressed include thermocouple temperatures, sweep gas data (flow rate, pressure, and moisture content), and fission product monitoring system (FPMS) data for each of the six capsules in the experiment. A total of 3,307,500 5-minute thermocouple and sweep gas data records were received and processed by NDMAS for this period. There are no AGR-2 data for cycle 150A because the experiment was removed from the reactor. Of these data, 82.2% were determined to be Qualified based on NDMAS accuracy testing and data validity assessment. There were 450,557 Failed temperature records due to thermocouple failures, and 138,528 Failed gas flow records due to gas flow cross-talk and leakage problems that occurred in the capsules after cycle 150A. For FPMS data, NDMAS received and processed preliminary release rate and release-to-birth rate ratio (R/B) data for the first three reactor cycles (cycles 149B, 150B, and 151B). This data consists of 45,983 release rate records and 45,235 R/B records for the 12 radionuclides reported. The qualification status of these FPMS data has been set to In Process until receipt of QA-approved data generator reports. All of the above data have been processed and tested using a SAS-based enterprise application software system, stored in a secure Structured Query Language database, and made available on the NDMAS Web portal (http://ndmas.inl.gov) for both internal and external VHTR project participants.

  14. Parasail: SIMD C library for global, semi-global, and local pairwise sequence alignments

    DOE Public Access Gateway for Energy & Science Beta (PAGES Beta)

    Daily, Jeffrey A.

    2016-02-10

    Sequence alignment algorithms are a key component of many bioinformatics applications. Though various fast Smith-Waterman local sequence alignment implementations have been developed for x86 CPUs, most are embedded into larger database search tools. In addition, fast implementations of Needleman-Wunsch global sequence alignment and its semi-global variants are not as widespread. This article presents the first software library for local, global, and semi-global pairwise intra-sequence alignments and improves the performance of previous intra-sequence implementations. As a result, a faster intra-sequence pairwise alignment implementation is described and benchmarked. Using a 375 residue query sequence a speed of 136 billion cell updates permore » second (GCUPS) was achieved on a dual Intel Xeon E5-2670 12-core processor system, the highest reported for an implementation based on Farrar’s ’striped’ approach. When using only a single thread, parasail was 1.7 times faster than Rognes’s SWIPE. For many score matrices, parasail is faster than BLAST. The software library is designed for 64 bit Linux, OS X, or Windows on processors with SSE2, SSE41, or AVX2. Source code is available from https://github.com/jeffdaily/parasail under the Battelle BSD-style license. In conclusion, applications that require optimal alignment scores could benefit from the improved performance. For the first time, SIMD global, semi-global, and local alignments are available in a stand-alone C library.« less

  15. SU-E-P-07: Evaluation of Productivity Systems for Radiation Therapy

    SciTech Connect (OSTI)

    Ramsey, C; Usynin, A [Thompson Cancer Survival Center, Knoxville, TN (United States)

    2014-06-01

    Purpose: Health systems throughout the United States are under increased financial pressure to reduce operating cost. As a result, productivity models developed by third-party consultants are being used to optimize staff to treatment volumes. The purpose of this study was to critically evaluate productivity systems for radiation oncology. Methods: Staffing efficiency was evaluated using multiple productivity models. The first model evaluated staffing levels using equal weighting of procedure codes and hours worked. A second productivity model was developed using hours worked by job class and relative value units for each procedure code. A third model was developed using the measured procedure times extracted from the electronic medical record, which tracks the wait and treatment times for each patient for each treatment fraction. A MatLab program was developed to query and analyze the daily treatment data. A model was then created to determine any theoretical gains in treatment productivity. Results: Productivity was evaluated for six radiation therapy departments operating nine linear accelerators delivering over 40,000 treatment fractions per year. Third party productivity models that do not take into consideration the unique nature of radiation therapy can be counterproductive. For example, other outpatient departments can compress their daily schedule to decrease the worked hours. This approach was tested using the treatment schedule evaluation tool developed as part of this study. It was determined that the maximum possible savings for treatment schedule compression was $32,000 per year per linac. All annual cost savings would be lost if only two patients per year choose to be treated elsewhere because of limited or restricted appointment times. Conclusion: The use of productivity models in radiation therapy can easily result in a loss of treatment revenue that is greater than any potential cost savings in reduced hours worked by staff.

  16. Modeling of fluidized-bed combustion of coal: Phase II, final reports. Volume VII. FBC Data-Base-Management System (FBC-DBMS) users manual

    SciTech Connect (OSTI)

    Louis, J.F.; Tung, S.E.

    1980-10-01

    The primary goal of the Fluidized Bed Combustor Data Base (FBCDB) is to establish a data repository for the express use of designers and research personnel involved in FBC development. FBCDB is implemented on MIT's 370/168 computer, using the Model 204 Data Base Management System (DBMS) developed by Computer Corporation of America. DBMS is a software that provides an efficient way of storing, retrieving, updating and manipulating data using an English-like query language. The primary content of FBCDB is a collection of data points defined by the value of a number of specific FBC variables. A user may interactively access the data base from a computer terminal at any location, retrieve, examine, and manipulate the data as well as produce tables or graphs of the results. More than 20 program segments are currently available in M204 User Language to simplify the user interface for the FBC design or research personnel. However, there are still many complex and advanced retrieving as well as applications programs to be written for this purpose. Although there are currently 71 entries, and about 2000 groups reposited in the system, this size of data is only an intermediate portion of our selection. The usefulness of the system at the present time is, therefore, limited. This version of FBCDB will be released on a limited scale to obtain review and comments. The document is intended as a reference guide to the use of FBCDB. It has been structured to introduce the user to the basics of FBCDB, summarize what the available segments in FBCDB can do, and give detailed information on the operation of FBCDB. This document represents a preliminary draft of a Users Manual. The draft will be updated when the data base system becomes fully implemented. Any suggestions as to how this manual may be improved will be appreciated.

  17. Modeling of fluidized-bed combustion of coal: Phase II, final reports. Volume VI. FBC-Data Base-Management-System (FBC-DBMS) development

    SciTech Connect (OSTI)

    Louis, J.F.; Tung, S.E.

    1980-10-01

    The primary goal of the Fluidized Bed Combustor Data Base, (FBCDB), situated in MIT's Energy laboratory, is to establish a data repository for the express use of designers and research personnel involved in FBC development. DBMS is a software that provides an efficient way of storing, retrieving, updating and manipulating data using an English-like query language. It is anticipated that the FBCDB would play an active and a direct role in the development of FBC technology as well as in the FBC commercial application. After some in-house experience and after a careful and extensive review of commercially available database systems, it was determined that the Model 204 DBMS by Computer Corporation of America was the most suitable to our needs. The setup of a prototype in-house database also allowed us to investigate and understand fully the particular problems involved in coordinating FBC development with a DBMS. Various difficult aspects were encountered and solutions had been sought. For instance, we found that it was necessary to rename the variables to avoid repetition as well as to increase usefulness of our database and, hence, we had designed a classification system for which variables were classified under category to achieve standardization of variable names. The primary content of FBCDB is a collection of data points defined by the value of a number of specific FBC variables. A user may interactively access the database from a computer terminal at any location, retrieve, examine, and manipulate the data as well as produce tables or graphs of the results.

  18. Supporting large-scale computational science

    SciTech Connect (OSTI)

    Musick, R., LLNL

    1998-02-19

    Business needs have driven the development of commercial database systems since their inception. As a result, there has been a strong focus on supporting many users, minimizing the potential corruption or loss of data, and maximizing performance metrics like transactions per second, or TPC-C and TPC-D results. It turns out that these optimizations have little to do with the needs of the scientific community, and in particular have little impact on improving the management and use of large-scale high-dimensional data. At the same time, there is an unanswered need in the scientific community for many of the benefits offered by a robust DBMS. For example, tying an ad-hoc query language such as SQL together with a visualization toolkit would be a powerful enhancement to current capabilities. Unfortunately, there has been little emphasis or discussion in the VLDB community on this mismatch over the last decade. The goal of the paper is to identify the specific issues that need to be resolved before large-scale scientific applications can make use of DBMS products. This topic is addressed in the context of an evaluation of commercial DBMS technology applied to the exploration of data generated by the Department of Energy`s Accelerated Strategic Computing Initiative (ASCI). The paper describes the data being generated for ASCI as well as current capabilities for interacting with and exploring this data. The attraction of applying standard DBMS technology to this domain is discussed, as well as the technical and business issues that currently make this an infeasible solution.

  19. Agua Caliente Wind/Solar Project at Whitewater Ranch

    SciTech Connect (OSTI)

    Hooks, Todd; Stewart, Royce

    2014-12-16

    Agua Caliente Band of Cahuilla Indians (ACBCI) was awarded a grant by the Department of Energy (DOE) to study the feasibility of a wind and/or solar renewable energy project at the Whitewater Ranch (WWR) property of ACBCI. Red Mountain Energy Partners (RMEP) was engaged to conduct the study. The ACBCI tribal lands in the Coachella Valley have very rich renewable energy resources. The tribe has undertaken several studies to more fully understand the options available to them if they were to move forward with one or more renewable energy projects. With respect to the resources, the WWR property clearly has excellent wind and solar resources. The DOE National Renewable Energy Laboratory (NREL) has continued to upgrade and refine their library of resource maps. The newer, more precise maps quantify the resources as among the best in the world. The wind and solar technology available for deployment is also being improved. Both are reducing their costs to the point of being at or below the costs of fossil fuels. Technologies for energy storage and microgrids are also improving quickly and present additional ways to increase the wind and/or solar energy retained for later use with the network management flexibility to provide power to the appropriate locations when needed. As a result, renewable resources continue to gain more market share. The transitioning to renewables as the major resources for power will take some time as the conversion is complex and can have negative impacts if not managed well. While the economics for wind and solar systems continue to improve, the robustness of the WWR site was validated by the repeated queries of developers to place wind and/or solar there. The robust resources and improving technologies portends toward WWR land as a renewable energy site. The business case, however, is not so clear, especially when the potential investment portfolio for ACBCI has several very beneficial and profitable alternatives.

  20. Practice Patterns of Radiotherapy in Cervical Cancer Among Member Groups of the Gynecologic Cancer Intergroup (GCIG)

    SciTech Connect (OSTI)

    Gaffney, David K. . E-mail: david.gaffney@hci.utah.edu; Du Bois, Andreas; Narayan, Kailash; Reed, Nick; Toita, Takafumi; Pignata, Sandro; Blake, Peter; Portelance, Lorraine; Sadoyze, Azmat; Poetter, Richard; Colombo, Alessandro; Randall, Marcus; Mirza, Mansoor R.; Trimble, Edward L.

    2007-06-01

    Purpose: The aim of this study was to describe radiotherapeutic practice of the treatment of cervical cancer in member groups of the Gynecologic Cancer Intergroup (GCIG). Methods and Materials: A survey was developed and distributed to the members of the GCIG focusing on details of radiotherapy practice. Different scenarios were queried including advanced cervical cancer, postoperative patients, and para-aortic-positive lymph node cases. Items focused on indications for radiation therapy, radiation fields, dose, use of chemotherapy, brachytherapy and others. The cooperative groups from North America were compared with the other groups to evaluate potential differences in radiotherapy doses. Results: A total of 39 surveys were returned from 13 different cooperative groups. For the treatment of advanced cervical cancer, external beam pelvic doses and total doses to point A were 47 + 3.5 Gy (mean + SD) and 79.1 + 7.9 Gy, respectively. Point A doses were not different between the North American cooperative groups compared with the others (p = 0.103). All groups used concomitant chemotherapy, with 30 of 36 respondents using weekly cisplatin. Of 33 respondents, 31 intervened for a low hemoglobin level. For a para-aortic field, the upper border was most commonly (15 of 24) at the T12-L1 interspace. Maintenance chemotherapy (after radiotherapy) was not performed by 68% of respondents. For vaginal brachytherapy after hysterectomy, 23 groups performed HDR brachytherapy and four groups used LDR brachytherapy. In the use of brachytherapy, there was no uniformity in dose prescription. Conclusions: Radiotherapy practices among member groups of the GCIG are similar in terms of both doses and use of chemotherapy.