Sample records for kib metadata metadata

  1. ARM - Metadata Management Tool

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    AFDC Printable Version Share this resource Send a link to EERE: Alternative Fuels Data Center Home Page to someone by E-mail Share EERE: Alternative Fuels Data Center Home Page on Facebook Tweet about EERE: Alternative Fuels Data Center Home Page on Twitter Bookmark EERE: Alternative1 First Use of Energy for All Purposes (Fuel and Nonfuel), 2002; Level: National5Sales for4,645 3,625govInstrumentstdmadap Documentation TDMADAP : XDCnarrowbandheat flux ARMMeasurements RelatedFormsMetadata

  2. A Translation Layer to Convey Preservation Metadata

    E-Print Network [OSTI]

    Di Iorio, Angela

    2009-01-01T23:59:59.000Z

    and knowledge base of preservation metadata management. Into the knowledge domain of preservation metadata management.knowledge; ? deepening into PREMIS knowledge, experience and implementation; ? improving the conditions for preservation metadata management

  3. Sharing Metadata: Building Collections and Communities

    E-Print Network [OSTI]

    Chandler, Robin L.; Westbrook, Bradley D.; Rundblad, Kevin

    2009-01-01T23:59:59.000Z

    Metadata: Building Collections  and Communities Robin afford to think about collections only in  the context of an  institution's collections. Sharing Metadata: Building 

  4. THE NEW ONLINE METADATA EDITOR FOR GENERATING STRUCTURED METADATA

    SciTech Connect (OSTI)

    Devarakonda, Ranjeet [ORNL] [ORNL; Shrestha, Biva [ORNL] [ORNL; Palanisamy, Giri [ORNL] [ORNL; Hook, Leslie A [ORNL] [ORNL; Killeffer, Terri S [ORNL] [ORNL; Boden, Thomas A [ORNL] [ORNL; Cook, Robert B [ORNL] [ORNL; Zolly, Lisa [United States Geological Service (USGS)] [United States Geological Service (USGS); Hutchison, Viv [United States Geological Service (USGS)] [United States Geological Service (USGS); Frame, Mike [United States Geological Service (USGS)] [United States Geological Service (USGS); Cialella, Alice [Brookhaven National Laboratory (BNL)] [Brookhaven National Laboratory (BNL); Lazer, Kathy [Brookhaven National Laboratory (BNL)] [Brookhaven National Laboratory (BNL)

    2014-01-01T23:59:59.000Z

    Nobody is better suited to describe data than the scientist who created it. This description about a data is called Metadata. In general terms, Metadata represents the who, what, when, where, why and how of the dataset [1]. eXtensible Markup Language (XML) is the preferred output format for metadata, as it makes it portable and, more importantly, suitable for system discoverability. The newly developed ORNL Metadata Editor (OME) is a Web-based tool that allows users to create and maintain XML files containing key information, or metadata, about the research. Metadata include information about the specific projects, parameters, time periods, and locations associated with the data. Such information helps put the research findings in context. In addition, the metadata produced using OME will allow other researchers to find these data via Metadata clearinghouses like Mercury [2][4]. OME is part of ORNL s Mercury software fleet [2][3]. It was jointly developed to support projects funded by the United States Geological Survey (USGS), U.S. Department of Energy (DOE), National Aeronautics and Space Administration (NASA) and National Oceanic and Atmospheric Administration (NOAA). OME s architecture provides a customizable interface to support project-specific requirements. Using this new architecture, the ORNL team developed OME instances for USGS s Core Science Analytics, Synthesis, and Libraries (CSAS&L), DOE s Next Generation Ecosystem Experiments (NGEE) and Atmospheric Radiation Measurement (ARM) Program, and the international Surface Ocean Carbon Dioxide ATlas (SOCAT). Researchers simply use the ORNL Metadata Editor to enter relevant metadata into a Web-based form. From the information on the form, the Metadata Editor can create an XML file on the server that the editor is installed or to the user s personal computer. Researchers can also use the ORNL Metadata Editor to modify existing XML metadata files. As an example, an NGEE Arctic scientist use OME to register their datasets to the NGEE data archive and allows the NGEE archive to publish these datasets via a data search portal (http://ngee.ornl.gov/data). These highly descriptive metadata created using OME allows the Archive to enable advanced data search options using keyword, geo-spatial, temporal and ontology filters. Similarly, ARM OME allows scientists or principal investigators (PIs) to submit their data products to the ARM data archive. How would OME help Big Data Centers like the Oak Ridge National Laboratory Distributed Active Archive Center (ORNL DAAC)? The ORNL DAAC is one of NASA s Earth Observing System Data and Information System (EOSDIS) data centers managed by the Earth Science Data and Information System (ESDIS) Project. The ORNL DAAC archives data produced by NASA's Terrestrial Ecology Program. The DAAC provides data and information relevant to biogeochemical dynamics, ecological data, and environmental processes, critical for understanding the dynamics relating to the biological, geological, and chemical components of the Earth's environment. Typically data produced, archived and analyzed is at a scale of multiple petabytes, which makes the discoverability of the data very challenging. Without proper metadata associated with the data, it is difficult to find the data you are looking for and equally difficult to use and understand the data. OME will allow data centers like the NGEE and ORNL DAAC to produce meaningful, high quality, standards-based, descriptive information about their data products in-turn helping with the data discoverability and interoperability. Useful Links: USGS OME: http://mercury.ornl.gov/OME/ NGEE OME: http://ngee-arctic.ornl.gov/ngeemetadata/ ARM OME: http://archive2.ornl.gov/armome/ Contact: Ranjeet Devarakonda (devarakondar@ornl.gov) References: [1] Federal Geographic Data Committee. Content standard for digital geospatial metadata. Federal Geographic Data Committee, 1998. [2] Devarakonda, Ranjeet, et al. "Mercury: reusable metadata management, data discovery and access system." Earth Science Informatics 3.1-2 (2010): 87-94. [

  5. PREMIS Data Dictionary for Preservation Metadata 

    E-Print Network [OSTI]

    Higgins, Sarah

    The Repositories Support Project Briefing Paper on ‘Metadata’ examines how metadata standards can be used to make it easier to find, use and manage digital objects stored in an institutional repository. Long-term preservation ...

  6. Functional Metadata Schema for Engineering Knowledge Management

    E-Print Network [OSTI]

    Mizoguchi, Riichiro

    Functional Metadata Schema for Engineering Knowledge Management Yoshinobu Kitamura Naoya Washio engineering knowledge about functionality representing design rationale because of the lack of rich common, ontology, metadata, functional model, knowledge management 1. INTRODUCTION In the engineering design

  7. Simplifying Video Editing Using Metadata

    E-Print Network [OSTI]

    Stevens, Scott M.

    Simplifying Video Editing Using Metadata Juan Casares, A. Chris Long, Brad A. Myers, Rishi Institute, Carnegie Mellon University1 ABSTRACT Digital video is becoming increasingly ubiquitous. However, editing video remains difficult for several reasons: it is a time-based medium, it has dual tracks

  8. Metadata creation in digital library projects

    E-Print Network [OSTI]

    Mercer, Holly

    2003-10-10T23:59:59.000Z

    Metadata Creation in Digital Library Projects Striking a Balance: Holly Mercer University of Kansas Libraries October 10, 2003 Brick and Click Libraries Metadata Creation in Digital Library Projects Striking a Balance: ?Digital library environment... ?Metadata ?Strategies ?Library?s role What are Digital Libraries? ? Licensed electronic resources ? Archival/manuscript collections ? Departmental/research center collections ? Networked resources ? Related files with a structure & user interface...

  9. apples western metadata: Topics by E-print Network

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Research Archive Summary: Instalment on the role of metadata within the digital curation life-cycle. Describes the increasingly important role of metadata for digital curation,...

  10. Extensions to Metadata Languages for Environmental Primary Data

    E-Print Network [OSTI]

    IBANEZ, LUIS M.

    2011-01-01T23:59:59.000Z

    Semantic Mapping . . . . . . 53 HEMP: Processing descriptionMetadata Processing: HEMP . . . . . 41 Hierarchical Scanner:for Metadata Processing: HEMP Environmental files with

  11. A Metadata-Rich File System

    SciTech Connect (OSTI)

    Ames, S; Gokhale, M B; Maltzahn, C

    2009-01-07T23:59:59.000Z

    Despite continual improvements in the performance and reliability of large scale file systems, the management of file system metadata has changed little in the past decade. The mismatch between the size and complexity of large scale data stores and their ability to organize and query their metadata has led to a de facto standard in which raw data is stored in traditional file systems, while related, application-specific metadata is stored in relational databases. This separation of data and metadata requires considerable effort to maintain consistency and can result in complex, slow, and inflexible system operation. To address these problems, we have developed the Quasar File System (QFS), a metadata-rich file system in which files, metadata, and file relationships are all first class objects. In contrast to hierarchical file systems and relational databases, QFS defines a graph data model composed of files and their relationships. QFS includes Quasar, an XPATH-extended query language for searching the file system. Results from our QFS prototype show the effectiveness of this approach. Compared to the defacto standard, the QFS prototype shows superior ingest performance and comparable query performance on user metadata-intensive operations and superior performance on normal file metadata operations.

  12. Metadata Value Chain for Open Access Journals

    E-Print Network [OSTI]

    Mercer, Holly; Dyas-Correia, Sharon

    2011-01-01T23:59:59.000Z

    In this session, Holly Mercer presented a case for finding global solutions to improve the metadata that is available for journals, particularly small, independent open access journals. She discussed the scholarly communication lifecycle...

  13. Metadata Value Chain for Open Access Journals 

    E-Print Network [OSTI]

    Mercer, Holly; Dyas-Correia, Sharon

    2011-01-01T23:59:59.000Z

    In this session, Holly Mercer presented a case for finding global solutions to improve the metadata that is available for journals, particularly small, independent open access journals. She discussed the scholarly ...

  14. All Teh Metadatas Re-Revisited

    E-Print Network [OSTI]

    Cowles, Kevin Esme; Critchlow, Matthew; Westbrook, Bradley

    2013-01-01T23:59:59.000Z

    storage • Explicit structural metadata  • Native REST API •Fedora REST API (partial) DAMS Manager Separate Java webappUses DAMS Repository REST API Functionality moved into the 

  15. Mining scientific data archives through metadata generation

    SciTech Connect (OSTI)

    Springmeyer, R.; Werner, N.; Long, J.

    1997-04-01T23:59:59.000Z

    Data analysis and management tools typically have not supported the documenting of data, so scientists must manually maintain all information pertaining to the context and history of their work. This metadata is critical to effective retrieval and use of the masses of archived data, yet little of it exists on-line or in an accessible format. Exploration of archived legacy data typically proceeds as a laborious process, using commands to navigate through file structures on several machines. This file-at-a-time approach needs to be replaced with a model that represents data as collections of interrelated objects. The tools that support this model must focus attention on data while hiding the complexity of the computational environment. This problem was addressed by developing a tool for exploring large amounts of data in UNIX directories via automatic generation of metadata summaries. This paper describes the model for metadata summaries of collections and the Data Miner tool for interactively traversing directories and automatically generating metadata that serves as a quick overview and index to the archived data. The summaries include thumbnail images as well as links to the data, related directories, and other metadata. Users may personalize the metadata by adding a title and abstract to the summary, which is presented as an HTML page viewed with a World Wide Web browser. We have designed summaries for 3 types of collections of data: contents of a single directory; virtual directories that represent relations between scattered files; and groups of related calculation files. By focusing on the scientists` view of the data mining task, we have developed techniques that assist in the ``detective work `` of mining without requiring knowledge of mundane details about formats and commands. Experiences in working with scientists to design these tools are recounted.

  16. Extracting Artifact Lifecycle Models from Metadata History

    E-Print Network [OSTI]

    Godfrey, Michael W.

    Extracting Artifact Lifecycle Models from Metadata History Olga Baysal, Oleksii Kononenko, Reid, Canada {obaysal, okononen, rtholmes, migod}@cs.uwaterloo.ca Abstract--Software developers and managers make decisions based on the understanding they have of their software systems. This understanding

  17. Collection Metadata Solutions for Digital Library Applications

    E-Print Network [OSTI]

    Janée, Greg

    Collection Metadata Solutions for Digital Library Applications Linda L. Hill and Greg Jane´ e Alexandria Digital Library Project, University of California, Santa Barbara, 1205 Girvetz, Santa Barbara, CA a digital library, collections may range from an ad hoc set of objects that serve a temporary purpose

  18. Component Metadata for Software Engineering Tasks

    E-Print Network [OSTI]

    Harrold, Mary Jean

    functions. In particular, she has no source code, no reliability or safety information, no in- formationComponent Metadata for Software Engineering Tasks Alessandro Orso1 , Mary Jean Harrold1 , and David of an application developer who wants to perform two different software engineering tasks on her application: gen

  19. DDI - A Metadata Standard for the Community

    E-Print Network [OSTI]

    Vardigan, Mary; Wackerow, Joachim

    2013-04-02T23:59:59.000Z

    Application using parts of DDI A Project of and for the Community Developed by a self-sustaining membership Alliance recently reconfigured with new Bylaws Two major development lines DDI Codebook DDI Lifecycle Metadata for both human and machine consumption... statistical institutes, and software developers. DDI Codebook reflects the content of a traditional archival codebook and encompasses high-level study information, detailed information on variables, and support for aggregate data and basic geography. DDI...

  20. alfalfa western metadata: Topics by E-print Network

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Metadata Extraction System Yang Liu 1;4 Elizabeth Shriberg 1;2 Andreas Stolcke 1;2 Dustin Hillard 3 Computer Technologies and Information Sciences Websites Summary:...

  1. E-Print Network 3.0 - automatic adaptive metadata Sample Search...

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Page: << < 1 2 3 4 5 > >> 1 Accessibility Commons: A Metadata Infrastructure for Web Accessibility Summary: to effective use of external metadata. Various research projects...

  2. LWA Station-Level Observing Procedure and Associated Metadata

    E-Print Network [OSTI]

    Ellingson, Steven W.

    ) and declination (DEC). · Solar Tracking (TRK SOL). This is a beam output mode in which the beam tracks the Sun of observation "metadata"; that is, data which is produced by MCS as part of the processes of scheduling and conducting the observation. Metadata is distinct from the primary instrument output, which is captured by MCS

  3. Metadata Working Group Activity Summary for 2009-2010 (Prepared by Steven Folsom, Chair of Metadata Working Group)

    E-Print Network [OSTI]

    Schweik, Charles M.

    Service - Inform the process for inclusion of local digital collections in a discovery service platform of metadata for digital objects within the UMass Libraries. The survey was sent to all Department Heads and a small number of other librarians who are known to work with digital objects and metadata. The questions

  4. Design and Implementation of a Metadata-rich File System

    SciTech Connect (OSTI)

    Ames, S; Gokhale, M B; Maltzahn, C

    2010-01-19T23:59:59.000Z

    Despite continual improvements in the performance and reliability of large scale file systems, the management of user-defined file system metadata has changed little in the past decade. The mismatch between the size and complexity of large scale data stores and their ability to organize and query their metadata has led to a de facto standard in which raw data is stored in traditional file systems, while related, application-specific metadata is stored in relational databases. This separation of data and semantic metadata requires considerable effort to maintain consistency and can result in complex, slow, and inflexible system operation. To address these problems, we have developed the Quasar File System (QFS), a metadata-rich file system in which files, user-defined attributes, and file relationships are all first class objects. In contrast to hierarchical file systems and relational databases, QFS defines a graph data model composed of files and their relationships. QFS incorporates Quasar, an XPATH-extended query language for searching the file system. Results from our QFS prototype show the effectiveness of this approach. Compared to the de facto standard, the QFS prototype shows superior ingest performance and comparable query performance on user metadata-intensive operations and superior performance on normal file metadata operations.

  5. Evolution of the ATLAS Metadata Interface (AMI)

    E-Print Network [OSTI]

    Odier, Jerome; The ATLAS collaboration; Fulachier, Jerome; Lambert, Fabian

    2015-01-01T23:59:59.000Z

    The ATLAS Metadata Interface (AMI) can be considered to be a mature application because it has existed for at least 10 years. Over the years, the number of users and the number of functions provided for these users has increased. It has been necessary to adapt the hardware infrastructure in a seamless way so that the Quality of Service remains high. We will describe the evolution of the application from the initial one, using single server with a MySQL backend database, to the current state, where we use a cluster of Virtual Machines on the French Tier 1 Cloud at Lyon, an ORACLE database backend also at Lyon, with replication to CERN using ORACLE streams behind a back-up server.

  6. Original articles The Stanford Digital Library metadata architecturec

    E-Print Network [OSTI]

    Gravano, Luis

    Original articles The Stanford Digital Library metadata architecturec Michelle Baldonado, Chen / Accepted: 14 January 1997 Abstract. The overall goal of the Stanford Digital Library project is to provide an infrastructure that aords interoperability among heterogeneous, autono- mous digital library services

  7. adaptive metadata generation: Topics by E-print Network

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    of PCs via Ethernet, InfiniBand and Fibers, and cloud storage via Internet. Instead, DBMS often Zhu, Yifeng 13 SmartStore: A New Metadata Organization Paradigm with...

  8. MATCH: Metadata Access Tool for Climate and Health Datasets

    DOE Data Explorer [Office of Scientific and Technical Information (OSTI)]

    MATCH is a searchable clearinghouse of publicly available Federal metadata (i.e. data about data) and links to datasets. Most metadata on MATCH pertain to geospatial data sets ranging from local to global scales. The goals of MATCH are to: 1) Provide an easily accessible clearinghouse of relevant Federal metadata on climate and health that will increase efficiency in solving research problems; 2) Promote application of research and information to understand, mitigate, and adapt to the health effects of climate change; 3) Facilitate multidirectional communication among interested stakeholders to inform and shape Federal research directions; 4) Encourage collaboration among traditional and non-traditional partners in development of new initiatives to address emerging climate and health issues. [copied from http://match.globalchange.gov/geoportal/catalog/content/about.page

  9. Database and Metadata Support of a Web-based Multimedia Digital Library for Medical Education

    E-Print Network [OSTI]

    Gruenwald, Le

    Libraries play an important role in education; they are beneficial to both instructors and students. Many System (IMS) standard for metadata representation and the National Library of Medicine's Medical Subject. We also use RDBMS to store metadata and support metadata queries. The National SMETE Library program

  10. Metadata Enrichment Services for the Europeana Digital Library

    E-Print Network [OSTI]

    Sebastiani, Fabrizio

    needs, thus allowing the tool to be effectively available to the multitude (2000+) of Europeana CPs. 1 to enrich the semantics of its own MRs prior to contributing them to Europeana. Different CPs are thus provenance. We view metadata enrichment as consisting of essentially two activities: clas- sification of MRs

  11. Metadata Extraction from Bibliographic Documents for Digital Library

    E-Print Network [OSTI]

    Paris-Sud XI, Université de

    1 Metadata Extraction from Bibliographic Documents for Digital Library A. Belaďd1 and D. Besagni2 1 for a digital library (DL), their automatic recognition presents an obvious interest. However as their origin. 1. Introduction The digital library (DL) [1] has become more and more a common tool for everyone

  12. METADATA GENERATION FOR CULTURAL HERITAGE: CREATIVE HISTORIES THE JOSEFSPLATZ

    E-Print Network [OSTI]

    METADATA GENERATION FOR CULTURAL HERITAGE: CREATIVE HISTORIES ­ THE JOSEFSPLATZ EXPERIENCE Brigitte Krenn1 , Gregor Sieber1 , Hans Petschar2 Abstract Creative Histories is a Cultural Heritage application using a full 3D-navigation model for PC and mobile phones to navigate historical sites across different

  13. ATLAS Metadata Infrastructure Evolution for Run 2 and Beyond

    E-Print Network [OSTI]

    van Gemmeren, Peter; The ATLAS collaboration; Malon, David; Vaniachine, Alexandre

    2015-01-01T23:59:59.000Z

    ATLAS developed and employed for Run 1 of the Large Hadron Collider a sophisticated infrastructure for metadata handling in event processing jobs. This infrastructure profits from a rich feature set provided by the ATLAS execution control framework, including standardized interfaces and invocation mechanisms for tools and services, segregation of transient data stores with concomitant object lifetime management, and mechanisms for handling occurrences asynchronous to the control framework’s state machine transitions. This metadata infrastructure is evolving and being extended for Run 2 to allow its use and reuse in downstream physics analyses, analyses that may or may not utilize the ATLAS control framework. At the same time, multiprocessing versions of the control framework and the requirements of future multithreaded frameworks are leading to redesign of components that use an incident-handling approach to asynchrony. The increased use of scatter-gather architectures, both local and distributed, requires ...

  14. Enabling Biodiversity Research with Open Source Workflow, GIS and Metadata Tools

    E-Print Network [OSTI]

    Grady, C.J.

    2011-11-16T23:59:59.000Z

    for conservation and management decisions • Challenges – Large extents (> 10,000 km2) – Fine resolution (< 1000 m2 ? 30m x 30m) –Many species (10,000 +) QGIS with Lifemapper MacroEcology plug?in Hexagonal Grid Mouse ‘Presence’ Sites / Species Plot VisTrails Reproducibility • Simple... process metadata • Process metadata extensions • Lifemapper client metadata reader Collaborations • KU Biodiversity Institute • NSF Cyber?Commons • Change Thinking • CI Team http://www.youtube.com/watch?v=VCFixtqlimg Future Directions • Publish...

  15. Towards a Quantitative Evaluation of Geospatial Metadata Quality in the Context of Semantic

    E-Print Network [OSTI]

    Towards a Quantitative Evaluation of Geospatial Metadata Quality in the Context of Semantic to facilitate the reuse of geospatial data in a distributed and heterogeneous environment. In this process, the provided geospatial metadata that are appropriate for the intended use may be incomplete or not appropriate

  16. Unsupervised Metadata Extraction in Scientific Digital Libraries Using A-Priori Domain-Specific

    E-Print Network [OSTI]

    Marchese, Maurizio

    Unsupervised Metadata Extraction in Scientific Digital Libraries Using A-Priori Domain to sustain the scalability of the semantic-enabled services of the future. Unsupervised infor- mation Libraries domain. We propose and present a novel approach focusing on the improvement in the metadata

  17. Schema-Independent and Schema-Friendly Scientific Metadata Management1 Scott Jensen and Beth Plale

    E-Print Network [OSTI]

    Plale, Beth

    . The activities provided to a scientist can include access to public data repositories, large-scale computational Computational science is creating a deluge of data, and being able to reuse this data requires detailed descriptive metadata. Scientific communities have developed detailed metadata schemas to describe data

  18. A metadata infrastructure using ISO standards Department of Computing, University of Surrey, UK

    E-Print Network [OSTI]

    Gillam, Lee

    A metadata infrastructure using ISO standards Lee Gillam Department of Computing, University descriptors that have been, or are being, defined through ISO. ISO standards can be used for the creation, either already in existence or currently emerging through ISO, for metadata. These standards include

  19. Evolution of the Architecture of the ATLAS Metadata Interface (AMI)

    E-Print Network [OSTI]

    Odier, Jerome; The ATLAS collaboration; Fulachier, Jerome; Lambert, Fabian

    2015-01-01T23:59:59.000Z

    The ATLAS Metadata Interface (AMI) is now a mature application. Over the years, the number of users and the number of provided functions has dramatically increased. It is necessary to adapt the hardware infrastructure in a seamless way so that the quality of service remains high. We describe the evolution from the beginning of the application life, using one server with a MySQL backend database, to the current state in which a cluster of virtual machines on the French Tier 1 cloud at Lyon, an Oracle database also at Lyon, with replication to Oracle at CERN and a back-up server are used.

  20. Metadata for PoroTomo Project Subtask 3.2 DAS at Garner Valley

    DOE Data Explorer [Office of Scientific and Technical Information (OSTI)]

    Chelsea Lancelle

    Metadata for the data collected at the NEES@UCSB Garner Valley Downhole Array field site on September 10-12, 2013 as part of the larger PoroTomo project.

  1. Metadata for PoroTomo Project Subtask 3.2 DAS at Garner Valley

    SciTech Connect (OSTI)

    Chelsea Lancelle

    2013-09-10T23:59:59.000Z

    Metadata for the data collected at the NEES@UCSB Garner Valley Downhole Array field site on September 10-12, 2013 as part of the larger PoroTomo project.

  2. Symmetric Active/Active Metadata Service for High Availability Parallel File Systems

    SciTech Connect (OSTI)

    He, X. [Tennessee Technological University; Ou, Li [Tennessee Technological University; Engelmann, Christian [ORNL; Chen, Xin [Tennessee Technological University; Scott, Stephen L [ORNL

    2009-01-01T23:59:59.000Z

    High availability data storage systems are critical for many applications as research and business become more data-driven. Since metadata management is essential to system availability, multiple metadata services are used to improve the availability of distributed storage systems. Past research focused on the active/standby model, where each active service has at least one redundant idle backup. However, interruption of service and even some loss of service state may occur during a fail-over depending on the used replication technique. In addition, the replication overhead for multiple metadata services can be very high. The research in this paper targets the symmetric active/active replication model, which uses multiple redundant service nodes running in virtual synchrony. In this model, service node failures do not cause a fail-over to a backup and there is no disruption of service or loss of service state. We further discuss a fast delivery protocol to reduce the latency of the needed total order broadcast. Our prototype implementation shows that metadata service high availability can be achieved with an acceptable performance trade-off using our symmetric active/active metadata service solution.

  3. Towards more transparent and reproducible omics studies through a common metadata checklist and data publications

    SciTech Connect (OSTI)

    Kolker, Eugene; Ozdemir, Vural; Martens , Lennart; Hancock, William S.; Anderson, Gordon A.; Anderson, Nathaniel; Aynacioglu, Sukru; Baranova, Ancha; Campagna, Shawn R.; Chen, Rui; Choiniere, John; Dearth, Stephen P.; Feng, Wu-Chun; Ferguson, Lynnette; Fox, Geoffrey; Frishman, Dmitrij; Grossman, Robert; Heath, Allison; Higdon, Roger; Hutz, Mara; Janko, Imre; Jiang, Lihua; Joshi, Sanjay; Kel, Alexander; Kemnitz, Joseph W.; Kohane, Isaac; Kolker, Natali; Lancet, Doron; Lee, Elaine; Li, Weizhong; Lisitsa, Andrey; Llerena, Adrian; MacNealy-Koch, Courtney; Marhsall, Jean-Claude; Masuzzo, Paolo; May, Amanda; Mias, George; Monroe, Matthew E.; Montague, Elizabeth; Monney, Sean; Nesvizhskii, Alexey; Noronha, Santosh; Omenn, Gilbert; Rajasimha, Harsha; Ramamoorthy, Preveen; Sheehan, Jerry; Smarr, Larry; Smith, Charles V.; Smith, Todd; Snyder, Michael; Rapole, Srikanth; Srivastava, Sanjeeva; Stanberry, Larissa; Stewart, Elizabeth; Toppo, Stefano; Uetz, Peter; Verheggen, Kenneth; Voy, Brynn H.; Warnich, Louise; Wilhelm, Steven W.; Yandl, Gregory

    2014-01-01T23:59:59.000Z

    Biological processes are fundamentally driven by complex interactions between biomolecules. Integrated high-throughput omics studies enable multifaceted views of cells, organisms, or their communities. With the advent of new post-genomics technologies omics studies are becoming increasingly prevalent yet the full impact of these studies can only be realized through data harmonization, sharing, meta-analysis, and integrated research,. These three essential steps require consistent generation, capture, and distribution of the metadata. To ensure transparency, facilitate data harmonization, and maximize reproducibility and usability of life sciences studies, we propose a simple common omics metadata checklist. The proposed checklist is built on the rich ontologies and standards already in use by the life sciences community. The checklist will serve as a common denominator to guide experimental design, capture important parameters, and be used as a standard format for stand-alone data publications. This omics metadata checklist and data publications will create efficient linkages between omics data and knowledge-based life sciences innovation and importantly, allow for appropriate attribution to data generators and infrastructure science builders in the post-genomics era. We ask that the life sciences community test the proposed omics metadata checklist and data publications and provide feedback for their use and improvement.

  4. MCM LTER METADATA FILE TITLE: Lake ice thickness in the McMurdo Dry Valleys

    E-Print Network [OSTI]

    Priscu, John C.

    MCM LTER METADATA FILE TITLE: Lake ice thickness in the McMurdo Dry Valleys ABSTRACT: Ice thickness was measured from the bottom of the ice cover to the piezometric water level and to the top of the ice cover-2360 achiuchiolo@montana.edu VARIABLES: Location Name, Location Code, Limno Run, Collection Date, z-water, z-ice, z

  5. Common Metadata for Climate Modelling Digital Repositories CIM-enabled OASIS

    E-Print Network [OSTI]

    Common Metadata for Climate Modelling Digital Repositories CIM-enabled OASIS CERFACS Technical-1.2.1 Scientific Digital Repositories DOCUMENT Deliverable D5.7 Month 36 Deliverable Title CIM-enabled OASIS describes the adaption of the OASIS4 coupler so that it can use a CIM XML file for its configuration instead

  6. Metadata Modeling for Quality of Service Management in Distributed Multimedia Systems1

    E-Print Network [OSTI]

    von Bochmann, Gregor

    Metadata Modeling for Quality of Service Management in Distributed Multimedia Systems1 Brigitte.Anne}@uqam.ca E-mail:{bochmann, hafid}@iro.umontreal.ca Abstract In distributed multimedia systems, since multimedia objects are voluminous and unstructured, manipulation, transfer and visualization of such objects

  7. Meta-Metadata: An Information Semantic Language and Software Architecture for Collection Visualization Application

    E-Print Network [OSTI]

    Mathur, Abhinav

    2011-02-22T23:59:59.000Z

    .3? Yahoo Search ...................................................................................................54? 7.4? Lines of Code Comparison...............................................................................55? 8.? USER STUDY... ACM portal Meta-Metadata semantic actions..................................................52? Figure 21 Information visualization from ACM portal using combinFormation .............53? Figure 22 Yahoo search XML and corresponding Meta...

  8. Intagorn et al. Harvesting Geospatial Knowledge from Social Metadata Harvesting Geospatial Knowledge from Social

    E-Print Network [OSTI]

    Lerman, Kristina

    Intagorn et al. Harvesting Geospatial Knowledge from Social Metadata Harvesting Geospatial Institute lerman@isi.edu ABSTRACT Up-to-date geospatial information can help crisis management community of the social photo-sharing site Flickr to learn geospatial concepts and relations. Our method leverages

  9. How to Improve Geospatial Data Usability: From Metadata to Quality-Aware GIS Community

    E-Print Network [OSTI]

    How to Improve Geospatial Data Usability: From Metadata to Quality-Aware GIS Community R. Devillers), Canada 3 ­ Industrial Research Chair in Geospatial Databases for Decision Support, Laval University, retrieve and analyze geospatial data. The field grew exponentially and, in the last two decades

  10. Fig. 1. Example of the Area Level Template. Using a Cruise Report to Generate XML Metadata

    E-Print Network [OSTI]

    Ware, Colin

    Fig. 1. Example of the Area Level Template. Using a Cruise Report to Generate XML Metadata Briana M, abstract and the process description. This information cannot be mined from the raw survey data. This paper illustrates how to generate a template from a Microsoft Word based cruise report that can be used

  11. paper submission to EuroDL'99, Paris Metadata for Photographs: from Digital Library to Multimedia

    E-Print Network [OSTI]

    Paris-Sud XI, Université de

    paper submission to EuroDL'99, Paris Metadata for Photographs: from Digital Library to Multimedia Libraries provide access to an increasing variety of digital multimedia information, including images, photos, sound and video. For many years Libraries and Digital Libraries have been relying

  12. Managing the History of Metadata in support for DB Archiving and Schema Evolution

    E-Print Network [OSTI]

    Zaniolo, Carlo

    Managing the History of Metadata in support for DB Archiving and Schema Evolution Carlo A. Curino1 the necessity to manage them and preserve the schema evolution history. In this paper, we describe the Panta Rhei Framework designed to provide powerful tools that: (i) facilitate schema evolution and guide

  13. Towards an Automatic Metadata Management Framework for Smart Oil Charalampos Chelmis1

    E-Print Network [OSTI]

    Prasanna, Viktor K.

    . Introduction Oil and gas organizations are in continuous pressure to investigate and employ innovativeTowards an Automatic Metadata Management Framework for Smart Oil Fields Charalampos Chelmis1 , Jing in various analysis, prediction and domain-specific procedures that result in even larger volumes of derived

  14. Migration of the ATLAS Metadata Interface (AMI) to Web 2.0 and cloud

    E-Print Network [OSTI]

    Odier, Jerome; The ATLAS collaboration; Fulachier, Jerome; Lambert, Fabian

    2015-01-01T23:59:59.000Z

    The ATLAS Metadata Interface (AMI), a mature application of more than 10 years of existence, is currently under adaptation to some recently available technologies. The web interfaces, which previously manipulated XML documents using XSL transformations, are being migrated to Asynchronous JavaScript (AJAX). Web development is considerably simplified by the introduction of a framework based on JQuery and Twitter Bootstrap. Finally, the AMI services are being migrated to an OpenStack cloud infrastructure.

  15. Migration of the ATLAS Metadata Interface (AMI) to Web 2.0 and cloud

    E-Print Network [OSTI]

    Odier, Jerome; The ATLAS collaboration; Fulachier, Jerome; Lambert, Fabian

    2015-01-01T23:59:59.000Z

    The ATLAS Metadata Interface (AMI) can be considered to be a mature application because it has existed for at least 10 years. Over the last year, we have been adapting the application to some recently available technologies. The web interface, which previously manipulated XML documents using XSL transformations, has been migrated to Asynchronous Java Script (AJAX). Web development has been considerably simplified by the development of a framework for AMI based on JQuery and Twitter Bootstrap. Finally there has been a major upgrade of the python web service client.

  16. The Genomes On Line Database (GOLD) in 2009: status of genomic and metagenomic projects and their associated metadata

    SciTech Connect (OSTI)

    Liolios, Konstantinos; Chen, Amy; Mavromatis, Konstantinos; Tavernarakis, Nektarios; Hugenholtz, Phil; Markowitz, Victor; Kyrpides, Nikos C.

    2009-09-01T23:59:59.000Z

    The Genomes On Line Database (GOLD) is a comprehensive resource for centralized monitoring of genome and metagenome projects worldwide. Both complete and ongoing projects, along with their associated metadata, can be accessed in GOLD through precomputed tables and a search page. As of September 2009, GOLD contains information for more than 5800 sequencing projects, of which 1100 have been completed and their sequence data deposited in a public repository. GOLD continues to expand, moving toward the goal of providing the most comprehensive repository of metadata information related to the projects and their organisms/environments in accordance with the Minimum Information about a (Meta)Genome Sequence (MIGS/MIMS) specification.

  17. The Genomes On Line Database (GOLD) in 2007: status of genomic and metagenomic projects and their associated metadata

    SciTech Connect (OSTI)

    Fenner, Marsha W; Liolios, Konstantinos; Mavromatis, Konstantinos; Tavernarakis, Nektarios; Kyrpides, Nikos C.

    2007-12-31T23:59:59.000Z

    The Genomes On Line Database (GOLD) is a comprehensive resource of information for genome and metagenome projects world-wide. GOLD provides access to complete and ongoing projects and their associated metadata through pre-computed lists and a search page. The database currently incorporates information for more than 2900 sequencing projects, of which 639 have been completed and the data deposited in the public databases. GOLD is constantly expanding to provide metadata information related to the project and the organism and is compliant with the Minimum Information about a Genome Sequence (MIGS) specifications.

  18. Habitat-Lite: A GSC case study based on free text terms for environmental metadata

    SciTech Connect (OSTI)

    Kyrpides, Nikos; Hirschman, Lynette; Clark, Cheryl; Cohen, K. Bretonnel; Mardis, Scott; Luciano, Joanne; Kottmann, Renzo; Cole, James; Markowitz, Victor; Kyrpides, Nikos; Field, Dawn

    2008-04-01T23:59:59.000Z

    There is an urgent need to capture metadata on the rapidly growing number of genomic, metagenomic and related sequences, such as 16S ribosomal genes. This need is a major focus within the Genomic Standards Consortium (GSC), and Habitat is a key metadata descriptor in the proposed 'Minimum Information about a Genome Sequence' (MIGS) specification. The goal of the work described here is to provide a light-weight, easy-to-use (small) set of terms ('Habitat-Lite') that captures high-level information about habitat while preserving a mapping to the recently launched Environment Ontology (EnvO). Our motivation for building Habitat-Lite is to meet the needs of multiple users, such as annotators curating these data, database providers hosting the data, and biologists and bioinformaticians alike who need to search and employ such data in comparative analyses. Here, we report a case study based on semi-automated identification of terms from GenBank and GOLD. We estimate that the terms in the initial version of Habitat-Lite would provide useful labels for over 60% of the kinds of information found in the GenBank isolation-source field, and around 85% of the terms in the GOLD habitat field. We present a revised version of Habitat-Lite and invite the community's feedback on its further development in order to provide a minimum list of terms to capture high-level habitat information and to provide classification bins needed for future studies.

  19. Proceedings of the USENIX Technical Conference, June, 2000. Journaling versus Soft Updates: Asynchronous Meta-data Protection in File

    E-Print Network [OSTI]

    Ganger, Greg

    and Soft Updates. Journaling systems use an auxiliary log to record meta-data operations and Soft Updates this synchronicity requirement is relaxed can journaling systems approach the performance of systems like Soft Updates (which also relaxes this requirement). Our asynchronous journaling and Soft Updates systems

  20. Metadata should help users to assess the quality (fitness for use) of geospatial data, thus reducing the risk of data

    E-Print Network [OSTI]

    Abstract Metadata should help users to assess the quality (fitness for use) of geospatial data in order to reduce user meta-uncertainty related to geospatial data quality, and then reduce the risks information at different levels of analysis. Using a multidimensional database approach, this paper proposes

  1. Integration of Phenotypic Metadata and Protein Similarity in Archaea Using a Spectral Bipartitioning Approach

    SciTech Connect (OSTI)

    Hooper, Sean D.; Anderson, Iain J; Pati, Amrita; Dalevi, Daniel; Mavromatis, Konstantinos; Kyrpides, Nikos C

    2009-01-01T23:59:59.000Z

    In order to simplify and meaningfully categorize large sets of protein sequence data, it is commonplace to cluster proteins based on the similarity of those sequences. However, it quickly becomes clear that the sequence flexibility allowed a given protein varies significantly among different protein families. The degree to which sequences are conserved not only differs for each protein family, but also is affected by the phylogenetic divergence of the source organisms. Clustering techniques that use similarity thresholds for protein families do not always allow for these variations and thus cannot be confidently used for applications such as automated annotation and phylogenetic profiling. In this work, we applied a spectral bipartitioning technique to all proteins from 53 archaeal genomes. Comparisons between different taxonomic levels allowed us to study the effects of phylogenetic distances on cluster structure. Likewise, by associating functional annotations and phenotypic metadata with each protein, we could compare our protein similarity clusters with both protein function and associated phenotype. Our clusters can be analyzed graphically and interactively online.

  2. Creating and Evaluating Metadata for a Large Legacy Thesis Collection: From 'Vocational Agriculture' (1922) to 'Microemulsion-mediated syntheses' (2004)

    E-Print Network [OSTI]

    Potvin, Sarah; Creel, James Silas

    2013-05-23T23:59:59.000Z

    was dependent on this body of MARC data. Over eight decades, though, we see shifts in A&M?s MARC records. As Surratt and Hill (2004) observe in their review of thesis and dissertation cataloging at Texas A&M through its history, the evolution of the MARC... Working Group, these standards aimed to prevent data loss when mapping from MODS to the Dublin Core standard of the DSpace TDL ETD repository. National ?? Networked Digital Library of Theses and Dissertations, ETD-MS: An Interoperability Metadata...

  3. The image database stores metadata and URL addresses to JPEG images stored on a web server.Each image on the web

    E-Print Network [OSTI]

    Hall, Sharon J.

    The image database stores metadata and URL addresses to JPEG images stored on a web server. This database integration would allow for the production of guides that could carry titles such as Flora for the production of both custom keys and natural language descriptions.This database structure was initially based

  4. Web Site Metadata

    E-Print Network [OSTI]

    Wilde, Erik; Roy, Anuradha

    2009-01-01T23:59:59.000Z

    Chuan Chang. Accessing the Deep Web. Communications of theand Alon Halevy. Google’s Deep Web Crawl. In Proceedings ofuseful for exposing the Deep Web [6, 8], for example, those

  5. landstatus_metadata_20100820.pdf

    National Nuclear Security Administration (NNSA)

    AFDC Printable Version Share this resource Send a link to EERE: Alternative Fuels Data Center Home Page to someone by E-mail Share EERE: Alternative Fuels Data Center Home Page on Facebook Tweet about EERE: Alternative Fuels Data Center Home Page on Twitter Bookmark EERE: Alternative1 First Use of Energy for All Purposes (Fuel and Nonfuel), 2002; Level: National5Sales for4,645 3,625 1,006 492 742EnergyOn AprilAElectronic Input Options Gary L. HirschOccurrencei-rapter | National|kcp |lafo |¡

  6. Metadata for Open Access E-Journals

    E-Print Network [OSTI]

    Mercer, Holly

    2011-04-14T23:59:59.000Z

    : NUMBER:Communicate the results A3 Face to face meetings Copyright restrictions, licences Publishers and infomediaries Funding for research communication Scientific data repositories New scientific knowledge Publication New empirical data... Informally communicated results Data in repositories ICT tools 1 Communicate the results informally 2 Communicate the results through publication 3 Share the data Bo-Christer Bj?rk, 2007 NODE: TITLE: NUMBER:Facilitate dissemination...

  7. Gearbox Reliability Collaborative: Gearbox Inspection Metadata

    SciTech Connect (OSTI)

    Munch, K.; McDade, M.

    2010-09-01T23:59:59.000Z

    NREL has developed a software application called GearFacts that will be distributed to Database Collaborative participants to capture failure information when gearboxes are rebuilt. This publication names and describes each field used in GearFacts and can be used as a reference to the software's design.

  8. A Translation Layer to Convey Preservation Metadata

    E-Print Network [OSTI]

    Di Iorio, Angela

    2009-01-01T23:59:59.000Z

    and Spoke Framework. Digital Library Federation, 26 Aprilis very common in the digital library community; MPEG21-DIDLof the Italian national Digital Library Portal and Cultural-

  9. New Metadata Organizer Streamlines JGI Data Management

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Archive (DnA) file system went live, started as a collaboration between the Sequence Data Management (SDM), Quality Assurance and Quality Control and Genome Assembly groups at JGI....

  10. Brady's Geothermal Field Seismic Network Metadata - Datasets...

    Open Energy Info (EERE)

    Relationship Dataset Dataset extent Map data OpenStreetMap contributors Tiles by MapQuest License Creative Commons Attribution 4.0 Open Data Author University of...

  11. New Metadata Organizer Streamlines JGI Data Management

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    AFDC Printable Version Share this resource Send a link to EERE: Alternative Fuels Data Center Home Page to someone by E-mail Share EERE: Alternative Fuels Data Center Home Page on Facebook Tweet about EERE: Alternative Fuels Data Center Home Page on Twitter Bookmark EERE: Alternative1 First Use of Energy for All Purposes (Fuel and Nonfuel), 2002; Level: National5Sales for4,645U.S. DOE Office of Science (SC)Integrated CodesTransparency VisitSilverNephelineNeuralNewIdeas Springgains inSkyNew

  12. Multi-Dimensional Hashed Indexed Metadata System

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    AFDC Printable Version Share this resource Send a link to EERE: Alternative Fuels Data Center Home Page to someone by E-mail Share EERE: Alternative Fuels Data Center Home Page on Facebook Tweet about EERE: Alternative Fuels Data Center Home Page on Twitter Bookmark EERE: Alternative1 First Use of Energy for All Purposes (Fuel and Nonfuel), 2002; Level: National5Sales for4,645U.S. DOEThe Bonneville PowerCherries 82981-1cnHighandSWPA / SPRA /Ml'.SolarUS Dept ofActing Chiefof Inks and Toners |

  13. Extensions to Metadata Languages for Environmental Primary Data

    E-Print Network [OSTI]

    IBANEZ, LUIS M.

    2011-01-01T23:59:59.000Z

    timber supply chain. Croatian Journal of Forest Engineer-timber supply chain. Croatian Journal of Forest Engineer-

  14. CT Scans of Cores Metadata, Barrow, Alaska 2015

    DOE Data Explorer [Office of Scientific and Technical Information (OSTI)]

    Katie McKnight; Tim Kneafsey; Craig Ulrich

    Individual ice cores were collected from Barrow Environmental Observatory in Barrow, Alaska, throughout 2013 and 2014. Cores were drilled along different transects to sample polygonal features (i.e. the trough, center and rim of high, transitional and low center polygons). Most cores were drilled around 1 meter in depth and a few deep cores were drilled around 3 meters in depth. Three-dimensional images of the frozen cores were constructed using a medical X-ray computed tomography (CT) scanner. TIFF files can be uploaded to ImageJ (an open-source imaging software) to examine soil structure and densities within each core.

  15. Dynamically exploiting available metadata for browsing and information retrieval

    E-Print Network [OSTI]

    Sinha, Vineet, 1978-

    2004-01-01T23:59:59.000Z

    Systems trying to help users deal with information overload need to be able to support the user in an information-centric manner, and need to support portions of the information which are structured -like creation dates ...

  16. Manakin Case Study: visualizing geospatial metadata and complex items

    E-Print Network [OSTI]

    Mikeal, Adam; Green, Cody; Maslov, Alexey; Phillips, Scott; Weimer, Kathy; Leggett, John

    2007-07-16T23:59:59.000Z

    :/handle.tamu.edu/1969.1/2490, and has been featured as an Editor’s Pick on Yahoo.com for its use of the Yahoo! Maps API. This presentation is the third in a set of thre about the Manakin project. This presentation discusses a specific use case of using Manakin...-based interface obvious choice for context Yahoo! Maps chosen for ease of API and aesthethic qualities Geographic coordinates were available for every item Map places every item into a geographic context Allows user to quickly determine coverage area...

  17. Extensions to Metadata Languages for Environmental Primary Data

    E-Print Network [OSTI]

    IBANEZ, LUIS M.

    2011-01-01T23:59:59.000Z

    Cli?s, NJ, 1995. [210] SK Saari, CD Campbell, J. Russell, IJCli?s, NJ, 1995. [210] SK Saari, CD Campbell, J. Russell, IJ

  18. Metadata of the chapter that will be visualized online

    E-Print Network [OSTI]

    and display video material with a quantum leap in quality in economic fashion. Here quality is measured

  19. Extensions to Metadata Languages for Environmental Primary Data

    E-Print Network [OSTI]

    IBANEZ, LUIS M.

    2011-01-01T23:59:59.000Z

    Climate Science Modelling Language 51. CSML Customizable85. EML Educational Modelling Language 86. EPML EPC MarkupLanguage Educational Modelling Language Rule Markup Language

  20. Striking a Balance : Metadata Creation in Digital Library Projects

    E-Print Network [OSTI]

    Mercer, Holly

    2003-10-10T23:59:59.000Z

    As institutions develop digital libraries, the work of describing these collections is increasingly taking place outside the library. Content creators and resource authors catalog digital content, yet may not have experience with descriptive...

  1. Exploiting Reflection and Metadata to build Mobile Computing Middleware

    E-Print Network [OSTI]

    Mascolo, Cecilia

    Emmerich and Cecilia Mascolo Dept. of Computer Science University College London Gower Street, London, WC1E 6BT, UK {L.Capra|W.Emmerich|C.Mascolo}@cs.ucl.ac.uk 1 Introduction Recent advances in wireless

  2. Embedding Metadata and Other Semantics In Word-Processing Documents

    E-Print Network [OSTI]

    Sefton, Peter; Barnes, Ian; Ward, Ron; Downing, Jim

    2008-12-08T23:59:59.000Z

    in marking up text inline. We are working on extending this to include other domains, including chemistry. One approach will be to run the 8 chemical semantics engine over a document as part of The-OREM while it still being authored, getting it to mark... is by semantically rich documents along the lines of the Datument (P. Murray-Rust & H. S. Rzepa 2004). But more importantly for users, the techniques we describe here are designed to assist in getting work done, by making it easier to submit their work...

  3. Scalable Parallel File System for Data and Metadata-intensive...

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Host: Rob Ross A critical purpose for parallel file systems used in high performance computing is to capture quickly and durably hold checkpoints of long running massive...

  4. LWA Station-Level Observing Procedure and Associated Metadata

    E-Print Network [OSTI]

    Ellingson, Steven W.

    output, which is captured by MCS data recorders (MCS-DRs). The latter is documented in the ICDs of the DP of data from the session-specified DP output, during which the parameters defining the observation (as a specified right ascension (RA) and declination (DEC). · Solar Tracking (TRK SOL). This is a beam output mode

  5. Using Characteristics of Computational Science Schemas for Workflow Metadata Management *

    E-Print Network [OSTI]

    Plale, Beth

    @cs.indiana.edu Abstract Computational science workflows are generating an ever-increasing volume of data products and is not static but instead is subject to frequent updates and additions. In contrast to general XML data of scientific workflows and data products [3]. As noted in a recent study funded by the UK e-Science Core

  6. Extensions to Metadata Languages for Environmental Primary Data

    E-Print Network [OSTI]

    IBANEZ, LUIS M.

    2011-01-01T23:59:59.000Z

    a posteriori”. The Shapefiles have been the most popularsuch as Excel files or Shapefiles. These few file types haveworking with spatial data, Shapefiles were and still are the

  7. Professional Worker Career Experience Survey (PWCES) Data and Metadata

    E-Print Network [OSTI]

    Rosenbloom, Joshua L.; Ash, Ronald A.

    2009-10-06T23:59:59.000Z

    qname,question,QPage,type,DataType,value,answer,Measuring,Source Q1,My current job title is,1,text,text,,,Work History,ITWF Q2,How long have you been in your current position?,1,text,numeric,,,,ITWF Q3,How long have you been with your current... employer?,1,text,numeric,,,Work History,ITWF Q4,My current career field is classified as,1,select,numeric,99,Select One,,ITWF Q4,,1,select,numeric,1,Management/Financial/Professional ,, Q4,,1,select,numeric,2,Computer/Mathematical ,, Q4,,1,select,numeric,3...

  8. http://factfinder.census.gov/servlet/MetadataBrowserServlet?typ

    National Nuclear Security Administration (NNSA)

    AFDC Printable Version Share this resource Send a link to EERE: Alternative Fuels Data Center Home Page to someone by E-mail Share EERE: Alternative Fuels Data Center Home Page on Facebook Tweet about EERE: Alternative Fuels Data Center Home Page on Twitter Bookmark EERE: Alternative1 First Use of Energy for All Purposes (Fuel and Nonfuel), 2002; Level: National5Sales for4,645 3,625 1,006 492 742EnergyOn AprilAElectronic Input Options Gary L. Hirsch SNLMaythemeansI MEDIAN HOUSEHOLD INCOME IN

  9. http://factfinder.census.gov/servlet/MetadataBrowserServlet?typ

    National Nuclear Security Administration (NNSA)

    AFDC Printable Version Share this resource Send a link to EERE: Alternative Fuels Data Center Home Page to someone by E-mail Share EERE: Alternative Fuels Data Center Home Page on Facebook Tweet about EERE: Alternative Fuels Data Center Home Page on Twitter Bookmark EERE: Alternative1 First Use of Energy for All Purposes (Fuel and Nonfuel), 2002; Level: National5Sales for4,645 3,625 1,006 492 742EnergyOn AprilAElectronic Input Options Gary L. Hirsch SNLMaythemeansI MEDIAN HOUSEHOLD INCOME

  10. ARM - Steps for Submitting Field Campaign Data and Metadata

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    AFDC Printable Version Share this resource Send a link to EERE: Alternative Fuels Data Center Home Page to someone by E-mail Share EERE: Alternative Fuels Data Center Home Page on Facebook Tweet about EERE: Alternative Fuels Data Center Home Page on Twitter Bookmark EERE: Alternative1 First Use of Energy for All Purposes (Fuel and Nonfuel), 2002; Level: National5Sales for4,645 3,625govInstrumentstdmadap Documentation TDMADAP :ProductsVaisalaAlaskaInstruments NSASteering Committee Related

  11. OpenEI:Projects/Tools Metadata | Open Energy Information

    Open Energy Info (EERE)

    AFDC Printable Version Share this resource Send a link to EERE: Alternative Fuels Data Center Home Page to someone by E-mail Share EERE: Alternative Fuels Data Center Home Page on Facebook Tweet about EERE: Alternative Fuels Data Center Home Page on Twitter Bookmark EERE: Alternative Fuels Data Center Home Page onYou are now leaving Energy.gov You are now leaving Energy.gov You are being directedAnnual SiteofEvaluatingGroup |JilinLuOpenNorthOlympia GreenThe community Energy Tools Strategic

  12. COLLOQUIUM: Cybersnooping: Collection and Analysis of Metadata and Content

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    AFDC Printable Version Share this resource Send a link to EERE: Alternative Fuels Data Center Home Page to someone by E-mail Share EERE: Alternative Fuels Data Center Home Page on Facebook Tweet about EERE: Alternative Fuels Data Center Home Page on Twitter Bookmark EERE: Alternative1 First Use of Energy for All Purposes (Fuel and Nonfuel), 2002; Level: National5Sales for4,645 3,625govInstrumentstdmadapInactiveVisiting the TWPSuccess Stories Siteandscience,Institute forPrinceton PlasmaLab|

  13. Integrating Metadata Standards to Support Long-Term Preservation of Digital Assets: Developing Best Practices for Expressing Preservation Metadata in a Container Format

    E-Print Network [OSTI]

    Guenther, Rebecca; Wolfe, Robert

    2009-01-01T23:59:59.000Z

    Standard (METS) in digital library applications as ausing these two digital library standards together presentsas an initiative of the Digital Library Federation. METS is

  14. Habitat-Lite: A GSC case study based on free text terms for environmental metadata

    E-Print Network [OSTI]

    Hirschman, Lynette

    2008-01-01T23:59:59.000Z

    soil, water, sediment, sludge, or air)? If so, then assignthermophilic methanogenic sludge waste water of papersediment ENVO:00002007 8 sludge ENVO:00002044 9 waste water

  15. The Nonparametric Metadata Dependent Relational Model Dae Il Kim daeil@cs.brown.edu

    E-Print Network [OSTI]

    Sudderth, Erik

    relational (NMDR) model, a Bayesian nonparametric stochastic block model for network data. The NMDR al- lows observed node relationships. Ret- rospective MCMC methods allow our sam- pler to work directly interests. Wang & Wong (1987) proposed the latent stochastic blockmodel (LSB) as a generalization of mixture

  16. Object-Oriented Database for Managing Building Modeling Components and Metadata: Preprint

    SciTech Connect (OSTI)

    Long, N.; Fleming, K.; Brackney, L.

    2011-12-01T23:59:59.000Z

    Building simulation enables users to explore and evaluate multiple building designs. When tools for optimization, parametrics, and uncertainty analysis are combined with analysis engines, the sheer number of discrete simulation datasets makes it difficult to keep track of the inputs. The integrity of the input data is critical to designers, engineers, and researchers for code compliance, validation, and building commissioning long after the simulations are finished. This paper discusses an application that stores inputs needed for building energy modeling in a searchable, indexable, flexible, and scalable database to help address the problem of managing simulation input data.

  17. FY 2008-2009 F&W Program Accords (MOA) Proposal Review Table 1. Proposal Metadata

    E-Print Network [OSTI]

    of rainbow trout, the investigation of how winter ecology of each present life history is associated rainbow trout over the study period. This critical missing information on life histories and ecology, mining, and cattle grazing and the impacts can be observed in the San Poil through cemented sediments

  18. A metadata approach for clinical data management in translational genomics studies in breast cancer

    E-Print Network [OSTI]

    Papatheodorou, Irene; Crichton, Charles; Morris, Lorna; Maccallum, Peter; METABRIC Group, Molecular Taxonomy of Breast Cancer International Consortium; Davies, Jim; Brenton, James D; Caldas, Carlos

    2009-11-30T23:59:59.000Z

    Laboratory, Wolfson Building, Parks Road, Oxford, OX1 3QD, UK Email: Irene Papatheodorou* - ivp21@cam.ac.uk; Charles Crichton - Charles.Crichton@comlab.ox.ac.uk; Lorna Morris - ljm59@cam.ac.uk; Peter Maccallum - peter.maccallum@cancer.org.uk; METABRIC Group... Description Framework (RDF) [3,4] and the Web Ontology Language (OWL) [5] are still very much under development, they are already being applied widely in business, government, and scientific contexts [6]. Two Cancer Informatics projects, caBIG [7-10] and Can...

  19. A metadata-aware application for remote scoring and exchange of tissue microarray images

    E-Print Network [OSTI]

    Morris, Lorna; Tsui, Andrew; Crichton, Charles; Harris, Steve; Maccallum, Peter H; Howat, William J; Davies, Jim; Brenton, James D; Caldas, Carlos

    2013-05-01T23:59:59.000Z

    , and analysis. Although auto- mated techniques for the quantification of staining in images are being developed [3] manual scoring is con-and exchange of tissue m Lorna Morris1, Andrew Tsui4, Charles Crichton4, Steve Har James D Brenton1,2,3 and Carlos Caldas1,2,3... . BMC Bioinformatics 2013, 14:147 Page 3 of 10 http://www.biomedcentral.com/1471-2105/14/147Table 1 Comparison of cancergrid common data elements informatics data exchange specification (DES) Cancergrid CDE name API DES XML Donor Tissue Block Identifier...

  20. Dublin Core Metadata Initiative, 1995-2009 Biblioteka Narodowa, tlumaczenie polskie, 2009

    E-Print Network [OSTI]

    Rodriguez, Carlos

    -terms/ Tlumaczenia : http://dublincore.org/resources/translations/ Status dokumentu : Zalecenie DCMI. Opis : Dokument

  1. DDI: Capturing metadata throughout the research process for preservation and discovery

    E-Print Network [OSTI]

    Thomas, Wendy

    2013-04-04T23:59:59.000Z

    .01 2001 Version 2.1 2005 Version 3.1 2009 Version 3.2 [2013] Version 1.02 2001 Version 1.3 2002 [Version 4 20??] Version 2.5 2012 Codebook Lifecycle [Version 2.6 20??] [Version 3.3 20??] Desired Areas of Coverage Simple survey Aggregate data Complex... bibliographic records S05-4 Learning DDI: Pack S05 Copyright © GESIS – Leibniz Institute for the Social Sciences, 2010 Published under Creative Commons Attribute-ShareAlike 3.0 Unported Continued use of DDI-Codebook Current users New users whose needs are met...

  2. "Weird Machines" in ELF: A Spotlight on the Underappreciated Metadata Rebecca Shapiro

    E-Print Network [OSTI]

    Smith, Sean W.

    setup stages as well as any time the RTLD is invoked. 1 Introduction The great threat model change/shellcode into the target system and tricks the target into passing control into it) to "malicious computation" driven terms to describe the execution model and computational power of these exploit programming techniques

  3. Incorporating Camera Metadata for Attended Region Detection and Consumer Photo Classification

    E-Print Network [OSTI]

    Fan, Jianping

    amount of space for the sky in an outdoor photo to make it looks balanced. The interesting subjects-1-59593-597-7/07/0007 ...$5.00. middle of the picture and keep it integrated. If environ- mental luminance is not sufficient

  4. Metadata for Nearshore Fish Atlas of Alaska Definitions of field names

    E-Print Network [OSTI]

    of bays; kelps (e.g., Saccharina latissima, Alaria marginata) often attached to bedrock faces Catch understory kelps growing as dense, low-lying mats on rocky substrates; dominant kelps (e.g., Saccharina latissima, Cymathere triplicata, Alaria spp.) Latitude: decimal degrees, WGS84 datum, northern hemisphere

  5. Meta-Metadata: An Information Semantic Language and Software Architecture for Collection Visualization Application 

    E-Print Network [OSTI]

    Mathur, Abhinav

    2011-02-22T23:59:59.000Z

    Information collection and discovery tasks involve aggregation and manipulation of information resources. An information resource is a location from which a human gathers data to contribute to his/her understanding of something significant...

  6. isCitedBy: A Metadata Scheme for DataCite

    E-Print Network [OSTI]

    Starr, Joan; Gastl, Angela

    2011-01-01T23:59:59.000Z

    Jan (2004): "Using digital library techniques — Registrationhtml for digital libraries" Springer LNCS 3232. doi:Group. At California Digital Library, she manages EZID, a

  7. MCM LTER METADATA FILE TITLE: Bacteria enumeration in lakes of the McMurdo Dry Valleys

    E-Print Network [OSTI]

    Priscu, John C.

    with a five-liter Niskin bottle during normal LTER limnological sampling. Sub-samples for bacteria enumeration are decanted into a 1 L amber Nalgene bottle. 18 mL bacteria samples are pipetted from the amber Nalgene bottle into new 20 ml glass scintillation vials for each depth. Bacteria samples are preserved by adding 0.5 ml

  8. Open Rigging Through XML: Character Setup Utilizing Metadata and Node Based Editing

    E-Print Network [OSTI]

    Kelly, Logan Scott

    2014-02-19T23:59:59.000Z

    Editor. . . . . . . . . . . . . . . . . . . . 15 5 An example MetaDataManagerNode connected to two MetaRootNodes in Maya. . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . . 20 6 The drop down menu provided for accessing the Rig... for the node are displayed in the Node Details Panel. . . . . 28 13 A Geometry Node attached to a Rig Node. . . . . . . . . . . . . . . . 29 14 A Global Node attached to a Rig Node. . . . . . . . . . . . . . . . . . 30 15 A Spine Node added to the node graph...

  9. Metadata for PoroTomo Project Subtask 3.2 DAS at Garner Valley...

    Open Energy Info (EERE)

    Subtask 3.2 DAS at Garner Valley. Preview Go to resource distributed acousti... fiber optics geothermal Additional Info Field Value Source http:gdr.openei.orgsubmissions465...

  10. AUTHOR'S PROOF Metadata of the article that will be visualized in OnlineFirst

    E-Print Network [OSTI]

    Williams II, Robert L.

    CDDRs. Moreover, the passive robot can also serve as an independent Cartesian metrology system Foot note information Category (7) ­ System Modelling/Simulation/Control/Computer­Aided Design/Robot-Direct-Driven Robot (CDDR) with Passive SCARA Support: Theory and Simulation 2 Journal Name Journal of Intelligent

  11. Metadata of the chapter that will be visualized online Chapter Title Seismic Monitoring of Nuclear Explosions

    E-Print Network [OSTI]

    Foulger, G. R.

    Administration, Beijing, China 6 Introduction 7 The original development of nuclear weapons, and their 8 first,000 10 nuclear test explosions were conducted. About 500 of 11 these were carried out in the atmosphere with some regional concentra- 14 tions, and aroused widespread public opposition to 15 nuclear testing

  12. Configuration of Smart Environments Made Simple Combining Visual Modeling with Semantic Metadata and Reasoning

    E-Print Network [OSTI]

    the configuration of smart homes to increase individual well-being, and reconfigurations of smart environments, and in particular the smart home by integrating ideas from pervasive computing and the Internet of Things (Io number of smart devices in their home and workplace environments and despite the necessary technologies

  13. Metadata of the chapter that will be visualized online Chapter Title Andean Glaciers

    E-Print Network [OSTI]

    Massachusetts at Amherst, University of

    into tropical glaciers, located in Venezuela, Colombia, 14 Ecuador, Peru, Bolivia and northernmost Chile, and 15 energy balance and their sensitivity to climate change 27 (climate variability and high altitude of the twentieth century was 53slightly less than 600 km2 . Glaciers in Bolivia can be 54found in two main mountain

  14. http://water.usgs.gov/GIS/metadata/usgswrd/XML/nv_dtw750nv_l.xm

    National Nuclear Security Administration (NNSA)

    AFDC Printable Version Share this resource Send a link to EERE: Alternative Fuels Data Center Home Page to someone by E-mail Share EERE: Alternative Fuels Data Center Home Page on Facebook Tweet about EERE: Alternative Fuels Data Center Home Page on Twitter Bookmark EERE: Alternative1 First Use of Energy for All Purposes (Fuel and Nonfuel), 2002; Level: National5Sales for4,645 3,625 1,006 492 742EnergyOn AprilAElectronic Input Options Gary L. Hirsch SNLMaythemeansIHome * About * Archives

  15. Metadata of the chapter that will be visualized online Chapter Title Assessing Habitat Fragmentation Effects on Primates: The Importance of

    E-Print Network [OSTI]

    Arroyo RodrĂ­guez, VĂ­ctor

    , Canada, H3A 2T7 Author Family Name Reyna-Hurtado Particle Given Name Rafael Suffix Organization ECOSUR-Campeche, Avenida Rancho s/n Address Poligono 2, Lerma, 24500El, Campeche, Mexico Author Family Name Fahrig Particle

  16. Metadata of the chapter that will be visualized online Chapter Title Climate Variability and High Altitude Temperature and Precipitation

    E-Print Network [OSTI]

    Vuille, Mathias

    in instrumentation, satellite telemetry, and 41power supply through solar panels have made high eleva- 42tion different from nearby 19 low elevations. Mountain-induced dynamic and thermo- 20 dynamic processes modify (Environmental Lapse 61Rate), although this rate varies by region, season, time of 62day, and by the type of air

  17. 208 Int. J. Metadata, Semantics and Ontologies, Vol. 5, No. 3, 2010 Querying structured information sources on the Web

    E-Print Network [OSTI]

    Freire, Juliana

    (UFRGS), Porto Alegre 91501-970, Brazil E-mail: mergen@inf.ufrgs.br *Corresponding author Juliana Freire-970, Brazil E-mail: heuser@inf.ufrgs.br Abstract: To provide access to heterogeneous data distributed over Grande do Sul, Porto Alegre, Brazil. He achieved his MSc title at the same institute. He spent part

  18. Slide18 | OSTI, US Dept of Energy, Office of Scientific and Technical...

    Office of Scientific and Technical Information (OSTI)

    and content managers a secure tool to quickly retrieve Agency metadata, add or edit resource records, and expedite the maintenance and quality control of the metadata and URLs....

  19. Slide16 | OSTI, US Dept of Energy, Office of Scientific and Technical...

    Office of Scientific and Technical Information (OSTI)

    and content managers a secure tool to quickly retrieve Agency metadata, add or edit resource records, and expedite the maintenance and quality control of the metadata and URLs...

  20. Slide15 | OSTI, US Dept of Energy, Office of Scientific and Technical...

    Office of Scientific and Technical Information (OSTI)

    Three Points on Nano Info Diffusion -Modeling - it's possible -Metadata - numeric data, unlike textual data, requires metadata to ensure access -Stewardship - numeric data...

  1. Slide21 | OSTI, US Dept of Energy, Office of Scientific and Technical...

    Office of Scientific and Technical Information (OSTI)

    IWGDD Policies Under Consideration Metadata - numeric data, unlike textual data, requires metadata to ensure access Not all data is worth preserving. Numeric databases need a...

  2. gulf_of_mexico_90mwindspeed_off

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    gulfofmexico90mwindspeedoff Metadata also available as Metadata: IdentificationInformation DataQualityInformation SpatialDataOrganizationInformation SpatialReferenceInf...

  3. Feed Feeds: Managing Feeds Using Feeds

    E-Print Network [OSTI]

    Wilde, Erik; Pesenson, Igor

    2008-01-01T23:59:59.000Z

    standards can be used to manage feed metadata, making feed management available through an established API.

  4. Lab 3: GIS Data Exploration Camp W.G. Williams

    E-Print Network [OSTI]

    Ramsey, R. Douglas

    should) be used to create, edit, and view metadata associated with shapefiles, raster datasets, and Arc

  5. DataONE currently holds a collection of over 125,000 metadata records linked to data sets in 10 environmental and earth science data reposito-

    E-Print Network [OSTI]

    California at Santa Barbara, University of

    analysis cover set analysis cover set forest hemlock species forest hemlock species m biomass cm m biomass discoverable through DataONE.org a t c a l i f o r n i a drinking assessment supply drinking assessment supply nitrogen co2 nitrogen co2 biomass primary biomass primary soil n organic soil n organic river marsh island

  6. A Metaphor Graphics Based Representation of Digital Libraries on the World Wide Web: Using the libViewer to Make Metadata Visible

    E-Print Network [OSTI]

    Rauber,Andreas

    A Metaphor Graphics Based Representation of Digital Libraries on the World Wide Web: Using the lib way to go. Most interfaces to digital libraries present themselves as various forms of sorted lists interface to digital libraries using metaphor graphics to display information on the el­ ements in a digital

  7. A Metaphor Graphics Based Representation of Digital Libraries on the World Wide Web: Using the libViewer to Make Metadata Visible

    E-Print Network [OSTI]

    Rauber,Andreas

    A Metaphor Graphics Based Representation of Digital Libraries on the World Wide Web: Using the lib way to go. Most interfaces to digital libraries present themselves as various forms of sorted lists interface to digital libraries using metaphor graphics to display information on the el- ements in a digital

  8. Home Page| Database Search| Online Tools| MetaData Query| Standard Criterion| Data Center| About BMICC| Login| | --> English News --> Biology News

    E-Print Network [OSTI]

    Sóbester, András

    pipelines, and methane naturally leaking from the seabed, could in future be detected using changes into the atmosphere of gases from the seabed. This applies both to gas extracted by the petrochemical industries and to the methane which is naturally released from the seabed." Natural leaks of methane gas can be damaging

  9. Slide07 | OSTI, US Dept of Energy, Office of Scientific and Technical...

    Office of Scientific and Technical Information (OSTI)

    Metadata Is a Must (when it comes to numeric data) Numeric databases and other non-text databases must have metadata to enable searchability and retrieval Numeric databases must...

  10. Slide08 | OSTI, US Dept of Energy, Office of Scientific and Technical...

    Office of Scientific and Technical Information (OSTI)

    Slide 8: Metadata Is a Must (when it comes to numeric data) -Numeric databases and other non-text databases must have metadata to enable searchability and retrieval -Numeric...

  11. iEMSs 2008:International Congress on Environmental Modelling and Software Integrating Sciences and Information Technology for Environmental Assessment and Decision Making

    E-Print Network [OSTI]

    Hoffman, Forrest M.

    to be based on results from integrated Earth System Models (ESMs), C-LAMP is helping to establish the metadata

  12. Slide14 | OSTI, US Dept of Energy, Office of Scientific and Technical...

    Office of Scientific and Technical Information (OSTI)

    * Multimedia presents special opportunities and challenges - lack of written transcripts, minimal metadata, scientifictechnicalmedical terminology, lengthy videos (>1 hour)...

  13. Slide12 | OSTI, US Dept of Energy, Office of Scientific and Technical...

    Office of Scientific and Technical Information (OSTI)

    Multimedia presents special opportunities and challenges - lack of written transcripts, minimal metadata, scientifictechnicalmedical terminology, lengthy videos (>1 hour)...

  14. Slide20 | OSTI, US Dept of Energy, Office of Scientific and Technical...

    Office of Scientific and Technical Information (OSTI)

    presents special opportunitiesand challenges - lack of written transcripts, minimal metadata, scientifictechnicalmedical terminology, lengthy videos (>1 hour) ScienceCinema...

  15. Advanced Hydrologic Prediction Service (AHPS) Website

    E-Print Network [OSTI]

    precipitation shapefiles will include metadata fields for both valid date and date/time the file was created

  16. Resources Annotation, Retrieval and Presentation: a Semantic Annotation Management System

    E-Print Network [OSTI]

    Paris-Sud XI, Université de

    for the management of metadata records, each attribute of the record being represented by one or more triples to the resource management system itself. Expressing metadata records in RDF allows us to see metadata as semanticResources Annotation, Retrieval and Presentation: a Semantic Annotation Management System M. Albert

  17. C Y B E R C E N T E R S E M I N A R S E R I E S Metadata and Provenance: Fins in the Sea of Data

    E-Print Network [OSTI]

    Fernández-Juricic, Esteban

    and Computation Practice and Experience. Presentation Abstract The data deluge and discussions of its implications, that it was a topic at the 2012 Davos World Economic Forum. The data deluge of interest to computer scientists, this deluge will be a catalyst for new scientific discovery that fuels advances in grand challenge questions

  18. Walter Cazzola, Sonia Pini, and Massimo Ancona. The Role of Design Information in Software Evolution. In Proceedings of the 2nd ECOOP Workshop on Reflection, AOP and Meta-Data for

    E-Print Network [OSTI]

    Cazzola, Walter

    life cycle have been widely studied in software system development. These methodologies can be used problem is the fact that models are shown like only an intermediate step in the software development life-cycle exist during all the software cycle. Often, the early stages of development, the specifications

  19. Slide02 | OSTI, US Dept of Energy, Office of Scientific and Technical...

    Office of Scientific and Technical Information (OSTI)

    Slide 2: Three Topics Relating to Nano Info Diffusion We see three complementary approaches to improve information sharing and awareness - Modeling - it's possible - Metadata -...

  20. Slide02 | OSTI, US Dept of Energy, Office of Scientific and Technical...

    Office of Scientific and Technical Information (OSTI)

    Three Topics Relating to Nano Info Diffusion We see three complementary approaches to improve information sharing and awareness Modeling - It's possible. Metadata - numeric data,...

  1. Slide14 | OSTI, US Dept of Energy, Office of Scientific and Technical...

    Office of Scientific and Technical Information (OSTI)

    In Summary: Three Points on Nano Info Diffusion In Summary: Three Points on Nano Info Diffusion. Link to larger image. Modeling - It's possible. Metadata - numeric data, unlike...

  2. File:Texas NOI for Storm Water Discharges Associated with Construction...

    Open Energy Info (EERE)

    NOI for Storm Water Discharges Associated with Construction Activities (TXR150000).pdf Jump to: navigation, search File File history File usage Metadata File:Texas NOI for Storm...

  3. atmyb44 enhances stomatal: Topics by E-print Network

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    simulation to optimize carbon dioxide (CO2) sequestration and enhance oil recovery (CO2-EOR) based on known 169 Metadata Models for Technology Enhanced Learning Computer...

  4. andongensis extract aloe: Topics by E-print Network

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Documents Articles, Books DocbookIndexer PDFBox PDF Metadata from content & context Fill in generated... Droogmans, Lieven; Bosman, Ben 2005-07-07 70 Point Source Extraction...

  5. aloe andongensis extract: Topics by E-print Network

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Documents Articles, Books DocbookIndexer PDFBox PDF Metadata from content & context Fill in generated... Droogmans, Lieven; Bosman, Ben 2005-07-07 70 Point Source Extraction...

  6. File:CDPHE Industrial Individual Wastewater Discharge Permit...

    Open Energy Info (EERE)

    Industrial Individual Wastewater Discharge Permit Application.pdf Jump to: navigation, search File File history File usage Metadata File:CDPHE Industrial Individual Wastewater...

  7. Slide19 | OSTI, US Dept of Energy, Office of Scientific and Technical...

    Office of Scientific and Technical Information (OSTI)

    Agency Content Managers Identify New Websites To Be CrawledIndexed The Metadata Input System "Add Record" page allows Alliance content managers to add new records using agency,...

  8. Slide12 | OSTI, US Dept of Energy, Office of Scientific and Technical...

    Office of Scientific and Technical Information (OSTI)

    Slide 12: Establishing Ground Rules for Metadata Enables Access to Data Overcome barriers of organizations -International Portal -Interagency Portal -Agency Portal -Lab or...

  9. Slide11 | OSTI, US Dept of Energy, Office of Scientific and Technical...

    Office of Scientific and Technical Information (OSTI)

    Establishing Ground Rules for Metadata Enables Access to Data Overcome barriers of organizations International Portal Interagency Portal Agency Portal Lab or Institute Portal Nano...

  10. al disciplinamiento social: Topics by E-print Network

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Information Technology) Prof. Dr. Andreas Rausch (Software Systems Engineering Zachmann, Gabriel 3 Intagorn et al. Harvesting Geospatial Knowledge from Social Metadata...

  11. National Geothermal Data System Design and Testing

    Office of Energy Efficiency and Renewable Energy (EERE) Indexed Site

    Technology, Nevada Bureau of Mines and Geology- University of Nevada-Reno, and Stanford University - Data providers are updating catalogues of their repositories with metadata...

  12. A light detection and ranging (lidar) study of the Sierra Nevada

    E-Print Network [OSTI]

    Phelps, Gary M. II

    2011-01-01T23:59:59.000Z

    1997). "The Stanford Digital Library metadata architecture."International Journal on Digital Libraries 1(2): 108-121.working web-based digital library for environmental

  13. Slide13 | OSTI, US Dept of Energy, Office of Scientific and Technical...

    Office of Scientific and Technical Information (OSTI)

    customized S&T information tools and services requested by individual DOE offices on a cost-recovery basis: * Information science and subject-matter expertise * Metadata and...

  14. File:App Commercial Leases and Easements or Amendment or Residential...

    Open Energy Info (EERE)

    App Commercial Leases and Easements or Amendment or Residential Coastal Easements HOA.pdf Jump to: navigation, search File File history File usage Metadata File:App Commercial...

  15. OSTI, US Dept of Energy, Office of Scientific and Technical Informatio...

    Office of Scientific and Technical Information (OSTI)

    transcripts, minimal metadata, scientifictechnicalmedical terminology, lengthy videos (>1 hour) Slide13 Slide13 Access to Multimedia-based Science & Technology A Case Study...

  16. OSTI, US Dept of Energy, Office of Scientific and Technical Informatio...

    Office of Scientific and Technical Information (OSTI)

    transcripts, minimal metadata, scientifictechnicalmedical terminology, lengthy videos (>1 hour) Slide15 Slide15 Access to Multimedia-based Science & Technology A Case Study...

  17. Slide11 | OSTI, US Dept of Energy, Office of Scientific and Technical...

    Office of Scientific and Technical Information (OSTI)

    no "full text" to search * Metadata, if available, is often minimal * Scientific, technical, and medical terminologyvocabulary * Videos can be long, often up to an hour or more...

  18. Alexandria Digital Library Project The ADEPT Digital Library

    E-Print Network [OSTI]

    Janée, Greg

    Alexandria Digital Library Project The ADEPT Digital Library Architecture Greg Janée gjanee@alexandria.ucsb.edu James Frew frew@bren.ucsb.edu #12;Alexandria Digital Library Project 2Greg Janée, James Frew · JCDL 2002-level metadata: buckets constraint types metadata mapping standard buckets Collection discovery Current

  19. Journal of Information Science, XX (X) 2009, pp. 122 DOI: 10.1177/0165551506nnnnnn 1 The Author(s), 2009, Reprints and Permissions: http://www.sagepub.co.uk/journalsPermissions.nav

    E-Print Network [OSTI]

    Menczer, Filippo

    increase of websites has generated an information deluge, creating an often confusing and overwhelming. As information deluge turns into data deluge, there is a need to add metadata to data, a process already in place. In turn metadata deluge will become another deluge, with no end in sight to this information abstraction

  20. National Environmental Public Health Tracking Network

    E-Print Network [OSTI]

    .9 available for use by grantees 9/4/2007 9/4/2007 Nationally Consistent Data and Measures Develop draft Metadata Services Provide training on and test Metadata Tool V0.9 for describing grantee data 9 of NCDM 10/31/2007 12/1/2007 Nationally Consistent Data and Measures Release to grantees the final

  1. Policy enabled information sharing system

    DOE Patents [OSTI]

    Jorgensen, Craig R.; Nelson, Brian D.; Ratheal, Steve W.

    2014-09-02T23:59:59.000Z

    A technique for dynamically sharing information includes executing a sharing policy indicating when to share a data object responsive to the occurrence of an event. The data object is created by formatting a data file to be shared with a receiving entity. The data object includes a file data portion and a sharing metadata portion. The data object is encrypted and then automatically transmitted to the receiving entity upon occurrence of the event. The sharing metadata portion includes metadata characterizing the data file and referenced in connection with the sharing policy to determine when to automatically transmit the data object to the receiving entity.

  2. Small file aggregation in a parallel computing system

    DOE Patents [OSTI]

    Faibish, Sorin; Bent, John M.; Tzelnic, Percy; Grider, Gary; Zhang, Jingwang

    2014-09-02T23:59:59.000Z

    Techniques are provided for small file aggregation in a parallel computing system. An exemplary method for storing a plurality of files generated by a plurality of processes in a parallel computing system comprises aggregating the plurality of files into a single aggregated file; and generating metadata for the single aggregated file. The metadata comprises an offset and a length of each of the plurality of files in the single aggregated file. The metadata can be used to unpack one or more of the files from the single aggregated file.

  3. On the Integration of Topic Modeling and Dictionary Learning Lingbo Li ll83@duke.edu

    E-Print Network [OSTI]

    Carin, Lawrence

    (LDA) (Blei et al., 2003), originally developed for text analysis, have been applied successfully as well as available meta-data such as image annotations (Barnard et al., 2003; Blei & Jordan, 2003; Blei

  4. Sharing Research Data; Building Social Science Institute for Quantitative Social Science

    E-Print Network [OSTI]

    international standards of preservation formatting, documentation, attribution, code, metadata ­ Care software via API · Full integration thus far: ~400 journals (out of 5000 using OJS) · Many more integrations in progress #12;An Integrated, Living Workflow Integration with data repositories (repository API

  5. Jason Karl, Ph.D. USDA ARS, Jornada Experimental Range

    E-Print Network [OSTI]

    .journalmap.org Twitter: @journalmap ·API Development ·Embeddable Searches/Results ·Additional search Filters ·Large-scale article geotagging ·Location- reporting standards ·Encoding location in metadata/XML Current ·Taylor

  6. TABLEFS: Embedding a NoSQL Database Inside the Local File System

    E-Print Network [OSTI]

    monolithic database management system (DBMS) bundles [31]. 1 Background Even in the era of big data, most local file systems [34]. Embedded Databases File system metadata is structured data, a natural fit

  7. Umbra: Efficient and scalable memory shadowing

    E-Print Network [OSTI]

    Zhao, Qin

    Shadow value tools use metadata to track properties of application data at the granularity of individual machine instructions. These tools provide effective means of monitoring and analyzing the runtime behavior of ...

  8. An immersive system for browsing and visualizing surveillance video

    E-Print Network [OSTI]

    DeCamp, Philip James

    HouseFly is an interactive data browsing and visualization system that synthesizes audio-visual recordings from multiple sensors, as well as the meta-data derived from those recordings, into a unified viewing experience. ...

  9. Los Alamos National Laboratory Associate Directorate for Theory, Simulation, and Computation (ADTSC) LA-UR 13-2083982 Parallel Log Structured File System (PLFS)

    E-Print Network [OSTI]

    Cooperative Research and Development Agreement (CRADA) to further enhance, design, build, test, and deploy, and metadata performance of our I/O workloads. As part of the LANL umbrella CRADA with EMC Corporation, LANL

  10. Patterns in the daily diary of the 41st president, George Bush

    E-Print Network [OSTI]

    Kumar, Shreyas

    2007-04-25T23:59:59.000Z

    This thesis explores interfaces for locating and comprehending patterns among time-based materials in digital libraries. Time-based digital library materials are like other digital library materials in that they are comprised of data and metadata...

  11. Socioeconomic Data and Applications Center (SEDAC) User Services

    E-Print Network [OSTI]

    Columbia University

    ://sedac.ciesin.columbia.edu/povmap to download global data sets of malnutrition and infant mortality rates as tables, shapefiles, and grids (1 and shapefiles; and maps, catalogs, and metadata. Project Highlights · Global and regional data concerning

  12. Calendar | OSTI, US Dept of Energy, Office of Scientific and...

    Office of Scientific and Technical Information (OSTI)

    itting241.1Metadata.ppt 2015-01-02 11:38 Restricted Announcement of U.S. Department of Energy (DOE) Classified and Unclassified Controlled Nuclear Information (UCNI) 2015-01-02...

  13. Using SAS to generate DDI-Codebook XML from Information Managed in Excel Spreadsheets

    E-Print Network [OSTI]

    Wright, Philip A.

    2013-04-02T23:59:59.000Z

    \\fem_variable_formats.txt' delimiter = '09'x MISSOVER DSD lrecl = 32767 firstobs = 2 ; input VARIABLE $ FORMAT $ ; run ; There is more than one way to import information from Excel into SAS External File Interface Import Wizard Proc Import Excel Library... into SAS Proc Import There is more than one way to import information from Excel into SAS proc import datafile = "C:\\worksheets\\female_metadata.xls" out = user.female_metadata dbms = excel ; range = "'FINAL SECTION J'" ; getnames = yes ; mixed...

  14. Next-Generation Search Engines for Information Retrieval

    SciTech Connect (OSTI)

    Devarakonda, Ranjeet [ORNL; Hook, Leslie A [ORNL; Palanisamy, Giri [ORNL; Green, James M [ORNL

    2011-01-01T23:59:59.000Z

    In the recent years, there have been significant advancements in the areas of scientific data management and retrieval techniques, particularly in terms of standards and protocols for archiving data and metadata. Scientific data is rich, and spread across different places. In order to integrate these pieces together, a data archive and associated metadata should be generated. Data should be stored in a format that can be retrievable and more importantly it should be in a format that will continue to be accessible as technology changes, such as XML. While general-purpose search engines (such as Google or Bing) are useful for finding many things on the Internet, they are often of limited usefulness for locating Earth Science data relevant (for example) to a specific spatiotemporal extent. By contrast, tools that search repositories of structured metadata can locate relevant datasets with fairly high precision, but the search is limited to that particular repository. Federated searches (such as Z39.50) have been used, but can be slow and the comprehensiveness can be limited by downtime in any search partner. An alternative approach to improve comprehensiveness is for a repository to harvest metadata from other repositories, possibly with limits based on subject matter or access permissions. Searches through harvested metadata can be extremely responsive, and the search tool can be customized with semantic augmentation appropriate to the community of practice being served. One such system, Mercury, a metadata harvesting, data discovery, and access system, built for researchers to search to, share and obtain spatiotemporal data used across a range of climate and ecological sciences. Mercury is open-source toolset, backend built on Java and search capability is supported by the some popular open source search libraries such as SOLR and LUCENE. Mercury harvests the structured metadata and key data from several data providing servers around the world and builds a centralized index. The harvested files are indexed against SOLR search API consistently, so that it can render search capabilities such as simple, fielded, spatial and temporal searches across a span of projects ranging from land, atmosphere, and ocean ecology. Mercury also provides data sharing capabilities using Open Archive Initiatives Protocol for Metadata Handling (OAI-PMH). In this paper we will discuss about the best practices for archiving data and metadata, new searching techniques, efficient ways of data retrieval and information display.

  15. System and method for integrating and accessing multiple data sources within a data warehouse architecture

    DOE Patents [OSTI]

    Musick, Charles R. (Castro Valley, CA); Critchlow, Terence (Livermore, CA); Ganesh, Madhaven (San Jose, CA); Slezak, Tom (Livermore, CA); Fidelis, Krzysztof (Brentwood, CA)

    2006-12-19T23:59:59.000Z

    A system and method is disclosed for integrating and accessing multiple data sources within a data warehouse architecture. The metadata formed by the present method provide a way to declaratively present domain specific knowledge, obtained by analyzing data sources, in a consistent and useable way. Four types of information are represented by the metadata: abstract concepts, databases, transformations and mappings. A mediator generator automatically generates data management computer code based on the metadata. The resulting code defines a translation library and a mediator class. The translation library provides a data representation for domain specific knowledge represented in a data warehouse, including "get" and "set" methods for attributes that call transformation methods and derive a value of an attribute if it is missing. The mediator class defines methods that take "distinguished" high-level objects as input and traverse their data structures and enter information into the data warehouse.

  16. ARM Standards Policy Committee Report

    SciTech Connect (OSTI)

    Cialella, A; Jensen, M; Koontz, A; McFarlane, S; McCoy, R; Monroe, J; Palanisamy, G; Perez, R; Sivaraman, C

    2012-09-19T23:59:59.000Z

    Data and metadata standards promote the consistent recording of information and are necessary to ensure the stability and high quality of Atmospheric Radiation Measurement (ARM) Climate Research Facility data products for scientific users. Standards also enable automated routines to be developed to examine data, which leads to more efficient operations and assessment of data quality. Although ARM Infrastructure agrees on the utility of data and metadata standards, there is significant confusion over the existing standards and the process for allowing the release of new data products with exceptions to the standards. The ARM Standards Policy Committee was initiated in March 2012 to develop a set of policies and best practices for ARM data and metadata standards.

  17. Database integration in a multimedia-modeling environment

    SciTech Connect (OSTI)

    Dorow, Kevin E.

    2002-09-02T23:59:59.000Z

    Integration of data from disparate remote sources has direct applicability to modeling, which can support Brownfield assessments. To accomplish this task, a data integration framework needs to be established. A key element in this framework is the metadata that creates the relationship between the pieces of information that are important in the multimedia modeling environment and the information that is stored in the remote data source. The design philosophy is to allow modelers and database owners to collaborate by defining this metadata in such a way that allows interaction between their components. The main parts of this framework include tools to facilitate metadata definition, database extraction plan creation, automated extraction plan execution / data retrieval, and a central clearing house for metadata and modeling / database resources. Cross-platform compatibility (using Java) and standard communications protocols (http / https) allow these parts to run in a wide variety of computing environments (Local Area Networks, Internet, etc.), and, therefore, this framework provides many benefits. Because of the specific data relationships described in the metadata, the amount of data that have to be transferred is kept to a minimum (only the data that fulfill a specific request are provided as opposed to transferring the complete contents of a data source). This allows for real-time data extraction from the actual source. Also, the framework sets up collaborative responsibilities such that the different types of participants have control over the areas in which they have domain knowledge-the modelers are responsible for defining the data relevant to their models, while the database owners are responsible for mapping the contents of the database using the metadata definitions. Finally, the data extraction mechanism allows for the ability to control access to the data and what data are made available.

  18. Building the International Lattice Data Grid

    E-Print Network [OSTI]

    G. Beckett; B. Joo; C. M. Maynard; D. Pleiter; O. Tatebe; T. Yoshie

    2009-10-09T23:59:59.000Z

    We present the International Lattice Data Grid (ILDG), a loosely federated grid of grids for sharing data from Lattice Quantum Chromodynamics (LQCD) simulations. The ILDG comprises of metadata, file format and web-service standards, which can be used to wrap regional data-grid interfaces, allowing seamless access to catalogues and data in a diverse set of collaborating regional grids. We discuss the technological underpinnings of the ILDG, primarily the metadata and the middleware, and offer a critique of its various aspects with the hindsight of the design work and the first full year of production.

  19. National energy information system version 1. 0 catalog of holdings

    SciTech Connect (OSTI)

    Everett, C.G.

    1980-06-01T23:59:59.000Z

    This report contains a description of Version 1.0 of the National Energy Information System (NEIS). When fully developed, NEIS will consist of: A collection of energy data organized in such a way as to be useful for analysis of energy issues. Metadata (i.e., data about the data) that document the characteristics and organization of all EIA data including the NEIS data, and provide ways of locating and accessing it. A set of procedures for gathering, updating, and disseminating NEIS data and metadata.

  20. Quality Classifiers for Open Source Software Repositories

    E-Print Network [OSTI]

    Tsatsaronis, George; Giakoumakis, Emmanouel A

    2009-01-01T23:59:59.000Z

    Open Source Software (OSS) often relies on large repositories, like SourceForge, for initial incubation. The OSS repositories offer a large variety of meta-data providing interesting information about projects and their success. In this paper we propose a data mining approach for training classifiers on the OSS meta-data provided by such data repositories. The classifiers learn to predict the successful continuation of an OSS project. The `successfulness' of projects is defined in terms of the classifier confidence with which it predicts that they could be ported in popular OSS projects (such as FreeBSD, Gentoo Portage).

  1. Implementing Self-Organising Virtual Enterprises Using Social Behaviour Nets

    E-Print Network [OSTI]

    Paris-Sud XI, Université de

    to support Internet-based virtual enterprises are currently characterised as having fixed process models, fixed meta-data models and a fixed set of users at project inception. This paper proposes the So to a combination of its own goals and state changes. The SoBeNet test-bed is based on the Java JADE environment

  2. Educating eScience Librarians Librarians are part of eScience project teams that

    E-Print Network [OSTI]

    Oakleaf, Megan

    ) knowledge of eScience/eResearch (2) science data and metadata management (3) enabling data technologies · Libraries are active players and contributors in data curation and preservation for discovery and long development and changes in eScience projects in which libraries participate · Leadership and management skills

  3. Data Working Group Determine if the University Libraries should accept broad responsibility for curating research data and, if

    E-Print Network [OSTI]

    Schweik, Charles M.

    be involved. Develop meaningful resources on data management for the University community. Activities include to evaluate preservation and metadata issues for research data, additional staff will be added members of the staff may be asked to join meetings when their specialized knowledge is needed. Information

  4. Master of Science Thesis Stockholm, Sweden 2007

    E-Print Network [OSTI]

    Maguire Jr., Gerald Q.

    . This thesis project addresses the problem of automatically distributing MSDS and the corresponding metadata. It presents the underlying technology selected for this project. It gives a brief introduction of this project underrätta en användare att det finns en nyare MSDS och sedan skicka den korresponderande filen och dess

  5. Abstract--The performance and efficiency of Web Services can be greatly increased in conversational and streaming message

    E-Print Network [OSTI]

    in conversational and streaming message exchanges by streaming the message exchange paradigm. In this paper, we describe our design and implementation of a novel approach to the message exchange optimization. This area message exchanges. The redundant or static massage parts are stored in shared metadata space ­ the Context

  6. Raising the Curtain or: getting unique and rare collections out to the public 

    E-Print Network [OSTI]

    Knowles, Claire; Renton, Scott

    2015-05-14T23:59:59.000Z

    This case study focuses on a project using a crowdsourced metadata ‘game’ to develop collections, and shows that once you have a process in place, it is not too creating engagement with your collections. The public can help solve your problem...

  7. NARSTO Quality Systems Science Center 2003-2004 Annual Report

    E-Print Network [OSTI]

    and Information Sharing Tool (DIST) · Web-based entry of metadata for archive documentation · Link data Angeles (5 data sets) ·Pittsburgh (RSMS data) ·SOS ·Atlanta 1999 Wind Profiler (UAH MIPS) ·South Carolina-AC05-00OR22725. #12;2 QSSC Activities for 2003-2004 Data Management and Archiving // External

  8. NARSTO Quality Systems Science Center 2006-2007 Annual Report

    E-Print Network [OSTI]

    and Information Sharing Tool (DIST) · Web-based entry of metadata for archive documentation · Link data Data (29 data files with plots) NARSTO EPA_SS_LOS_ANGELES Aethalometer Elemental Carbon Data (6 data files with plots) In the Near-Term Working Queue EPA_SS_FRESNO data EPA_SS_ST_LOIUS data EPA_SS_LOS_ANGELES

  9. NARSTO Quality Systems Science Center 2007-2008 Annual Report

    E-Print Network [OSTI]

    . · Maintained Data and Information Sharing Tool (DIST) · Web-based entry of metadata for archive documentation with plots) Completes archiving Pittsburgh data. NARSTO EPA_SS_LOS_ANGELES Aethalometer Elemental Carbon Data under contract DE-AC05-00OR22725. #12;QSSC Activities for 2007-2008 Data Management and Archiving

  10. This Agreement is between the author (Author) and ProQuest LLC, through its UMI Dissertation Publishing business (ProQuest/UMI). Under this Agreement, Author grants ProQuest/UMI certain rights to preserve, archive and publish the dissertation or thesis, a

    E-Print Network [OSTI]

    Taylor, Jerry

    1 This Agreement is between the author (Author) and ProQuest LLC, through its UMI® Dissertation to preserve, archive and publish the dissertation or thesis, abstract, and index terms (the Work) provided the abstract, bibliography and other metadata in the ProQuest Dissertations and Theses database (PQDT

  11. Figure 1. IDVL interface showing 12 documents returned for "El Nio" query along with different multimedia

    E-Print Network [OSTI]

    Wactlar, Howard D.

    to evaluate them. We aggregate and integrate video content on-demand to enable summarization and visualization that provides the users with coherent responses to their queries in a useful broader context, perhaps library, i.e., metadata, to return a set of video documents in response to a user query, as shown

  12. Automated Information Management via Abductive Logic Agents

    E-Print Network [OSTI]

    Toni, Francesca

    Automated Information Management via Abductive Logic Agents F.Toni 1 Imperial College of Science representation of networks and meta-data via abductive logic programs and make use of the reasoning engine abductive logic pro- grams are equipped with for checking as well as enforcing the properties represented

  13. Annotation of the Shallow and the Deep Siegfried Handschuh

    E-Print Network [OSTI]

    Staab, Steffen

    Annotation of the Shallow and the Deep Web Siegfried Handschuh˝ and Steffen Staab˝ ľ ˝ Institute a framework,CREAM, that allows the creation of semantic annotation on the Shallow and the Deep Web. Hence with the simultaneous creation of metadata, and the deep annotation. 1 Introduction The semantic web ­ the web

  14. Feeding TEL: Building an Ecosystem Around BuRST

    E-Print Network [OSTI]

    Hammerton, James

    Feeding TEL: Building an Ecosystem Around BuRST to Convey Publication Metadata Peter Kraker1, Austria {pkraker, afessl, phoefler, slind}@know-center.at Abstract. In this paper we present an ecosystem of this ecosystem, semantically enriched RSS feeds are used for dissemination. These feeds are complemented

  15. Towards Better Utilization of NEXRAD Data in Hydrology: an Overview of Hydro-NEXRAD

    E-Print Network [OSTI]

    Lawrence, Ramon

    Towards Better Utilization of NEXRAD Data in Hydrology: an Overview of Hydro-NEXRAD Witold F metadata extraction and management techniques are required. The authors describe and discuss the Hydro of the Hydro-NEXRAD project is to increase the use of NEXRAD data in hydrologic research. The project

  16. DiLiA The Digital Library Assistant Kathrin Eichler, Holmer Hemsen, Gnter Neumann, Norbert Reithinger, Sven

    E-Print Network [OSTI]

    Neumann, Günter

    to enter a search term for searching in the digital libraries metadata. In addition, the user can add items the search in digital libraries in sev- eral dimensions. In the project advanced information visualisation meth- ods are developed for user controlled interactive search. The interaction model has been designed

  17. Introduction to ORNL DAAC THREDDS Data Server

    E-Print Network [OSTI]

    Introduction to ORNL DAAC THREDDS Data Server #12;Outline · THREDDS Data Server (TDS) ­ How to UseCDF Subset Service · OGC WCS · HTTP Server · ORNL DAAC THREDDS Data Server #12;THREDDS Data Server (TDS) · The THREDDS Data Server (TDS) is a web server that provides metadata and data access for scientific data sets

  18. NEW BSE EMPHASIS: Geospatial Engineering Bachelor of Science in Engineering (BSE) Geospatial Engineering

    E-Print Network [OSTI]

    NEW BSE EMPHASIS: Geospatial Engineering Bachelor of Science in Engineering (BSE) ­ Geospatial Engineering The Geospatial Engineering pathway was defined to study geospatial concepts that include and metadata formats, and visualization. The approved plan reflects state-of-the-art geospatial research

  19. Envisioning uncertainty in geospatial information Kathryn Blackmond Laskey a,*, Edward J. Wright b

    E-Print Network [OSTI]

    Valtorta, Marco

    Envisioning uncertainty in geospatial information Kathryn Blackmond Laskey a,*, Edward J. Wright b in revised form 17 May 2009 Accepted 20 May 2009 Available online 21 June 2009 Keywords: Geospatial reasoning Geographic information systems Geospatial metadata Probabilistic ontologies Multi-entity Bayesian networks

  20. Proceedings of the 2014 Conference on Empirical Methods in Natural Language Processing (EMNLP), pages 17521757, October 25-29, 2014, Doha, Qatar. c 2014 Association for Computational Linguistics

    E-Print Network [OSTI]

    is a distribution over words. Topic models such as latent Dirichlet alloca- tion (Blei et al., 2003, LDA) and its, and summarization of the otherwise unstructured corpora (Blei, 2012; Blei, 2014). In addition to exploratory data jointly capture both the text and associated metadata such as a continuous response variable (Blei and Mc

  1. Proc. of the 12th Int. Conference on Digital Audio Effects (DAFx-09), Como, Italy, September 1-4, 2009

    E-Print Network [OSTI]

    Blei, David M.

    . of Computer Science Princeton University Princeton, New Jersey, USA mdhoffma@cs.princeton.edu David M. Blei Dept. of Computer Science Princeton University Princeton, New Jersey, USA blei@cs.princeton.edu Perry R in different combinations, and without hand-generated metadata David M. Blei is supported by ONR 175-6343, NSF

  2. FIA-12-0001- In the Matter of Kevin Leary

    Broader source: Energy.gov [DOE]

    OHA issued a decision denying a FOIA appeal filed by Kevin Leary of a final determination issued by the DOE Richland Operations Office (RO). Mr. Leary requested copies of e-mails, the associated metadata, and instant messages for Margot Voogd, Larry Romine and Bryan Foley, from August 1, 2005, through January 31, 2009. RO informed Mr. Leary that no documents existed.

  3. FIA-12-0001- In the Matter of Kevin Leary

    Broader source: Energy.gov [DOE]

    OHA issued a decision denying a FOIA appeal filed by Kevin Leary of a final determination issued by the DOE Richland Operations Office (RO). Mr. Leary requested copies of e-mails, the associated metadata, and instant messages for Margot Voogd, Larry Romine and Bryan Foley, from August 1, 2005, through January 31, 2009.

  4. Montage on the Grid Please send comments to gurmeet@isi.eduor deelman@isi.edu 1

    E-Print Network [OSTI]

    Deelman, Ewa

    of the Montage Project. 2. Porting Montage to Chimera/Pegasus Chimera is a virtual data system for representing, querying and automating data derivations [3] . Chimera provides a framework for representing a set in the data_dir and creating a metadata table called images.tbl can be expressed as the following chimera

  5. Facilitating Document Annotation using Content and Querying Value

    E-Print Network [OSTI]

    Hristidis, Vagelis

    content management software (e.g., Microsoft SharePoint), allow users to share documents and annotate1 Facilitating Document Annotation using Content and Querying Value Eduardo J. Ruiz #1 , Vagelis the generation of the structured metadata by identifying documents that are likely to contain informa- tion

  6. CSISSLCenter for Spatial Information Science and Systems

    E-Print Network [OSTI]

    Christian, Eric

    Page 1 CSISSLCenter for Spatial Information Science and Systems 03/19/2008 Survey of ISO Standards for Geospatial Metadata Liping Di ldi@gmu.edu Center for Spatial Information Science and Systems George Mason University #12;Page 2 CSISS Center for Spatial Information Science and Systems 03/19/2008 Introduction · ISO

  7. Exploratory Patent Search with Faceted Search and Configurable Entity Mining

    E-Print Network [OSTI]

    Tzitzikas, Yannis

    Exploratory Patent Search with Faceted Search and Configurable Entity Mining Pavlos Fafalios1 Searching for patents is usually a recall-oriented problem and depending on the patent search type, quite. We propose an exploratory strategy for patent search that exploits the metadata already available

  8. SPOCS User Guide

    SciTech Connect (OSTI)

    Curtis, Darren S.; Phillips, Aaron R.; McCue, Lee Ann

    2013-04-15T23:59:59.000Z

    SPOCS implements a graph-based ortholog prediction method to generate a simple tab-delimited table of orthologs, and in addition, html files that provide a visualization of the ortholog/paralog relationships to which gene/protein expression metadata may be overlaid.

  9. AgNIC Pre-conference 2011 “If It’s Digital and In Google – Then They Will Come”

    E-Print Network [OSTI]

    McGeachin, Robert B.

    2011-05-17T23:59:59.000Z

    AgNIC offers a two hour pre-conference on how to scan and digitize print items, make derivative versions, including archival PDF/A, for easy access on the Internet, create basic Dublin Core metadata, and how to add digital objects to a digital...

  10. A Survey of Data Provenance Techniques Yogesh L. Simmhan, Beth Plale, Dennis Gannon

    E-Print Network [OSTI]

    culminates with an identification of open research problems in the field. 1 Introduction The growing number Sky Survey (SDSS) [2] generate terabytes of data whose complexity is managed by data grids. This data for these data mining tasks to have metadata describing the data properties [3]. Provenance is one kind

  11. Proof Delivery Form Proceedings of the International Astronomical Union

    E-Print Network [OSTI]

    Martens, Petrus C.

    .1017/S1743921313010867 Hemispheric Patterns in Filament Chirality1 and Sigmoid Shape over the Solar Cycle of Mathematical Sciences, Durham University,7 Science Laboratories, South Road, Durham DH1 3LE, UK,8 email a full solar cycle of filament metadata at hand it was easy to verify the supposedly known17 hemispheric

  12. DEVELOPING A SPATIAL DATA INFRASTRUCTURE FOR THE HANFORD SITE

    SciTech Connect (OSTI)

    RUSH SF

    2009-11-06T23:59:59.000Z

    Summary of this report is: (1) aggressive implementation of metadata; (2) higher confidence in spatial data and organizational structure; (3) improved data sharing between Hanford and neighboring government agencies; and (4) improved data sharing and management reduce unnecessary cost to DOE and the American taxpayer.

  13. 9/21/05 John Eul9/21/05 John Eul --PrincipalPrincipal Business SolutionsBusiness Solutions

    E-Print Network [OSTI]

    Brock, David

    Exposes Data to All Authorized Supports Legacy SystemsSupports Legacy Systems Neutral Set of API'sNeutral Set of API's No Single Point of FailureNo Single Point of Failure Minimizes Duplicate ProcessorsWeb Metadata #12;M LanguageM Language Adaptable to Existing StandardsAdaptable to Existing Standards Harmonize

  14. Creating Enriched YouTube Media Fragments With NERD Using Timed-Text

    E-Print Network [OSTI]

    Quartly, Graham

    Introduction New W3C standards such as HTML5, Media Fragment URI and the Ontology for Media Resources have their closed captions [4]. The captions are ana- lyzed by 3 web services (AlchemyAPI, OpenCalais and Zemanta) The system first extracts the metadata and the subtitles if available using the YouTube API: GET api

  15. File: AHISFinalReport011108 Page 1 of 9 Printed: 4/15/08 4:19 PM Final Report to Arizona Water Institute

    E-Print Network [OSTI]

    Fay, Noah

    between different metadata standards (three for the purpose of this project). The team has met Java API client to query web services The new search interface can be viewed here: http the data access module for the water data based upon the standard web services defined by CUAHSI

  16. Schattauer 2013 Development and Validation of a

    E-Print Network [OSTI]

    component be reliable, but, in addition, the connectivity between components must be reliable. Objective) and is the connectivity between components reliable (e.g., communication protocols and the metadata necessary for data developers and to healthcare stakeholders seeking to invest in connected electronic systems for prehospital

  17. LISA VI, Pune, India, February 2010 -Holl: Small Data Archives & Libraries 1 Small Data Archives & Libraries

    E-Print Network [OSTI]

    Holl, András

    & photo/microfilm) budget yearly report Regulation what to archive? proprietary period operational rules & science ready data + documents, etc. raw data: archiving by data acquisition pipeline science ready data 2010 - Holl: Small Data Archives & Libraries 10 Metadata Images fix the data acquisition pipeline

  18. GIS and Digital Geospatial Data at TAMU Libraries - White Paper

    E-Print Network [OSTI]

    Weimer, Kathy

    2008-08-22T23:59:59.000Z

    , generate metadata and create online exhibits. They may also collaborate on campus wide, local/regional or national research projects. Environmental Scan - GIS at Texas A&M, 2005 GIS came to the TAMU Libraries in the mid-1990s. Recognizing...

  19. Richard Marciano Chien-Yi Hou

    E-Print Network [OSTI]

    Minsker, Barbara S.

    the multiple storage locations. · open source, BSD license #12;How iRODS is used? · A data grid for sharing Repository for UNC collections · Regional collaboration ­ RENCI Data Grid linking resources across North · Micro-service · Rule #12;8 User Can Search, Access, Add and Manage Data & Metadata *Access data with Web

  20. Non-Intrusive Load Identification for Smart Outlets Sean Barker, Mohamed Musthag, David Irwin, and Prashant Shenoy

    E-Print Network [OSTI]

    Shenoy, Prashant

    broad sectors of energy consumption--industry and transportation [1]. As a result, even small@ecs.umass.edu Abstract--An increasing interest in energy-efficiency com- bined with the decreasing cost of embedded, and then manually update the outlets meta-data in software whenever a new device is plugged into the outlet. Correct

  1. ENGO 435 Remote Sensing LAB 1: INTRODUCTION TO REMOTE SENSING

    E-Print Network [OSTI]

    Habib, Ayman

    ENGO 435 ­ Remote Sensing Fall 2003 LAB 1: INTRODUCTION TO REMOTE SENSING Due: By 4:30 pm on September 30th Objectives: To familiarize you with: Remote sensing resources on the Internet. Some remote (Metadata). Part A: Remote Sensing Resources There is a wealth of remote sensing information and data

  2. Hardcopy Uncontrolled NOAA NESDIS

    E-Print Network [OSTI]

    Kuligowski, Bob

    TITLE: DG-11.3: METADATA DOCUMENT GUIDELINE VERSION 3.0 AUTHORS: Ken Jensen (Raytheon Information.0 New Documentation Guideline (DG-11.4) adapted from SPSRB draft guidelines by Ken Jensen (Raytheon Information Solutions) New Document 05/05/2006 1.1 Revision by Ken Jensen (Raytheon Information Solutions

  3. S-DSM for Heterogeneous Machine Architectures Eduardo Pinheiro, DeQing Chen, Sandhya Dwarkadas,

    E-Print Network [OSTI]

    Scott, Michael L.

    , or data repositories. For processes that hap- pen to share a high-bandwidth, low-latency network, In- ter conversion to and from a common wire format when transmitting data and updates between ma- chines, and maintains an elabo- rate collection of metadata to support fast data access and updates. 1 Introduction

  4. SDSM for Heterogeneous Machine Architectures Eduardo Pinheiro, DeQing Chen, Sandhya Dwarkadas,

    E-Print Network [OSTI]

    Dwarkadas, Sandhya

    , or data repositories. For processes that hap­ pen to share a high­bandwidth, low­latency network, In­ ter conversion to and from a common wire format when transmitting data and updates between ma­ chines, and maintains an elabo­ rate collection of metadata to support fast data access and updates. 1 Introduction

  5. CistromeMap: A knowledgebase and web server for ChIP-Seq and DNase-Seq studies in mouse and human

    E-Print Network [OSTI]

    Liu, Xiaole Shirley

    Bo Qin1, , Meng Zhou2, , Ying Ge1, , Len Taing3,4 , Tao Liu4,3 , Qian Wang1 , Su Wang1 , JunshengIP-Seq and DNase-Seq data in mouse and human. We have also manually curated metadata to ensure annotation consistency, and developed a user-friendly display matrix for quick navigation and retrieval of data

  6. Semantic search integration to climate data

    SciTech Connect (OSTI)

    Devarakonda, Ranjeet [ORNL; Palanisamy, Giri [ORNL; Pouchard, Line Catherine [ORNL; Shrestha, Biva [ORNL

    2014-01-01T23:59:59.000Z

    In this paper we present how research projects at Oak Ridge National Laboratory are using Semantic Search capabilities to help scientists perform their research. We will discuss how the Mercury metadata search system, with the help of the semantic search capability, is being used to find, retrieve, and link climate change data. DOI: 10.1109/CTS.2014.6867639

  7. Undefined 0 (0) 1 1 Fiction Literature as Linked Open Data

    E-Print Network [OSTI]

    Hyvönen, Eero

    in an instance of the SAHA RDF-based metadata editor [5]. SAHA is used by the librarians for maintaining. The relevant web addresses are: SAHA Browser & Editor View http://saha.kirjastot.fi/kirjasampo/index.shtml RDF Export http://saha.kirjastot.fi/kirjasampo/export.shtml SPARQL Endpoint http://saha

  8. UMA ENABLED ENVIRONMENT FOR MOBILE MEDIA USING MPEG-21 CLIENT AND SERVER TECHNOLOGY

    E-Print Network [OSTI]

    Nřrvĺg, Kjetil

    in the MPEG-21 "Digital Item" [1], gives a standardized representation of a media resource and its metadata for transaction of a digital media resource. This unit is named the `Digital Item'. The Digital Item is considered as the `what', and is a standardized representation of digital content denoted the media resource. This Digital

  9. EPrints & Preservation David Tarrant

    E-Print Network [OSTI]

    Southampton, University of

    and scriptable : Core Objectives ·Lower the barrier for depositors while improving metadata quality and ultimate? ·File migration to avert risks found by analysis. ·Movement of file to new storage. #12;Preservation;Mock up Transformation Interface Transformation? Tool Preservation Level PPT -> PPTX PPT -> PDF

  10. Serdica Journal of Computing, Volume 6, Number 1, 2012, pp. 5966. manuscript Article disassembly new ways to handle information in publications

    E-Print Network [OSTI]

    Holl, András

    Serdica Journal of Computing, Volume 6, Number 1, 2012, pp. 5966. manuscript Article disassembly blocks. For scientific journal articles, such blocks include figures and tables, among others. Providing published in different journals. These metadata and copyrights should be supplied by the publishers, who

  11. Range Temporal Search on Sensor-rich Mobile Videos Preliminaries

    E-Print Network [OSTI]

    Shahabi, Cyrus

    Range Temporal Search on Sensor-rich Mobile Videos Preliminaries FOV model - A video as a set of video frames - Each video frame is modeled a Field Of View (FOV) based on the metadata (location, orientation, timestamp, etc.) of the video. - Note the size of is the maximum visible distance

  12. An IoT Gateway Centric Architecture to Provide Novel M2M Services Soumya Kanti Datta, Christian Bonnet, Navid Nikaein

    E-Print Network [OSTI]

    Gesbert, David

    include smart home, e-health, smart grid and more. The resulting eco system is referred to Internet interaction between mobile clients and smart/legacy things (sensors and actuators) via a wireless gateway) managing connection with non-smart things connected over modbus, (iii) associate metadata to sensor

  13. ISO 191** Standards for NASA SiriJodha Singh Khalsa, Lead

    E-Print Network [OSTI]

    Christian, Eric

    ISO 191** Standards for NASA SiriJodha Singh Khalsa, Lead MENDS Tiger Team to support ISO metadata ­ GCMD and ECHO both acGvely working on ISO ­ Missions and data centers looking at ISO · FGDC has endorsed ISO 19115 ­ OMB mandated all

  14. Dear Data Managers and Providers: We have not revised the 2000 edition of the Data Management

    E-Print Network [OSTI]

    Dear Data Managers and Providers: We have not revised the 2000 edition of the Data Management the Data and Metadata Reporting Guidance materials, including the data exchange standard template, under plans on the drawing board for revising all of our quality management documents in the near future

  15. ACEEE Int. J. on Network Security, Vol. 01, No. 03, Dec 2010 2010 ACEEE

    E-Print Network [OSTI]

    Paris-Sud XI, Université de

    ACEEE Int. J. on Network Security, Vol. 01, No. 03, Dec 2010 © 2010 ACEEE DOI: 01.IJNS.01.03.85 23 using the concept of application metadata. Later, we draw a qualitative and simulation based summarization for the proposed model. We finally evaluate the results and draw up a series of key conclusions

  16. Storing files in a parallel computing system based on user-specified parser function

    DOE Patents [OSTI]

    Faibish, Sorin; Bent, John M; Tzelnic, Percy; Grider, Gary; Manzanares, Adam; Torres, Aaron

    2014-10-21T23:59:59.000Z

    Techniques are provided for storing files in a parallel computing system based on a user-specified parser function. A plurality of files generated by a distributed application in a parallel computing system are stored by obtaining a parser from the distributed application for processing the plurality of files prior to storage; and storing one or more of the plurality of files in one or more storage nodes of the parallel computing system based on the processing by the parser. The plurality of files comprise one or more of a plurality of complete files and a plurality of sub-files. The parser can optionally store only those files that satisfy one or more semantic requirements of the parser. The parser can also extract metadata from one or more of the files and the extracted metadata can be stored with one or more of the plurality of files and used for searching for files.

  17. ETD Management in the Texas Digital Library: lessons Learned from a Demonstrator

    E-Print Network [OSTI]

    Mikeal, Adam

    2008-07-24T23:59:59.000Z

    management, repository interfaces, subision workflows, and data preservation each constitute significant projects in their own right, and these responsibilities are delegated to various groups within TDL. Metadata Profile ETDs became a focus very... modifying base repository code. Preservation Network A core component of TDL’s mision, data preservation ocupies an important place in the long-ter plans of TDL services. While stil in the early stages of development, TDL is curently investing in a...

  18. Research Data Management - Managing your digital research data : Example presentation to Department of Chemistry 2012

    E-Print Network [OSTI]

    Collins, Anna

    2015-02-06T23:59:59.000Z

    to standards “I guess it makes sense for a robot to read an e-book [401]” by brianjmatis on flickr Make material findable • Comprehensive descriptive metadata allows relevant material to be discovered more easily • Related materials can be located... of project sponsor (eg industrial applications of research) • Thesis contains significant quantity of 3rd party copyright material Discuss your options with your supervisor Consult digital repository website for information Make plans early...

  19. AgNIC Pre-conference 2009 “If It’s Digital and In Google – Then They Will Come”

    E-Print Network [OSTI]

    McGeachin, Robert B.

    2009-04-21T23:59:59.000Z

    preservation Scanners Digitization standards Planning a project Mechanics / Demo of scanning Why Digitize ? Make an archival preserved copy Make material accessible to the world More usage of materials Make Metadata Records Harvestable to Other Indexing... Customers don?t like to use microfilm! Digitization Concepts Digital image ? representation of a two dimensional image in digital bits of ones and zeroes AKA Raster image -- raster graphics image or bitmap, is a data structure representing a rectangular grid...

  20. PRIDE Surveillance Projects Data Packaging Project, Information Package Specification Version 1.0

    SciTech Connect (OSTI)

    Kelleher, D.M.; Shipp, R. L.; Mason, J. D.

    2009-09-28T23:59:59.000Z

    This document contains a specification for a standard XML document format called an information package that can be used to store information and the context required to understand and use that information in information management systems and other types of information archives. An information package consists of packaged information, a set of information metadata that describes the packaged information, and an XML signature that protects the packaged information. The information package described in this specification was designed to be used to store Department of Energy (DOE) and National Nuclear Security Administration (NNSA) information and includes the metadata required for that information: a unique package identifier, information marking that conforms to DOE and NNSA requirements, and access control metadata. Information package metadata can also include information search terms, package history, and notes. Packaged information can be text content, binary content, and the contents of files and other containers. A single information package can contain multiple types of information. All content not in a text form compatible with XML must be in a text encoding such as base64. Package information is protected by a digital XML signature that can be used to determine whether the information has changed since it was signed and to identify the source of the information. This specification has been tested but has not been used to create production information packages. The authors expect that gaps and unclear requirements in this specification will be identified as this specification is used to create information packages and as information stored in information packages is used. The authors expect to issue revised versions of this specification as needed to address these issues.

  1. A Framework for Re-Purposing Textbooks Using Learning Outcomes/Methodology, Device Characteristics, Representation and User Dimensions

    E-Print Network [OSTI]

    Ciftci, Tolga

    2013-05-01T23:59:59.000Z

    -book standard HTML Hypertext Markup Language ID3 A metadata container format used to store information related to mp3 audio files in this thesis. Mobipocket An e-book format TEI Text Encoding Initiative XML Extensible Markup Language XSLT Extensible..................................................................... 81 Figure 39: TemplateWithXpaths tag used to generate a filename. .................................. 83 Figure 40: id3Selector tag example. ................................................................................. 84 Figure 41: attr...

  2. Knowledge Discovery Framework for the Virtual Observatory

    E-Print Network [OSTI]

    Thomas, Brian; Huang, Zenping; Teuben, Peter

    2015-01-01T23:59:59.000Z

    We describe a framework that allows a scientist-user to easily query for information across all Virtual Observatory (VO) repositories and pull it back for analysis. This framework hides the gory details of meta-data remediation and data formatting from the user, allowing them to get on with search, retrieval and analysis of VO data as if they were drawn from a single source using a science based terminology rather than a data-centric one.

  3. GOLD: The Genomes Online Database

    DOE Data Explorer [Office of Scientific and Technical Information (OSTI)]

    Kyrpides, Nikos; Liolios, Dinos; Chen, Amy; Tavernarakis, Nektarios; Hugenholtz, Philip; Markowitz, Victor; Bernal, Alex

    Since its inception in 1997, GOLD has continuously monitored genome sequencing projects worldwide and has provided the community with a unique centralized resource that integrates diverse information related to Archaea, Bacteria, Eukaryotic and more recently Metagenomic sequencing projects. As of September 2007, GOLD recorded 639 completed genome projects. These projects have their complete sequence deposited into the public archival sequence databases such as GenBank EMBL,and DDBJ. From the total of 639 complete and published genome projects as of 9/2007, 527 were bacterial, 47 were archaeal and 65 were eukaryotic. In addition to the complete projects, there were 2158 ongoing sequencing projects. 1328 of those were bacterial, 59 archaeal and 771 eukaryotic projects. Two types of metadata are provided by GOLD: (i) project metadata and (ii) organism/environment metadata. GOLD CARD pages for every project are available from the link of every GOLD_STAMP ID. The information in every one of these pages is organized into three tables: (a) Organism information, (b) Genome project information and (c) External links. [The Genomes On Line Database (GOLD) in 2007: Status of genomic and metagenomic projects and their associated metadata, Konstantinos Liolios, Konstantinos Mavromatis, Nektarios Tavernarakis and Nikos C. Kyrpides, Nucleic Acids Research Advance Access published online on November 2, 2007, Nucleic Acids Research, doi:10.1093/nar/gkm884]

    The basic tables in the GOLD database that can be browsed or searched include the following information:

    • Gold Stamp ID
    • Organism name
    • Domain
    • Links to information sources
    • Size and link to a map, when available
    • Chromosome number, Plas number, and GC content
    • A link for downloading the actual genome data
    • Institution that did the sequencing
    • Funding source
    • Database where information resides
    • Publication status and information

    (Specialized Interface)

  4. National Energy Information System basic concepts

    SciTech Connect (OSTI)

    Not Available

    1980-07-01T23:59:59.000Z

    The basic concepts on which the National Energy Information System (NEIS) rests are described and clarified. Identified are the current state of, as well as future information gathering activities of the system. The NEIS was originally created under Congressional mandate to collect, process, and disseminate data useful for analysis of energy supply and consumption issues. An overview of NEIS, data descriptions, metadata (directories, documentation), and procedures of the NEIS system are discussed in a question-and-answer format.

  5. Paso del Norte Watershed Council Coordinated Water Resources Database Project

    E-Print Network [OSTI]

    Brown, Christopher; Sheng, Zhuping; Rich, Matt

    data elements/variables Heavy metals Source(s) of data Field Study Spatial extent El Paso, Cd. Juarez Data gathered or updated 2002-2003 Frequency of data One Time Format of digital file Excel spreadsheet Restrictions on use None...@infolnk.net Contact address 4145 Benjamin Franklin and 4158 Estocolmo Pronaf circuit Cd. Juarez, Chihuahua Contact FAX number (656) 611-1270 Paso del Norte Watershed Council PDNWC Contact: Alfredo Granados Ph.D. Metadata form for USACE and EPWU Coordinated...

  6. Building Component Library: An Online Repository to Facilitate Building Energy Model Creation; Preprint

    SciTech Connect (OSTI)

    Fleming, K.; Long, N.; Swindler, A.

    2012-05-01T23:59:59.000Z

    This paper describes the Building Component Library (BCL), the U.S. Department of Energy's (DOE) online repository of building components that can be directly used to create energy models. This comprehensive, searchable library consists of components and measures as well as the metadata which describes them. The library is also designed to allow contributors to easily add new components, providing a continuously growing, standardized list of components for users to draw upon.

  7. BIOINFORMATICS APPLICATIONS NOTE Vol. 28 no. 10 2012, pages 14111412

    E-Print Network [OSTI]

    Liu, Xiaole Shirley

    Taing3,4, Tao Liu3,4, Qian Wang1, Su Wang1, Junsheng Chen1, Lingling Shen5, Xikun Duan1, Sheng'en Hu1 of the publicly available ChIP- Seq and DNase-Seq data in mouse and human. We have also manually curated metadata to ensure annotation consistency, and developed a user-friendly display matrix for quick navigation

  8. Beyond the Feed and Board: Holistic Principles for Expressive Web Curation

    E-Print Network [OSTI]

    Lupfer, Nicholas

    2014-12-02T23:59:59.000Z

    WORK We combine a holistic outlook on creative expression with recent related work from human computer interaction. We look at prior techniques for gathering infor- mation in the form of clippings. We examine the connections between summaries, metadata... for the degree of MASTER OF SCIENCE Chair of Committee, Andruid Kerne Committee Members, Steve M. Smith Tracy Hammond Head of Department, Dilma Da Silva December 2014 Major Subject: Computer Science Copyright 2014 Nicholas Lane Lupfer ABSTRACT We develop holistic...

  9. Co-located Collaborative Information-based Ideation through Embodied Cross-Surface Curation

    E-Print Network [OSTI]

    Fei, Shenfeng

    2014-12-15T23:59:59.000Z

    Clippings with Metadata . . . . . . . . . . . 24 2.3 Cross-surface Portals . . . . . . . . . . . . . . . . . . . . . . . . . . . 25 2.3.1 Creating Cross-surface Portals with NFC Tags . . . . . . . . . 25 2.3.2 Transferring Rich Bookmarks through Active... Portals . . . . . 29 2.3.3 Implementation . . . . . . . . . . . . . . . . . . . . . . . . . . 31 2.4 Personal Surface: Private Curation of Rich Bookmarks . . . . . . . . 34 2.4.1 Creating and Storing Information as Rich Bookmarks . . . . . 34 2.4.2 Private...

  10. Accepted for Publication Simulation modeling combined with decision control can

    E-Print Network [OSTI]

    Accepted for Publication ABSTRACT Simulation modeling combined with decision control can offer) with Model Predictive Control (MPC) paradigms using a Knowledge Interchange Broker (KIB). This environment uses the KIB to compose discrete event simulation and model predictive control models. This approach

  11. The Earth System Grid Center for Enabling Technologies: Focusing Technologies on Climate Datasets and Resource Needs

    SciTech Connect (OSTI)

    Williams, Dean N. [Lawrence Livermore National Lab. (LLNL), Livermore, CA (United States)

    2007-09-26T23:59:59.000Z

    This report discusses a project that used prototyping technology to access and analyze climate data. This project was initially funded under the DOE’s Next Generation Internet (NGI) program, with follow-on support from BER and the Mathematical, Information, and Computational Sciences (MICS) office. In this prototype, we developed Data Grid technologies for managing the movement and replication of large datasets, and applied these technologies in a practical setting (i.e., an ESG-enabled data browser based on current climate data analysis tools), achieving cross-country transfer rates of more than 500 Mb/s. Having demonstrated the potential for remotely accessing and analyzing climate data located at sites across the U.S., we won the “Hottest Infrastructure” award in the Network Challenge event. While the ESG I prototype project substantiated a proof of concept (“Turning Climate Datasets into Community Resources”), the SciDAC Earth System Grid (ESG) II project made this a reality. Our efforts targeted the development of metadata technologies (standard schema, XML metadata extraction based on netCDF, and a Metadata Catalog Service), security technologies (Web-based user registration and authentication, and community authorization), data transport technologies (GridFTPenabled OPeNDAP-G for high-performance access, robust multiple file transport and integration with mass storage systems, and support for dataset aggregation and subsetting), as well as web portal technologies to provide interactive access to climate data holdings. At this point, the technology was in place and assembled, and ESG II was poised to make a substantial impact on the climate modelling community.

  12. Integrating Interactive Visualizations in the Search Process of Digital Libraries and IR Systems

    E-Print Network [OSTI]

    Hienert, Daniel; Schaer, Philipp; Mayr, Philipp

    2012-01-01T23:59:59.000Z

    Interactive visualizations for exploring and retrieval have not yet become an integral part of digital libraries and information retrieval systems. We have integrated a set of interactive graphics in a real world social science digital library. These visualizations support the exploration of search queries, results and authors, can filter search results, show trends in the database and can support the creation of new search queries. The use of weighted brushing supports the identification of related metadata for search facets. We discuss some use cases of the combination of IR systems and interactive graphics. In a user study we verify that users can gain insights from statistical graphics intuitively and can adopt interaction techniques.

  13. ChemXSeer Digital Library Gaussian Search

    E-Print Network [OSTI]

    Lahiri, Shibamouli; Nangia, Shikha; Mitra, Prasenjit; Giles, C Lee; Mueller, Karl T

    2011-01-01T23:59:59.000Z

    We report on the Gaussian file search system designed as part of the ChemXSeer digital library. Gaussian files are produced by the Gaussian software [4], a software package used for calculating molecular electronic structure and properties. The output files are semi-structured, allowing relatively easy access to the Gaussian attributes and metadata. Our system is currently capable of searching Gaussian documents using a boolean combination of atoms (chemical elements) and attributes. We have also implemented a faceted browsing feature on three important Gaussian attribute types - Basis Set, Job Type and Method Used. The faceted browsing feature enables a user to view and process a smaller, filtered subset of documents.

  14. The semantic architecture of the World-Wide Molecular Matrix (WWMM)

    E-Print Network [OSTI]

    Murray-Rust, Peter; Adams, Sam E; Downing, Jim; Townsend, Joe A; Zhang, Yong

    2011-10-14T23:59:59.000Z

    but would be structured information. We were grateful to receive early funding from the eScience pro- ject ("Molecular Standards for the Grid”[7]) but have been somewhat frustrated by the top-heavy concentra- tion on CPU performance, bulk storage of un... Eye and Quixote systems are excellent example of systems that can succeed as distributed WWMM repositories. In CrystalEye the uniqueness is determined by the biblio- graphic data of the publication (or the metadata from the creators). In Quixote a calculation...

  15. Intelligent Information Retrieval

    E-Print Network [OSTI]

    Michael J. Kurtz; Guenther Eichhorn; Alberto Accomazzi; Carolyn Grant; Edwin Henneken; Stephen S. Murray

    2005-10-31T23:59:59.000Z

    Since it was first announced at ADASS 2 the Smithsonian/NASA Astrophysics System Abstract Service (ADS) has played a central role in the information seeking behavior of astronomers. Central to the ability of the ADS to act as a search and discovery tool is its role as metadata agregator. Over the past 13 years the ADS has introduced many new techniques to facilitate information retrieval, broadly defined. We discuss some of these developments; with particular attention to how the ADS might interact with the virtual observatory, and to the new myADS-arXiv customized open access virtual journal. The ADS is at http://ads.harvard.edu

  16. Comprehensive Citation Across the Data Life Cycle Using DDI

    E-Print Network [OSTI]

    Hoyle, Larry; Vardigan, Mary; Hume, Sam; Ionescu, Sanda; Greenfield, Jay; Iverson, Jeremy; Kunze, John; Radler, Barry; Weibel, Stuart; Witt, Michael C.; Thomas, Wendy

    2014-11-17T23:59:59.000Z

    data CITE.TOPICS2_XA; set CITE.TOPICS2; run; title 'Contents of the Revised Dataset';proc datasets lib=CITE nolist ; modify TOPICS2_XA ; XATTR SET DS Study_FundingInformation='Participant travel and accomodations funded by NSF grant 1448107... table was exported as this SAS dataset Study_FundingInformation Participant travel and accomodations funded by NSF grant 1448107 Study_KindOfData SAS Text Miner Topics results table. Derived topics descriptions, dataset includes metadata in SAS extended...

  17. Framework for Physics Computation

    SciTech Connect (OSTI)

    Schwan, Karsten [Georgia Institute of Technology

    2012-01-13T23:59:59.000Z

    The Georgia Tech team has been working in collaboration with ORNL and Rutgers on improved I/O for petascale fusion codes, specifically, to integrate staging methods into the ADIOS framework. As part of this on-going work, we have released the DataTap server as part of the ADIOS release, and we have been working on improving the ‘in situ’ processing capabilities of the ADIOS framework. In particular, we have been moving forward with a design that adds additional metadata to describe the data layout and structure of data that is being moved for I/O purposes, building on the FFS type system developed in our past research

  18. Proposal for a quantity based data model in the Virtual Observatory

    E-Print Network [OSTI]

    Brian Thomas; Edward Shaya

    2003-12-23T23:59:59.000Z

    We propose the beginnings of a data model for the Virtual Observatory (VO) built up from simple ``quantity'' objects. In this paper we present how an object-oriented, domain (or namespace)-scoped simple quantity may be used to describe astronomical data. Our model is designed around the requirements that it be searchable and serve as a transport mechanism for all types of VO data and meta-data. In this paper we describe this model in terms of an OWL ontology and UML diagrams. An XML schema is available online.

  19. Tutoring System for Dance Learning

    E-Print Network [OSTI]

    Kannan, Rajkumar; Ramadoss, Balakrishnan

    2010-01-01T23:59:59.000Z

    Recent advances in hardware sophistication related to graphics display, audio and video devices made available a large number of multimedia and hypermedia applications. These multimedia applications need to store and retrieve the different forms of media like text, hypertext, graphics, still images, animations, audio and video. Dance is one of the important cultural forms of a nation and dance video is one such multimedia types. Archiving and retrieving the required semantics from these dance media collections is a crucial and demanding multimedia application. This paper summarizes the difference dance video archival techniques and systems. Keywords: Multimedia, Culture Media, Metadata archival and retrieval systems, MPEG-7, XML.

  20. Digitization Best Practices

    SciTech Connect (OSTI)

    Xue, Fei [Los Alamos National Laboratory; Holtkamp, Irma S. [Los Alamos National Laboratory; Knudson, Frances L. [Los Alamos National Laboratory

    2012-07-31T23:59:59.000Z

    This project involved performing tests and documenting results to determine best practices for digitizing older print documents. The digitization process is complicated, especially when original documents exhibit non-standard fonts and are faded. Tests focused on solutions to improve high quality scanning, increase OCR accuracy, and efficiently use embedded metadata. Results are summarized. From the test results on the right sides, we know that when we plan to digitize documents, we should balance Quantity and Quality based on our expectation, and then make final decision for the digitization process.

  1. ARM - Publications: Science Team Meeting Documents

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    AFDC Printable Version Share this resource Send a link to EERE: Alternative Fuels Data Center Home Page to someone by E-mail Share EERE: Alternative Fuels Data Center Home Page on Facebook Tweet about EERE: Alternative Fuels Data Center Home Page on Twitter Bookmark EERE: Alternative1 First Use of Energy for All Purposes (Fuel and Nonfuel), 2002; Level: National5Sales for4,645U.S. DOE Office of ScienceandMesa- PolarizationgovCampaignsSummergovFieldgovDataPIIrradiance Using RemotelyUseMetadata

  2. ARM - Publications: Science Team Meeting Documents

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    AFDC Printable Version Share this resource Send a link to EERE: Alternative Fuels Data Center Home Page to someone by E-mail Share EERE: Alternative Fuels Data Center Home Page on Facebook Tweet about EERE: Alternative Fuels Data Center Home Page on Twitter Bookmark EERE: Alternative1 First Use of Energy for All Purposes (Fuel and Nonfuel), 2002; Level: National5Sales for4,645U.S. DOE Office of ScienceandMesa- PolarizationgovCampaignsSummergovFieldgovDataPIIrradiance Using RemotelyUseMetadataA

  3. File:08COaBulkTransmissionSitingProcess.pdf | Open Energy Information

    Open Energy Info (EERE)

    AFDC Printable Version Share this resource Send a link to EERE: Alternative Fuels Data Center Home Page to someone by E-mail Share EERE: Alternative Fuels Data Center Home Page on Facebook Tweet about EERE: Alternative Fuels Data Center Home Page on Twitter Bookmark EERE: Alternative Fuels Data Center Home Page onYou are now leaving Energy.gov You are now leaving Energy.gov You are8COaBulkTransmissionSitingProcess.pdf Jump to: navigation, search File File history File usage Metadata

  4. Blind Date: Using proper motions to determine the ages of historical images

    E-Print Network [OSTI]

    Jonathan T. Barron; David W. Hogg; Dustin Lang; Sam Roweis

    2008-05-06T23:59:59.000Z

    Astrometric calibration is based on patterns of cataloged stars and therefore effectively assumes a particular epoch, which can be substantially incorrect for historical images. With the known proper motions of stars we can "run back the clock" to an approximation of the night sky in any given year, and in principle the year that best fits stellar patterns in any given image is an estimate of the year in which that image was taken. In this paper we use 47 scanned photographic images of M44 spanning years 1910-1975 to demonstrate this technique. We use only the pixel information in the images; we use no prior information or meta-data about image pointing, scale, orientation, or date. Blind Date returns date meta-data for the input images. It also improves the astrometric calibration of the image because the final astrometric calibration is performed at the appropriate epoch. The accuracy and reliability of Blind Date are functions of image size, pointing, angular resolution, and depth; performance is related to the sum of proper-motion signal-to-noise ratios for catalog stars measured in the input image. All of the science-quality images and 85 percent of the low-quality images in our sample of photographic plate images of M44 have their dates reliably determined to within a decade, many to within months.

  5. The SAMGrid database server component: its upgraded infrastructure and future development path

    SciTech Connect (OSTI)

    Loebel-Carpenter, L.; White, S.; Baranovski, A.; Garzoglio, G.; Herber, R.; Illingworth, R.; Kennedy, R.; Kreymer, A.; Kumar, A.; Lueking, L.; Lyon, A.; Merritt, W.; Terekhov, I.; Trumbo, J.; Veseli, S.; /Fermilab; Burgon-Lyon, M.; St. Denis, R.; /Glasgow U.; Belforte, S.; /INFN, Trieste; Kerzel, U.; /Karlsruhe U.; Bartsch, V.; Leslie, M.; /Oxford

    2004-12-01T23:59:59.000Z

    The SAMGrid Database Server encapsulates several important services, such as accessing file metadata and replica catalog, keeping track of the processing information, as well as providing the runtime support for SAMGrid station services. Recent deployment of the SAMGrid system for CDF has resulted in unification of the database schema used by CDF and D0, and the complexity of changes required for the unified metadata catalog has warranted a complete redesign of the DB Server. We describe here the architecture and features of the new server. In particular, we discuss the new CORBA infrastructure that utilizes python wrapper classes around IDL structs and exceptions. Such infrastructure allows us to use the same code on both server and client sides, which in turn results in significantly improved code maintainability and easier development. We also discuss future integration of the new server with an SBIR II project which is directed toward allowing the DB Server to access distributed databases, implemented in different DB systems and possibly using different schema.

  6. LAGOVirtual: A Collaborative Environment for the Large Aperture GRB Observatory

    E-Print Network [OSTI]

    R. Camacho; R. Chacon; G. Diaz; C. Guada; V. Hamar; H. Hoeger; A. Melfo; L. A. Nunez; Y. Perez; C. Quintero; M. Rosales; R. Torrens; the LAGO Collaboration

    2009-12-12T23:59:59.000Z

    We present the LAGOVirtual Project: an ongoing project to develop platform to collaborate in the Large Aperture GRB Observatory (LAGO). This continental-wide observatory is devised to detect high energy (around 100 GeV) component of Gamma Ray Bursts, by using the single particle technique in arrays of Water Cherenkov Detectors (WCD) at high mountain sites (Chacaltaya, Bolivia, 5300 m a.s.l., Pico Espejo, Venezuela, 4750 m a.s.l., Sierra Negra, Mexico, 4650 m a.s.l). This platform will allow LAGO collaboration to share data, and computer resources through its different sites. This environment has the possibility to generate synthetic data by simulating the showers through AIRES application and to store/preserve distributed data files collected by the WCD at the LAGO sites. The present article concerns the implementation of a prototype of LAGO-DR adapting DSpace, with a hierarchical structure (i.e. country, institution, followed by collections that contain the metadata and data files), for the captured/simulated data. This structure was generated by using the community, sub-community, collection, item model; available at the DSpace software. Each member institution-country of the project has the appropriate permissions on the system to publish information (descriptive metadata and associated data files). The platform can also associate multiple files to each item of data (data from the instruments, graphics, postprocessed-data, etc.).

  7. LHAPDF6: parton density access in the LHC precision era

    E-Print Network [OSTI]

    Andy Buckley; James Ferrando; Stephen Lloyd; Karl Nordstrom; Ben Page; Martin Ruefenacht; Marek Schoenherr; Graeme Watt

    2015-03-05T23:59:59.000Z

    The Fortran LHAPDF library has been a long-term workhorse in particle physics, providing standardised access to parton density functions for experimental and phenomenological purposes alike, following on from the venerable PDFLIB package. During Run 1 of the LHC, however, several fundamental limitations in LHAPDF's design have became deeply problematic, restricting the usability of the library for important physics-study procedures and providing dangerous avenues by which to silently obtain incorrect results. In this paper we present the LHAPDF 6 library, a ground-up re-engineering of the PDFLIB/LHAPDF paradigm for PDF access which removes all limits on use of concurrent PDF sets, massively reduces static memory requirements, offers improved CPU performance, and fixes fundamental bugs in multi-set access to PDF metadata. The new design, restricted for now to interpolated PDFs, uses centralised numerical routines and a powerful cascading metadata system to decouple software releases from provision of new PDF data and allow completely general parton content. More than 200 PDF sets have been migrated from LHAPDF 5 to the new universal data format, via a stringent quality control procedure. LHAPDF 6 is supported by many Monte Carlo generators and other physics programs, in some cases via a full set of compatibility routines, and is recommended for the demanding PDF access needs of LHC Run 2 and beyond.

  8. NREL's Building Component Library for Use with Energy Models

    DOE Data Explorer [Office of Scientific and Technical Information (OSTI)]

    The Building Component Library (BCL) is the U.S. Department of Energy’s comprehensive online searchable library of energy modeling building blocks and descriptive metadata. Novice users and seasoned practitioners can use the freely available and uniquely identifiable components to create energy models and cite the sources of input data, which will increase the credibility and reproducibility of their simulations. The BCL contains components which are the building blocks of an energy model. They can represent physical characteristics of the building such as roofs, walls, and windows, or can refer to related operational information such as occupancy and equipment schedules and weather information. Each component is identified through a set of attributes that are specific to its type, as well as other metadata such as provenance information and associated files. The BCL also contains energy conservation measures (ECM), referred to as measures, which describe a change to a building and its associated model. For the BCL, this description attempts to define a measure for reproducible application, either to compare it to a baseline model, to estimate potential energy savings, or to examine the effects of a particular implementation. The BCL currently contains more than 30,000 components and measures. A faceted search mechanism has been implemented on the BCL that allows users to filter through the search results using various facets. Facet categories include component and measure types, data source, and energy modeling software type. All attributes of a component or measure can also be used to filter the results.

  9. ASTROMETRY.NET: BLIND ASTROMETRIC CALIBRATION OF ARBITRARY ASTRONOMICAL IMAGES

    SciTech Connect (OSTI)

    Lang, Dustin; Mierle, Keir; Roweis, Sam [Department of Computer Science, University of Toronto, 6 King's College Road, Toronto, Ontario M5S 3G4 (Canada); Hogg, David W.; Blanton, Michael [Center for Cosmology and Particle Physics, Department of Physics, New York University, 4 Washington Place, New York, NY 10003 (United States)], E-mail: dstn@astro.princeton.edu

    2010-05-15T23:59:59.000Z

    We have built a reliable and robust system that takes as input an astronomical image, and returns as output the pointing, scale, and orientation of that image (the astrometric calibration or World Coordinate System information). The system requires no first guess, and works with the information in the image pixels alone; that is, the problem is a generalization of the 'lost in space' problem in which nothing-not even the image scale-is known. After robust source detection is performed in the input image, asterisms (sets of four or five stars) are geometrically hashed and compared to pre-indexed hashes to generate hypotheses about the astrometric calibration. A hypothesis is only accepted as true if it passes a Bayesian decision theory test against a null hypothesis. With indices built from the USNO-B catalog and designed for uniformity of coverage and redundancy, the success rate is >99.9% for contemporary near-ultraviolet and visual imaging survey data, with no false positives. The failure rate is consistent with the incompleteness of the USNO-B catalog; augmentation with indices built from the Two Micron All Sky Survey catalog brings the completeness to 100% with no false positives. We are using this system to generate consistent and standards-compliant meta-data for digital and digitized imaging from plate repositories, automated observatories, individual scientific investigators, and hobbyists. This is the first step in a program of making it possible to trust calibration meta-data for astronomical data of arbitrary provenance.

  10. Blind Date: Using proper motions to determine the ages of historical images

    E-Print Network [OSTI]

    Barron, Jonathan T; Lang, Dustin; Roweis, Sam

    2008-01-01T23:59:59.000Z

    Astrometric calibration is based on patterns of cataloged stars and therefore effectively assumes a particular epoch, which can be substantially incorrect for historical images. With the known proper motions of stars we can "run back the clock" to an approximation of the night sky in any given year, and in principle the year that best fits stellar patterns in any given image is an estimate of the year in which that image was taken. In this paper we use 47 scanned photographic images of M44 spanning years 1910-1975 to demonstrate this technique. We use only the pixel information in the images; we use no prior information or meta-data about image pointing, scale, orientation, or date. Blind Date returns date meta-data for the input images. It also improves the astrometric calibration of the image because the final astrometric calibration is performed at the appropriate epoch. The accuracy and reliability of Blind Date are functions of image size, pointing, angular resolution, and depth; performance is related t...

  11. Phase III: Implementation and Operation of the repository

    SciTech Connect (OSTI)

    None

    1998-01-01T23:59:59.000Z

    In October all three Sun servers (3000, Ultra/II) were loaded, and we were able to demonstrate the complete system ? databases, web front-page, and GeoTrek ? to a large audience. All systems required for the web-based metadata catalog are in place and operational. Version 54 of the PetroTrek/GeoTrek software was loaded in December and is being tested. It is expected this version will be placed in the production environment for beta testing in January. Installation of the Bureau of Economic Geology (BEG) databases on the IS Houston server was completed and GeoTrek was loaded on BEG?s server in Austin. Installation of the entire MMS well data set (35,000 records) was completed and it will be moved to the production environment for beta testing in January. The Eastern Gulf Region PTTC interior salt basin data set was prepared and it will be loaded in January. The Gulf of Mexico PGS seismic data set was prepared and loaded in December. The web-site front pages for the NGDRS GeoTrek Metadata Catalog are 70 percent completed and write-ups for the tutorials are scheduled to be completed in January. Ten users identified to be beta testers are actively testing the system.

  12. Gas Hydrate Research Database and Web Dissemination Channel

    SciTech Connect (OSTI)

    Micheal Frenkel; Kenneth Kroenlein; V Diky; R.D. Chirico; A. Kazakow; C.D. Muzny; M. Frenkel

    2009-09-30T23:59:59.000Z

    To facilitate advances in application of technologies pertaining to gas hydrates, a United States database containing experimentally-derived information about those materials was developed. The Clathrate Hydrate Physical Property Database (NIST Standard Reference Database {number_sign} 156) was developed by the TRC Group at NIST in Boulder, Colorado paralleling a highly-successful database of thermodynamic properties of molecular pure compounds and their mixtures and in association with an international effort on the part of CODATA to aid in international data sharing. Development and population of this database relied on the development of three components of information-processing infrastructure: (1) guided data capture (GDC) software designed to convert data and metadata into a well-organized, electronic format, (2) a relational data storage facility to accommodate all types of numerical and metadata within the scope of the project, and (3) a gas hydrate markup language (GHML) developed to standardize data communications between 'data producers' and 'data users'. Having developed the appropriate data storage and communication technologies, a web-based interface for both the new Clathrate Hydrate Physical Property Database, as well as Scientific Results from the Mallik 2002 Gas Hydrate Production Research Well Program was developed and deployed at http://gashydrates.nist.gov.

  13. GlusterFS One Storage Server to Rule Them All

    SciTech Connect (OSTI)

    Boyer, Eric B. [Los Alamos National Laboratory; Broomfield, Matthew C. [Los Alamos National Laboratory; Perrotti, Terrell A. [Los Alamos National Laboratory

    2012-07-30T23:59:59.000Z

    GlusterFS is a Linux based distributed file system, designed to be highly scalable and serve many clients. Some reasons to use GlusterFS are: No centralized metadata server, Scalability, Open Source, Dynamic and live service modifications, Can be used over Infiniband or Ethernet, Can be tuned for speed and/or resilience and Flexible administration. It's useful for enterprise environments - virtualization; high performance computing (HPC) and it works with Mac, Linux and Windows clients. Conclusions are: (1) GlusterFS proved to have widespread capabilities as a virtual file system; (2) Scalability is very dependent upon the underlying hardware; (3) Lack of built-in encryption and security paradigm; and (4) Best suited in a general purpose computing environment.

  14. DMA engine for repeating communication patterns

    DOE Patents [OSTI]

    Chen, Dong (Croton on Hudson, NY); Gara, Alan G. (Mount Kisco, NY); Giampapa, Mark E. (Irvington, NY); Heidelberger, Philip (Cortlandt Manor, NY); Steinmacher-Burow, Burkhard (Esslingen, DE); Vranas, Pavlos (Danville, CA)

    2010-09-21T23:59:59.000Z

    A parallel computer system is constructed as a network of interconnected compute nodes to operate a global message-passing application for performing communications across the network. Each of the compute nodes includes one or more individual processors with memories which run local instances of the global message-passing application operating at each compute node to carry out local processing operations independent of processing operations carried out at other compute nodes. Each compute node also includes a DMA engine constructed to interact with the application via Injection FIFO Metadata describing multiple Injection FIFOs where each Injection FIFO may containing an arbitrary number of message descriptors in order to process messages with a fixed processing overhead irrespective of the number of message descriptors included in the Injection FIFO.

  15. Disk storage management for LHCb based on Data Popularity estimator

    E-Print Network [OSTI]

    Hushchyn, Mikhail; Ustyuzhanin, Andrey

    2015-01-01T23:59:59.000Z

    This paper presents a system providing recommendations for optimizing the LHCb data storage. The LHCb data storage system is a hybrid system. All datasets are kept as archives on magnetic tapes. The most popular datasets are kept on disks. The recommendation system takes the dataset usage history and metadata (size, type, configuration etc.) to generate a recommendation report. In this article present how we use machine learning algorithms to predict future data popularity. Using these predictions it is possible to estimate which datasets should be removed from disk. We use regression algorithms and time series analysis to find the optimal number of replicas for datasets that are kept on disk. Based on the data popularity and the number of replicas optimization, the recommendation system minimizes a loss function to find the optimal data distribution. The loss function represents all requirements for data distribution in the data storage system. We demonstrate how the recommendation system helps to save disk ...

  16. Reducing Concurrency Bottlenecks in Parallel I/O Workloads

    SciTech Connect (OSTI)

    Manzanares, Adam C. [Los Alamos National Laboratory; Bent, John M. [Los Alamos National Laboratory; Wingate, Meghan [Los Alamos National Laboratory

    2011-01-01T23:59:59.000Z

    To enable high performance parallel checkpointing we introduced the Parallel Log Structured File System (PLFS). PLFS is middleware interposed on the file system stack to transform concurrent writing of one application file into many non-concurrently written component files. The promising effectiveness of PLFS makes it important to examine its performance for workloads other than checkpoint capture, notably the different ways that state snapshots may be later read, to make the case for using PLFS in the Exascale I/O stack. Reading a PLFS file involved reading each of its component files. In this paper we identify performance limitations on broader workloads in an early version of PLFS, specifically the need to build and distribute an index for the overall file, and the pressure on the underlying parallel file system's metadata server, and show how PLFS's decomposed components architecture can be exploited to alleviate bottlenecks in the underlying parallel file system.

  17. Ontology Language to Support Description of Experiment Control System Semantics, Collaborative Knowledge-Base Design and Ontology Reuse

    SciTech Connect (OSTI)

    Vardan Gyurjyan, D Abbott, G Heyes, E Jastrzembski, B Moffit, C Timmer, E Wolin

    2009-10-01T23:59:59.000Z

    In this paper we discuss the control domain specific ontology that is built on top of the domain-neutral Resource Definition Framework (RDF). Specifically, we will discuss the relevant set of ontology concepts along with the relationships among them in order to describe experiment control components and generic event-based state machines. Control Oriented Ontology Language (COOL) is a meta-data modeling language that provides generic means for representation of physics experiment control processes and components, and their relationships, rules and axioms. It provides a semantic reference frame that is useful for automating the communication of information for configuration, deployment and operation. COOL has been successfully used to develop a complete and dynamic knowledge-base for experiment control systems, developed using the AFECS framework.

  18. Handling of Multimedia Files in the Invenio Software

    E-Print Network [OSTI]

    Oltmanns, Björn; Schiefer, Bernhard

    ‘Handling of multimedia files in the Invenio Software’ is motivated by the need for integration of multimedia files in the open-source, large-scale digital library software Invenio, developed and used at CERN, the European Organisation for Nuclear Research. In the last years, digital assets like pictures, presentations podcasts and videos became abundant in these systems and digital libraries have grown out of their classic role of only storing bibliographical metadata. The thesis focuses on digital video as a type of multimedia and covers the complete workflow of handling video material in the Invenio software: from the ingestion of digital video material to its processing on to the storage and preservation and finally the streaming and presentation of videos to the user. The potential technologies to realise a video submission workflow are discussed in-depth and evaluated towards system integration with Invenio. The focus is set on open and free technologies, which can be redistributed with the Inve...

  19. Establishment of the Northeast Coastal Watershed Geospatial Data Network (NECWGDN)

    SciTech Connect (OSTI)

    Hannigan, Robyn [University of Massachusetts Boston] [University of Massachusetts Boston

    2014-02-17T23:59:59.000Z

    The goals of NECWGDN were to establish integrated geospatial databases that interfaced with existing open-source (water.html) environmental data server technologies (e.g., HydroDesktop) and included ecological and human data to enable evaluation, prediction, and adaptation in coastal environments to climate- and human-induced threats to the coastal marine resources within the Gulf of Maine. We have completed the development and testing of a "test bed" architecture that is compatible with HydroDesktop and have identified key metadata structures that will enable seamless integration and delivery of environmental, ecological, and human data as well as models to predict threats to end-users. Uniquely this database integrates point as well as model data and so offers capacities to end-users that are unique among databases. Future efforts will focus on the development of integrated environmental-human dimension models that can serve, in near real time, visualizations of threats to coastal resources and habitats.

  20. Check Your Data Freedom: A Taxonomy to Assess Life Science Database Openness

    E-Print Network [OSTI]

    De Rosnay, Melanie Dulong

    2012-01-01T23:59:59.000Z

    Molecular biology data are subject to terms of use that vary widely between databases and curating institutions. This research presents a taxonomy of contractual and technical restrictions applicable to databases in life science. It builds upon research led by Science Commons demonstrating why open data and the freedom to integrate facilitate innovation and how this openness can be achieved. The taxonomy describes technical and legal restrictions applicable to life science databases, and its metadata have been used to assess terms of use of databases hosted by Life Science Resource Name (LSRN) Schema. While a few public domain policies are standardized, most terms of use are not harmonized, difficult to understand and impose controls that prevent others from effectively reusing data. Identifying a small number of restrictions allows one to quickly appreciate which databases are open. A checklist for data openness is proposed in order to assist database curators who wish to make their data more open to make su...

  1. Learning from FITS: Limitations in use in modern astronomical research

    E-Print Network [OSTI]

    Thomas, Brian; Economou, Frossie; Greenfield, Perry; Hirst, Paul; Berry, David S; Bray, Erik; Gray, Norman; Muna, Demitri; Turner, James; de Val-Borro, Miguel; Santander-Vela, Juande; Shupe, David; Good, John; Berriman, G Bruce; Kitaeff, Slava; Fay, Jonathan; Laurino, Omar; Alexov, Anastasia; Landry, Walter; Masters, Joe; Brazier, Adam; Schaaf, Reinhold; Edwards, Kevin; Redman, Russell O; Marsh, Thomas R; Streicher, Ole; Norris, Pat; Pascual, Sergio; Davie, Matthew; Droettboom, Michael; Robitaille, Thomas; Campana, Riccardo; Hagen, Alex; Hartogh, Paul; Klaes, Dominik; Craiga, Matthew W; Homeier, Derek

    2015-01-01T23:59:59.000Z

    The Flexible Image Transport System (FITS) standard has been a great boon to astronomy, allowing observatories, scientists and the public to exchange astronomical information easily. The FITS standard, however, is showing its age. Developed in the late 1970s, the FITS authors made a number of implementation choices that, while common at the time, are now seen to limit its utility with modern data. The authors of the FITS standard could not anticipate the challenges which we are facing today in astronomical computing. Difficulties we now face include, but are not limited to, addressing the need to handle an expanded range of specialized data product types (data models), being more conducive to the networked exchange and storage of data, handling very large datasets, and capturing significantly more complex metadata and data relationships. There are members of the community today who find some or all of these limitations unworkable, and have decided to move ahead with storing data in other formats. If this frag...

  2. Designing a future Conditions Database based on LHC experience

    E-Print Network [OSTI]

    Formica, Andrea; The ATLAS collaboration; Gallas, Elizabeth; Govi, Giacomo; Lehmann Miotto, Giovanna; Pfeiffer, Andreas

    2015-01-01T23:59:59.000Z

    The ATLAS and CMS Conditions Database infrastructures have served each of the respective experiments well through LHC Run 1, providing efficient access to a wide variety of conditions information needed in online data taking and offline processing and analysis. During the long shutdown between Run 1 and Run 2, we have taken various measures to improve our systems for Run 2. In some cases, a drastic change was not possible because of the relatively short time scale to prepare for Run 2. In this process, and in the process of comparing to the systems used by other experiments, we realized that for Run 3, we should consider more fundamental changes and possibilities. We seek changes which would streamline conditions data management, improve monitoring tools, better integrate the use of metadata, incorporate analytics to better understand conditions usage, as well as investigate fundamental changes in the storage technology, which might be more efficient while minimizing maintenance of the data as well as simplif...

  3. SPOCS: Software for Predicting and Visualizing Orthology/Paralogy Relationships Among Genomes

    SciTech Connect (OSTI)

    Curtis, Darren S.; Phillips, Aaron R.; Callister, Stephen J.; Conlan, Sean; McCue, Lee Ann

    2013-10-15T23:59:59.000Z

    At the rate that prokaryotic genomes can now be generated, comparative genomics studies require a flexible method for quickly and accurately predicting orthologs among the rapidly changing set of genomes available. SPOCS implements a graph-based ortholog prediction method to generate a simple tab-delimited table of orthologs and in addition, html files that provide a visualization of the predicted ortholog/paralog relationships to which gene/protein expression metadata may be overlaid. AVAILABILITY AND IMPLEMENTATION: A SPOCS web application is freely available at http://cbb.pnnl.gov/portal/tools/spocs.html. Source code for Linux systems is also freely available under an open source license at http://cbb.pnnl.gov/portal/software/spocs.html; the Boost C++ libraries and BLAST are required.

  4. DOE Geothermal Data Repository - Tethering Data to Information: Preprint

    SciTech Connect (OSTI)

    Weers, J.; Anderson, A.

    2014-02-01T23:59:59.000Z

    Data are not inherently information. Without context, data are just numbers, figures, names, or points on a line. By assigning context to data, we can validate ideas, form opinions, and generate knowledge. This is an important distinction to information scientists, as we recognize that the context in which we keep our data plays a big part in generating its value. The mechanisms used to assign this context often include their own data, supplemental to the data being described and defining semantic relationships, commonly referred to as metadata. This paper provides the status of the DOE Geothermal Data Repository (DOE GDR), including recent efforts to tether data submissions to information, discusses the important distinction between data and information, outlines a path to generate useful knowledge from raw data, and details the steps taken in order to become a node on the National Geothermal Data System (NGDS).

  5. A Framework for the Systematic Collection of Open Source Intelligence

    SciTech Connect (OSTI)

    Pouchard, Line Catherine [ORNL] [ORNL; Trien, Joseph P [ORNL] [ORNL; Dobson, Jonathan D [ORNL] [ORNL

    2009-01-01T23:59:59.000Z

    Following legislative directions, the Intelligence Community has been mandated to make greater use of Open Source Intelligence (OSINT). Efforts are underway to increase the use of OSINT but there are many obstacles. One of these obstacles is the lack of tools helping to manage the volume of available data and ascertain its credibility. We propose a unique system for selecting, collecting and storing Open Source data from the Web and the Open Source Center. Some data management tasks are automated, document source is retained, and metadata containing geographical coordinates are added to the documents. Analysts are thus empowered to search, view, store, and analyze Web data within a single tool. We present ORCAT I and ORCAT II, two implementations of the system.

  6. Report of the Event Tag Review and Recommendation Group

    SciTech Connect (OSTI)

    ATLAS Group; Assamagan, K.A.; Barberis, D.; Bentvelsen, S.; Brooijmans, G.; Cranmer, K.; Cranshaw, J.; Dell'Acqua, A.; Farbin, A.; Froidevaux, D.; Gianotti, F.; Hinchliffe, I.; LeCompte, T.; Maeno, T.; Malon, D.; Paige, F.; Polesello, G.; Quarrie, D.; Rousseau, D.; Schaffer, R.D.; Smizanska, M.; Unal, G.; Voss, K.; Wielers, M.

    2006-04-12T23:59:59.000Z

    In order to facilitate access to the large volumes of data (multiple petabytes per year) which will be produced during data taking and Monte Carlo production at ATLAS, work has proceeded on building a system of event-level metadata to allow selections of a subset of events to use as input to an analysis. This was included in the ATLAS Computing Model and was first studied and implemented by the Physics Analysis Tools group based on the decisions of the ESD/AOD Task Force. They used tools developed and supported by the CERN IT group and the ATLAS Database group. During 2005 this structure was put through various tests and evaluations. Also, work by physicists on reconstruction and analysis led to an improved understanding of the requirements on the TAG. This report addresses the effect of these new inputs on the previous work with regard to content and the infrastructure needed to support it.

  7. Fault-tolerant dynamic task graph scheduling

    SciTech Connect (OSTI)

    Kurt, Mehmet C.; Krishnamoorthy, Sriram; Agrawal, Kunal; Agrawal, Gagan

    2014-11-16T23:59:59.000Z

    In this paper, we present an approach to fault tolerant execution of dynamic task graphs scheduled using work stealing. In particular, we focus on selective and localized recovery of tasks in the presence of soft faults. We elicit from the user the basic task graph structure in terms of successor and predecessor relationships. The work stealing-based algorithm to schedule such a task graph is augmented to enable recovery when the data and meta-data associated with a task get corrupted. We use this redundancy, and the knowledge of the task graph structure, to selectively recover from faults with low space and time overheads. We show that the fault tolerant design retains the essential properties of the underlying work stealing-based task scheduling algorithm, and that the fault tolerant execution is asymptotically optimal when task re-execution is taken into account. Experimental evaluation demonstrates the low cost of recovery under various fault scenarios.

  8. Final Report for "Analyzing and visualizing next generation climate data"

    SciTech Connect (OSTI)

    Pletzer, Alexander

    2012-11-13T23:59:59.000Z

    The project "Analyzing and visualizing next generation climate data" adds block-structured (mosaic) grid support, parallel processing, and 2D/3D curvilinear interpolation to the open-source UV-CDAT climate data analysis tool. Block structured grid support complies to the Gridspec extension submitted to the Climate and Forecast metadata conventions. It contains two parts: aggregation of data spread over multiple mosaic tiles (M-SPEC) and aggregation of temporal data stored in different files (F-SPEC). Together, M-SPEC and F-SPEC allow users to interact with data stored in multiple files as if the data were in a single file. For computational expensive tasks, a flexible, multi-dimensional, multi-type distributed array class allows users to process data in parallel using remote memory access. Both nodal and cell based interpolation is supported; users can choose between different interpolation libraries including ESMF and LibCF depending on the their particular needs.

  9. Using architectures for semantic interoperability to create journal clubs for emergency response

    SciTech Connect (OSTI)

    Powell, James E [Los Alamos National Laboratory; Collins, Linn M [Los Alamos National Laboratory; Martinez, Mark L B [Los Alamos National Laboratory

    2009-01-01T23:59:59.000Z

    In certain types of 'slow burn' emergencies, careful accumulation and evaluation of information can offer a crucial advantage. The SARS outbreak in the first decade of the 21st century was such an event, and ad hoc journal clubs played a critical role in assisting scientific and technical responders in identifying and developing various strategies for halting what could have become a dangerous pandemic. This research-in-progress paper describes a process for leveraging emerging semantic web and digital library architectures and standards to (1) create a focused collection of bibliographic metadata, (2) extract semantic information, (3) convert it to the Resource Description Framework /Extensible Markup Language (RDF/XML), and (4) integrate it so that scientific and technical responders can share and explore critical information in the collections.

  10. Radio Astronomy Data Model for Single-Dish Multiple-Feed Telescopes, and Robledo Archive Architecture

    E-Print Network [OSTI]

    J. D. Santander-Vela; E. Garcia; J. F. Gomez; L. Verdes-Montenegro; S. Leon; R. Gutierrez; C. Rodrigo; O. Morata; E. Solano; O. Suarez

    2008-10-02T23:59:59.000Z

    All the effort that the astrophysical community has put into the development of the Virtual Observatory (VO) has surpassed the non-return point: the VO is a reality today, and an initiative that will self-sustain, and to which all archival projects must adhere. We have started the design of the scientific archive for the DSS-63 70-m antenna at NASA's DSN station in Robledo de Chavela (Madrid). Here we show how we can use all VO proposed data models to build a VO-compliant single-dish, multiple-feed, radio astronomical archive data model (RADAMS) suitable for the archival needs of the antenna. We also propose an exhaustive list of Universal Content Descriptors (UCDs) and FITS keywords for all relevant metadata. We will further refine this data model with the experience that we will gain from that implementation.

  11. The QuarkNet/Grid collaborative learning e-lab

    SciTech Connect (OSTI)

    Bardeen, Marjorie; Gilbert, Eric; Jordan, Thomas; Nepywoda, Paul; Quigg, Elizabeth; /Fermilab; Wilde, Mike; /Argonne; Zhao, Yong; /Chicago U.

    2004-12-01T23:59:59.000Z

    We describe a case study that uses grid computing techniques to support the collaborative learning of high school students investigating cosmic rays. Students gather and upload science data to our e-Lab portal. They explore those data using techniques from the GriPhyN collaboration. These techniques include virtual data transformations, workflows, metadata cataloging and indexing, data product provenance and persistence, as well as job planners. Students use web browsers and a custom interface that extends the GriPhyN Chiron portal to perform all of these tasks. They share results in the form of online posters and ask each other questions in this asynchronous environment. Students can discover and extend the research of other students, modeling the processes of modern large-scale scientific collaborations. Also, the e-Lab portal provides tools for teachers to guide student work throughout an investigation.

  12. Revising the worksheet with L3: a language and environment foruser-script interaction

    SciTech Connect (OSTI)

    Hohn, Michael H.

    2008-01-22T23:59:59.000Z

    This paper describes a novel approach to the parameter anddata handling issues commonly found in experimental scientific computingand scripting in general. The approach is based on the familiarcombination of scripting language and user interface, but using alanguage expressly designed for user interaction and convenience. The L3language combines programming facilities of procedural and functionallanguages with the persistence and need-based evaluation of data flowlanguages. It is implemented in Python, has access to all Pythonlibraries, and retains almost complete source code compatibility to allowsimple movement of code between the languages. The worksheet interfaceuses metadata produced by L3 to provide selection of values through thescriptit self and allow users to dynamically evolve scripts withoutre-running the prior versions. Scripts can be edited via text editors ormanipulated as structures on a drawing canvas. Computed values are validscripts and can be used further in other scripts via simplecopy-and-paste operations. The implementation is freely available underan open-source license.

  13. Calculating a checksum with inactive networking components in a computing system

    DOE Patents [OSTI]

    Aho, Michael E; Chen, Dong; Eisley, Noel A; Gooding, Thomas M; Heidelberger, Philip; Tauferner, Andrew T

    2014-12-16T23:59:59.000Z

    Calculating a checksum utilizing inactive networking components in a computing system, including: identifying, by a checksum distribution manager, an inactive networking component, wherein the inactive networking component includes a checksum calculation engine for computing a checksum; sending, to the inactive networking component by the checksum distribution manager, metadata describing a block of data to be transmitted by an active networking component; calculating, by the inactive networking component, a checksum for the block of data; transmitting, to the checksum distribution manager from the inactive networking component, the checksum for the block of data; and sending, by the active networking component, a data communications message that includes the block of data and the checksum for the block of data.

  14. OpenMSI: A High-Performance Web-Based Platform for Mass Spectrometry Imaging

    SciTech Connect (OSTI)

    Rubel, Oliver; Greiner, Annette; Cholia, Shreyas; Louie, Katherine; Bethel, E. Wes; Northen, Trent R.; Bowen, Benjamin P.

    2013-10-02T23:59:59.000Z

    Mass spectrometry imaging (MSI) enables researchers to directly probe endogenous molecules directly within the architecture of the biological matrix. Unfortunately, efficient access, management, and analysis of the data generated by MSI approaches remain major challenges to this rapidly developing field. Despite the availability of numerous dedicated file formats and software packages, it is a widely held viewpoint that the biggest challenge is simply opening, sharing, and analyzing a file without loss of information. Here we present OpenMSI, a software framework and platform that addresses these challenges via an advanced, high-performance, extensible file format and Web API for remote data access (http://openmsi.nersc.gov). The OpenMSI file format supports storage of raw MSI data, metadata, and derived analyses in a single, self-describing format based on HDF5 and is supported by a large range of analysis software (e.g., Matlab and R) and programming languages (e.g., C++, Fortran, and Python). Careful optimization of the storage layout of MSI data sets using chunking, compression, and data replication accelerates common, selective data access operations while minimizing data storage requirements and are critical enablers of rapid data I/O. The OpenMSI file format has shown to provide >2000-fold improvement for image access operations, enabling spectrum and image retrieval in less than 0.3 s across the Internet even for 50 GB MSI data sets. To make remote high-performance compute resources accessible for analysis and to facilitate data sharing and collaboration, we describe an easy-to-use yet powerful Web API, enabling fast and convenient access to MSI data, metadata, and derived analysis results stored remotely to facilitate high-performance data analysis and enable implementation of Web based data sharing, visualization, and analysis.

  15. Integration experiences and performance studies of A COTS parallel archive systems

    SciTech Connect (OSTI)

    Chen, Hsing-bung [Los Alamos National Laboratory; Scott, Cody [Los Alamos National Laboratory; Grider, Bary [Los Alamos National Laboratory; Torres, Aaron [Los Alamos National Laboratory; Turley, Milton [Los Alamos National Laboratory; Sanchez, Kathy [Los Alamos National Laboratory; Bremer, John [Los Alamos National Laboratory

    2010-01-01T23:59:59.000Z

    Current and future Archive Storage Systems have been asked to (a) scale to very high bandwidths, (b) scale in metadata performance, (c) support policy-based hierarchical storage management capability, (d) scale in supporting changing needs of very large data sets, (e) support standard interface, and (f) utilize commercial-off-the-shelf(COTS) hardware. Parallel file systems have been asked to do the same thing but at one or more orders of magnitude faster in performance. Archive systems continue to move closer to file systems in their design due to the need for speed and bandwidth, especially metadata searching speeds such as more caching and less robust semantics. Currently the number of extreme highly scalable parallel archive solutions is very small especially those that will move a single large striped parallel disk file onto many tapes in parallel. We believe that a hybrid storage approach of using COTS components and innovative software technology can bring new capabilities into a production environment for the HPC community much faster than the approach of creating and maintaining a complete end-to-end unique parallel archive software solution. In this paper, we relay our experience of integrating a global parallel file system and a standard backup/archive product with a very small amount of additional code to provide a scalable, parallel archive. Our solution has a high degree of overlap with current parallel archive products including (a) doing parallel movement to/from tape for a single large parallel file, (b) hierarchical storage management, (c) ILM features, (d) high volume (non-single parallel file) archives for backup/archive/content management, and (e) leveraging all free file movement tools in Linux such as copy, move, ls, tar, etc. We have successfully applied our working COTS Parallel Archive System to the current world's first petaflop/s computing system, LANL's Roadrunner, and demonstrated its capability to address requirements of future archival storage systems.

  16. Provenance In Sensor Data Management: A Cohesive, Independent Solution

    SciTech Connect (OSTI)

    Hensley, Zachary P [ORNL] [ORNL; Sanyal, Jibonananda [ORNL] [ORNL; New, Joshua Ryan [ORNL] [ORNL

    2014-01-01T23:59:59.000Z

    In today's information-driven workplaces, data is constantly undergoing transformations and being moved around. The typical business-as-usual approach is to use email attachments, shared network locations, databases, and now, the cloud. More often than not, there are multiple versions of the data sitting in different locations and users of this data are confounded by the lack of metadata describing its provenance, or in other words, its lineage. Our project is aimed to solve this issue in the context of sensor data. The Oak Ridge National Laboratory's Building Technologies Research and Integration Center has reconfigurable commercial buildings deployed on the Flexible Research Platforms (FRPs). These FRPs are instrumented with a large number of sensors which measure a number of variables such as HVAC efficiency, relative humidity, and temperature gradients across doors, windows, and walls. Sub-minute resolution data from hundreds of channels is acquired. This sensor data, traditionally, was saved to a shared network location which was accessible to a number of scientists for performing complicated simulation and analysis tasks. The sensor data also participates in elaborate quality assurance exercises as a result of inherent faults. Sometimes, faults are induced to observe building behavior. It became apparent that proper scientific controls required not just managing the data acquisition and delivery, but to also manage the metadata associated with temporal subsets of the sensor data. We built a system named ProvDMS, or Provenance Data Management System for the FRPs, which would both allow researchers to retrieve data of interest as well as trace data lineage. This provides researchers a one-stop shop for comprehensive views of various data transformation allowing researchers to effectively trace their data to its source so that experiments, and derivations of experiments, may be reused and reproduced without much overhead of the repeatability of experiments that use it. Using these traces, researchers can determine exactly what happens to data as it moves through its life cycle.

  17. Integration experiments and performance studies of a COTS parallel archive system

    SciTech Connect (OSTI)

    Chen, Hsing-bung [Los Alamos National Laboratory; Scott, Cody [Los Alamos National Laboratory; Grider, Gary [Los Alamos National Laboratory; Torres, Aaron [Los Alamos National Laboratory; Turley, Milton [Los Alamos National Laboratory; Sanchez, Kathy [Los Alamos National Laboratory; Bremer, John [Los Alamos National Laboratory

    2010-06-16T23:59:59.000Z

    Current and future Archive Storage Systems have been asked to (a) scale to very high bandwidths, (b) scale in metadata performance, (c) support policy-based hierarchical storage management capability, (d) scale in supporting changing needs of very large data sets, (e) support standard interface, and (f) utilize commercial-off-the-shelf (COTS) hardware. Parallel file systems have been asked to do the same thing but at one or more orders of magnitude faster in performance. Archive systems continue to move closer to file systems in their design due to the need for speed and bandwidth, especially metadata searching speeds such as more caching and less robust semantics. Currently the number of extreme highly scalable parallel archive solutions is very small especially those that will move a single large striped parallel disk file onto many tapes in parallel. We believe that a hybrid storage approach of using COTS components and innovative software technology can bring new capabilities into a production environment for the HPC community much faster than the approach of creating and maintaining a complete end-to-end unique parallel archive software solution. In this paper, we relay our experience of integrating a global parallel file system and a standard backup/archive product with a very small amount of additional code to provide a scalable, parallel archive. Our solution has a high degree of overlap with current parallel archive products including (a) doing parallel movement to/from tape for a single large parallel file, (b) hierarchical storage management, (c) ILM features, (d) high volume (non-single parallel file) archives for backup/archive/content management, and (e) leveraging all free file movement tools in Linux such as copy, move, Is, tar, etc. We have successfully applied our working COTS Parallel Archive System to the current world's first petafiop/s computing system, LANL's Roadrunner machine, and demonstrated its capability to address requirements of future archival storage systems.

  18. Gap Assessment (FY 13 Update)

    SciTech Connect (OSTI)

    Getman, Dan

    2013-09-30T23:59:59.000Z

    To help guide its future data collection efforts, The DOE GTO funded a data gap analysis in FY2012 to identify high potential hydrothermal areas where critical data are needed. This analysis was updated in FY2013 and the resulting datasets are represented by this metadata. The original process was published in FY 2012 and is available here: https://pangea.stanford.edu/ERE/db/GeoConf/papers/SGW/2013/Esposito.pdf Though there are many types of data that can be used for hydrothermal exploration, five types of exploration data were targeted for this analysis. These data types were selected for their regional reconnaissance potential, and include many of the primary exploration techniques currently used by the geothermal industry. The data types include: 1. well data 2. geologic maps 3. fault maps 4. geochemistry data 5. geophysical data To determine data coverage, metadata for exploration data (including data type, data status, and coverage information) were collected and catalogued from nodes on the National Geothermal Data System (NGDS). It is the intention of this analysis that the data be updated from this source in a semi-automated fashion as new datasets are added to the NGDS nodes. In addition to this upload, an online tool was developed to allow all geothermal data providers to access this assessment and to directly add metadata themselves and view the results of the analysis via maps of data coverage in Geothermal Prospector (http://maps.nrel.gov/gt_prospector). A grid of the contiguous U.S. was created with 88,000 10-km by 10-km grid cells, and each cell was populated with the status of data availability corresponding to the five data types. Using these five data coverage maps and the USGS Resource Potential Map, sites were identified for future data collection efforts. These sites signify both that the USGS has indicated high favorability of occurrence of geothermal resources and that data gaps exist. The uploaded data are contained in two data files for each data category. The first file contains the grid and is in the SHP file format (shape file.) Each populated grid cell represents a 10k area within which data is known to exist. The second file is a CSV (comma separated value) file that contains all of the individual layers that intersected with the grid. This CSV can be joined with the map to retrieve a list of datasets that are available at any given site. The attributes in the CSV include: 1. grid_id : The id of the grid cell that the data intersects with 2. title: This represents the name of the WFS service that intersected with this grid cell 3. abstract: This represents the description of the WFS service that intersected with this grid cell 4. gap_type: This represents the category of data availability that these data fall within. As the current processing is pulling data from NGDS, this category universally represents data that are available in the NGDS and are ready for acquisition for analytic purposes. 5. proprietary_type: Whether the data are considered proprietary 6. service_type: The type of service 7. base_url: The service URL

  19. Environmental effects of marine energy development around the world. Annex IV Final Report

    SciTech Connect (OSTI)

    Copping, Andrea; Hanna, Luke; Whiting, Johnathan; Geerlofs, Simon; Grear, Molly; Blake, Kara (Pacific Northwest National Laboratory, Richland, WA (United States)); Coffey, Anna; Massaua, Meghan; Brown-Saracino, Jocelyn; Battey, Hoyt (US Dept. of Energy, Washington, DC (United States))

    2013-01-15T23:59:59.000Z

    Annex IV is an international collaborative project to examine the environmental effects of marine energy devices among countries through the International Energy Agency’s Ocean Energy Systems Initiative (OES). The U.S. Department of Energy (DOE) serves as the Operating Agent for the Annex, in partnership with the Bureau of Ocean Energy Management (BOEM; formerly the Minerals Management Service), the Federal Energy Regulatory Commission (FERC), and National Oceanographic and Atmospheric Administration (NOAA). Numerous ocean energy technologies and devices are being developed around the world, and the few data that exist about the environmental effects of these technologies are dispersed among countries and developers. The purpose of Annex IV is to facilitate efficient government oversight of the development of ocean energy systems by compiling and disseminating information about the potential environmental effects of marine energy technologies and to identify methods of monitoring for these effects. Beginning in 2010, this three-year effort produced a publicly available searchable online database of environmental effects information (Tethys). It houses scientific literature pertaining to the environmental effects of marine energy systems, as well as metadata on international ocean energy projects and research studies. Two experts’ workshops were held in Dublin, Ireland (September 2010 and October 2012) to engage with international researchers, developers, and regulators on the scope and outcomes of the Annex IV project. Metadata and information stored in the Tethys database and feedback obtained from the two experts’ workshops were used as resources in the development of this report. This Annex IV final report contains three case studies of specific interactions of marine energy devices with the marine environment that survey, compile, and analyze the best available information in one coherent location. These case studies address 1) the physical interactions between animals and tidal turbines; 2) the acoustic impact of marine energy devices on marine animals; and 3) the effects of energy removal on physical systems. Each case study contains a description of environmental monitoring efforts and research studies, lessons learned, and analysis of remaining information gaps. The information collected through the Annex IV effort and referenced in this report, can be accessed on the Tethys database at http://mhk.pnnl.gov/wiki/index.php/Tethys_ Home.

  20. National Geothermal Data System: Interactive Assessment of Geothermal Energy Potential in the U.S.

    SciTech Connect (OSTI)

    Allison, Lee [Executive Office of the State of Arizona (Arizona Geological Survey); Richard, Stephen [Executive Office of the State of Arizona (Arizona Geological Survey); Clark, Ryan; Patten, Kim; Love, Diane; Coleman, Celia; Chen, Genhan; Matti, Jordan; Pape, Estelle; Musil, Leah

    2012-01-30T23:59:59.000Z

    Geothermal-relevant geosciences data from all 50 states (www.stategeothermaldata.org), federal agencies, national labs, and academic centers are being digitized and linked in a distributed online network via the U.S. Department of Energy-funded National Geothermal Data System (NGDS) to foster geothermal energy exploration and development through use of interactive online ‘mashups,’data integration, and applications. Emphasis is first to make as much information as possible accessible online, with a long range goal to make data interoperable through standardized services and interchange formats. An initial set of thirty geoscience data content models is in use or under development to define a standardized interchange format: aqueous chemistry, borehole temperature data, direct use feature, drill stem test, earthquake hypocenter, fault feature, geologic contact feature, geologic unit feature, thermal/hot spring description, metadata, quaternary fault, volcanic vent description, well header feature, borehole lithology log, crustal stress, gravity, heat flow/temperature gradient, permeability, and feature descriptions data like developed geothermal systems, geologic unit geothermal properties, permeability, production data, rock alteration description, rock chemistry, and thermal conductivity. Map services are also being developed for isopach maps, aquifer temperature maps, and several states are working on geothermal resource overview maps. Content models are developed preferentially from existing community use in order to encourage widespread adoption and promulgate minimum metadata quality standards. Geoscience data and maps from other NGDS participating institutions, or “nodes” (USGS, Southern Methodist University, Boise State University Geothermal Data Coalition) are being supplemented with extensive land management and land use resources from the Western Regional Partnership (15 federal agencies and 5 Western states) to provide access to a comprehensive, holistic set of data critical to geothermal energy development. As of September 2011, we have over 34,000 records registered in the system catalog, and 234,942 data resources online, along with scores of Web services to deliver integrated data to the desktop for free downloading or online use. The data exchange mechanism is built on the U.S. Geoscience Information Network (USGIN, http://usgin.org and http://lab.usgin.org) protocols and standards developed as a partnership of the Association of American State Geologists (AASG) and U.S. Geological Survey.

  1. A computer system for access to distributed genome mapping data. Progress report

    SciTech Connect (OSTI)

    Marr, T.G.

    1992-02-05T23:59:59.000Z

    Development of a computer system for access to distributed genome mapping data is continuing. This effort is to develop software which accesses multiple databases and retrieves data which contain information useful for accelerating mapping human chromosomes. For example, the molecular sequence databases (GenBank, EMBL Data Library, PIR, SwissProt) which contain data required for the development of oligonucleotides for probing DNA as well as for extracting data for primer pair development for PCR-based methods. It is also to develop software which qualitatively integrates the following mapping data: (1) markers regionally localized using cytogenetic methods, (2) polymorphic markers ordered by genetic linkage analysis, (3) clones ordered by various ``finger-printing`` methods, (4) fragments ordered by long-range restriction mapping, (5) single genomic fragments or clones that have STSs assigned to them, (6) nucleotide sequences, (7) the associated metadata such as the submitting investigator`s name, location, etc; the source organism; the chromosome the element is from; the chromosomal location is whatever detail is available.

  2. A computer system for access to distributed genome mapping data

    SciTech Connect (OSTI)

    Marr, T.G.

    1992-02-05T23:59:59.000Z

    Development of a computer system for access to distributed genome mapping data is continuing. This effort is to develop software which accesses multiple databases and retrieves data which contain information useful for accelerating mapping human chromosomes. For example, the molecular sequence databases (GenBank, EMBL Data Library, PIR, SwissProt) which contain data required for the development of oligonucleotides for probing DNA as well as for extracting data for primer pair development for PCR-based methods. It is also to develop software which qualitatively integrates the following mapping data: (1) markers regionally localized using cytogenetic methods, (2) polymorphic markers ordered by genetic linkage analysis, (3) clones ordered by various finger-printing'' methods, (4) fragments ordered by long-range restriction mapping, (5) single genomic fragments or clones that have STSs assigned to them, (6) nucleotide sequences, (7) the associated metadata such as the submitting investigator's name, location, etc; the source organism; the chromosome the element is from; the chromosomal location is whatever detail is available.

  3. Knowledge Integration to Make Decisions About Complex Systems: Sustainability of Energy Production from Agriculture

    ScienceCinema (OSTI)

    Francesco Danuso

    2010-01-08T23:59:59.000Z

    A major bottleneck for improving the governance of complex systems, rely on our ability to integrate different forms of knowledge into a decision support system (DSS). Preliminary aspects are the classification of different types of knowledge (a priori or general, a posteriori or specific, with uncertainty, numerical, textual, algorithmic, complete/incomplete, etc.), the definition of ontologies for knowledge management and the availability of proper tools like continuous simulation models, event driven models, statistical approaches, computational methods (neural networks, evolutionary optimization, rule based systems etc.) and procedure for textual documentation. Following these views at University of Udine, a computer language (SEMoLa, Simple, Easy Modelling Language) for knowledge integration has been developed.  SEMoLa can handle models, data, metadata and textual knowledge; it implements and extends the system dynamics ontology (Forrester, 1968; Jřrgensen, 1994) in which systems are modelled by the concepts of material, group, state, rate, parameter, internal and external events and driving variables. As an example, a SEMoLa model to improve management and sustainability (economical, energetic, environmental) of the agricultural farms is presented. The model (X-Farm) simulates a farm in which cereal and forage yield, oil seeds, milk, calves and wastes can be sold or reused. X-Farm is composed by integrated modules describing fields (crop and soil), feeds and materials storage, machinery management, manpower  management, animal husbandry, economic and energetic balances, seed oil extraction, manure and wastes management, biogas production from animal wastes and biomasses.

  4. The NASA Astrophysics Data System: Architecture

    E-Print Network [OSTI]

    A. Accomazzi; G. Eichhorn; M. J. Kurtz; C. S. Grant; S. S. Murray

    2000-02-04T23:59:59.000Z

    The powerful discovery capabilities available in the ADS bibliographic services are possible thanks to the design of a flexible search and retrieval system based on a relational database model. Bibliographic records are stored as a corpus of structured documents containing fielded data and metadata, while discipline-specific knowledge is segregated in a set of files independent of the bibliographic data itself. The creation and management of links to both internal and external resources associated with each bibliography in the database is made possible by representing them as a set of document properties and their attributes. To improve global access to the ADS data holdings, a number of mirror sites have been created by cloning the database contents and software on a variety of hardware and software platforms. The procedures used to create and manage the database and its mirrors have been written as a set of scripts that can be run in either an interactive or unsupervised fashion. The ADS can be accessed at http://adswww.harvard.edu

  5. THE MADISON SQUARE GARDEN DISPERSION STUDY (MSG05) METEOROLOGICAL DATA DESCRIPTION.

    SciTech Connect (OSTI)

    REYNOLDS, R.M.

    2006-10-01T23:59:59.000Z

    MSG05 was a study of atmospheric transport and dispersion in the deep urban canyons of Midtown New York City, in the area of Madison Square Garden. This downtown area is considered to be a prime target for terrorist activities, and has one of the largest commuter populations in the world. Little is known about air flow and hazardous gas dispersion in such scenarios, since previous urban field experiments have focused on small to medium sized cities with much smaller street canyons. On March 10 and 14, 2005, a series of Perfluorocarbon Tracer (PFT) tracers were released and tracked with about 30 sampling stations at radial distances of about 0.2 and 0.4 km, with vertical profiles near a 250 m tall building (One Penn Plaza). Meteorological stations collected wind data in the MSG vicinity, at street level and rooftop level. MSG05 is expected to provide useful information on rapid vertical dispersion will assist in planning for more extensive studies. This data release is being made available to a restricted group of key scientists who have worked on the project. Part of the QA program involves feedback from scientists and modelers who are working on this study. This document describes the meteorological component of the project. The file organization and metadata are detailed so that a researcher can work with the data sets.

  6. A Bioinformatician's Guide to Metagenomics

    SciTech Connect (OSTI)

    Kunin, Victor; Copeland, Alex; Lapidus, Alla; Mavromatis, Konstantinos; Hugenholtz, Philip

    2008-08-01T23:59:59.000Z

    As random shotgun metagenomic projects proliferate and become the dominant source of publicly available sequence data, procedures for best practices in their execution and analysis become increasingly important. Based on our experience at the Joint Genome Institute, we describe step-by-step the chain of decisions accompanying a metagenomic project from the viewpoint of a bioinformatician. We guide the reader through a standard workflow for a metagenomic project beginning with pre-sequencing considerations such as community composition and sequence data type that will greatly influence downstream analyses. We proceed with recommendations for sampling and data generation including sample and metadata collection, community profiling, construction of shotgun libraries and sequencing strategies. We then discuss the application of generic sequence processing steps (read preprocessing, assembly, and gene prediction and annotation) to metagenomic datasets by contrast to genome projects. Different types of data analyses particular to metagenomes are then presented including binning, dominant population analysis and gene-centric analysis. Finally data management systems and issues are presented and discussed. We hope that this review will assist bioinformaticians and biologists in making better-informed decisions on their journey during a metagenomic project.

  7. Automatic Generation of Data Types for Classification of Deep Web Sources

    SciTech Connect (OSTI)

    Ngu, A H; Buttler, D J; Critchlow, T J

    2005-02-14T23:59:59.000Z

    A Service Class Description (SCD) is an effective meta-data based approach for discovering Deep Web sources whose data exhibit some regular patterns. However, it is tedious and error prone to create an SCD description manually. Moreover, a manually created SCD is not adaptive to the frequent changes of Web sources. It requires its creator to identify all the possible input and output types of a service a priori. In many domains, it is impossible to exhaustively list all the possible input and output data types of a source in advance. In this paper, we describe machine learning approaches for automatic generation of the data types of an SCD. We propose two different approaches for learning data types of a class of Web sources. The Brute-Force Learner is able to generate data types that can achieve high recall, but with low precision. The Clustering-based Learner generates data types that have a high precision rate, but with a lower recall rate. We demonstrate the feasibility of these two learning-based solutions for automatic generation of data types for citation Web sources and presented a quantitative evaluation of these two solutions.

  8. A Greenhouse-Gas Information System: Monitoring and Validating Emissions Reporting and Mitigation

    SciTech Connect (OSTI)

    Jonietz, Karl K. [Los Alamos National Laboratory; Dimotakis, Paul E. [JPL/CAL Tech; Rotman, Douglas A. [Lawrence Livermore National Laboratory; Walker, Bruce C. [Sandia National Laboratory

    2011-09-26T23:59:59.000Z

    This study and report focus on attributes of a greenhouse-gas information system (GHGIS) needed to support MRV&V needs. These needs set the function of such a system apart from scientific/research monitoring of GHGs and carbon-cycle systems, and include (not exclusively): the need for a GHGIS that is operational, as required for decision-support; the need for a system that meets specifications derived from imposed requirements; the need for rigorous calibration, verification, and validation (CV&V) standards, processes, and records for all measurement and modeling/data-inversion data; the need to develop and adopt an uncertainty-quantification (UQ) regimen for all measurement and modeling data; and the requirement that GHGIS products can be subjected to third-party questioning and scientific scrutiny. This report examines and assesses presently available capabilities that could contribute to a future GHGIS. These capabilities include sensors and measurement technologies; data analysis and data uncertainty quantification (UQ) practices and methods; and model-based data-inversion practices, methods, and their associated UQ. The report further examines the need for traceable calibration, verification, and validation processes and attached metadata; differences between present science-/research-oriented needs and those that would be required for an operational GHGIS; the development, operation, and maintenance of a GHGIS missions-operations center (GMOC); and the complex systems engineering and integration that would be required to develop, operate, and evolve a future GHGIS.

  9. National Computational Infrastructure for Lattice Gauge Theory SciDAC-2 Closeout Report

    SciTech Connect (OSTI)

    Sun, Xian-He

    2013-08-01T23:59:59.000Z

    As part of this project work, researchers from Vanderbilt University, Fermi National Laboratory and Illinois Institute of technology developed a real-time cluster fault-tolerant cluster monitoring framework. This framework is open source and is available for download upon request. This work has also been used at Fermi Laboratory, Vanderbilt University and Mississippi State University across projects other than LQCD. The goal for the scientific workflow project is to investigate and develop domain-specific workflow tools for LQCD to help effectively orchestrate, in parallel, computational campaigns consisting of many loosely-coupled batch processing jobs. Major requirements for an LQCD workflow system include: a system to manage input metadata, e.g. physics parameters such as masses, a system to manage and permit the reuse of templates describing workflows, a system to capture data provenance information, a systems to manage produced data, a means of monitoring workflow progress and status, a means of resuming or extending a stopped workflow, fault tolerance features to enhance the reliability of running workflows. Requirements for an LQCD workflow system are available in documentation.

  10. Toward a Data Scalable Solution for Facilitating Discovery of Science Resources

    SciTech Connect (OSTI)

    Weaver, Jesse R.; Castellana, Vito G.; Morari, Alessandro; Tumeo, Antonino; Purohit, Sumit; Chappell, Alan R.; Haglin, David J.; Villa, Oreste; Choudhury, Sutanay; Schuchardt, Karen L.; Feo, John T.

    2014-12-31T23:59:59.000Z

    Science is increasingly motivated by the need to process larger quantities of data. It is facing severe challenges in data collection, management, and processing, so much so that the computational demands of “data scaling” are competing with, and in many fields surpassing, the traditional objective of decreasing processing time. Example domains with large datasets include astronomy, biology, genomics, climate/weather, and material sciences. This paper presents a real-world use case in which we wish to answer queries pro- vided by domain scientists in order to facilitate discovery of relevant science resources. The problem is that the metadata for these science resources is very large and is growing quickly, rapidly increasing the need for a data scaling solution. We propose a system – SGEM – designed for answering graph-based queries over large datasets on cluster architectures, and we re- port performance results for queries on the current RDESC dataset of nearly 1.4 billion triples, and on the well-known BSBM SPARQL query benchmark.

  11. Knowledge Integration to Make Decisions About Complex Systems: Sustainability of Energy Production from Agriculture

    SciTech Connect (OSTI)

    Danuso, Francesco (University of Udine) [University of Udine

    2008-06-18T23:59:59.000Z

    A major bottleneck for improving the governance of complex systems, rely on our ability to integrate different forms of knowledge into a decision support system (DSS). Preliminary aspects are the classification of different types of knowledge (a priori or general, a posteriori or specific, with uncertainty, numerical, textual, algorithmic, complete/incomplete, etc.), the definition of ontologies for knowledge management and the availability of proper tools like continuous simulation models, event driven models, statistical approaches, computational methods (neural networks, evolutionary optimization, rule based systems etc.) and procedure for textual documentation. Following these views at University of Udine, a computer language (SEMoLa, Simple, Easy Modelling Language) for knowledge integration has been developed. SEMoLa can handle models, data, metadata and textual knowledge; it implements and extends the system dynamics ontology (Forrester, 1968; Joergensen, 1994) in which systems are modeled by the concepts of material, group, state, rate, parameter, internal and external events and driving variables. As an example, a SEMoLa model to improve management and sustainability (economical, energetic, environmental) of the agricultural farms is presented. The model (X-Farm) simulates a farm in which cereal and forage yield, oil seeds, milk, calves and wastes can be sold or reused. X-Farm is composed by integrated modules describing fields (crop and soil), feeds and materials storage, machinery management, manpower management, animal husbandry, economic and energetic balances, seed oil extraction, manure and wastes management, biogas production from animal wastes and biomasses.

  12. PyDecay/GraphPhys: A Unified Language and Storage System for Particle Decay Process Descriptions

    SciTech Connect (OSTI)

    Dunietz, Jesse N.; /MIT /SLAC

    2011-06-22T23:59:59.000Z

    To ease the tasks of Monte Carlo (MC) simulation and event reconstruction (i.e. inferring particle-decay events from experimental data) for long-term BaBar data preservation and analysis, the following software components have been designed: a language ('GraphPhys') for specifying decay processes, common to both simulation and data analysis, allowing arbitrary parameters on particles, decays, and entire processes; an automated visualization tool to show graphically what decays have been specified; and a searchable database storage mechanism for decay specifications. Unlike HepML, a proposed XML standard for HEP metadata, the specification language is designed not for data interchange between computer systems, but rather for direct manipulation by human beings as well as computers. The components are interoperable: the information parsed from files in the specification language can easily be rendered as an image by the visualization package, and conversion between decay representations was implemented. Several proof-of-concept command-line tools were built based on this framework. Applications include building easier and more efficient interfaces to existing analysis tools for current projects (e.g. BaBar/BESII), providing a framework for analyses in future experimental settings (e.g. LHC/SuperB), and outreach programs that involve giving students access to BaBar data and analysis tools to give them a hands-on feel for scientific analysis.

  13. Knowledge Integration to Make Decisions About Complex Systems: Sustainability of Energy Production from Agriculture

    SciTech Connect (OSTI)

    Francesco Danuso

    2008-06-18T23:59:59.000Z

    A major bottleneck for improving the governance of complex systems, rely on our ability to integrate different forms of knowledge into a decision support system (DSS). Preliminary aspects are the classification of different types of knowledge (a priori or general, a posteriori or specific, with uncertainty, numerical, textual, algorithmic, complete/incomplete, etc.), the definition of ontologies for knowledge management and the availability of proper tools like continuous simulation models, event driven models, statistical approaches, computational methods (neural networks, evolutionary optimization, rule based systems etc.) and procedure for textual documentation. Following these views at University of Udine, a computer language (SEMoLa, Simple, Easy Modelling Language) for knowledge integration has been developed.  SEMoLa can handle models, data, metadata and textual knowledge; it implements and extends the system dynamics ontology (Forrester, 1968; Jřrgensen, 1994) in which systems are modelled by the concepts of material, group, state, rate, parameter, internal and external events and driving variables. As an example, a SEMoLa model to improve management and sustainability (economical, energetic, environmental) of the agricultural farms is presented. The model (X-Farm) simulates a farm in which cereal and forage yield, oil seeds, milk, calves and wastes can be sold or reused. X-Farm is composed by integrated modules describing fields (crop and soil), feeds and materials storage, machinery management, manpower  management, animal husbandry, economic and energetic balances, seed oil extraction, manure and wastes management, biogas production from animal wastes and biomasses.

  14. EPICS Version 4 - Implementing Complex Data Types

    SciTech Connect (OSTI)

    Marty Kraimer,; John dalesio

    2012-11-27T23:59:59.000Z

    Through phase 1 and phase 2 SBIR grants, s fully functional I/O Controller and communication protocol for version 4 of EPICS is completed. This new software architecture provides a flexible and extendible architecture. Version 4 is implemented fully in Java. The performance metrics look promising. The final portion of phase 2 is to optimize the communication mechanisms. Subsequent work on different aspects of this are required to provide a viable solutions in various areas. Version 3 of EPICS is able to provide a platform for implementing channel based control, because the channel and attributes for time stamping, alarm, display and control were narrow, well defined, and complete. To extend EPICS functionality beyond this, it is necessary to define attributes needed for archive data, array, image data, and directory services. The proper handling of several array types enables the development of middle layer servers such as orbit and bump control in accelerators. Phase 1 should produce a well defined, reviewed, and agreed upon definition of the metadata required for these services. A Phase 2 grant would provide tools that implemented archiving, general array, imaging, and directory applications.

  15. SciDAC - Center for Simulation of Wave Interactions with MHD -- General Atomics Support of ORNL Collaboration

    SciTech Connect (OSTI)

    Abla, G

    2012-11-09T23:59:59.000Z

    The Center for Simulation of Wave Interactions with Magnetohydrodynamics (SWIM) project is dedicated to conduct research on integrated multi-physics simulations. The Integrated Plasma Simulator (IPS) is a framework that was created by the SWIM team. It provides an integration infrastructure for loosely coupled component-based simulations by facilitating services for code execution coordination, computational resource management, data management, and inter-component communication. The IPS framework features improving resource utilization, implementing application-level fault tolerance, and support of the concurrent �multi-tasking� execution model. The General Atomics (GA) team worked closely with other team members on this contract, and conducted research in the areas of computational code monitoring, meta-data management, interactive visualization, and user interfaces. The original website to monitor SWIM activity was developed in the beginning of the project. Due to the amended requirements, the software was redesigned and a revision of the website was deployed into production in April of 2010. Throughout the duration of this project, the SWIM Monitoring Portal (http://swim.gat.com:8080/) has been a critical production tool for supporting the project�s physics goals.

  16. A Comparison of Techniques for Detecting Abnormal Change in Blogs

    SciTech Connect (OSTI)

    Furuta, Dr. Richard Keith [Texas A& M University] [Texas A& M University; ShipmanIII, Dr. Frank Major [Texas A& M University] [Texas A& M University; Bogen, Paul Logasa [ORNL] [ORNL

    2012-01-01T23:59:59.000Z

    Distributed collections are made of metadata entries that contain references to artifacts not controlled by the collection curators. These collections often have limited forms of change; for digital distributed collections, primarily creation and deletion of additional resources. However, there exists a class of digital collection that undergoes additional kinds of change. These collections consist of resources that are distributed across the Internet and brought together via hyperlinking. Resources in these collections can be expected to change as time goes on. Part of the difficulty in maintaining these collections is determining if a changed page is still a valid member of the collection. Others have tried to address this by defining a maximum allowed threshold of change, however, these methods treat change as a potential problem and treat web content as static despite its intrinsic dynamicism. Instead we acknowledge change on the web as a normal part of a web document and determine the difference between what a maintainer expects a page to do and what it actually does. In this work we evaluate options for extractors and analyzers from a suite of techniques against a human-generated ground-truth set of blog changes. The results of this work show a statistically significant improvement over traditional threshold techniques for our collection.

  17. Operation of the Australian Store.Synchrotron for macromolecular crystallography

    SciTech Connect (OSTI)

    Meyer, Grischa R. [Monash University, Clayton, Victoria 3800 (Australia); Aragăo, David; Mudie, Nathan J.; Caradoc-Davies, Tom T. [Australian Synchrotron, 800 Blackburn Road, Clayton, Victoria 3168 (Australia); McGowan, Sheena; Bertling, Philip J.; Groenewegen, David; Quenette, Stevan M. [Monash University, Clayton, Victoria 3800 (Australia); Bond, Charles S. [The University of Western Australia, 35 Stirling Highway, Crawley 6009, Western Australia (Australia); Buckle, Ashley M. [Monash University, Clayton, Victoria 3800 (Australia); Androulakis, Steve, E-mail: steve.androulakis@monash.edu [Monash Bioinformatics Platform, Monash University, Clayton, Victoria 3800 (Australia)

    2014-10-01T23:59:59.000Z

    The Store.Synchrotron service, a fully functional, cloud computing-based solution to raw X-ray data archiving and dissemination at the Australian Synchrotron, is described. The Store.Synchrotron service, a fully functional, cloud computing-based solution to raw X-ray data archiving and dissemination at the Australian Synchrotron, is described. The service automatically receives and archives raw diffraction data, related metadata and preliminary results of automated data-processing workflows. Data are able to be shared with collaborators and opened to the public. In the nine months since its deployment in August 2013, the service has handled over 22.4 TB of raw data (?1.7 million diffraction images). Several real examples from the Australian crystallographic community are described that illustrate the advantages of the approach, which include real-time online data access and fully redundant, secure storage. Discoveries in biological sciences increasingly require multidisciplinary approaches. With this in mind, Store.Synchrotron has been developed as a component within a greater service that can combine data from other instruments at the Australian Synchrotron, as well as instruments at the Australian neutron source ANSTO. It is therefore envisaged that this will serve as a model implementation of raw data archiving and dissemination within the structural biology research community.

  18. Transforming Epidemiology for 21st Century Medicine and Public Health

    SciTech Connect (OSTI)

    Khoury, Muin J [National Institutes of Health; Lam, Tram Kim [National Institutes of Health; Ioannidis, John [Stanford University; Hartge, Patricia [National Institutes of Health; Spitz, Margaret R. [Baylor College of Medicine, Huston; Buring, Julie E. [Brigham and Women's Hospital; Chanock, Stephen J. [National Institutes of Health; Tourassi, Georgia [ORNL; Zauber, Ann [Memorial Sloan-Kettering Cancer Center; Schully, Sheri D [National Institutes of Health

    2013-01-01T23:59:59.000Z

    n 2012, the National Cancer Institute (NCI) engaged the scientific community to provide a vision for cancer epidemiology in the 21st century. Eight overarching thematic recommendations, with proposed corresponding actions for consideration by funding agencies, professional societies, and the research community emerged from the collective intellectual discourse. The themes are (i) extending the reach of epidemiology beyond discovery and etiologic research to include multilevel analysis, intervention evaluation, implementation, and outcomes research; (ii) transforming the practice of epidemiology by moving toward more access and sharing of protocols, data, metadata, and specimens to foster collaboration, to ensure reproducibility and replication, and accelerate translation; (iii) expanding cohort studies to collect exposure, clinical, and other information across the life course and examining multiple health-related endpoints; (iv) developing and validating reliable methods and technologies to quantify exposures and outcomes on a massive scale, and to assess concomitantly the role of multiple factors in complex diseases; (v) integrating big data science into the practice of epidemiology; (vi) expanding knowledge integration to drive research, policy, and practice; (vii) transforming training of 21st century epidemiologists to address interdisciplinary and translational research; and (viii) optimizing the use of resources and infrastructure for epidemiologic studies. These recommendations can transform cancer epidemiology and the field of epidemiology, in general, by enhancing transparency, interdisciplinary collaboration, and strategic applications of new technologies. They should lay a strong scientific foundation for accelerated translation of scientific discoveries into individual and population health benefits.

  19. Japanese Virtual Observatory (JVO) as an advanced astronomical research enviroment

    E-Print Network [OSTI]

    Y. Shirasaki; M. Tanaka; S. Kawanomoto; S. Honda; M. Ohishi; Y. Mizumoto; N. Yasuda; Y. Masunaga; Y. Ishihara; J. Tsutsumi; H. Nakamoto; Y. Kobayashi; M. Sakamoto

    2006-04-28T23:59:59.000Z

    We present the design and implementation of the Japanese Virtual Observatory (JVO) system. JVO is a portal site to various kinds of astronomical resources distributed all over the world. We have developed five components for constructing the portal: (1) registry, (2) data service, (3) workflow system, (4) data analysis service (5) portal GUI. Registry services are used for publishing and searching data services in the VO, and they are constructed using an OAI-PMH metadata harvesting protocol and a SOAP web service protocol so that VO standard architecture is applied. Data services are developed based on the Astronomical Data Query Language (ADQL) which is an international VO standard and an extension of the standard SQL. The toolkit for building the ADQL-based service is released to the public on the JVO web site. The toolkit also provides the protocol translation from a Simple Image Access Protocol (SIAP) to ADQL protocol, so that both the VO standard service can be constructed using our toolkit. In order to federate the distributed databases and analysis services, we have designed a workflow language which is described in XML and developed execution system of the workflow. We have succeeded to connect to a hundred of data resources of the world as of April 2006. We have applied this system to the study of QSO environment by federating a QSO database, a Subaru Suprim-Cam database, and some analysis services such a SExtractor and HyperZ web services. These experiences are described is this paper.

  20. The CMS integration grid testbed

    SciTech Connect (OSTI)

    Graham, Gregory E.

    2004-08-26T23:59:59.000Z

    The CMS Integration Grid Testbed (IGT) comprises USCMS Tier-1 and Tier-2 hardware at the following sites: the California Institute of Technology, Fermi National Accelerator Laboratory, the University of California at San Diego, and the University of Florida at Gainesville. The IGT runs jobs using the Globus Toolkit with a DAGMan and Condor-G front end. The virtual organization (VO) is managed using VO management scripts from the European Data Grid (EDG). Gridwide monitoring is accomplished using local tools such as Ganglia interfaced into the Globus Metadata Directory Service (MDS) and the agent based Mona Lisa. Domain specific software is packaged and installed using the Distribution After Release (DAR) tool of CMS, while middleware under the auspices of the Virtual Data Toolkit (VDT) is distributed using Pacman. During a continuous two month span in Fall of 2002, over 1 million official CMS GEANT based Monte Carlo events were generated and returned to CERN for analysis while being demonstrated at SC2002. In this paper, we describe the process that led to one of the world's first continuously available, functioning grids.

  1. A Comprehensive Collection of Systems Biology Data Characterizing the Host Response to Viral Infection

    SciTech Connect (OSTI)

    Aevermann, Brian; Pickett, Brett E.; Kumar, Sanjeev; Klem, Edward B.; Agnihothram, Sudhakar; Askovich, Peter S.; Bankhead, Armand; Bolles, Meagan; Carter, Victoria; Chang, Jean H.; Clauss, Therese RW; Dash, Pradyot; Diercks, Alan H.; Eisfeld, Amie J.; Ellis, Amy L.; Fan, Shufang; Ferris, Martin T.; Gralinski, Lisa; Green, Richard; Gritsenko, Marina A.; Hatta, Masato; Heegel, Robert A.; Jacobs, Jon M.; Jeng, Sophia; Josset, Laurence; Kaiser, Shari M.; Kelly, Sarah; Law, Gale L.; Li, Chengjun; Li, Jiangning; Long, Casey; Luna, Maria L.; Matzke, Melissa M.; McDermott, Jason E.; Menachery, Vineet; Metz, Thomas O.; Mitchell, Hugh D.; Monroe, Matthew E.; Navarro, Garnet; Neumann, Gabriele; Podyminogin, Rebecca L.; Purvine, Samuel O.; Rosenberger, Carrie; Sanders, Catherine J.; Schepmoes, Athena A.; Shukla, Anil K.; Sims, Amy; Sova, Pavel; Tam, Vincent C.; Tchitchek, Nicholas; Thomas, Paul G.; Tilton, Susan C.; Totura, Allison L.; Wang, Jing; Webb-Robertson, Bobbie-Jo M.; Wen, Ji; Weiss, Jeffrey M.; Yang, Feng; Yount, Boyd; Zhang, Qibin; Mcweeney, Shannon K.; Smith, Richard D.; Waters, Katrina M.; Kawaoka, Yoshihiro; Baric, Ralph; Aderem, Alan; Katze, Michael G.; Scheuermann, Richard H.

    2014-10-14T23:59:59.000Z

    The Systems Biology for Infectious Diseases Research program was established by the U.S. National Institute of Allergy and Infectious Diseases to investigate host-pathogen interactions at a systems level. This program generated 47 transcriptomic and proteomic datasets from 30 studies that investigate in vivo and in vitro host responses to viral infections. Human pathogens in the Orthomyxoviridae and Coronaviridae families, especially pandemic H1N1 and avian H5N1 influenza A viruses and severe acute respiratory syndrome coronavirus (SARS-CoV), were investigated. Study validation was demonstrated via experimental quality control measures and meta-analysis of independent experiments performed under similar conditions. Primary assay results are archived at the GEO and PeptideAtlas public repositories, while processed statistical results together with standardized metadata are publically available at the Influenza Research Database (www.fludb.org) and the Virus Pathogen Resource (www.viprbrc.org). By comparing data from mutant versus wild-type virus and host strains, RNA versus protein differential expression, and infection with genetically similar strains, these data can be used to further investigate genetic and physiological determinants of host responses to viral infection.

  2. Synchrotron Radiation Circular Dichroism (SRCD) Spectroscopy - An Enhanced Method for Examining Protein Conformations and Protein Interactions

    SciTech Connect (OSTI)

    B Wallace; R Janes

    2011-12-31T23:59:59.000Z

    CD (circular dichroism) spectroscopy is a well-established technique in structural biology. SRCD (synchrotron radiation circular dichroism) spectroscopy extends the utility and applications of conventional CD spectroscopy (using laboratory-based instruments) because the high flux of a synchrotron enables collection of data at lower wavelengths (resulting in higher information content), detection of spectra with higher signal-to-noise levels and measurements in the presence of absorbing components (buffers, salts, lipids and detergents). SRCD spectroscopy can provide important static and dynamic structural information on proteins in solution, including secondary structures of intact proteins and their domains, protein stability, the differences between wild-type and mutant proteins, the identification of natively disordered regions in proteins, and the dynamic processes of protein folding and membrane insertion and the kinetics of enzyme reactions. It has also been used to effectively study protein interactions, including protein-protein complex formation involving either induced-fit or rigid-body mechanisms, and protein-lipid complexes. A new web-based bioinformatics resource, the Protein Circular Dichroism Data Bank (PCDDB), has been created which enables archiving, access and analyses of CD and SRCD spectra and supporting metadata, now making this information publicly available. To summarize, the developing method of SRCD spectroscopy has the potential for playing an important role in new types of studies of protein conformations and their complexes.

  3. Bioenergy Science Center KnowledgeBase

    DOE Data Explorer [Office of Scientific and Technical Information (OSTI)]

    Syed, M. H.; Karpinets, T. V.; Parang, M.; Leuze, M. R.; Park, B. H.; Hyatt, D.; Brown, S. D.; Moulton, S. Galloway, M.D.; Uberbacher, E. C.

    The challenge of converting cellulosic biomass to sugars is the dominant obstacle to cost effective production of biofuels in s capable of significant enough quantities to displace U. S. consumption of fossil transportation fuels. The BioEnergy Science Center (BESC) tackles this challenge of biomass recalcitrance by closely linking (1) plant research to make cell walls easier to deconstruct, and (2) microbial research to develop multi-talented biocatalysts tailor-made to produce biofuels in a single step. [from the 2011 BESC factsheet] The BioEnergy Science Center (BESC) is a multi-institutional, multidisciplinary research (biological, chemical, physical and computational sciences, mathematics and engineering) organization focused on the fundamental understanding and elimination of biomass recalcitrance. The BESC Knowledgebase and its associated tools is a discovery platform for bioenergy research. It consists of a collection of metadata, data, and computational tools for data analysis, integration, comparison and visualization for plants and microbes in the center.The BESC Knowledgebase (KB) and BESC Laboratory Information Management System (LIMS) enable bioenergy researchers to perform systemic research. [http://bobcat.ornl.gov/besc/index.jsp

  4. Identifying Differences in Cultural Behavior in Online Groups

    SciTech Connect (OSTI)

    Gregory, Michelle L.; Engel, David W.; Bell, Eric B.; Mcgrath, Liam R.

    2012-07-23T23:59:59.000Z

    We have developed methods to identify online communities, or groups, using a combination of structural information variables and content information variables from weblog posts and their comments to build a characteristic footprint for groups. We have worked with both explicitly connected groups and 'abstract' groups, in which the connection between individuals is in interest (as determined by content based features) and behavior (metadata based features) as opposed to explicit links. We find that these variables do a good job at identifying groups, placing members within a group, and helping determine the appropriate granularity for group boundaries. The group footprint can then be used to identify differences between the online groups. In the work described here we are interested in determining how an individual's online behavior is influenced by their membership in more than one group. For example, individuals belong to a certain culture; they may belong as well to a demographic group, and other 'chosen' groups such as churches or clubs. There is a plethora of evidence surrounding the culturally sensitive adoption, use, and behavior on the Internet. In this work we begin to investigate how culturally defined internet behaviors may influence behaviors of subgroups. We do this through a series of experiments in which we analyze the interaction between culturally defined behaviors and the behaviors of the subgroups. Our goal is to (a) identify if our features can capture cultural distinctions in internet use, and (b) determine what kinds of interaction there are between levels and types of groups.

  5. LAGOVirtual: A Collaborative Environment for the Large Aperture GRB Observatory

    E-Print Network [OSTI]

    Camacho, R; Diaz, G; Guada, C; Hamar, V; Hoeger, H; Melfo, A; Nunez, L A; Perez, Y; Quintero, C; Rosales, M; Torrens, R

    2009-01-01T23:59:59.000Z

    We present the LAGOVirtual Project: an ongoing project to develop platform to collaborate in the Large Aperture GRB Observatory (LAGO). This continental-wide observatory is devised to detect high energy (around 100 GeV) component of Gamma Ray Bursts, by using the single particle technique in arrays of Water Cherenkov Detectors (WCD) at high mountain sites (Chacaltaya, Bolivia, 5300 m a.s.l., Pico Espejo, Venezuela, 4750 m a.s.l., Sierra Negra, Mexico, 4650 m a.s.l). This platform will allow LAGO collaboration to share data, and computer resources through its different sites. This environment has the possibility to generate synthetic data by simulating the showers through AIRES application and to store/preserve distributed data files collected by the WCD at the LAGO sites. The present article concerns the implementation of a prototype of LAGO-DR adapting DSpace, with a hierarchical structure (i.e. country, institution, followed by collections that contain the metadata and data files), for the captured/simulate...

  6. OES-IA Annex IV: Environmental Effects of Marine and Hydrokinetic Devices - Report from the Experts’ Workshop September 27th – 28th 2010 Clontarf Castle, Dublin Ireland

    SciTech Connect (OSTI)

    Copping, Andrea E.; O'Toole, Michael J.

    2010-12-02T23:59:59.000Z

    An experts' workshop was convened in Dublin Ireland September 27th – 28th 2010 in support of IEA Ocean Energy Systems Implementing Agreement Annex IV. PNNL was responsible for organizing the content of the workshop, overseeing the contractors (Irish Marine Institute) hosting the event, presenting material on Annex IV and materials applicable to the workshop intent. PNNL is also overseeing a contractor (Wave Energy Center/University of Plymouth – WEC/UP) in the collection and analysis of the Annex IV data. Fifty-eight experts from 8 countries attended the workshop by invitation, spending two days discussing the needs of Annex IV. Presentations by DOE (background on Annex IV), PNNL (process for developing Annex IV; presentation of the draft database for PNNL project, plans for incorporating Annex IV data), WEC/UP on the environmental effect matrix, and four MHK developers (two from the UK, one from Ireland and one from Sweden; each discussing their own projects and lessons learned for measuring and mitigating environmental effects, as well as interactions with consenting [permitting] processes) helped provide background. The workshop participants worked part of the time in the large group and most of the time in four smaller breakout groups. Participants engaged in the process and provided a wealth of examples of MHK environmental work, particularly in the European nations. They provided practical and actionable advice on the following: • Developing the Annex IV database, with specific uses and audiences • Strong consensus that we should collect detailed metadata on available data sets, rather than attempting to draw in copious datasets. The participants felt there would then be an opportunity to then ask for specific set of data as needed, with specific uses and ownership of the data specified. This is particularly important as many data collected, particularly in Europe but also in Canada, are proprietary; developers were not comfortable with the idea of handing over all their environmental effects data, but all said they would entertain the request if they specifics were clear. • The recommendation was to collect metadata via an online interactive form, taking no more than one hour to complete. • Although the idea of cases representing the “best practices” was recognized as useful, the participants pointed out that there are currently so few MHK projects in the water, that any and all projects were appropriate to highlight as “cases”. There was also discomfort at the implication that “best practices” implied “lesser practices”; this being unhelpful to a new and emerging industry. • Workshop participants were asked if they were willing to continue to engage in the Annex IV process; all expressed willingness. The workshop was successful in adequately addressing its objectives and through participation and interaction in the breakout sessions around the various topics. As a result of the workshop, many delegates are now better informed and have a greater understanding of the potential environmental effects of MHK devices on the marine environment. There is now a greater sense of understanding of the issues involved and consensus by those regulators, developers and scientists who attended the workshop. A strong network has also been built over the two days between European and US/Canadian technical experts in wave and tidal energy.

  7. Expansion and Enhacement of the Wyoming Coalbed Methane Clearinghouse Website to the Wyoming Energy Resources Information Clearinghouse.

    SciTech Connect (OSTI)

    Hulme, Diana; Hamerlinck, Jeffrey; Bergman, Harold; Oakleaf, Jim

    2010-03-26T23:59:59.000Z

    Energy development is expanding across the United States, particularly in western states like Wyoming. Federal and state land management agencies, local governments, industry and non-governmental organizations have realized the need to access spatially-referenced data and other non-spatial information to determine the geographical extent and cumulative impacts of expanding energy development. The Wyoming Energy Resources Information Clearinghouse (WERIC) is a web-based portal which centralizes access to news, data, maps, reports and other information related to the development, management and conservation of Wyomingâ??s diverse energy resources. WERIC was established in 2006 by the University of Wyomingâ??s Ruckelshaus Institute of Environment and Natural Resources (ENR) and the Wyoming Geographic Information Science Center (WyGISC) with funding from the US Department of Energy (DOE) and the US Bureau of Land Management (BLM). The WERIC web portal originated in concept from a more specifically focused website, the Coalbed Methane (CBM) Clearinghouse. The CBM Clearinghouse effort focused only on coalbed methane production within the Powder River Basin of northeast Wyoming. The CBM Clearinghouse demonstrated a need to expand the effort statewide with a comprehensive energy focus, including fossil fuels and renewable and alternative energy resources produced and/or developed in Wyoming. WERIC serves spatial data to the greater Wyoming geospatial community through the Wyoming GeoLibrary, the WyGISC Data Server and the Wyoming Energy Map. These applications are critical components that support the Wyoming Energy Resources Information Clearinghouse (WERIC). The Wyoming GeoLibrary is a tool for searching and browsing a central repository for metadata. It provides the ability to publish and maintain metadata and geospatial data in a distributed environment. The WyGISC Data Server is an internet mapping application that provides traditional GIS mapping and analysis functionality via the web. It is linked into various state and federal agency spatial data servers allowing users to visualize multiple themes, such as well locations and core sage grouse areas, in one domain. Additionally, this application gives users the ability to download any of the data being displayed within the web map. The Wyoming Energy Map is the newest mapping application developed directly from this effort. With over a 100 different layers accessible via this mapping application, it is the most comprehensive Wyoming energy mapping application available. This application also provides the public with the ability to create cultural and wildlife reports based on any location throughout Wyoming and at multiple scales. The WERIC website also allows users to access links to federal, state, and local natural resource agency websites and map servers; research documents about energy; and educational information, including information on upcoming energy-relate conferences. The WERIC website has seen significant use by energy industry consultants, land management agencies, state and local decision-makers, non-governmental organizations and the public. Continued service to these sectors is desirable but some challenges remain in keeping the WERIC site viable. The most pressing issue is finding the human and financial resources to keep the site continually updated. Initially, the concept included offering users the ability to maintain the site themselves; however, this has proven not to be a viable option since very few people contributed. Without user contributions, the web page relied on already committed university staff to publish and link to the appropriate documents and web-pages. An option that is currently being explored to address this issue is development of a partnership with the University of Wyoming, School of Energy Resources (SER). As part of their outreach program, SER may be able to contribute funding for a full-time position dedicated to maintenance of WERIC.

  8. NA-42 TI Shared Software Component Library FY2011 Final Report

    SciTech Connect (OSTI)

    Knudson, Christa K.; Rutz, Frederick C.; Dorow, Kevin E.

    2011-07-21T23:59:59.000Z

    The NA-42 TI program initiated an effort in FY2010 to standardize its software development efforts with the long term goal of migrating toward a software management approach that will allow for the sharing and reuse of code developed within the TI program, improve integration, ensure a level of software documentation, and reduce development costs. The Pacific Northwest National Laboratory (PNNL) has been tasked with two activities that support this mission. PNNL has been tasked with the identification, selection, and implementation of a Shared Software Component Library. The intent of the library is to provide a common repository that is accessible by all authorized NA-42 software development teams. The repository facilitates software reuse through a searchable and easy to use web based interface. As software is submitted to the repository, the component registration process captures meta-data and provides version control for compiled libraries, documentation, and source code. This meta-data is then available for retrieval and review as part of library search results. In FY2010, PNNL and staff from the Remote Sensing Laboratory (RSL) teamed up to develop a software application with the goal of replacing the aging Aerial Measuring System (AMS). The application under development includes an Advanced Visualization and Integration of Data (AVID) framework and associated AMS modules. Throughout development, PNNL and RSL have utilized a common AMS code repository for collaborative code development. The AMS repository is hosted by PNNL, is restricted to the project development team, is accessed via two different geographic locations and continues to be used. The knowledge gained from the collaboration and hosting of this repository in conjunction with PNNL software development and systems engineering capabilities were used in the selection of a package to be used in the implementation of the software component library on behalf of NA-42 TI. The second task managed by PNNL is the development and continued maintenance of the NA-42 TI Software Development Questionnaire. This questionnaire is intended to help software development teams working under NA-42 TI in documenting their development activities. When sufficiently completed, the questionnaire illustrates that the software development activities recorded incorporate significant aspects of the software engineering lifecycle. The questionnaire template is updated as comments are received from NA-42 and/or its development teams and revised versions distributed to those using the questionnaire. PNNL also maintains a list of questionnaire recipients. The blank questionnaire template, the AVID and AMS software being developed, and the completed AVID AMS specific questionnaire are being used as the initial content to be established in the TI Component Library. This report summarizes the approach taken to identify requirements, search for and evaluate technologies, and the approach taken for installation of the software needed to host the component library. Additionally, it defines the process by which users request access for the contribution and retrieval of library content.

  9. Daymet: Daily Surface Weather Data on a 1-km Grid for North America, Version 2.

    SciTech Connect (OSTI)

    Thornton, Peter E [ORNL; Thornton, Michele M [ORNL; Mayer, Benjamin W [ORNL; Wilhelmi, Nate [National Center for Atmospheric Research (NCAR); Wei, Yaxing [ORNL; Devarakonda, Ranjeet [ORNL; Cook, Robert B [ORNL

    2014-01-01T23:59:59.000Z

    More information: http://daymet.ornl.gov Presenter: Ranjeet Devarakonda Environmental Sciences Division Oak Ridge National Laboratory (ORNL) Daymet: Daily Surface Weather Data and Climatological Summaries provides gridded estimates of daily weather parameters for North America, including daily continuous surfaces of minimum and maximum temperature, precipitation occurrence and amount, humidity, shortwave radiation, snow water equivalent, and day length. The current data product (Version 2) covers the period January 1, 1980 to December 31, 2013 [1]. The prior product (Version 1) only covered from 1980-2008. Data are available on a daily time step at a 1-km x 1-km spatial resolution in Lambert Conformal Conic projection with a spatial extent that covers the conterminous United States, Mexico, and Southern Canada as meteorological station density allows. Daymet data can be downloaded from 1) the ORNL Distributed Active Archive Center (DAAC) search and order tools (http://daac.ornl.gov/cgi-bin/cart/add2cart.pl?add=1219) or directly from the DAAC FTP site (http://daac.ornl.gov/cgi-bin/dsviewer.pl?ds_id=1219) and 2) the Single Pixel Tool [2] and THREDDS (Thematic Real-time Environmental Data Services) Data Server [3]. The Single Pixel Data Extraction Tool allows users to enter a single geographic point by latitude and longitude in decimal degrees. A routine is executed that translates the (lon, lat) coordinates into projected Daymet (x,y) coordinates. These coordinates are used to access the Daymet database of daily-interpolated surface weather variables. Daily data from the nearest 1 km x 1 km Daymet grid cell are extracted from the database and formatted as a table with one column for each Daymet variable and one row for each day. All daily data for selected years are returned as a single (long) table, formatted for display in the browser window. At the top of this table is a link to the same data in a simple comma-separated text format, suitable for import into a spreadsheet or other data analysis software. The Single Pixel Data Extraction Tool also provides the option to download multiple coordinates programmatically. A multiple extractor script is freely available to download at http://daymet.ornl.gov/files/daymet.zip. The ORNL DAAC s THREDDS data server (TDS) provides customized visualization and access to Daymet time series of North American mosaics. Users can subset and download Daymet data via a variety of community standards, including OPeNDAP, NetCDF Subset service, and Open Geospatial Consortium (OGC) Web Map/Coverage Service. The ORNL DAAC TDS also exposes Daymet metadata through its ncISO service to facilitate harvesting Daymet metadata records into 3rd party catalogs. References: [1] Thornton, P.E., M.M. Thornton, B.W. Mayer, N. Wilhelmi, Y. Wei, R. Devarakonda, and R.B. Cook. 2014. Daymet: Daily Surface Weather Data on a 1-km Grid for North America, Version 2. Data set. Available on-line [http://daac.ornl.gov] from Oak Ridge National Laboratory Distributed Active Archive Center, Oak Ridge, Tennessee, USA. [2] Devarakonda R., et al. 2012. Daymet: Single Pixel Data Extraction Tool. Available on-line [http://daymet.ornl.go/singlepixel.html]. [3] Wei Y., et al. 2014. Daymet: Thematic Real-time Environmental Data Services. Available on-line [http://daymet.ornl.gov/thredds_tiles.html].

  10. The Epsomitic Phototrophic Microbial Mat of Hot Lake, Washington: Community Structural Responses to Seasonal Cycling

    SciTech Connect (OSTI)

    Lindemann, Stephen R.; Moran, James J.; Stegen, James C.; Renslow, Ryan S.; Hutchison, Janine R.; Cole, Jessica K.; Dohnalkova, Alice; Tremblay, Julien; Singh, Kanwar; Malfatti, Stephanie; Chen, Feng; Tringe, Susannah; Beyenal, Haluk; Fredrickson, Jim K.

    2013-11-13T23:59:59.000Z

    Phototrophic microbial mats are compact ecosystems composed of highly interactive organisms in which energy and element cycling take place over millimeter-to-centimeter-scale distances. Although microbial mats are common in hypersaline environments, they have not been extensively characterized in systems dominated by divalent ions. Hot Lake is a meromictic, epsomitic lake that occupies a small, endorheic basin in north-central Washington. The lake harbors a benthic, phototrophic mat that assembles each spring, disassembles each fall, and is subject to greater than tenfold variation in salinity (primarily Mg2+ and SO2?4) and irradiation over the annual cycle. We examined spatiotemporal variation in the mat community at five time points throughout the annual cycle with respect to prevailing physicochemical parameters by amplicon sequencing of the V4 region of the 16S rRNA gene coupled to near-full-length 16S RNA clone sequences. The composition of these microbial communities was relatively stable over the seasonal cycle and included dominant populations of Cyanobacteria, primarily a group IV cyanobacterium (Leptolyngbya), and Alphaproteobacteria (specifically, members of Rhodobacteraceae and Geminicoccus). Members of Gammaproteobacteria (e.g., Thioalkalivibrio and Halochromatium) and Deltaproteobacteria (e.g., Desulfofustis) that are likely to be involved in sulfur cycling peaked in summer and declined significantly by mid-fall, mirroring larger trends in mat community richness and evenness. Phylogenetic turnover analysis of abundant phylotypes employing environmental metadata suggests that seasonal shifts in light variability exert a dominant influence on the composition of Hot Lake microbial mat communities. The seasonal development and organization of these structured microbial mats provide opportunities for analysis of the temporal and physical dynamics that feed back to community function.

  11. Tethys and Annex IV Progress Report for FY 2012

    SciTech Connect (OSTI)

    Hanna, Luke A.; Butner, R. Scott; Whiting, Jonathan M.; Copping, Andrea E.

    2013-09-01T23:59:59.000Z

    The marine and hydrokinetic (MHK) environmental Impacts Knowledge Management System, dubbed “Tethys” after the mythical Greek titaness of the seas, is being developed by the Pacific Northwest National Laboratory (PNNL) to support the U.S. Department of Energy’s Wind and Water Power Program (WWPP). Functioning as a smart database, Tethys enables its users to identify key words or terms to help gather, organize and make available information and data pertaining to the environmental effects of MHK and offshore wind (OSW) energy development. By providing and categorizing relevant publications within a simple and searchable database, Tethys acts as a dissemination channel for information and data which can be utilized by regulators, project developers and researchers to minimize the environmental risks associated with offshore renewable energy developments and attempt to streamline the permitting process. Tethys also houses a separate content-related Annex IV data base with identical functionality to the Tethys knowledge base. Annex IV is a collaborative project among member nations of the International Energy Agency (IEA) Ocean Energy Systems – Implementing Agreement (OES-IA) that examines the environmental effects of ocean energy devices and projects. The U.S. Department of Energy leads the Annex IV working with federal partners such as the Federal Energy Regulatory Commission (FERC), the Bureau of Ocean Energy Management (BOEM), and the National Oceanic Atmospheric Administration (NOAA). While the Annex IV database contains technical reports and journal articles, it is primarily focused on the collection of project site and research study metadata forms (completed by MHK researchers and developers around the world, and collected by PNNL) which provide information on environmental studies and the current progress of the various international MHK developments in the Annex IV member nations. The purpose of this report is to provide a summary of the content, accessibility and functionality enhancements made to the Annex IV and Tethys knowledge bases in FY12.

  12. LLNL Genomic Assessment: Viral and Bacterial Sequencing Needs for TMTI, Task 1.4.2 Report

    SciTech Connect (OSTI)

    Slezak, T; Borucki, M; Lam, M; Lenhoff, R; Vitalis, E

    2010-01-26T23:59:59.000Z

    Good progress has been made on both bacterial and viral sequencing by the TMTI centers. While access to appropriate samples is a limiting factor to throughput, excellent progress has been made with respect to getting agreements in place with key sources of relevant materials. Sharing of sequenced genomes funded by TMTI has been extremely limited to date. The April 2010 exercise should force a resolution to this, but additional managerial pressures may be needed to ensure that rapid sharing of TMTI-funded sequencing occurs, regardless of collaborator constraints concerning ultimate publication(s). Policies to permit TMTI-internal rapid sharing of sequenced genomes should be written into all TMTI agreements with collaborators now being negotiated. TMTI needs to establish a Web-based system for tracking samples destined for sequencing. This includes metadata on sample origins and contributor, information on sample shipment/receipt, prioritization by TMTI, assignment to one or more sequencing centers (including possible TMTI-sponsored sequencing at a contributor site), and status history of the sample sequencing effort. While this system could be a component of the AFRL system, it is not part of any current development effort. Policy and standardized procedures are needed to ensure appropriate verification of all TMTI samples prior to the investment in sequencing. PCR, arrays, and classical biochemical tests are examples of potential verification methods. Verification is needed to detect miss-labeled, degraded, mixed or contaminated samples. Regular QC exercises are needed to ensure that the TMTI-funded centers are meeting all standards for producing quality genomic sequence data.

  13. AUTOMATING GROUNDWATER SAMPLING AT HANFORD THE NEXT STEP

    SciTech Connect (OSTI)

    CONNELL CW; CONLEY SF; HILDEBRAND RD; CUNNINGHAM DE; R_D_Doug_Hildebrand@rl.gov; DeVon_E_Cunningham@rl.gov

    2010-01-21T23:59:59.000Z

    Historically, the groundwater monitoring activities at the Department of Energy's Hanford Site in southeastern Washington State have been very "people intensive." Approximately 1500 wells are sampled each year by field personnel or "samplers." These individuals have been issued pre-printed forms showing information about the well(s) for a particular sampling evolution. This information is taken from 2 official electronic databases: the Hanford Well information System (HWIS) and the Hanford Environmental Information System (HEIS). The samplers used these hardcopy forms to document the groundwater samples and well water-levels. After recording the entries in the field, the samplers turned the forms in at the end of the day and other personnel posted the collected information onto a spreadsheet that was then printed and included in a log book. The log book was then used to make manual entries of the new information into the software application(s) for the HEIS and HWIS databases. A pilot project for automating this extremely tedious process was lauched in 2008. Initially, the automation was focused on water-level measurements. Now, the effort is being extended to automate the meta-data associated with collecting groundwater samples. The project allowed electronic forms produced in the field by samplers to be used in a work flow process where the data is transferred to the database and electronic form is filed in managed records - thus eliminating manually completed forms. Elimating the manual forms and streamlining the data entry not only improved the accuracy of the information recorded, but also enhanced the efficiency and sampling capacity of field office personnel.

  14. SU-E-I-48: The Behavior of AEC in Scan Regions Outside the Localizer Radiograph FOV: An In Phantom Study of CT Systems From Four Vendors

    SciTech Connect (OSTI)

    Supanich, M [Rush University Medical Center, Chicago, IL (United States); Bevins, N [Henry Ford Health System, Detroit, MI (United States)

    2014-06-01T23:59:59.000Z

    Purpose: This review of scanners from 4 major manufacturers examines the clinical impact of performing CT scans that extend into areas of the body that were not acquired in the CT localizer radiograph. Methods: Anthropomorphic chest and abdomen phantoms were positioned together on the tables of CT scanners from 4 different vendors. All of the scanners offered an Automatic Exposure Control (AEC) option with both lateral and axial tube current modulation. A localizer radiograph was taken covering the entire extent of both phantoms and then the scanner's Chest-Abdomen-Pelvis (CAP) study was performed with the clinical AEC settings employed and the scan and reconstruction range extending from the superior portion of the chest phantom through the inferior portion of the abdomen phantom. A new study was then initiated with a localizer radiograph extending the length of the chest phantom (not covering the abdomen phantom). The same CAP protocol and AEC settings were then used to scan and reconstruct the entire length of both phantoms. Scan parameters at specific locations in the abdomen phantom from both studies were investigated using the information contained in the DICOM metadata of the reconstructed images. Results: The AEC systems on all scanners utilized different tube current settings in the abdomen phantom for the scan completed without the full localizer radiograph. The AEC system behavior was also scanner dependent with the default manual tube current, the maximum tube current and the tube current at the last known position observed as outcomes. Conclusion: The behavior of the AEC systems of CT scanners in regions not covered by the localizer radiograph is vendor dependent. To ensure optimal image quality and radiation exposure it is important to include the entire planned scan region in the localizer radiograph.

  15. DEVELOPING THE NATIONAL GEOTHERMAL DATA SYSTEM ADOPTION OF CKAN FOR DOMESTIC & INTERNATIONAL DATA DEPLOYMENT

    SciTech Connect (OSTI)

    Clark, Ryan J. [Arizona Geological Survey; Kuhmuench, Christoph [Siemens Corporation; Richard, Stephen M. [Arizona Geological Survey

    2013-01-01T23:59:59.000Z

    The National Geothermal Data System (NGDS) De- sign and Testing Team is developing NGDS software currently referred to as the “NGDS Node-In-A-Box”. The software targets organizations or individuals who wish to host at least one of the following: • an online repository containing resources for the NGDS; • an online site for creating metadata to register re- sources with the NGDS • NDGS-conformant Web APIs that enable access to NGDS data (e.g., WMS, WFS, WCS); • NDGS-conformant Web APIs that support dis- covery of NGDS resources via catalog service (e.g. CSW) • a web site that supports discovery and under- standing of NGDS resources A number of different frameworks for development of this online application were reviewed. The NGDS Design and Testing Team determined to use CKAN (http://ckan.org/), because it provides the closest match between out of the box functionality and NGDS node-in-a-box requirements. To achieve the NGDS vision and goals, this software development project has been inititated to provide NGDS data consumers with a highly functional inter- face to access the system, and to ease the burden on data providers who wish to publish data in the sys- tem. It is important to note that this software package constitutes a reference implementation. The NGDS software is based on open standards, which means other server software can make resources available, and other client applications can utilize NGDS data. A number of international organizations have ex- pressed interest in the NGDS approach to data access. The CKAN node implementation can provide a sim- ple path for deploying this technology in other set- tings.

  16. A Reflection on the Structure and Process of the Web of Data

    E-Print Network [OSTI]

    Marko A. Rodriguez

    2009-08-04T23:59:59.000Z

    The Web community has introduced a set of standards and technologies for representing, querying, and manipulating a globally distributed data structure known as the Web of Data. The proponents of the Web of Data envision much of the world's data being interrelated and openly accessible to the general public. This vision is analogous in many ways to the Web of Documents of common knowledge, but instead of making documents and media openly accessible, the focus is on making data openly accessible. In providing data for public use, there has been a stimulated interest in a movement dubbed Open Data. Open Data is analogous in many ways to the Open Source movement. However, instead of focusing on software, Open Data is focused on the legal and licensing issues around publicly exposed data. Together, various technological and legal tools are laying the groundwork for the future of global-scale data management on the Web. As of today, in its early form, the Web of Data hosts a variety of data sets that include encyclopedic facts, drug and protein data, metadata on music, books and scholarly articles, social network representations, geospatial information, and many other types of information. The size and diversity of the Web of Data is a demonstration of the flexibility of the underlying standards and the overall feasibility of the project as a whole. The purpose of this article is to provide a review of the technological underpinnings of the Web of Data as well as some of the hurdles that need to be overcome if the Web of Data is to emerge as the defacto medium for data representation, distribution, and ultimately, processing.

  17. National Geoscience Data Repository System, Phase II. Final report, January 30, 1995--January 28, 1997

    SciTech Connect (OSTI)

    NONE

    1998-04-01T23:59:59.000Z

    The American Geological Institute (AGI) has completed Phase II of a project to establish a National Geoscience Data Repository System (NGDRS). The project`s primary objectives are to preserve geoscience data in jeopardy of being destroyed and to make that data available to those who have a need to use it in future investigations. These data are available for donation to the public as a result of the downsizing that has occurred in the major petroleum and mining companies in the United States for the past decade. In recent years, these companies have consolidated domestic operations, sold many of their domestic properties and relinquished many of their leases. The scientific data associated with those properties are no longer considered to be useful assets and are consequently in danger of being lost forever. The national repository project will make many of these data available to the geoscience community for the first time. To address this opportunity, AGI sought support from the Department of Energy (DOE) in 1994 to initiate the NGDRS Phase I feasibility study to determine the types and quantity of data that companies would be willing to donate. The petroleum and mining companies surveyed indicated that they were willing to donate approximately five million well logs, one hundred million miles of seismic reflection data, millions of linear feet of core and cuttings, and a variety of other types of scientific data. Based on the positive results of the Phase I study, AGI undertook Phase II of the program in 1995. Funded jointly by DOE and industry, Phase II encompasses the establishment of standards for indexing and cataloging of geoscience data and determination of the costs of transferring data from the private sector to public-sector data repositories. Pilot projects evaluated the feasibility of the project for transfer of different data types and creation of a Web-based metadata supercatalog and browser.

  18. Tracking Provenance in ORNL's Flexible Research Platforms

    SciTech Connect (OSTI)

    Hensley, Zachary P [ORNL; Sanyal, Jibonananda [ORNL; New, Joshua Ryan [ORNL

    2013-08-01T23:59:59.000Z

    Provenance is dened as information about the origin of objects, a concept that applies to both physical and digital objects and often overlaps both. The use of provenance in systems designed for research is an important but forgotten feature. Provenance allows for proper and exact tracking of information, its use, its lineage, its derivations and other metadata that are important for correctly adhering to the scien- tic method. In our project's prescribed use of provenance, researchers can determine detailed information about the use of sensor data in their experiments on ORNL's Flexible Research Platforms (FRPs). Our project's provenance system, Provenance Data Management System (ProvDMS), tracks information starting with the creation of information by an FRP sensor. The system determines station information, sensor information, and sensor channel information. The system allows researchers to derive generations of experiments from the sensor data and tracks their hierarchical flow. Key points can be seen in the history of the information as part of the information's workflow. The concept of provenance and its usage in science is relatively new and while used in other cases around the world, our project's provenance diers in a key area. To keep track of provenance, most systems must be designed or redesigned around the new provenance system. Our system is designed as a cohesive but sepa- rate entity and allows for researchers to continue using their own methods of analysis without being constrained in their ways in order to track the provenance. We have designed ProvDMS using a lightweight provenance library, Core Provenance Library (CPL) v.6 In addition to keeping track of sensor data experiments and its provenance, ProvDMS also provides a web-enabled visualization of the inheritance.

  19. Geographic Information System (GIS) Emergency Support for the May 2000 Cerro Grande Wildfire, Los Alamos, New Mexico, USA

    SciTech Connect (OSTI)

    C.R.Mynard; G.N.Keating; P.M.Rich; D.R. Bleakly

    2003-05-01T23:59:59.000Z

    In May 2000 the Cerro Grande wildfire swept through Los Alamos, New Mexico, burning approximately 17,400 ha (43,000 acres) and causing evacuation of Los Alamos National Laboratory (LANL) and the communities of Los Alamos and White Rock. An integral part of emergency response during the fire was the use of geographic information system (GIS) technology, which continues to be used in support of post-fire restoration and environmental monitoring. During the fire Laboratory GIS staff and volunteers from other organizations worked to produce maps and provide support for emergency managers, including at an emergency GIS facility in Santa Fe. Subsequent to the fire, Laboratory GIS teams supported the multiagency Burned Area Emergency Rehabilitation (BAER) team to provide GIS data and maps for planning mitigation efforts. The GIS teams continue to help researchers, operations personnel, and managers deal with the tremendous changes caused by the fire. Much of the work is under the auspices of the Cerro Grande Rehabilitation Project (CGRP) to promote recovery from fire damage, improve information exchange, enhance emergency management, and conduct mitigation activities. GIS efforts during the fire provided important lessons about institutional matters, working relationships, and emergency preparedness. These lessons include the importance of (1) an integrated framework for assessing natural and human hazards in a landscape context; (2) a strong GIS capability for emergency response; (3) coordinated emergency plans for GIS operations; (4) a method for employees to report their whereabouts and receive authoritative information during an evacuation; (5) GIS data that are complete, backed-up, and available during an emergency; (6) adaptation of GIS to the circumstances of the emergency; (7) better coordination in the GIS community; (8) better integration of GIS into LANL operations; and (9) a central data warehouse for data and metadata. These lessons are important for planning future directions of GIS at LANL. Growing maturity of GIS is expected to lead to standardization and a better-integrated, more-coordinated approach to data sharing and emergency management at LANL, and within DOE, in accord with the federal government's increasing focus on electronic communication for its organizational and public interactions.

  20. Advanced I/O for large-scale scientific applications.

    SciTech Connect (OSTI)

    Klasky, Scott (Oak Ridge National Laboratory, Oak Ridge, TN); Schwan, Karsten (Georgia Institute of Technology, Atlanta, GA); Oldfield, Ron A.; Lofstead, Gerald F., II (Georgia Institute of Technology, Atlanta, GA)

    2010-01-01T23:59:59.000Z

    As scientific simulations scale to use petascale machines and beyond, the data volumes generated pose a dual problem. First, with increasing machine sizes, the careful tuning of IO routines becomes more and more important to keep the time spent in IO acceptable. It is not uncommon, for instance, to have 20% of an application's runtime spent performing IO in a 'tuned' system. Careful management of the IO routines can move that to 5% or even less in some cases. Second, the data volumes are so large, on the order of 10s to 100s of TB, that trying to discover the scientifically valid contributions requires assistance at runtime to both organize and annotate the data. Waiting for offline processing is not feasible due both to the impact on the IO system and the time required. To reduce this load and improve the ability of scientists to use the large amounts of data being produced, new techniques for data management are required. First, there is a need for techniques for efficient movement of data from the compute space to storage. These techniques should understand the underlying system infrastructure and adapt to changing system conditions. Technologies include aggregation networks, data staging nodes for a closer parity to the IO subsystem, and autonomic IO routines that can detect system bottlenecks and choose different approaches, such as splitting the output into multiple targets, staggering output processes. Such methods must be end-to-end, meaning that even with properly managed asynchronous techniques, it is still essential to properly manage the later synchronous interaction with the storage system to maintain acceptable performance. Second, for the data being generated, annotations and other metadata must be incorporated to help the scientist understand output data for the simulation run as a whole, to select data and data features without concern for what files or other storage technologies were employed. All of these features should be attained while maintaining a simple deployment for the science code and eliminating the need for allocation of additional computational resources.

  1. NATIONAL GEOTHERMAL DATA SYSTEM: AN EXEMPLAR OF OPEN ACCESS TO DATA

    SciTech Connect (OSTI)

    Blackman, Harold [Arizona Geological Survey; Blackman, Harold M. [Arizona Geological Survey; Blackman, Harold M. [Arizona Geological Survey; Blackman, Harold [Boise State University; Blackman, Harold [United States Department of Energy; Blackman, Harold

    2013-10-01T23:59:59.000Z

    The formal launch of National Geothermal Data System (NGDS – www.geothermaldata.org) in 2014 will provide open access to technical geothermal-relevant data from all of the Department of Energy- sponsored geothermal development and research projects and geologic data from all 50 states. By making data easily discoverable and accessible this system will open new exploration opportunities and shorten project development. The prototype data system currently includes multiple data nodes, and nationwide data online and available to the public, indexed through a single catalog under construction at http://search.geothermaldata.org. Data from state geological surveys and partners includes more than 5 million records online, including 1.48 million well headers (oil and gas, water, geothermal), 732,000 well logs, and 314,000 borehole temperatures and is growing rapidly. There are over 250 Web services and another 138 WMS (Web Map Services) registered in the system as of August, 2013. Additional data record is being added by companion projects run by Boise State University, Southern Methodist University, and the USGS. The National Renewable Energy Laboratory is managing the Geothermal Data Repository, an NGDS node that will be a clearinghouse for data from hundreds of DOE-funded geothermal projects. NGDS is built on the US Geoscience Information Network (USGIN) data integration framework, which is a joint undertaking of the USGS and the Association of American State Geologists (AASG). NGDS is fully compliant with the White House Executive Order of May 2013, requiring all federal agencies to make their data holdings publicly accessible online in open source, interoperable formats with common core and extensible metadata. The National Geothermal Data System is being designed, built, deployed, and populated primarily with grants from the US Department of Energy, Geothermal Technologies Office. To keep this operational system sustainable after the original implementation will require four core elements: continued serving of data and applications by providers; maintenance of system operations; a governance structure; and an effective business model. Each of these presents a number of challenges currently under consideration.

  2. National Computational Infrastructure for LatticeGauge Theory SciDAC-2 Closeout Report

    SciTech Connect (OSTI)

    Bapty, Theodore; Dubey, Abhishek

    2013-07-18T23:59:59.000Z

    As part of the reliability project work, researchers from Vanderbilt University, Fermi National Laboratory and Illinois Institute of technology developed a real-time cluster fault-tolerant cluster monitoring framework. The goal for the scientific workflow project is to investigate and develop domain-specific workflow tools for LQCD to help effectively orchestrate, in parallel, computational campaigns consisting of many loosely-coupled batch processing jobs. Major requirements for an LQCD workflow system include: a system to manage input metadata, e.g. physics parameters such as masses, a system to manage and permit the reuse of templates describing workflows, a system to capture data provenance information, a systems to manage produced data, a means of monitoring workflow progress and status, a means of resuming or extending a stopped workflow, fault tolerance features to enhance the reliability of running workflows. In summary, these achievements are reported: • Implemented a software system to manage parameters. This includes a parameter set language based on a superset of the JSON data-interchange format, parsers in multiple languages (C++, Python, Ruby), and a web-based interface tool. It also includes a templating system that can produce input text for LQCD applications like MILC. • Implemented a monitoring sensor framework in software that is in production on the Fermilab USQCD facility. This includes equipment health, process accounting, MPI/QMP process tracking, and batch system (Torque) job monitoring. All sensor data are available from databases, and various query tools can be used to extract common data patterns and perform ad hoc searches. Common batch system queries such as job status are available in command line tools and are used in actual workflow-based production by a subset of Fermilab users. • Developed a formal state machine model for scientific workflow and reliability systems. This includes the use of Vanderbilt’s Generic Modeling Envirnment (GME) tool for code generation for the production of user APIs, code stubs, testing harnesses, and model correctness verification. It is used for creating wrappers around LQCD applications so that they can be integrated into existing workflow systems such as Kepler. • Implemented a database system for tracking the state of nodes and jobs managed by the Torque batch systems used at Fermilab. This robust system and various canned queuries are used for many tasks, including monitoring the health of the clusters, managing allocated projects, producing accounting reports, and troubleshooting nodes and jobs.

  3. Strategic Plan for Nuclear Energy -- Knowledge Base for Advanced Modeling and Simulation (NE-KAMS)

    SciTech Connect (OSTI)

    Kimberlyn C. Mousseau

    2011-10-01T23:59:59.000Z

    The Nuclear Energy Computational Fluid Dynamics Advanced Modeling and Simulation (NE-CAMS) system is being developed at the Idaho National Laboratory (INL) in collaboration with Bettis Laboratory, Sandia National Laboratory (SNL), Argonne National Laboratory (ANL), Utah State University (USU), and other interested parties with the objective of developing and implementing a comprehensive and readily accessible data and information management system for computational fluid dynamics (CFD) verification and validation (V&V) in support of nuclear energy systems design and safety analysis. The two key objectives of the NE-CAMS effort are to identify, collect, assess, store and maintain high resolution and high quality experimental data and related expert knowledge (metadata) for use in CFD V&V assessments specific to the nuclear energy field and to establish a working relationship with the U.S. Nuclear Regulatory Commission (NRC) to develop a CFD V&V database, including benchmark cases, that addresses and supports the associated NRC regulations and policies on the use of CFD analysis. In particular, the NE-CAMS system will support the Department of Energy Office of Nuclear Energy Advanced Modeling and Simulation (NEAMS) Program, which aims to develop and deploy advanced modeling and simulation methods and computational tools for reliable numerical simulation of nuclear reactor systems for design and safety analysis. Primary NE-CAMS Elements There are four primary elements of the NE-CAMS knowledge base designed to support computer modeling and simulation in the nuclear energy arena as listed below. Element 1. The database will contain experimental data that can be used for CFD validation that is relevant to nuclear reactor and plant processes, particularly those important to the nuclear industry and the NRC. Element 2. Qualification standards for data evaluation and classification will be incorporated and applied such that validation data sets will result in well-defined, well-characterized data. Element 3. Standards will be established for the design and operation of experiments for the generation of new validation data sets that are to be submitted to NE-CAMS that addresses the completeness and characterization of the dataset. Element 4. Standards will be developed for performing verification and validation (V&V) to establish confidence levels in CFD analyses of nuclear reactor processes; such processes will be acceptable and recognized by both CFD experts and the NRC.

  4. National Geothermal Data System: A Geothermal Data System for Exploration and Development

    SciTech Connect (OSTI)

    Allison, Lee [Executive Office of the State of Arizona (Arizona Geological Survey); Richard, Stephen [Executive Office of the State of Arizona (Arizona Geological Survey); Patten, Kim [Executive Office of the State of Arizona (Arizona Geological Survey); Love, Diane [Executive Office of the State of Arizona (Arizona Geological Survey); Coleman, Celia [Executive Office of the State of Arizona (Arizona Geological Survey); Chen, Genhan [Executive Office of the State of Arizona (Arizona Geological Survey)

    2012-09-30T23:59:59.000Z

    Geothermal-relevant geosciences data from all 50 states (www.stategeothermaldata.org), federal agencies, national labs, and academic centers are being digitized and linked in a distributed online network funded by the U.S. Department of Energy Geothermal Data System (GDS) to foster geothermal energy exploration and development through use of interactive online ‘mashups,’data integration, and applications. Emphasis is first to make as much information as possible accessible online, with a long range goal to make data interoperable through standardized services and interchange formats. A growing set of more than thirty geoscience data content models is in use or under development to define standardized interchange formats for: aqueous chemistry, borehole temperature data, direct use feature, drill stem test, seismic event hypocenter, fault feature, geologic contact feature, geologic unit feature, thermal/hot spring description, metadata, quaternary fault, volcanic vent description, well header feature, borehole lithology log, crustal stress, gravity, heat flow/temperature gradient, permeability, and feature description data like developed geothermal systems, geologic unit geothermal characterization, permeability, production data, rock alteration description, rock chemistry, and thermal conductivity. Map services are also being developed for isopach maps, aquifer temperature maps, and several states are working on geothermal resource overview maps. Content models are developed based on existing community datasets to encourage widespread adoption and promulgate content quality standards. Geoscience data and maps from other GDS participating institutions, or “nodes” (e.g., U.S. Geological Survey, Southern Methodist University, Oregon Institute of Technology, Stanford University, the University of Utah) are being supplemented with extensive land management and land use resources from the Western Regional Partnership (15 federal agencies and 5 Western states) to provide access to a comprehensive, holistic set of data critical to geothermal energy development. As of May 2012 , we have nearly 37,000 records registered in the system catalog, and 550,075 data resources online, along with hundreds of Web services to deliver integrated data to the desktop for free downloading or online use. The data exchange mechanism is built on the U.S. Geoscience Information Network (USGIN, http://usgin.org and http://lab.usgin.org) protocols and standards developed as a partnership of the Association of American State Geologists (AASG) and U.S. Geological Survey (USGS). Keywords Data

  5. The AmeriFlux Data Activity and Data System: An Evolving Collection of Data Management Techniques, Tools, Products and Services

    SciTech Connect (OSTI)

    Boden, Thomas A [ORNL; Krassovski, Misha B [ORNL; Yang, Bai [ORNL

    2013-01-01T23:59:59.000Z

    The Carbon Dioxide Information Analysis Center (CDIAC) at Oak Ridge National Laboratory (ORNL), USA has provided scientific data management support for the U.S. Department of Energy and international climate change science since 1982. Over this period, climate change science has expanded from research focusing on basic understanding of geochemical cycles, particularly the carbon cycle, to integrated research addressing climate change impacts, vulnerability, adaptation, and mitigation. Interests in climate change data and information worldwide have grown remarkably and, as a result, so have demands and expectations for CDIAC s data systems. To meet the growing demands, CDIAC s strategy has been to design flexible data systems using proven technologies blended with new, evolving technologies and standards. CDIAC development teams are multidisciplinary and include computer science and information technology expertise, but also scientific expertise necessary to address data quality and documentation issues and to identify data products and system capabilities needed by climate change scientists. CDIAC has learned there is rarely a single commercial tool or product readily available to satisfy long-term scientific data system requirements (i.e., one size does not fit all and the breadth and diversity of environmental data are often too complex for easy use with commercial products) and typically deploys a variety of tools and data products in an effort to provide credible data freely to users worldwide. Like many scientific data management applications, CDIAC s data systems are highly customized to satisfy specific scientific usage requirements (e.g., developing data products specific for model use) but are also designed to be flexible and interoperable to take advantage of new software engineering techniques, standards (e.g., metadata standards) and tools and to support future Earth system data efforts (e.g., ocean acidification). CDIAC has provided data management support for numerous long-term measurement projects crucial to climate change science. One current example is the AmeriFlux measurement network. AmeriFlux provides continuous measurements from forests, grasslands, wetlands, and croplands in North, Central, and South America and offers important insight about carbon cycling in terrestrial ecosystems. We share our approaches in satisfying the challenges of delivering AmeriFlux data worldwide to benefit others with similar challenges handling climate change data, further heighten awareness and use of an outstanding ecological data resource, and highlight expanded software engineering applications being used for climate change measurement data.

  6. Min-max control design for large angle maneuvers of flexible spacecraft 

    E-Print Network [OSTI]

    Kim, Cheolho

    1991-01-01T23:59:59.000Z

    vector a is a =a, b, ! 1 j ff 0 dsi ? J u;dm ? eij ka j fikdm 1 . f 1 ? e;. kib J ukdm ? e e kiv ia ukdm M''J M B B B B + ui + e|jkivjuk + eqk~j (&k + uk) + eigpep Jkaga~j(~k + uk) bi (2. 1. 15) where ds, is obtained from Eq. (2. 1. 10) os...) into the form l P aP =a, i2; = Bftqd?+ + eijkdkpq + Nspq(lp (2. 1. 18) Np ? Np (tt' M )rj +2 . sq~s (tt M )rj MB B Np (ql M )1v 9 + e 's*r1v B Np ? )(iv;1vs ? 6, sa1, 1v?)qlq+ xr(a1;a1s ? b;ra1?a2?) b, with (1j = 1, 2, . . . , M). From the Newton...

  7. PETASCALE DATA STORAGE INSTITUTE (PDSI) Final Report

    SciTech Connect (OSTI)

    Gibson, Garth [Carnegie Mellon University

    2012-11-26T23:59:59.000Z

    Petascale computing infrastructures for scientific discovery make petascale demands on information storage capacity, performance, concurrency, reliability, availability, and manageability. The Petascale Data Storage Institute focuses on the data storage problems found in petascale scientific computing environments, with special attention to community issues such as interoperability, community buy-in, and shared tools. The Petascale Data Storage Institute is a collaboration between researchers at Carnegie Mellon University, National Energy Research Scientific Computing Center, Pacific Northwest National Laboratory, Oak Ridge National Laboratory, Sandia National Laboratory, Los Alamos National Laboratory, University of Michigan, and the University of California at Santa Cruz. Because the Institute focuses on low level files systems and storage systems, its role in improving SciDAC systems was one of supporting application middleware such as data management and system-level performance tuning. In retrospect, the Petascale Data Storage Institute’s most innovative and impactful contribution is the Parallel Log-structured File System (PLFS). Published in SC09, PLFS is middleware that operates in MPI-IO or embedded in FUSE for non-MPI applications. Its function is to decouple concurrently written files into a per-process log file, whose impact (the contents of the single file that the parallel application was concurrently writing) is determined on later reading, rather than during its writing. PLFS is transparent to the parallel application, offering a POSIX or MPI-IO interface, and it shows an order of magnitude speedup to the Chombo benchmark and two orders of magnitude to the FLASH benchmark. Moreover, LANL production applications see speedups of 5X to 28X, so PLFS has been put into production at LANL. Originally conceived and prototyped in a PDSI collaboration between LANL and CMU, it has grown to engage many other PDSI institutes, international partners like AWE, and has a large team at EMC supporting and enhancing it. PLFS is open sourced with a BSD license on sourceforge. Post PDSI funding comes from NNSA and industry sources. Moreover, PLFS has spin out half a dozen or more papers, partnered on research with multiple schools and vendors, and has projects to transparently 1) dis- tribute metadata over independent metadata servers, 2) exploit drastically non-POSIX Hadoop storage for HPC POSIX applications, 3) compress checkpoints on the fly, 4) batch delayed writes for write speed, 5) compress read-back indexes and parallelize their redistribution, 6) double-buffer writes in NAND Flash storage to decouple host blocking during checkpoint from disk write time in the storage system, 7) pack small files into a smaller number of bigger containers. There are two large scale open source Linux software projects that PDSI significantly incubated, though neither were initated in PDSI. These are 1) Ceph, a UCSC parallel object storage research project that has continued to be a vehicle for research, and has become a released part of Linux, and 2) Parallel NFS (pNFS) a portion of the IETF’s NFSv4.1 that brings the core data parallelism found in Lustre, PanFS, PVFS, and Ceph to the industry standard NFS, with released code in Linux 3.0, and its vendor offerings, with products from NetApp, EMC, BlueArc and RedHat. Both are fundamentally supported and advanced by vendor companies now, but were critcally transferred from research demonstration to viable product with funding from PDSI, in part. At this point Lustre remains the primary path to scalable IO in Exascale systems, but both Ceph and pNFS are viable alternatives with different fundamental advantages. Finally, research community building was a big success for PDSI. Through the HECFSIO workshops and HECURA project with NSF PDSI stimulated and helped to steer leveraged funding of over $25M. Through the Petascale (now Parallel) Data Storage Workshop series, www.pdsw.org, colocated with SCxy each year, PDSI created and incubated five offerings of this high-attendance workshop. The

  8. Report of Official Foreign Travel to Germany, May 16-June 1, 2001

    SciTech Connect (OSTI)

    J. D. Mason

    2001-06-18T23:59:59.000Z

    The Department of Energy (DOE) and associated agencies have moved rapidly toward electronic production, management, and dissemination of scientific and technical information. The World-Wide Web (WWW) has become a primary means of information dissemination. Electronic commerce (EC) is becoming the preferred means of procurement. DOE, like other government agencies, depends on and encourages the use of international standards in data communications. Like most government agencies, DOE has expressed a preference for openly developed standards over proprietary designs promoted as ''standards'' by vendors. In particular, there is a preference for standards developed by organizations such as the International Organization for Standardization (ISO) and the American National Standards Institute (ANSI) that use open, public processes to develop their standards. Among the most widely adopted international standards is the Standard Generalized Markup Language (SGML, ISO 8879:1986, FIPS 152), to which DOE long ago made a commitment. Besides the official commitment, which has resulted in several specialized projects, DOE makes heavy use of coding derived from SGML: Most documents on the WWW are coded in HTML (Hypertext Markup Language), which is an application of SGML. The World-Wide Web Consortium (W3C), with the backing of major software houses like Adobe, IBM, Microsoft, Netscape, Oracle, and Sun, is promoting XML (eXtensible Markup Language), a class of SGML applications, for the future of the WWW and the basis for EC. In support of DOE's use of these standards, I have served since 1985 as Chairman of the international committee responsible for SGML and related standards, ISO/IEC JTC1/SC34 (SC34) and its predecessor organizations. During my May 2001 trip, I chaired the spring 2001 meeting of SC34 in Berlin, Germany. I also attended XML Europe 2001, a major conference on the use of SGML and XML sponsored by the Graphic Communications Association (GCA), and chaired a meeting of the International SGML/XML Users' Group (ISUG). In addition to the widespread use of the WWW among DOE's plants and facilities in Oak Ridge and among DOE sites across the nation, there have been several past and present SGML- and XML-based projects at the Y-12 National Security Complex (Y-12). Our local project team has done SGML and XML development at Y-12 and Oak Ridge National Laboratory (ORNL) since the late 1980s. SGML is a component of the Weapons Records Archiving and Preservation (WRAP) project at Y-12 and is the format for catalog metadata chosen for weapons records by the Nuclear Weapons Information Group (NWIG). The ''Ferret'' system for automated classification analysis uses XML to structure its knowledge base. The Ferret team also provides XML consulting to OSTI and DOE Headquarters, particularly the National Nuclear Security Administration (NNSA). Supporting standards development allows DOE and Y-12 the opportunity both to provide input into the process and to benefit from contact with some of the leading experts in the subject matter. Oak Ridge has been for some years the location to which other DOE sites turn for expertise in SGML, XML, and related topics.

  9. A greenhouse-gas information system monitoring and validating emissions reporting and mitigation

    SciTech Connect (OSTI)

    Jonietz, Karl K [Los Alamos National Laboratory; Dimotakis, Paul E [JPL/CAL TECH; Roman, Douglas A [LLNL; Walker, Bruce C [SNL

    2011-09-26T23:59:59.000Z

    Current GHG-mitigating regimes, whether internationally agreed or self-imposed, rely on the aggregation of self-reported data, with limited checks for consistency and accuracy, for monitoring. As nations commit to more stringent GHG emissions-mitigation actions and as economic rewards or penalties are attached to emission levels, self-reported data will require independent confirmation that they are accurate and reliable, if they are to provide the basis for critical choices and actions that may be required. Supporting emissions-mitigation efforts and agreements, as well as monitoring energy- and fossil-fuel intensive national and global activities would be best achieved by a process of: (1) monitoring of emissions and emission-mitigation actions, based, in part, on, (2) (self-) reporting of pertinent bottom-up inventory data, (3) verification that reported data derive from and are consistent with agreed-upon processes and procedures, and (4) validation that reported emissions and emissions-mitigation action data are correct, based on independent measurements (top-down) derived from a suite of sensors in space, air, land, and, possibly, sea, used to deduce and attribute anthropogenic emissions. These data would be assessed and used to deduce and attribute measured GHG concentrations to anthropogenic emissions, attributed geographically and, to the extent possible, by economic sector. The validation element is needed to provide independent assurance that emissions are in accord with reported values, and should be considered as an important addition to the accepted MRV process, leading to a MRV&V process. This study and report focus on attributes of a greenhouse-gas information system (GHGIS) needed to support MRV&V needs. These needs set the function of such a system apart from scientific/research monitoring of GHGs and carbon-cycle systems, and include (not exclusively): the need for a GHGIS that is operational, as required for decision-support; the need for a system that meets specifications derived from imposed requirements; the need for rigorous calibration, verification, and validation (CV&V) standards, processes, and records for all measurement and modeling/data-inversion data; the need to develop and adopt an uncertainty-quantification (UQ) regimen for all measurement and modeling data; and the requirement that GHGIS products can be subjected to third-party questioning and scientific scrutiny. This report examines and assesses presently available capabilities that could contribute to a future GHGIS. These capabilities include sensors and measurement technologies; data analysis and data uncertainty quantification (UQ) practices and methods; and model-based data-inversion practices, methods, and their associated UQ. The report further examines the need for traceable calibration, verification, and validation processes and attached metadata; differences between present science-/research-oriented needs and those that would be required for an operational GHGIS; the development, operation, and maintenance of a GHGIS missions-operations center (GMOC); and the complex systems engineering and integration that would be required to develop, operate, and evolve a future GHGIS. Present monitoring systems would be heavily relied on in any GHGIS implementation at the outset and would likely continue to provide valuable future contributions to GHGIS. However, present monitoring systems were developed to serve science/research purposes. This study concludes that no component or capability presently available is at the level of technological maturity and readiness required for implementation in an operational GHGIS today. However, purpose-designed and -built components could be developed and implemented in support of a future GHGIS. The study concludes that it is possible to develop and provide a capability-driven prototype GHGIS, as part of a Phase-1 effort, within three years from project-funding start, that would make use of and integrate existing sensing and system capabilities. As part of a Phase-2 effort, a requirem

  10. Collaborative Systemwide Monitoring and Evaluation Project (CSMEP) - Year 5 : Annual Report for FY 2008.

    SciTech Connect (OSTI)

    Marmorek, David R.; Porter, Marc; Pickard, Darcy; Wieckowski, Katherine

    2008-11-19T23:59:59.000Z

    The Collaborative Systemwide Monitoring and Evaluation Project (CSMEP) is a coordinated effort to improve the quality, consistency, and focus of fish population and habitat data to answer key monitoring and evaluation questions relevant to major decisions in the Columbia River Basin. CSMEP was initiated by the Columbia Basin Fish and Wildlife Authority (CBFWA) in October 2003. The project is funded by the Bonneville Power Administration (BPA) through the Northwest Power and Conservation Council's Fish and Wildlife Program (NPCC). CSMEP is a major effort of the federal state and Tribal fish and wildlife managers to develop regionally integrated monitoring and evaluation (M&E) across the Columbia River Basin. CSMEP has focused its work on five monitoring domains: status and trends monitoring of populations and action effectiveness monitoring of habitat, harvest, hatcheries, and the hydrosystem. CSMEP's specific goals are to: (1) interact with federal, state and tribal programmatic and technical entities responsible for M&E of fish and wildlife, to ensure that work plans developed and executed under this project are well integrated with ongoing work by these entities; (2) document, integrate, and make available existing monitoring data on listed salmon, steelhead, bull trout and other fish species of concern; (3) critically assess strengths and weaknesses of these data for answering key monitoring questions; and (4) collaboratively design, implement and evaluate improved M&E methods with other programmatic entities in the Pacific Northwest. During FY2008 CSMEP biologists continued their reviews of the strengths and weaknesses (S&W) of existing subbasin inventory data for addressing monitoring questions about population status and trends at different spatial and temporal scales. Work was focused on Lower Columbia Chinook and steelhead, Snake River fall Chinook, Upper Columbia Spring Chinook and steelhead, and Middle Columbia River Chinook and steelhead. These FY2008 data assessments and others assembled over the years of the CSMEP project can be accessed on the CBFWA public website. The CSMEP web database (http://csmep.streamnet.org/) houses metadata inventories from S&W assessments of Columbia River Basin watersheds that were completed prior to FY2008. These older S&W assessments are maintained by StreamNet, but budget cutbacks prevented us from adding the new FY2008 assessments into the database. Progress was made in FY2008 on CSMEP's goals of collaborative design of improved M&E methods. CSMEP convened two monitoring design workshops in Portland (December 5 and 6, 2007 and February 11 and 12, 2008) to continue exploration of how best to integrate the most robust features of existing M&E programs with new approaches. CSMEP continued to build on this information to develop improved designs and analytical tools for monitoring the status and trends of fish populations and the effectiveness of hatchery and hydrosystem recovery actions within the Columbia River Basin. CSMEP did not do any new work on habitat or harvest effectiveness monitoring designs in FY2008 due to budget cutbacks. CSMEP presented the results of the Snake Basin Pilot Study to the Independent Scientific Review Panel (ISRP) in Portland on December 7, 2008. This study is the finalization of CSMEP's pilot exercise of developing design alternatives across different M&E domains within the Snake River Basin spring/summer Chinook ESU. This work has been summarized in two linked reports (CSMEP 2007a and CSMEP 2007b). CSMEP participants presented many of the analyses developed for the Snake Basin Pilot work at the Western Division American Fisheries Society (AFS) conference in Portland on May 4 to 7, 2008. For the AFS conference CSMEP organized a symposium on regional monitoring and evaluation approaches. A presentation on CSMEP's Cost Integration Database Tool and Salmon Viability Monitoring Simulation Model developed for the Snake Basin Pilot Study was also given to the Pacific Northwest Aquatic monitoring Partnership (PNAMP) stee

  11. National Geothermal Data System: Transforming the Discovery, Access, and Analytics of Data for Geothermal Exploration

    SciTech Connect (OSTI)

    Patten, Kim [Arizona Geological Survey

    2013-05-01T23:59:59.000Z

    Compendium of Papers from the 38th Workshop on Geothermal Reservoir Engineering Stanford University, Stanford, California February 11-13, 2013 The National Geothermal Data System (NGDS) is a distributed, interoperable network of data collected from state geological surveys across all fifty states and the nation’s leading academic geothermal centers. The system serves as a platform for sharing consistent, reliable, geothermal-relevant technical data with users of all types, while supplying tools relevant for their work. As aggregated data supports new scientific findings, this content-rich linked data ultimately broadens the pool of knowledge available to promote discovery and development of commercial-scale geothermal energy production. Most of the up-front risks associated with geothermal development stem from exploration and characterization of subsurface resources. Wider access to distributed data will, therefore, result in lower costs for geothermal development. NGDS is on track to become fully operational by 2014 and will provide a platform for custom applications for accessing geothermal relevant data in the U.S. and abroad. It is being built on the U.S. Geoscience Information Network (USGIN) data integration framework to promote interoperability across the Earth sciences community. The basic structure of the NGDS employs state-of-the art informatics to advance geothermal knowledge. The following four papers comprising this Open-File Report are a compendium of presentations, from the 38th Annual Workshop on Geothermal Reservoir Engineering, taking place February 11-13, 2013 at Stanford University, Stanford, California. “NGDS Geothermal Data Domain: Assessment of Geothermal Community Data Needs,” outlines the efforts of a set of nationwide data providers to supply data for the NGDS. In particular, data acquisition, delivery, and methodology are discussed. The paper addresses the various types of data and metadata required and why simple links to existing data are insufficient for promoting geothermal exploration. Authors of this paper are Arlene Anderson, US DOE Geothermal Technologies Office, David Blackwell, Southern Methodist University (SMU), Cathy Chickering (SMU), Toni Boyd, Oregon Institute of Technology’s GeoHeat Center, Roland Horne, Stanford University, Matthew MacKenzie, Uberity, Joe Moore, University of Utah, Duane Nickull, Uberity, Stephen Richard, Arizona Geological Survey, and Lisa Shevenell, University of Nevada, Reno. “NGDS User Centered Design: Meeting the Needs of the Geothermal Community,” discusses the user- centered design approach taken in the development of a user interface solution for the NGDS. The development process is research based, highly collaborative, and incorporates state-of-the-art practices to ensure a quality user interface for the widest and greatest utility. Authors of this paper are Harold Blackman, Boise State University, Suzanne Boyd, Anthro-Tech, Kim Patten, Arizona Geological Survey, and Sam Zheng, Siemens Corporate Research. “Fueling Innovation and Adoption by Sharing Data on the DOE Geothermal Data Repository Node on the National Geothermal Data System,” describes the motivation behind the development of the Geothermal Data Repository (GDR) and its role in the NGDS. This includes the benefits of using the GDR to share geothermal data of all types and DOE’s data submission process. Authors of this paper are Jon Weers, National Renewable Energy Laboratory and Arlene Anderson, US DOE Geothermal Technologies Office. Finally, “Developing the NGDS Adoption of CKAN for Domestic & International Data Deployment,” provides an overview of the “Node-In-A-Box” software package designed to provide data consumers with a highly functional interface to access the system, and to ease the burden on data providers who wish to publish data in the system. It is important to note that this software package constitutes a reference implementation and that the NGDS architecture is based on open standards, which means other server software can make resources available, a