National Library of Energy BETA

Sample records for visitor reception computer

  1. Visitor Hanford Computer Access Request - Hanford Site

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Visitor Hanford Computer Access Request Visitor Hanford Computer Access Request Visitor Hanford Computer Access Request Visitor Hanford Computer Access Request Email Email Page |...

  2. Computing and Computational Sciences Directorate - Visitor Information

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Home › About Us Visitor Information Entering ORNL ORNL welcomes visitors to the Laboratory. However, because of increased security requirements, we've made some changes in how the site is accessed. Bethel Valley Road, which is the main access route to Oak Ridge National Laboratory from both directions, is now closed to the public. If you are planning a visit to ORNL, your host will arrange for you to proceed past entrance stations on Bethel Valley Road leading to the Laboratory's Visitor

  3. Visitors

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Visitors Upon arrival, all non-LLNL visitors touring the Facility must first report to Beth Mariotti (ext. 4-3396). All visitors, including employees of LLNL, are required to be escorted by a member of Facility Operations or an individual approved by Operations. Visitors tours can only be conducted when no safety hazards (i.e., laser, high voltage, radiation) are present

  4. Visitors

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Visitors /community/visitors/_assets/images/icon-visitors-orange-paper-plane.jpg Visitors Coming to Los Alamos? Need badging information? Need local resources or housing? This is the place. Badging, Badge Office» Bradbury Science Museum» Education Office Housing» Living in Los Alamos» TOP STORIES - highlights of our science, people, technologies close Connections Laboratory helps Ohkay Owingeh find math, science tutors; giving students a chance to work, learn and earn; Los Alamos, Van Andel

  5. Visitors

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Alamos and Northern New Mexico. News Releases Science Briefs Photos Picture of the Week Social Media Videos Fact Sheets Publications PHOTOS BY TOPIC Careers Community Visitors...

  6. Visitor Information

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Visitor Info Visitor Information NERSC is located in Shyh Wang Hall-also known as the computational research and theory (CRT) building (Bldg. 59)-on the campus of Lawrence Berkeley...

  7. SA 2712-CVC;CONTRACTOR/VISITOR COMPUTERS TEMPORARY ONE TIME EXCEPTION...

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    SA 2712-CVC (9-2009) Supersedes (6-2009) issue Page 2 of 2 ContractorVisitor Computers Temporary, One-Time Exception to SNL Policy This form must be completed and signed...

  8. Visitor Information

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Visitor Info Visitor Information NERSC is located in Shyh Wang Hall-also known as the computational research and theory (CRT) building (Bldg. 59)-on the campus of Lawrence Berkeley National Laboratory in Berkeley, California. You can find information about visiting the lab, including an interactive campus map, transportation and lodging suggestions at http://www.lbl.gov/visit/. Your host must arrange site access prior to your visit. Last edited: 2016-01-06 16:48:46

  9. ARM - Historical Visitor Statistics

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Visitor Statistics 2016 Quarterly Reports First Quarter (PDF) Second Quarter (PDF) Third Quarter (PDF) Fourth Quarter (PDF) Past Quarterly Reports Historical Statistics Field Campaigns Operational Visitors and Accounts Data Archive and Usage (October 1995 - Present) Historical Visitor Statistics As a national user facility, ARM is required to report facility use for actual visitors and for active user research computer and Archive accounts and only unique scientific users will be counted. A

  10. Visitor Information

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Visitor Information Visiting the Transportation Research and Analysis Computing Center For guests who wish to visit and meet with the technical staff, please contact the TRACC service desk or the Technical Point of Contact assigned to you for registration, scheduling, and confirmation, including Foreign Visits and Assignments. Registration with your host is required for entry into the TRACC facilities. A photo I.D., such as a driver's license or passport, is also required for facility access. If

  11. Visitor Information | The Ames Laboratory

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Visitor Information Contents Internet Access for Visitors Wireless Network Access for Visitors Captive Portal Service for Visitors Rules of Behavior Internet Access for Visitors...

  12. ALS Visitors

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Quick Facts ALS Visitors Print ALS staff members host a variety of scientific, educational, government, and community-related tours each month. April 2014 Congresswoman Jackie Speier, who represents southern San Francisco and northern Peninsula communities, recently spent an afternoon at Berkeley Lab. In addition to an overview of the Lab provided by Lab Director Paul Alivisatos, Speier toured the FLEXLAB and the Advanced Light Source. At the ALS she spoke with Beamline Scientist Ken Goldberg

  13. BPA Visitor Center

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    and is located at BPA headquarters in Portland, Oregon at 905 NE 11th Ave. (Public Transit @ the Lloyd Center MAX Station). Many of the publications available in the Visitor...

  14. Visitors | The Ames Laboratory

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Visitors Visitors are welcome at Ames Laboratory. As a U.S. Department of Energy research facility, Ames Laboratory is subject to security conditions established by the Department of Homeland Security. To make sure that you are complying with the current security conditions, please check with the Plant Protection Desk on the ground floor level of the Technical and Administrative Services Facility (TASF) building. Protection personnel can help you locate a specific staff member or direct you to a

  15. Summit Opening Reception

    Broader source: Energy.gov [DOE]

    Join Summit attendees in California C & D for an opening reception featuring kick-off speakers Minh Le, Director of the SunShot Initiative, and Tom Tait, mayor of the City of Anaheim. All...

  16. Craig Thomas Discovery & Visitor Center

    High Performance Buildings Database

    Moose, WY Grand Teton National Park's rugged landscape and stunning array of wildlife attract nearly three million visitors every year, making it one of our most popular national parks. A new Grand Teton National Park visitor center near the park's headquarters north of Jackson, Wyoming, replaces an outdated building, educates an increased number of visitors, and inspires further exploration of this extraordinary landscape. The project site is located along the Snake River, between a riparian forest and a sagebrush meadow.

  17. campus-visitor-map

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    MC-212 Pollard Aud. MC- 210 MC- 120 MC-130 MC-100 Main Campus Map Building Room/Office Contact Name + Number Visitor Map You are here. Emergency Assembly Point Entrance Buildings Sidewalks T o O R A U W a y South Campus SC-1 SC-1 Annex B et he l Va lle y R d SC-13 SC-9 SC-16 SC-300 SC-10 SC-15 SC-200 SC-100 Building Room/Office Contact Name + Number Directions from Main Campus: * Go back out to ORAU Way * Turn left onto ORAU Way * Turn left at S Illinois Ave/TN-62 (2.9 mi) * Bear slight right

  18. DOE - Office of Legacy Management -- Visitors Center

    Office of Legacy Management (LM)

    Ohio > Fernald > Visitors Center > Visitors Center Fernald Preserve, Ohio Visitors Center Fernald Visitors Center Visitors Center Directions Event Calendar Community Meeting Room, Program Shelter, and Resource Room Program/Speaker Request Brochures Fact Sheets Technical Papers BioBlitz Geocaching Pets Policy The Fernald Preserve Visitors Center is a 10,000-square-foot green building that celebrates the rich and varied history of the Fernald site. Information on the site's natural,

  19. Visitor Control / Site Access - Hanford Site

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    About Us Hanford Site Wide Programs Visitor Control Site Access About Us About Hanford Cleanup Hanford History Hanford Site Wide Programs Contact Us Visitor Control Site Access...

  20. NREL Visitors Center Closing Temporarily for Remodeling

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Visitors Center Closing Temporarily for Remodeling Media may contact: George Douglas, 303-275-4096 email: George Douglas Golden, Colo., Sept. 6, 2000 - The Visitors Center at the U.S. Department of Energy's (DOE) National Renewable Energy Laboratory (NREL) will be closed Friday, Sept. 8 through Friday, Sept. 22 for the installation of new floor tile and carpet. Visitors Center staff expects to re-open the building for normal operation on Monday, Sept. 25. Contact the Visitors Center volunteer

  1. ASCR Committees of Visitors | U.S. DOE Office of Science (SC)

    Office of Science (SC) Website

    Committees of Visitors Advanced Scientific Computing Research (ASCR) ASCR Home About Organization Chart .pdf file (40KB) Staff ASCR Budget ASCR Committees of Visitors Directions Jobs Research Facilities Science Highlights Benefits of ASCR Funding Opportunities Advanced Scientific Computing Advisory Committee (ASCAC) Community Resources Contact Information Advanced Scientific Computing Research U.S. Department of Energy SC-21/Germantown Building 1000 Independence Ave., SW Washington, DC 20585 P:

  2. Visitors

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    of our science, people, technologies close Grape Growers Association enlivens agriculture Entrepreneurial Networking Program unites New Mexico grape growers for successful...

  3. DOE - NNSA/NFO -- Visitor Information

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Visitor Information NNSA/NFO Language Options U.S. DOE/NNSA - Nevada Field Office Visitor Information The Nevada Field Office welcomes visitors to its facilities, but if you plan to visit, there are several things you will need to keep in mind. You must present valid photo identification as well as the name of the person(s) you will be visiting upon entering the facility. All persons and vehicles entering the North Las Vegas Complex are subject to search. Official Visitors are those who have

  4. Alternative Fuels Data Center: Propane Buses Shuttle Visitors in Maine

    Alternative Fuels and Advanced Vehicles Data Center [Office of Energy Efficiency and Renewable Energy (EERE)]

    Propane Buses Shuttle Visitors in Maine to someone by E-mail Share Alternative Fuels Data Center: Propane Buses Shuttle Visitors in Maine on Facebook Tweet about Alternative Fuels Data Center: Propane Buses Shuttle Visitors in Maine on Twitter Bookmark Alternative Fuels Data Center: Propane Buses Shuttle Visitors in Maine on Google Bookmark Alternative Fuels Data Center: Propane Buses Shuttle Visitors in Maine on Delicious Rank Alternative Fuels Data Center: Propane Buses Shuttle Visitors in

  5. Directions and Visitor Information | Argonne National Laboratory

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Directory Argonne National Laboratory About Safety News Careers Education Community Diversity Directory Energy Environment Security User Facilities Science Work with Argonne Visiting Argonne Site Access Policy Map Argonne-University of Chicago Shuttle Map of Argonne Commitment to Safety Argonne is dedicated to protecting visitors to our campus. Every visitor watches a two-minute training video about how to protect oneself and others while on campus. External dosimetry The External Dosimetry

  6. Visitor Security Requirements | Savannah River Ecology Laboratory

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Visitor Security Requirements NOTE: These requirements do not apply to events held at the SREL Conference Center, which is located outside the security barricade. Before you arrive: I. Obtaining a visitor's badge II. Prohibited articles III. Electronics policy IV. Vehicle inspections When you arrive: Proceed to the badge office (map). Present two forms of identification from the list of acceptable documents, including one photo ID, to receive your badge. You will not receive a badge without

  7. Exchange Visitors Program | Department of Energy

    Office of Energy Efficiency and Renewable Energy (EERE) Indexed Site

    Exchange Visitors Program Exchange Visitors Program The Department of Energy program offices engage in a variety of cooperative international exchanges and program initiatives which involve sponsoring foreign national professionals to work on projects in the U.S. The Office of Travel Management provides assistance to DOE Program Elements on employer-sponsored visa issues, and requirements. DOE J-1 Visas - provide foreign nationals with opportunities to participate in educational and cultural

  8. PIA - HSS Electronic Visitor Management System (HSEVMS) | Department of

    Office of Energy Efficiency and Renewable Energy (EERE) Indexed Site

    Energy HSS Electronic Visitor Management System (HSEVMS) PIA - HSS Electronic Visitor Management System (HSEVMS) PIA - HSS Electronic Visitor Management System (HSEVMS) PDF icon PIA - HSS Electronic Visitor Management System (HSEVMS) More Documents & Publications PIA - INL SECURITY INFORMATION MANAGEMENT SYSTEM BUSINESS ENCLAVE PIA - INL PeopleSoft - Human Resource System PIA - Human Resources - Personal Information Change Request - Idaho National Engineering Laboratory

  9. Support for non-locking parallel reception of packets belonging to a single memory reception FIFO

    DOE Patents [OSTI]

    Chen, Dong; Heidelberger, Philip; Salapura, Valentina; Senger, Robert M.; Steinmacher-Burow, Burkhard; Sugawara, Yutaka

    2011-01-27

    A method and apparatus for distributed parallel messaging in a parallel computing system. A plurality of DMA engine units are configured in a multiprocessor system to operate in parallel, one DMA engine unit for transferring a current packet received at a network reception queue to a memory location in a memory FIFO (rmFIFO) region of a memory. A control unit implements logic to determine whether any prior received packet destined for that rmFIFO is still in a process of being stored in the associated memory by another DMA engine unit of the plurality, and prevent the one DMA engine unit from indicating completion of storing the current received packet in the reception memory FIFO (rmFIFO) until all prior received packets destined for that rmFIFO are completely stored by the other DMA engine units. Thus, there is provided non-locking support so that multiple packets destined for a single rmFIFO are transferred and stored in parallel to predetermined locations in a memory.

  10. Committees of Visitors | U.S. DOE Office of Science (SC)

    Office of Science (SC) Website

    Committees of Visitors Deputy Director for Science Programs Deputy Director Home Mission & Functions Deputy Director Biography Organization Staff Presentations & Testimony Federal Advisory Committees Committees of Visitors ASCR Committees of Visitors BES Committees of Visitors BER Committees of Visitors FES Committees of Visitors HEP Committees of Visitors NP Committees of Visitors WDTS Committees of Visitors Contact Information Deputy Director for Science Programs U.S. Department of

  11. ASCR Committees of Visitors | U.S. DOE Office of Science (SC)

    Office of Science (SC) Website

    ASCR Committees of Visitors Deputy Director for Science Programs Deputy Director Home Mission & Functions Deputy Director Biography Organization Staff Presentations & Testimony Federal Advisory Committees Committees of Visitors ASCR Committees of Visitors BES Committees of Visitors BER Committees of Visitors FES Committees of Visitors HEP Committees of Visitors NP Committees of Visitors WDTS Committees of Visitors Contact Information Deputy Director for Science Programs U.S. Department

  12. BER Committees of Visitors | U.S. DOE Office of Science (SC)

    Office of Science (SC) Website

    BER Committees of Visitors Deputy Director for Science Programs Deputy Director Home Mission & Functions Deputy Director Biography Organization Staff Presentations & Testimony Federal Advisory Committees Committees of Visitors ASCR Committees of Visitors BES Committees of Visitors BER Committees of Visitors FES Committees of Visitors HEP Committees of Visitors NP Committees of Visitors WDTS Committees of Visitors Contact Information Deputy Director for Science Programs U.S. Department of

  13. FES Committees of Visitors | U.S. DOE Office of Science (SC)

    Office of Science (SC) Website

    FES Committees of Visitors Deputy Director for Science Programs Deputy Director Home Mission & Functions Deputy Director Biography Organization Staff Presentations & Testimony Federal Advisory Committees Committees of Visitors ASCR Committees of Visitors BES Committees of Visitors BER Committees of Visitors FES Committees of Visitors HEP Committees of Visitors NP Committees of Visitors WDTS Committees of Visitors Contact Information Deputy Director for Science Programs U.S. Department of

  14. HEP Committees of Visitors | U.S. DOE Office of Science (SC)

    Office of Science (SC) Website

    HEP Committees of Visitors Deputy Director for Science Programs Deputy Director Home Mission & Functions Deputy Director Biography Organization Staff Presentations & Testimony Federal Advisory Committees Committees of Visitors ASCR Committees of Visitors BES Committees of Visitors BER Committees of Visitors FES Committees of Visitors HEP Committees of Visitors NP Committees of Visitors WDTS Committees of Visitors Contact Information Deputy Director for Science Programs U.S. Department of

  15. WDTS Committees of Visitors | U.S. DOE Office of Science (SC)

    Office of Science (SC) Website

    WDTS Committees of Visitors Deputy Director for Science Programs Deputy Director Home Mission & Functions Deputy Director Biography Organization Staff Presentations & Testimony Federal Advisory Committees Committees of Visitors ASCR Committees of Visitors BES Committees of Visitors BER Committees of Visitors FES Committees of Visitors HEP Committees of Visitors NP Committees of Visitors WDTS Committees of Visitors Contact Information Deputy Director for Science Programs U.S. Department

  16. NP Committees of Visitors | U.S. DOE Office of Science (SC)

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    NP Committees of Visitors Deputy Director for Science Programs Deputy Director Home Mission & Functions Deputy Director Biography Organization Staff Presentations & Testimony Federal Advisory Committees Committees of Visitors ASCR Committees of Visitors BES Committees of Visitors BER Committees of Visitors FES Committees of Visitors HEP Committees of Visitors NP Committees of Visitors WDTS Committees of Visitors Contact Information Deputy Director for Science Programs U.S. Department of

  17. Jefferson Lab Visitor's Center - Schedule a Tour

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Regional Map Virtual Tour Schedule a Tour Newport News Resources Places to Visit Tourist Attractions Restaurants Climate and Weather Tour Safety Jefferson Lab Environmental, Health & Safety Manual Chapter 3230: Communication of Hazards to Visitors Jefferson Lab is classified as a low-hazard facility. This means the hazards encountered here are comparable to those typically found in an industrial setting. There are a few places on-site where minimal personal protective equipment (PPE) is

  18. PRIVACY IMPACT ASSESSMENT: Visitor Dosimeter Badge Tracking

    Energy Savers [EERE]

    PRIVACY IMPACT ASSESSMENT: Visitor Dosimeter Badge Tracking PIA Template Version 3 - May, 2009 Department of Energy Privacy Impact Assessment (PIA) Guidance is provided in the template. See DOE Order 206.1, Department of Energy Privacy Program, Appendix A, Privacy Impact Assessments, for requirements and additional guidance for conducting a PIA: http://www.directives.doe.gov/pdfs/doe/doetexUneword/206/o2061.pdf Please complete electronically: no hand-written submissions will be accepted. This

  19. ORO Office Safeguards and Security Clearance Tracking System and Visitor

    Office of Energy Efficiency and Renewable Energy (EERE) Indexed Site

    Control System PIA, Oak Ridge Operations Office | Department of Energy Office Safeguards and Security Clearance Tracking System and Visitor Control System PIA, Oak Ridge Operations Office ORO Office Safeguards and Security Clearance Tracking System and Visitor Control System PIA, Oak Ridge Operations Office ORO Office Safeguards and Security Clearance Tracking System and Visitor Control System PIA, Oak Ridge Operations Office PDF icon ORO Office Safeguards and Security Clearance Tracking

  20. W7 printing from VISITOR network

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    W7 printing from VISITOR network Last Modified by Jerry Camuso, 02/12/2015. 1 . From the "start menu", click on "Devices and Printers" 2. Click on "Add a printer" and choose "Add a local printer". 3 . Choose "create a new port" and pick "Standard TCP/IP Port" from the drop down menu. 4. Type the following for the next box and click "Next": Hostname or IP address: PRINTSERV.slac.stanford.edu Port name: This should be the DNS

  1. Fernald Preserve Attracts 50,000 Visitors | Department of Energy

    Office of Energy Efficiency and Renewable Energy (EERE) Indexed Site

    Fernald Preserve Attracts 50,000 Visitors Fernald Preserve Attracts 50,000 Visitors July 10, 2014 - 3:48pm Addthis What does this project do? Goal 4. Optimize the use of land and assets. 8205D-1997.jpg The Fernald Preserve Visitors Center opened in August 2008. Gary Studer, a longtime resident of the greater Cincinnati area, is a regular visitor to the Fernald Preserve in Ohio, enjoying activities like trail walks and bird watching. On Saturday, March 1, 2014, Gary became the 50,000th recipient

  2. BER Committees of Visitors | U.S. DOE Office of Science (SC)

    Office of Science (SC) Website

    Visitors Biological and Environmental Research (BER) BER Home About Organization Chart .pdf file (73KB) Staff BER Budget BER Committees of Visitors Directions Jobs Research...

  3. BES Committees of Visitors | U.S. DOE Office of Science (SC)

    Office of Science (SC) Website

    Committees of Visitors Basic Energy Sciences (BES) BES Home About Organization Chart .pdf file (47KB) Staff BES Budget BES Committees of Visitors Directions Jobs Organizational...

  4. How ORISE is Making a Difference: Virtual Community Reception Center

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Virtual Community Reception Center 3-D virtual training equips health departments for mass casualty radiation emergencies Virtual Community Reception Center The Virtual Community Reception Center. Click Image to launch video. Following a mass casualty radiation emergency, public health professionals will play a crucial role in assessing radiation exposures, screening for radioactive contamination, and prioritizing people for further care. This process-population monitoring-will be conducted in

  5. Zion's New Visitor Center a Model of Energy Efficiency

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Most park-goers will pass through the visitor center, where energy efficient features will save about 14,000 a year. Notable energy saving measures include: Trombe walls: Most of ...

  6. Fernald Preserve Visitors Center Grand Opening and LEED Platinum

    Energy Savers [EERE]

    Certification | Department of Energy Fernald Preserve Visitors Center Grand Opening and LEED Platinum Certification Fernald Preserve Visitors Center Grand Opening and LEED Platinum Certification October 16, 2008 - 4:14pm Addthis Remarks as Prepared for Acting Deputy Secretary Kupfer Thank you, Mike, for that introduction and to both you and Jane for hosting this event. You both have been instrumental in the dramatic transformation of this site. We made a commitment more than a decade ago to

  7. Village Power Exhibit Featured at NREL's Visitors Center

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Village Power Exhibit Featured at NREL's Visitors Center "Powering Our Lives, Powering Our World Energy" Expo Slated for February-April For more information contact: Gary Schmitz, 303-275-4050 email: Gary Schmitz Golden, Colo., Feb. 16, 2001 - An exhibit that demonstrates how renewable energy sources can bring needed electricity to developing nations worldwide opens Feb. 20 at the Visitors Center of the U.S. Department of Energy's National Renewable Energy Laboratory. The Village Power

  8. Illustrious Visitors Good Hosts Real Celebrity | Jefferson Lab

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Illustrious Visitors; Good Hosts! Real Celebrity. June 15, 2015 On a regular basis, the DOE Site Office asks us to update our list of illustrious visitors; often, there is nothing to say. Other times we are inundated. The past few months has seen a fine selection of events hosted by Jefferson Lab, and Jefferson Science Associates and its parent Southeastern Universities Research Association. This past week, Joachim Mnich, a member of the DESY Directorate and Chairman of ICFA, the International

  9. Reception of longitudinal vector potential radiation with a plasma antenna

    SciTech Connect (OSTI)

    Zimmerman, Robert K. Jr.

    2013-07-28

    To help resolve the long-running debate between physicists and engineers regarding the existence of the magnetic vector potential, herewith we describe an experiment demonstrating reception of time-harmonic vector potential radiation at 1.3 GHz.

  10. Interagency Visitor Center at Santa Monica Mountains National Recreation Area

    High Performance Buildings Database

    Calabasas, CA This project was to develop the first visitor center for the Santa Monica Mountains National Recreation Area located in the Los Angeles, California area. The previous visitor center was across from a shopping mall in rental space at park headquarters in Thousand Oaks. The new facility is centrally located in the park at a much more appropriate natural and cultural resource setting. It is a partnership project with the Mountains Recreation and Conservation Authority, which is a local land conservation and park agency. It is also a joint facility with California State Parks.

  11. BER Committees of Visitors | U.S. DOE Office of Science (SC)

    Office of Science (SC) Website

    BER Committees of Visitors Biological and Environmental Research Advisory Committee (BERAC) BERAC Home Meetings Members Charges/Reports Charter .pdf file (135KB) BER Committees of Visitors Federal Advisory Committees BER Home BER Committees of Visitors Print Text Size: A A A FeedbackShare Page Biological and Environmental Research Advisory Committee (BERAC) » The links below provide an archive of Biological and Environmental Research (BER) Committees of Visitors (COV) reports and responses.

  12. FES Committees of Visitors | U.S. DOE Office of Science (SC)

    Office of Science (SC) Website

    FES Committees of Visitors Fusion Energy Sciences Advisory Committee (FESAC) FESAC Home Meetings Members Charges/Reports Charter .pdf file (140KB) FES Committees of Visitors Federal Advisory Committees FES Home FES Committees of Visitors Print Text Size: A A A FeedbackShare Page Fusion Energy Sciences Advisory Committee (FESAC) » The links below provide an archive of Fusion Energy Sciences (FES) Committees of Visitors (COV) reports and responses. 2014 FESAC COV Report on FES Program .pdf file

  13. computers

    National Nuclear Security Administration (NNSA)

    Each successive generation of computing system has provided greater computing power and energy efficiency.

    CTS-1 clusters will support NNSA's Life Extension Program and...

  14. Los Alamos notifies employees, visitors of possible beryllium exposure

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Possible beryllium exposure Los Alamos notifies employees, visitors of possible beryllium exposure Current and former Laboratory employees and contractors who either worked at or visited the facility have been-or are in the process of being-notified. January 28, 2009 Los Alamos National Laboratory sits on top of a once-remote mesa in northern New Mexico with the Jemez mountains as a backdrop to research and innovation covering multi-disciplines from bioscience, sustainable energy sources, to

  15. Highly Asynchronous VisitOr Queue Graph Toolkit

    Energy Science and Technology Software Center (OSTI)

    2012-10-01

    HAVOQGT is a C++ framework that can be used to create highly parallel graph traversal algorithms. The framework stores the graph and algorithmic data structures on external memory that is typically mapped to high performance locally attached NAND FLASH arrays. The framework supports a vertex-centered visitor programming model. The frameworkd has been used to implement breadth first search, connected components, and single source shortest path.

  16. computers

    National Nuclear Security Administration (NNSA)

    California.

    Retired computers used for cybersecurity research at Sandia National...

  17. Computer

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    I. INTRODUCTION This paper presents several computational tools required for processing images of a heavy ion beam and estimating the magnetic field within a plasma. The...

  18. JLab Registration/International Services - Researcher/Visitor ID

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Requirements Researcher/Visitor ID Requirements CHANGE IN ACCEPTABLE ID DOCUMENTS FOR JLAB ACCESS U.S. Citizens must bring a valid Government issued ID card that contains a photo such as a passport or valid driver's license. Non-driver photo identification cards issued by the Department of Motor Vehicles can be used as proof of identification. Lawful Permanent Resident of the United States must bring their Green Card or passport with valid I-551 stamp AND a valid government issued

  19. More than 480 visitors from 23 states take site tour during Secret City

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    Festival | Y-12 National Security Complex More than 480 visitors from ... More than 480 visitors from 23 states take site tour during Secret City Festival Posted: June 25, 2014 - 8:34am Tours of Y-12 on June 13 brought more than 480 visitors. Y-12 employee and tour guide Bill Wilburn, left, chats with Teresa Neas, and her husband Kim, whose mother, Audeane Bowers-Neas, worked at Y-12 in the 1940s. Tours of Y-12 on June 13 brought more than 480 visitors. Tourists from 23 states and the

  20. NP Committees of Visitors | U.S. DOE Office of Science (SC)

    Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    NP Committees of Visitors Nuclear Science Advisory Committee (NSAC) NSAC Home Meetings Members Charges/Reports Charter .pdf file (78KB) NP Committees of Visitors Federal Advisory Committees NP Home NP Committees of Visitors Print Text Size: A A A FeedbackShare Page DOE/NSF Nuclear Science Advisory Committee (NSAC) » The links below provide an archive of Nuclear Physics (NP) Committees of Visitors (COV) reports and responses. 2013 NSAC COV Report on NP Program .pdf file (1.8MB) NP Response to

    1. HEP Committees of Visitors | U.S. DOE Office of Science (SC)

      Office of Science (SC) Website

      Panel (HEPAP) HEPAP Home Meetings 2015 HEPAP Membership ChargesReports Charter .pdf file (44KB) HEP Committees of Visitors Federal Advisory Committees HEP Home HEP...

    2. ASCR Committees of Visitors | U.S. DOE Office of Science (SC...

      Office of Science (SC) Website

      (ASCAC) ASCAC Home Meetings Members ChargesReports ASCAC Charter 2015 - signed .pdf file (134KB) ASCR Committees of Visitors Federal Advisory Committees ASCR Home ASCR...

    3. BES Committees of Visitors | U.S. DOE Office of Science (SC)

      Office of Science (SC) Website

      (BESAC) BESAC Home Meetings BESAC 2015-2016 Membership ChargesReports Charter .pdf file (128KB) BES Committees of Visitors Federal Advisory Committees BES Home BES...

    4. Acclaim for Three Leaders at Annual NREL Stakeholders Reception - News

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Releases | NREL Acclaim for Three Leaders at Annual NREL Stakeholders Reception December 9, 2003 Golden, Colo. - The U.S. Department of Energy's National Renewable Energy Laboratory (NREL) will honor three distinguished Coloradoans today for their leadership in promoting renewable energy and energy efficiency. Lola Spradley, speaker, Colorado House of Representatives. Spradley has set a positive course for the Legislature in promoting the use of renewable energy in the state. Cal Marsella,

    5. DMA shared byte counters in a parallel computer

      DOE Patents [OSTI]

      Chen, Dong; Gara, Alan G.; Heidelberger, Philip; Vranas, Pavlos

      2010-04-06

      A parallel computer system is constructed as a network of interconnected compute nodes. Each of the compute nodes includes at least one processor, a memory and a DMA engine. The DMA engine includes a processor interface for interfacing with the at least one processor, DMA logic, a memory interface for interfacing with the memory, a DMA network interface for interfacing with the network, injection and reception byte counters, injection and reception FIFO metadata, and status registers and control registers. The injection FIFOs maintain memory locations of the injection FIFO metadata memory locations including its current head and tail, and the reception FIFOs maintain the reception FIFO metadata memory locations including its current head and tail. The injection byte counters and reception byte counters may be shared between messages.

    6. Computing

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Office of Advanced Scientific Computing Research in the Department of Energy Office of Science under contract number DE-AC02-05CH11231. ! Application and System Memory Use, Configuration, and Problems on Bassi Richard Gerber Lawrence Berkeley National Laboratory NERSC User Services ScicomP 13 Garching bei München, Germany, July 17, 2007 ScicomP 13, July 17, 2007, Garching Overview * About Bassi * Memory on Bassi * Large Page Memory (It's Great!) * System Configuration * Large Page

    7. Computations

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Computations - Sandia Energy Energy Search Icon Sandia Home Locations Contact Us Employee Locator Energy & Climate Secure & Sustainable Energy Future Stationary Power Energy Conversion Efficiency Solar Energy Wind Energy Water Power Supercritical CO2 Geothermal Natural Gas Safety, Security & Resilience of the Energy Infrastructure Energy Storage Nuclear Power & Engineering Grid Modernization Battery Testing Nuclear Fuel Cycle Defense Waste Management Programs Advanced Nuclear

    8. FES Committees of Visitors | U.S. DOE Office of Science (SC)

      Office of Science (SC) Website

      Committees of Visitors Fusion Energy Sciences (FES) FES Home About Organization Chart .pdf file (80KB) Staff FES Budget FES Committees of Visitors Directions Jobs Fusion and Plasmas Research Facilities Science Highlights Benefits of FES Funding Opportunities Fusion Energy Sciences Advisory Committee (FESAC) Community Resources Contact Information Fusion Energy Sciences U.S. Department of Energy SC-24/Germantown Building 1000 Independence Ave., SW Washington, DC 20585 P: (301) 903-4941 F: (301)

    9. HEP Committees of Visitors | U.S. DOE Office of Science (SC)

      Office of Science (SC) Website

      Committees of Visitors High Energy Physics (HEP) HEP Home About Organization Chart .pdf file (141KB) Staff HEP Budget HEP Committees of Visitors Directions Jobs University Research & National Labs Research Facilities Science Highlights Benefits of HEP Funding Opportunities Advisory Committees Community Resources Contact Information High Energy Physics U.S. Department of Energy SC-25/Germantown Building 1000 Independence Ave., SW Washington, DC 20585 P: (301) 903-3624 F: (301) 903-2597 E:

    10. NP Committees of Visitors | U.S. DOE Office of Science (SC)

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Committees of Visitors Nuclear Physics (NP) NP Home About Organization Chart .pdf file (31KB) Staff NP Budget NP Committees of Visitors Directions Jobs Labs & Universities Nuclear Physics Related Brochures Research Facilities Science Highlights Benefits of NP Funding Opportunities Nuclear Science Advisory Committee (NSAC) Community Resources Contact Information Nuclear Physics U.S. Department of Energy SC-26/Germantown Building 1000 Independence Ave., SW Washington, DC 20585 P: (301)

    11. U.S. Department of Energy Solar Decathlon Visitors Guide 2015

      SciTech Connect (OSTI)

      2015-09-03

      The U.S. Department of Energy 2015 Visitors Guide is a free, hard-copy publication distributed free to those attending the Solar Decathlon event. The publications' objectives are to serve as the primary information resource for those in attendance, and to deliver a compelling message about the Solar Decathlon's success as a proven workforce development program and its role in educating students and the public about clean energy products and design solutions. The U.S. Department of Energy 2015 Visitors Guide SD15 Visitors Guide goals are to guide attendees through the Solar Decathlon village; List and explain the 10 contests; educate attendees about the participating teams and their competition houses; provide access to more information on the Solar Decathlon website through the use of QR codes; and acknowledge the support of all event sponsors.

    12. Computing Videos

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Computing Videos Computing

    13. Y-12 has almost 500 visitors from 23 states tour site during Secret City

      National Nuclear Security Administration (NNSA)

      Festival | National Nuclear Security Administration has almost 500 visitors from 23 states tour site during Secret City Festival | National Nuclear Security Administration Facebook Twitter Youtube Flickr RSS People Mission Managing the Stockpile Preventing Proliferation Powering the Nuclear Navy Emergency Response Recapitalizing Our Infrastructure Countering Nuclear Terrorism About Our Programs Our History Who We Are Our Leadership Our Locations Budget Our Operations Library Bios

    14. ALS Visitors

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      February 7. After presentations about climate change research from Margaret Torn, Billy Collins, and others, the group toured the ALS with Director Roger Falcone, and stopped by...

    15. ALS Visitors

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      7. After presentations about climate change research from Margaret Torn, Billy Collins, and others, the group toured the ALS with Director Roger Falcone, and stopped by...

    16. Carrier-frequency synchronization system for improved amplitude modulation and television broadcast reception

      DOE Patents [OSTI]

      Smith, Stephen F.; Moore, James A.

      2003-05-13

      Systems and methods are described for carrier-frequency synchronization for improved AM and TV broadcast reception. A method includes synchronizing a carrier frequency of a broadcast signal with a remote reference frequency. An apparatus includes a reference signal receiver; a phase comparator coupled to the reference signal receiver; a voltage controlled oscillator coupled to the phase comparator; and a radio frequency output coupled to the voltage controlled oscillator.

    17. High Energy Theory Workshops and Visitors at the Michigan Center for Theoretical Physics FY14

      SciTech Connect (OSTI)

      Pierce, Aaron T.

      2014-04-01

      The workshop was held from September 23-25, 2013 on the University of Michigan campus. Local organizers were Dragan Huterer, Katherine Freese, and Heidi Wu (University of Michigan). Marilena Lo Verde (University of Chicago) also served as an external organizer. This workshop sought to gather experimentalists and theorists to discuss and define directions in cosmology research after the 1st year release of Planck data. The workshop included 35 invited (non-U-M) cosmologists, most of them relatively junior. The workshop was notable for spirited discussion of various theoretical ideas and experimental developments, and particularly on how one could test theory with ongoing and future experiments. In our follow-up poll, 95% of participants reported that interactions with other participants at the workshop may lead to further collaboration. Most participants (again about 95%) reported that they are very satisfied with the quality of the program, information they received, and the logistical support. Slides are available on line at: http://www.umich.edu/~mctp/SciPrgPgs/events/2013/CAP13/program.html. The YHET visitor program invited weekly young visitors to the University of Michigan campus to present their work. This year 23 participants came under the program. Slides are available on line for talks when applicable: http://mctp.physics.lsa.umich.edu/brown-bag-seminar-history/winter 2014 and http://mctp.physics.lsa.umich.edu/brown-bag-seminar-history/fall-2013.

    18. High Energy Theory Workshops and Visitors at the Michigan Center for Theoretical Physics FY15

      SciTech Connect (OSTI)

      Pierce, Aaron T.

      2015-09-18

      The String theory workshop was held from March 4-7, 2015 on the University of Michigan campus. Local organizers were Gordon Kane and Aaron Pierce. Piyush Kumar (Yale), Jim Halverson (KITP), Bobby Acharya (ICTP) and Sven Krippendorf (Oxford) served as external organizers.The meeting focused on the status of work to project 10 or 11 dimensional string/M theories onto our 4 spacetime dimensions (compactification). The workshop had 31 participants, half from outside the U.S. Participants were encouraged to focus on predictions for recent and forthcoming data, particularly for Higgs physics and LHC and dark matter, rather than on the traditional approach of embedding the Standard Model particles and forces. The Higgs boson sympoosium was locally organized by James Wells (chair), Aaron Pierce and Jianming Qian. Additional input in the early stages by Stefan Pokorski (Warsaw) who was unable to attend in the end. The workshop consistent of 22 talks from experts around the world, both theoretical and experimental. Experimentalists summarized the current state of knowledge of the Higgs boson and its varients. The theory talks ranged from technical calculations of Standard Model processes to speculative novel ideas. The YHET visitor program invited weekly young visitors to the University of Michigan campus to present their work. This year 24 participants came under the program, with 17 of them receiving at least partial support for their visits.

    19. Advanced Scientific Computing Advisory Committee (ASCAC) Homepage | U.S.

      Office of Science (SC) Website

      DOE Office of Science (SC) ASCAC Home Advanced Scientific Computing Advisory Committee (ASCAC) ASCAC Home Meetings Members Charges/Reports ASCAC Charter 2015 - signed .pdf file (134KB) ASCR Committees of Visitors Federal Advisory Committees ASCR Home Exascale Advisory Committee Report .pdf file (2.1MB) The Opportunities and Challenges of Exascale Computing The Exascale initiative will be significant and transformative for Department of Energy missions. The ASCAC Subcommitte report is

    20. Visitor center at the Antelope Valley California Poppy Reserve, Lancaster, California

      SciTech Connect (OSTI)

      Colyer, R.D.; Freeman, S.P.

      1981-01-01

      The Antelope Valley California Poppy Reserve contains the largest remaining stand of the California Poppy (Eschschozia Californica), the state flower of California. To welcome the thousands of people viewing the desert wildflowers each spring, the State of California decided to build a visitor/interpretive center. This building deals primarily with the question of fit; a building's fit aesthetically with its site and the fit of a building's design response to the climate of the site. In this case, both aspects of this question led the client and architects to seek an earth sheltered solution using materials at least metaphorically indigenous to the region. On both a technical and formal level, this building seeks to fit the unique climate and historical heritage of its site.

    1. Carrier phase synchronization system for improved amplitude modulation and television broadcast reception

      DOE Patents [OSTI]

      Smith, Stephen F. [Loudon, TN; Moore, James A. [Powell, TN

      2009-09-08

      Systems and methods are described for carrier phase synchronization for improved AM and TV broadcast reception. A method includes synchronizing the phase of a carrier frequency of a broadcast signal with the phase of a remote reference frequency. An apparatus includes a receiver to detect the phase of a reference signal; a phase comparator coupled to the reference signal-phase receiver; a voltage controlled oscillator coupled to the phase comparator; and a phase-controlled radio frequency output coupled to the voltage controlled oscillator.

    2. Carrier phase synchronization system for improved amplitude modulation and television broadcast reception

      DOE Patents [OSTI]

      Smith, Stephen F.; Moore, James A.

      2011-02-01

      Systems and methods are described for carrier phase synchronization for improved AM and TV broadcast reception. A method includes synchronizing the phase of a carrier frequency of a broadcast signal with the phase of a remote reference frequency. An apparatus includes a receiver to detect the phase of a reference signal; a phase comparator coupled to the reference signal-phase receiver; a voltage controlled oscillator coupled to the phase comparator; and a phase-controlled radio frequency output coupled to the voltage controlled oscillator.

    3. Solar space heating for the visitors' center, Stephens College, Columbia, Missouri. Final report

      SciTech Connect (OSTI)

      Henley, Marion

      1980-06-01

      This document is the final report of the solar energy system located at the Visitors' Center on the Stephens College Campus, Columbia, Missouri. The system is installed in a four-story, 15,000 square foot building designed to include the college's Admission Office, nine guest rooms for overnight lodging for official guests of the college, a two-story art gallery, and a Faculty Lounge. The solar energy system is an integral design of the building and utilizes 176 Honeywell/Lennox hydronic flat-plate collectors which use a 50% water-ethylene glycol solution and water-to-water heat exchanger. Solar heated water is stored in a 5000 gallon water storage tank located in the basement equipment room. A natural gas fired hot water boiler supplies hot water when the solar energy heat supply fails to meet the demand. The designed solar contribution is 71% of the heating load. The demonstration period for this project ends June 30, 1984.

    4. Visiting Us | Argonne Leadership Computing Facility

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Visitor Registration Form Visitor Registration Form As a national laboratory, formal registration is REQUIRED for all visitors coming to Argonne National Laboratory. Visitors must fill out the visitor registration form. If you require hotel accommodations while you are onsite, please contact the Argonne Guest House. You will need to bring your driver's license for entry to the Laboratory. If you are from a DOE Laboratory, please bring your badge with you. If you will be arriving Sunday, please

    5. Computation & Simulation > Theory & Computation > Research >...

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      it. Click above to view. computational2 computational3 In This Section Computation & Simulation Computation & Simulation Extensive combinatorial results and ongoing basic...

    6. Computer Science Research Institute 2005 annual report of activities.

      SciTech Connect (OSTI)

      Watts, Bernadette M.; Collis, Samuel Scott; Ceballos, Deanna Rose; Womble, David Eugene

      2008-04-01

      This report summarizes the activities of the Computer Science Research Institute (CSRI) at Sandia National Laboratories during the period January 1, 2005 to December 31, 2005. During this period, the CSRI hosted 182 visitors representing 83 universities, companies and laboratories. Of these, 60 were summer students or faculty. The CSRI partially sponsored 2 workshops and also organized and was the primary host for 3 workshops. These 3 CSRI sponsored workshops had 105 participants, 78 from universities, companies and laboratories, and 27 from Sandia. Finally, the CSRI sponsored 12 long-term collaborative research projects and 3 Sabbaticals.

    7. Computer Science Research Institute 2003 annual report of activities.

      SciTech Connect (OSTI)

      DeLap, Barbara J.; Womble, David Eugene; Ceballos, Deanna Rose

      2006-03-01

      This report summarizes the activities of the Computer Science Research Institute (CSRI) at Sandia National Laboratories during the period January 1, 2003 to December 31, 2003. During this period the CSRI hosted 164 visitors representing 78 universities, companies and laboratories. Of these 78 were summer students or faculty members. The CSRI partially sponsored 5 workshops and also organized and was the primary host for 3 workshops. These 3 CSRI sponsored workshops had 178 participants--137 from universities, companies and laboratories, and 41 from Sandia. Finally, the CSRI sponsored 18 long-term collaborative research projects and 5 Sabbaticals.

    8. Compute nodes

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Compute nodes Compute nodes Click here to see more detailed hierachical map of the topology of a compute node. Last edited: 2016-02-01 08:07:08

    9. Computing Information

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      here you can find information relating to: Obtaining the right computer accounts. Using NIC terminals. Using BooNE's Computing Resources, including: Choosing your desktop....

    10. Computer System,

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      undergraduate summer institute http:isti.lanl.gov (Educational Prog) 2016 Computer System, Cluster, and Networking Summer Institute Purpose The Computer System,...

    11. Computing and Computational Sciences Directorate - Computer Science...

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      AWARD Winners: Jess Gehin; Jackie Isaacs; Douglas Kothe; Debbie McCoy; Bonnie Nestor; John Turner; Gilbert Weigand Organization(s): Nuclear Technology Program; Computing and...

    12. Computing Sciences

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Division The Computational Research Division conducts research and development in mathematical modeling and simulation, algorithm design, data storage, management and...

    13. Computing Resources

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Cluster-Image TRACC RESEARCH Computational Fluid Dynamics Computational Structural Mechanics Transportation Systems Modeling Computing Resources The TRACC Computational Clusters With the addition of a new cluster called Zephyr that was made operational in September of this year (2012), TRACC now offers two clusters to choose from: Zephyr and our original cluster that has now been named Phoenix. Zephyr was acquired from Atipa technologies, and it is a 92-node system with each node having two AMD

    14. Compute Nodes

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Compute Nodes Compute Nodes Quad CoreAMDOpteronprocessor Compute Node Configuration 9,572 nodes 1 quad-core AMD 'Budapest' 2.3 GHz processor per node 4 cores per node (38,288 total cores) 8 GB DDR3 800 MHz memory per node Peak Gflop rate 9.2 Gflops/core 36.8 Gflops/node 352 Tflops for the entire machine Each core has their own L1 and L2 caches, with 64 KB and 512KB respectively 2 MB L3 cache shared among the 4 cores Compute Node Software By default the compute nodes run a restricted low-overhead

    15. Jefferson Lab Visitor's Center

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    16. Computer Science

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Cite Seer Department of Energy provided open access science research citations in chemistry, physics, materials, engineering, and computer science IEEE Xplore Full text...

    17. Computer Security

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Computer Security All JLF participants must fully comply with all LLNL computer security regulations and procedures. A laptop entering or leaving B-174 for the sole use by a US citizen and so configured, and requiring no IP address, need not be registered for use in the JLF. By September 2009, it is expected that computers for use by Foreign National Investigators will have no special provisions. Notify maricle1@llnl.gov of all other computers entering, leaving, or being moved within B 174. Use

    18. Computing and Computational Sciences Directorate - Divisions

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      CCSD Divisions Computational Sciences and Engineering Computer Sciences and Mathematics Information Technolgoy Services Joint Institute for Computational Sciences National Center for Computational Sciences

    19. Virtual Community Reception Center

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

    20. Compute Nodes

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Compute Nodes Compute Nodes There are currently 2632 nodes available on PDSF. The compute (batch) nodes at PDSF are heterogenous, reflecting the periodic procurement of new nodes (and the eventual retirement of old nodes). From the user's perspective they are essentially all equivalent except that some have more memory per job slot. If your jobs have memory requirements beyond the default maximum of 1.1GB you should specify that in your job submission and the batch system will run your job on an

    1. Compute Nodes

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Nodes Quad CoreAMDOpteronprocessor Compute Node Configuration 9,572 nodes 1 quad-core AMD 'Budapest' 2.3 GHz processor per node 4 cores per node (38,288 total cores) 8 GB...

    2. Exascale Computing

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Computing Exascale Computing CoDEx Project: A Hardware/Software Codesign Environment for the Exascale Era The next decade will see a rapid evolution of HPC node architectures as power and cooling constraints are limiting increases in microprocessor clock speeds and constraining data movement. Applications and algorithms will need to change and adapt as node architectures evolve. A key element of the strategy as we move forward is the co-design of applications, architectures and programming

    3. LHC Computing

      SciTech Connect (OSTI)

      Lincoln, Don

      2015-07-28

      The LHC is the world’s highest energy particle accelerator and scientists use it to record an unprecedented amount of data. This data is recorded in electronic format and it requires an enormous computational infrastructure to convert the raw data into conclusions about the fundamental rules that govern matter. In this video, Fermilab’s Dr. Don Lincoln gives us a sense of just how much data is involved and the incredible computer resources that makes it all possible.

    4. Computational mechanics

      SciTech Connect (OSTI)

      Goudreau, G.L.

      1993-03-01

      The Computational Mechanics thrust area sponsors research into the underlying solid, structural and fluid mechanics and heat transfer necessary for the development of state-of-the-art general purpose computational software. The scale of computational capability spans office workstations, departmental computer servers, and Cray-class supercomputers. The DYNA, NIKE, and TOPAZ codes have achieved world fame through our broad collaborators program, in addition to their strong support of on-going Lawrence Livermore National Laboratory (LLNL) programs. Several technology transfer initiatives have been based on these established codes, teaming LLNL analysts and researchers with counterparts in industry, extending code capability to specific industrial interests of casting, metalforming, and automobile crash dynamics. The next-generation solid/structural mechanics code, ParaDyn, is targeted toward massively parallel computers, which will extend performance from gigaflop to teraflop power. Our work for FY-92 is described in the following eight articles: (1) Solution Strategies: New Approaches for Strongly Nonlinear Quasistatic Problems Using DYNA3D; (2) Enhanced Enforcement of Mechanical Contact: The Method of Augmented Lagrangians; (3) ParaDyn: New Generation Solid/Structural Mechanics Codes for Massively Parallel Processors; (4) Composite Damage Modeling; (5) HYDRA: A Parallel/Vector Flow Solver for Three-Dimensional, Transient, Incompressible Viscous How; (6) Development and Testing of the TRIM3D Radiation Heat Transfer Code; (7) A Methodology for Calculating the Seismic Response of Critical Structures; and (8) Reinforced Concrete Damage Modeling.

    5. Computing and Computational Sciences Directorate - Contacts

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Home › About Us Contacts Jeff Nichols Associate Laboratory Director Computing and Computational Sciences Becky Verastegui Directorate Operations Manager Computing and Computational Sciences Directorate Michael Bartell Chief Information Officer Information Technologies Services Division Jim Hack Director, Climate Science Institute National Center for Computational Sciences Shaun Gleason Division Director Computational Sciences and Engineering Barney Maccabe Division Director Computer Science

    6. Compute Nodes

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Compute Nodes Compute Nodes MC-proc.png Compute Node Configuration 6,384 nodes 2 twelve-core AMD 'MagnyCours' 2.1-GHz processors per node (see die image to the right and schematic below) 24 cores per node (153,216 total cores) 32 GB DDR3 1333-MHz memory per node (6,000 nodes) 64 GB DDR3 1333-MHz memory per node (384 nodes) Peak Gflop/s rate: 8.4 Gflops/core 201.6 Gflops/node 1.28 Peta-flops for the entire machine Each core has its own L1 and L2 caches, with 64 KB and 512KB respectively One 6-MB

    7. Computational mechanics

      SciTech Connect (OSTI)

      Raboin, P J

      1998-01-01

      The Computational Mechanics thrust area is a vital and growing facet of the Mechanical Engineering Department at Lawrence Livermore National Laboratory (LLNL). This work supports the development of computational analysis tools in the areas of structural mechanics and heat transfer. Over 75 analysts depend on thrust area-supported software running on a variety of computing platforms to meet the demands of LLNL programs. Interactions with the Department of Defense (DOD) High Performance Computing and Modernization Program and the Defense Special Weapons Agency are of special importance as they support our ParaDyn project in its development of new parallel capabilities for DYNA3D. Working with DOD customers has been invaluable to driving this technology in directions mutually beneficial to the Department of Energy. Other projects associated with the Computational Mechanics thrust area include work with the Partnership for a New Generation Vehicle (PNGV) for ''Springback Predictability'' and with the Federal Aviation Administration (FAA) for the ''Development of Methodologies for Evaluating Containment and Mitigation of Uncontained Engine Debris.'' In this report for FY-97, there are five articles detailing three code development activities and two projects that synthesized new code capabilities with new analytic research in damage/failure and biomechanics. The article this year are: (1) Energy- and Momentum-Conserving Rigid-Body Contact for NIKE3D and DYNA3D; (2) Computational Modeling of Prosthetics: A New Approach to Implant Design; (3) Characterization of Laser-Induced Mechanical Failure Damage of Optical Components; (4) Parallel Algorithm Research for Solid Mechanics Applications Using Finite Element Analysis; and (5) An Accurate One-Step Elasto-Plasticity Algorithm for Shell Elements in DYNA3D.

    8. Computing Resources

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Resources This page is the repository for sundry items of information relevant to general computing on BooNE. If you have a question or problem that isn't answered here, or a suggestion for improving this page or the information on it, please mail boone-computing@fnal.gov and we'll do our best to address any issues. Note about this page Some links on this page point to www.everything2.com, and are meant to give an idea about a concept or thing without necessarily wading through a whole website

    9. RECIPIENT

      Office of Energy Efficiency and Renewable Energy (EERE) Indexed Site

      ... purposes; parking ; cafeteria services; education and training; visitor reception; ... the cost of testing the initial turbines and providing advice and mentoring to the RTCs. ...

    10. Advanced Scientific Computing Research

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Advanced Scientific Computing Research Advanced Scientific Computing Research Discovering, developing, and deploying computational and networking capabilities to analyze, model,...

    11. Sandia Energy - Computational Science

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Computational Science Home Energy Research Advanced Scientific Computing Research (ASCR) Computational Science Computational Sciencecwdd2015-03-26T13:35:2...

    12. Computer System,

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      System, Cluster, and Networking Summer Institute New Mexico Consortium and Los Alamos National Laboratory HOW TO APPLY Applications will be accepted JANUARY 5 - FEBRUARY 13, 2016 Computing and Information Technology undegraduate students are encouraged to apply. Must be a U.S. citizen. * Submit a current resume; * Offcial University Transcript (with spring courses posted and/or a copy of spring 2016 schedule) 3.0 GPA minimum; * One Letter of Recommendation from a Faculty Member; and * Letter of

    13. Computing Events

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Events Computing Events Spotlighting the most advanced scientific and technical applications in the world! Featuring exhibits of the latest and greatest technologies from industry, academia and government research organizations; many of these technologies will be seen for the first time in Denver. Supercomputing Conference 13 Denver, Colorado November 17-22, 2013 Spotlighting the most advanced scientific and technical applications in the world, SC13 will bring together the international

    14. Computing and Computational Sciences Directorate - Computer Science and

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Mathematics Division Computer Science and Mathematics Division The Computer Science and Mathematics Division (CSMD) is ORNL's premier source of basic and applied research in high-performance computing, applied mathematics, and intelligent systems. Our mission includes basic research in computational sciences and application of advanced computing systems, computational, mathematical and analysis techniques to the solution of scientific problems of national importance. We seek to work

    15. Computing at JLab

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Jefferson Lab Jefferson Lab Home Search Contact JLab Computing at JLab ---------------------- Accelerator Controls CAD CDEV CODA Computer Center High Performance Computing Scientific Computing JLab Computer Silo maintained by webmaster@jlab.org

    16. Computing Resources | Argonne Leadership Computing Facility

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Computing Resources Mira Cetus and Vesta Visualization Cluster Data and Networking Software JLSE Computing Resources Theory and Computing Sciences Building Argonne's Theory and Computing Sciences (TCS) building houses a wide variety of computing systems including some of the most powerful supercomputers in the world. The facility has 25,000 square feet of raised computer floor space and a pair of redundant 20 megavolt amperes electrical feeds from a 90 megawatt substation. The building also

    17. High Performance Computing

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      HPC INL Logo Home High-Performance Computing INL's high-performance computing center provides general use scientific computing capabilities to support the lab's efforts in advanced...

    18. Computer Architecture Lab

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      User Defined Images Archive APEX Home R & D Exascale Computing CAL Computer Architecture Lab The goal of the Computer Architecture Laboratory (CAL) is engage in...

    19. Executing scatter operation to parallel computer nodes by repeatedly broadcasting content of send buffer partition corresponding to each node upon bitwise OR operation

      DOE Patents [OSTI]

      Archer, Charles J. (Rochester, MN); Ratterman, Joseph D. (Rochester, MN)

      2009-11-06

      Executing a scatter operation on a parallel computer includes: configuring a send buffer on a logical root, the send buffer having positions, each position corresponding to a ranked node in an operational group of compute nodes and for storing contents scattered to that ranked node; and repeatedly for each position in the send buffer: broadcasting, by the logical root to each of the other compute nodes on a global combining network, the contents of the current position of the send buffer using a bitwise OR operation, determining, by each compute node, whether the current position in the send buffer corresponds with the rank of that compute node, if the current position corresponds with the rank, receiving the contents and storing the contents in a reception buffer of that compute node, and if the current position does not correspond with the rank, discarding the contents.

    20. Solar Decathlon 2011: The Visitors Perspective

      Broader source: Energy.gov [DOE]

      Ride along as we visit the 2011 Solar Decathlon competition and explore the Solar Village with fellow sightseers!

    1. Fermilab | Science at Fermilab | Computing | Grid Computing

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      In the early 2000s, members of Fermilab's Computing Division looked ahead to experiments like those at the Large Hadron Collider, which would collect more data than any computing ...

    2. Mira Computational Readiness Assessment | Argonne Leadership Computing

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Facility INCITE Program 5 Checks & 5 Tips for INCITE Mira Computational Readiness Assessment ALCC Program Director's Discretionary (DD) Program Early Science Program INCITE 2016 Projects ALCC 2015 Projects ESP Projects View All Projects Publications ALCF Tech Reports Industry Collaborations Mira Computational Readiness Assessment Assess your project's computational readiness for Mira A review of the following computational readiness points in relation to scaling, porting, I/O, memory

    3. LANSCE | User Resources | Visitor Registration

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Visit Registration Users must register online at least: 4 weeks prior to experiment for U.S. citizens 2 months prior to experiment for foreign nationals. Lujan Center WNR Visit Registration Visit Registration

    4. Sandia Energy - Computations

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Computations Home Transportation Energy Predictive Simulation of Engines Reacting Flow Applied Math & Software Computations ComputationsAshley Otero2015-10-30T02:18:51+00:00...

    5. Computer hardware fault administration

      DOE Patents [OSTI]

      Archer, Charles J. (Rochester, MN); Megerian, Mark G. (Rochester, MN); Ratterman, Joseph D. (Rochester, MN); Smith, Brian E. (Rochester, MN)

      2010-09-14

      Computer hardware fault administration carried out in a parallel computer, where the parallel computer includes a plurality of compute nodes. The compute nodes are coupled for data communications by at least two independent data communications networks, where each data communications network includes data communications links connected to the compute nodes. Typical embodiments carry out hardware fault administration by identifying a location of a defective link in the first data communications network of the parallel computer and routing communications data around the defective link through the second data communications network of the parallel computer.

    6. Molecular Science Computing | EMSL

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      computational and state-of-the-art experimental tools, providing a cross-disciplinary environment to further research. Additional Information Computing user policies Partners...

    7. Applied & Computational Math

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      & Computational Math - Sandia Energy Energy Search Icon Sandia Home Locations Contact Us ... Twitter Google + Vimeo GovDelivery SlideShare Applied & Computational Math HomeEnergy ...

    8. advanced simulation and computing

      National Nuclear Security Administration (NNSA)

      Each successive generation of computing system has provided greater computing power and energy efficiency.

      CTS-1 clusters will support NNSA's Life Extension Program and...

    9. NERSC Computer Security

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Security NERSC Computer Security NERSC computer security efforts are aimed at protecting NERSC systems and its users' intellectual property from unauthorized access or...

    10. Cosmic Reionization On Computers | Argonne Leadership Computing...

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      its Cosmic Reionization On Computers (CROC) project, using the Adaptive Refinement Tree (ART) code as its main simulation tool. An important objective of this research is to make...

    11. Computing and Computational Sciences Directorate - Information...

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      cost-effective, state-of-the-art computing capabilities for research and development. ... communicates and manages strategy, policy and finance across the portfolio of IT assets. ...

    12. Parallel computing works

      SciTech Connect (OSTI)

      Not Available

      1991-10-23

      An account of the Caltech Concurrent Computation Program (C{sup 3}P), a five year project that focused on answering the question: Can parallel computers be used to do large-scale scientific computations '' As the title indicates, the question is answered in the affirmative, by implementing numerous scientific applications on real parallel computers and doing computations that produced new scientific results. In the process of doing so, C{sup 3}P helped design and build several new computers, designed and implemented basic system software, developed algorithms for frequently used mathematical computations on massively parallel machines, devised performance models and measured the performance of many computers, and created a high performance computing facility based exclusively on parallel computers. While the initial focus of C{sup 3}P was the hypercube architecture developed by C. Seitz, many of the methods developed and lessons learned have been applied successfully on other massively parallel architectures.

    13. Computers-BSA.ppt

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Computers! Boy Scout Troop 405! What is a computer?! Is this a computer?! Charles Babbage: Father of the Computer! 1830s Designed mechanical calculators to reduce human error. *Input device *Memory to store instructions and results *A processors *Output device! Vacuum Tube! Edison 1883 & Lee de Forest 1906 discovered that "vacuum tubes" could serve as electrical switches and amplifiers A switch can be ON (1)" or OFF (0) Electronic computers use Boolean (George Bool 1850) logic

    14. Computational Fluid Dynamics

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      scour-tracc-cfd TRACC RESEARCH Computational Fluid Dynamics Computational Structural Mechanics Transportation Systems Modeling Computational Fluid Dynamics Overview of CFD: Video Clip with Audio Computational fluid dynamics (CFD) research uses mathematical and computational models of flowing fluids to describe and predict fluid response in problems of interest, such as the flow of air around a moving vehicle or the flow of water and sediment in a river. Coupled with appropriate and prototypical

    15. Theory & Computation > Research > The Energy Materials Center...

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Theory & Computation In This Section Computation & Simulation Theory & Computation Computation & Simulation...

    16. Polymorphous computing fabric

      DOE Patents [OSTI]

      Wolinski, Christophe Czeslaw (Los Alamos, NM); Gokhale, Maya B. (Los Alamos, NM); McCabe, Kevin Peter (Los Alamos, NM)

      2011-01-18

      Fabric-based computing systems and methods are disclosed. A fabric-based computing system can include a polymorphous computing fabric that can be customized on a per application basis and a host processor in communication with said polymorphous computing fabric. The polymorphous computing fabric includes a cellular architecture that can be highly parameterized to enable a customized synthesis of fabric instances for a variety of enhanced application performances thereof. A global memory concept can also be included that provides the host processor random access to all variables and instructions associated with the polymorphous computing fabric.

    17. Fermilab | Science at Fermilab | Computing | High-performance Computing

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Lattice QCD Farm at the Grid Computing Center at Fermilab. Lattice QCD Farm at the Grid Computing Center at Fermilab. Computing High-performance Computing A workstation computer can perform billions of multiplication and addition operations each second. High-performance parallel computing becomes necessary when computations become too large or too long to complete on a single such machine. In parallel computing, computations are divided up so that many computers can work on the same problem at

    18. Scientific Computing at Los Alamos National Laboratory (Conference...

      Office of Scientific and Technical Information (OSTI)

      for visitors (e.g., Bill Gates visit ; 2014-06-30 - 2014-06-30 ; los alamos, New Mexico, United States Research Org: Los Alamos National Laboratory (LANL) Sponsoring Org:...

    19. Computers in Commercial Buildings

      U.S. Energy Information Administration (EIA) Indexed Site

      Government-owned buildings of all types, had, on average, more than one computer per person (1,104 computers per thousand employees). They also had a fairly high ratio of...

    20. Computers for Learning

      Broader source: Energy.gov [DOE]

      Through Executive Order 12999, the Computers for Learning Program was established to provide Federal agencies a quick and easy system for donating excess and surplus computer equipment to schools...

    1. Cognitive Computing for Security.

      SciTech Connect (OSTI)

      Debenedictis, Erik; Rothganger, Fredrick; Aimone, James Bradley; Marinella, Matthew; Evans, Brian Robert; Warrender, Christina E.; Mickel, Patrick

      2015-12-01

      Final report for Cognitive Computing for Security LDRD 165613. It reports on the development of hybrid of general purpose/ne uromorphic computer architecture, with an emphasis on potential implementation with memristors.

    2. Getting Computer Accounts

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Computer Accounts When you first arrive at the lab, you will be presented with lots of forms that must be read and signed in order to get an ID and computer access. You must ensure...

    3. Advanced Scientific Computing Research

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Advanced Scientific Computing Research Advanced Scientific Computing Research Discovering, developing, and deploying computational and networking capabilities to analyze, model, simulate, and predict complex phenomena important to the Department of Energy. Get Expertise Pieter Swart (505) 665 9437 Email Pat McCormick (505) 665-0201 Email Dave Higdon (505) 667-2091 Email Fulfilling the potential of emerging computing systems and architectures beyond today's tools and techniques to deliver

    4. Computational Structural Mechanics

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      load-2 TRACC RESEARCH Computational Fluid Dynamics Computational Structural Mechanics Transportation Systems Modeling Computational Structural Mechanics Overview of CSM Computational structural mechanics is a well-established methodology for the design and analysis of many components and structures found in the transportation field. Modern finite-element models (FEMs) play a major role in these evaluations, and sophisticated software, such as the commercially available LS-DYNA® code, is

    5. Computing and Computational Sciences Directorate - Information Technology

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Computational Sciences and Engineering The Computational Sciences and Engineering Division (CSED) is ORNL's premier source of basic and applied research in the field of data sciences and knowledge discovery. CSED's science agenda is focused on research and development related to knowledge discovery enabled by the explosive growth in the availability, size, and variability of dynamic and disparate data sources. This science agenda encompasses data sciences as well as advanced modeling and

    6. BNL ATLAS Grid Computing

      ScienceCinema (OSTI)

      Michael Ernst

      2010-01-08

      As the sole Tier-1 computing facility for ATLAS in the United States and the largest ATLAS computing center worldwide Brookhaven provides a large portion of the overall computing resources for U.S. collaborators and serves as the central hub for storing,

    7. Computing environment logbook

      DOE Patents [OSTI]

      Osbourn, Gordon C; Bouchard, Ann M

      2012-09-18

      A computing environment logbook logs events occurring within a computing environment. The events are displayed as a history of past events within the logbook of the computing environment. The logbook provides search functionality to search through the history of past events to find one or more selected past events, and further, enables an undo of the one or more selected past events.

    8. Mathematical and Computational Epidemiology

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Mathematical and Computational Epidemiology Search Site submit Contacts | Sponsors Mathematical and Computational Epidemiology Los Alamos National Laboratory change this image and alt text Menu About Contact Sponsors Research Agent-based Modeling Mixing Patterns, Social Networks Mathematical Epidemiology Social Internet Research Uncertainty Quantification Publications People Mathematical and Computational Epidemiology (MCEpi) Quantifying model uncertainty in agent-based simulations for

    9. Scalable optical quantum computer

      SciTech Connect (OSTI)

      Manykin, E A; Mel'nichenko, E V [Institute for Superconductivity and Solid-State Physics, Russian Research Centre 'Kurchatov Institute', Moscow (Russian Federation)

      2014-12-31

      A way of designing a scalable optical quantum computer based on the photon echo effect is proposed. Individual rare earth ions Pr{sup 3+}, regularly located in the lattice of the orthosilicate (Y{sub 2}SiO{sub 5}) crystal, are suggested to be used as optical qubits. Operations with qubits are performed using coherent and incoherent laser pulses. The operation protocol includes both the method of measurement-based quantum computations and the technique of optical computations. Modern hybrid photon echo protocols, which provide a sufficient quantum efficiency when reading recorded states, are considered as most promising for quantum computations and communications. (quantum computer)

    10. COMPUTATIONAL SCIENCE CENTER

      SciTech Connect (OSTI)

      DAVENPORT, J.

      2005-11-01

      The Brookhaven Computational Science Center brings together researchers in biology, chemistry, physics, and medicine with applied mathematicians and computer scientists to exploit the remarkable opportunities for scientific discovery which have been enabled by modern computers. These opportunities are especially great in computational biology and nanoscience, but extend throughout science and technology and include, for example, nuclear and high energy physics, astrophysics, materials and chemical science, sustainable energy, environment, and homeland security. To achieve our goals we have established a close alliance with applied mathematicians and computer scientists at Stony Brook and Columbia Universities.

    11. Sandia Energy - High Performance Computing

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      High Performance Computing Home Energy Research Advanced Scientific Computing Research (ASCR) High Performance Computing High Performance Computingcwdd2015-03-18T21:41:24+00:00...

    12. COMPUTATIONAL SCIENCE CENTER

      SciTech Connect (OSTI)

      DAVENPORT, J.

      2006-11-01

      Computational Science is an integral component of Brookhaven's multi science mission, and is a reflection of the increased role of computation across all of science. Brookhaven currently has major efforts in data storage and analysis for the Relativistic Heavy Ion Collider (RHIC) and the ATLAS detector at CERN, and in quantum chromodynamics. The Laboratory is host for the QCDOC machines (quantum chromodynamics on a chip), 10 teraflop/s computers which boast 12,288 processors each. There are two here, one for the Riken/BNL Research Center and the other supported by DOE for the US Lattice Gauge Community and other scientific users. A 100 teraflop/s supercomputer will be installed at Brookhaven in the coming year, managed jointly by Brookhaven and Stony Brook, and funded by a grant from New York State. This machine will be used for computational science across Brookhaven's entire research program, and also by researchers at Stony Brook and across New York State. With Stony Brook, Brookhaven has formed the New York Center for Computational Science (NYCCS) as a focal point for interdisciplinary computational science, which is closely linked to Brookhaven's Computational Science Center (CSC). The CSC has established a strong program in computational science, with an emphasis on nanoscale electronic structure and molecular dynamics, accelerator design, computational fluid dynamics, medical imaging, parallel computing and numerical algorithms. We have been an active participant in DOES SciDAC program (Scientific Discovery through Advanced Computing). We are also planning a major expansion in computational biology in keeping with Laboratory initiatives. Additional laboratory initiatives with a dependence on a high level of computation include the development of hydrodynamics models for the interpretation of RHIC data, computational models for the atmospheric transport of aerosols, and models for combustion and for energy utilization. The CSC was formed to bring together researchers in these areas and to provide a focal point for the development of computational expertise at the Laboratory. These efforts will connect to and support the Department of Energy's long range plans to provide Leadership class computing to researchers throughout the Nation. Recruitment for six new positions at Stony Brook to strengthen its computational science programs is underway. We expect some of these to be held jointly with BNL.

    13. ARM - Facility News Article

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      facility use by total visitor days and facility to track actual visitors and active user research computer accounts. Historical data show an apparent relationship between the...

    14. Edison Electrifies Scientific Computing

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Edison Electrifies Scientific Computing Edison Electrifies Scientific Computing NERSC Flips Switch on New Flagship Supercomputer January 31, 2014 Contact: Margie Wylie, mwylie@lbl.gov, +1 510 486 7421 The National Energy Research Scientific Computing (NERSC) Center recently accepted "Edison," a new flagship supercomputer designed for scientific productivity. Named in honor of American inventor Thomas Alva Edison, the Cray XC30 will be dedicated in a ceremony held at the Department of

    15. Energy Aware Computing

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Partnerships Shifter: User Defined Images Archive APEX Home » R & D » Energy Aware Computing Energy Aware Computing Dynamic Frequency Scaling One means to lower the energy required to compute is to reduce the power usage on a node. One way to accomplish this is by lowering the frequency at which the CPU operates. However, reducing the clock speed increases the time to solution, creating a potential tradeoff. NERSC continues to examine how such methods impact its operations and its

    16. Personal Computer Inventory System

      Energy Science and Technology Software Center (OSTI)

      1993-10-04

      PCIS is a database software system that is used to maintain a personal computer hardware and software inventory, track transfers of hardware and software, and provide reports.

    17. Applied Computer Science

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Results from a climate simulation computed using the Model for Prediction Across Scales (MPAS) code. This visualization shows the temperature of ocean currents using a green and ...

    18. Announcement of Computer Software

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      All Other Editions Are Obsolete UNITED STATES DEPARTMENT OF ENERGY ANNOUNCEMENT OF COMPUTER SOFTWARE OMB Control Number 1910-1400 (OMB Burden Disclosure Statement is on last...

    19. 60 Years of Computing | Department of Energy

      Office of Energy Efficiency and Renewable Energy (EERE) Indexed Site

      60 Years of Computing 60 Years of Computing

    20. Information Science, Computing, Applied Math

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Capabilities Information Science, Computing, Applied Math science-innovationassetsimagesicon-science.jpg Information Science, Computing, Applied Math National security ...

    1. Computer Processor Allocator

      Energy Science and Technology Software Center (OSTI)

      2004-03-01

      The Compute Processor Allocator (CPA) provides an efficient and reliable mechanism for managing and allotting processors in a massively parallel (MP) computer. It maintains information in a database on the health. configuration and allocation of each processor. This persistent information is factored in to each allocation decision. The CPA runs in a distributed fashion to avoid a single point of failure.

    2. SC e-journals, Computer Science

      Office of Scientific and Technical Information (OSTI)

      & Mathematical Organization Theory Computational Complexity Computational Economics Computational Management ... Technology EURASIP Journal on Information Security ...

    3. Computers as tools

      SciTech Connect (OSTI)

      Eriksson, I.V.

      1994-12-31

      The following message was recently posted on a bulletin board and clearly shows the relevance of the conference theme: {open_quotes}The computer and digital networks seem poised to change whole regions of human activity -- how we record knowledge, communicate, learn, work, understand ourselves and the world. What`s the best framework for understanding this digitalization, or virtualization, of seemingly everything? ... Clearly, symbolic tools like the alphabet, book, and mechanical clock have changed some of our most fundamental notions -- self, identity, mind, nature, time, space. Can we say what the computer, a purely symbolic {open_quotes}machine,{close_quotes} is doing to our thinking in these areas? Or is it too early to say, given how much more powerful and less expensive the technology seems destinated to become in the next few decades?{close_quotes} (Verity, 1994) Computers certainly affect our lives and way of thinking but what have computers to do with ethics? A narrow approach would be that on the one hand people can and do abuse computer systems and on the other hand people can be abused by them. Weli known examples of the former are computer comes such as the theft of money, services and information. The latter can be exemplified by violation of privacy, health hazards and computer monitoring. Broadening the concept from computers to information systems (ISs) and information technology (IT) gives a wider perspective. Computers are just the hardware part of information systems which also include software, people and data. Information technology is the concept preferred today. It extends to communication, which is an essential part of information processing. Now let us repeat the question: What has IT to do with ethics? Verity mentioned changes in {open_quotes}how we record knowledge, communicate, learn, work, understand ourselves and the world{close_quotes}.

    4. Applications of Parallel Computers

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Computers Applications of Parallel Computers UCB CS267 Spring 2015 Tuesday & Thursday, 9:30-11:00 Pacific Time Applications of Parallel Computers, CS267, is a graduate-level course offered at the University of California, Berkeley. The course is being taught by UC Berkeley professor and LBNL Faculty Scientist Jim Demmel. CS267 is broadcast live over the internet and all NERSC users are invited to monitor the broadcast course, but course credit is available only to student registered for the

    5. Quantum steady computation

      SciTech Connect (OSTI)

      Castagnoli, G. )

      1991-08-10

      This paper reports that current conceptions of quantum mechanical computers inherit from conventional digital machines two apparently interacting features, machine imperfection and temporal development of the computational process. On account of machine imperfection, the process would become ideally reversible only in the limiting case of zero speed. Therefore the process is irreversible in practice and cannot be considered to be a fundamental quantum one. By giving up classical features and using a linear, reversible and non-sequential representation of the computational process - not realizable in classical machines - the process can be identified with the mathematical form of a quantum steady state. This form of steady quantum computation would seem to have an important bearing on the notion of cognition.

    6. Cloud computing security.

      SciTech Connect (OSTI)

      Shin, Dongwan; Claycomb, William R.; Urias, Vincent E.

      2010-10-01

      Cloud computing is a paradigm rapidly being embraced by government and industry as a solution for cost-savings, scalability, and collaboration. While a multitude of applications and services are available commercially for cloud-based solutions, research in this area has yet to fully embrace the full spectrum of potential challenges facing cloud computing. This tutorial aims to provide researchers with a fundamental understanding of cloud computing, with the goals of identifying a broad range of potential research topics, and inspiring a new surge in research to address current issues. We will also discuss real implementations of research-oriented cloud computing systems for both academia and government, including configuration options, hardware issues, challenges, and solutions.

    7. Theory, Modeling and Computation

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      modeling and simulation will be enhanced not only by the wealth of data available from MaRIE but by the increased computational capacity made possible by the advent of extreme...

    8. Argonne Leadership Computing Facility

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      a n n u a l r e p o r t 2 0 1 2 Argonne Leadership Computing Facility Director's Message .............................................................................................................................1 About ALCF ......................................................................................................................................... 2 IntroDuCIng MIrA Introducing Mira

    9. Applied Computer Science

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      ADTSC » CCS » CCS-7 Applied Computer Science Innovative co-design of applications, algorithms, and architectures in order to enable scientific simulations at extreme scale Leadership Group Leader Linn Collins Email Deputy Group Leader (Acting) Bryan Lally Email Climate modeling visualization Results from a climate simulation computed using the Model for Prediction Across Scales (MPAS) code. This visualization shows the temperature of ocean currents using a green and blue color scale. These

    10. Computational Earth Science

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      6 Computational Earth Science We develop and apply a range of high-performance computational methods and software tools to Earth science projects in support of environmental health, cleaner energy, and national security. Contact Us Group Leader Carl Gable Deputy Group Leader Gilles Bussod Email Profile pages header Search our Profile pages Hari Viswanathan inspects a microfluidic cell used to study the extraction of hydrocarbon fuels from a complex fracture network. EES-16's Subsurface Flow

    11. Computational Physics and Methods

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      2 Computational Physics and Methods Performing innovative simulations of physics phenomena on tomorrow's scientific computing platforms Growth and emissivity of young galaxy hosting a supermassive black hole as calculated in cosmological code ENZO and post-processed with radiative transfer code AURORA. image showing detailed turbulence simulation, Rayleigh-Taylor Turbulence imaging: the largest turbulence simulations to date Advanced multi-scale modeling Turbulence datasets Density iso-surfaces

    12. Compute Reservation Request Form

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Compute Reservation Request Form Compute Reservation Request Form Users can request a scheduled reservation of machine resources if their jobs have special needs that cannot be accommodated through the regular batch system. A reservation brings some portion of the machine to a specific user or project for an agreed upon duration. Typically this is used for interactive debugging at scale or real time processing linked to some experiment or event. It is not intended to be used to guarantee fast

    13. New TRACC Cluster Computer

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      TRACC Cluster Computer With the addition of a new cluster called Zephyr that was made operational in September of this year (2012), TRACC now offers two clusters to choose from: Zephyr and our original cluster that has now been named Phoenix. Zephyr was acquired from Atipa technologies, and it is a 92-node system with each node having two AMD 16 core, 2.3 GHz, 32 GB processors. See also Computing Resources.

    14. Advanced Simulation and Computing

      National Nuclear Security Administration (NNSA)

      NA-ASC-117R-09-Vol.1-Rev.0 Advanced Simulation and Computing PROGRAM PLAN FY09 October 2008 ASC Focal Point Robert Meisner, Director DOE/NNSA NA-121.2 202-586-0908 Program Plan Focal Point for NA-121.2 Njema Frazier DOE/NNSA NA-121.2 202-586-5789 A Publication of the Office of Advanced Simulation & Computing, NNSA Defense Programs i Contents Executive Summary ----------------------------------------------------------------------------------------------- 1 I. Introduction

    15. Computing | Department of Energy

      Office of Energy Efficiency and Renewable Energy (EERE) Indexed Site

      Computing Computing Fun fact: Most systems require air conditioning or chilled water to cool super powerful supercomputers, but the Olympus supercomputer at Pacific Northwest National Laboratory is cooled by the location's 65 degree groundwater. Traditional cooling systems could cost up to $61,000 in electricity each year, but this more efficient setup uses 70 percent less energy. | Photo courtesy of PNNL. Fun fact: Most systems require air conditioning or chilled water to cool super powerful

    16. Can Cloud Computing Address the Scientific Computing Requirements for DOE

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Researchers? Well, Yes, No and Maybe Can Cloud Computing Address the Scientific Computing Requirements for DOE Researchers? Well, Yes, No and Maybe Can Cloud Computing Address the Scientific Computing Requirements for DOE Researchers? Well, Yes, No and Maybe January 30, 2012 Jon Bashor, Jbashor@lbl.gov, +1 510-486-5849 Magellan1.jpg Magellan at NERSC After a two-year study of the feasibility of cloud computing systems for meeting the ever-increasing computational needs of scientists,

    17. Computing and Computational Sciences Directorate - Joint Institute for

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Computational Sciences Joint Institute for Computational Sciences To help realize the full potential of new-generation computers for advancing scientific discovery, the University of Tennessee (UT) and Oak Ridge National Laboratory (ORNL) have created the Joint Institute for Computational Sciences (JICS). JICS combines the experience and expertise in theoretical and computational science and engineering, computer science, and mathematics in these two institutions and focuses these skills on

    18. Computing and Computational Sciences Directorate - National Center for

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Computational Sciences Home National Center for Computational Sciences The National Center for Computational Sciences (NCCS), formed in 1992, is home to two of Oak Ridge National Laboratory's (ORNL's) high-performance computing projects-the Oak Ridge Leadership Computing Facility (OLCF) and the National Climate-Computing Research Center (NCRC). The OLCF (www.olcf.ornl.gov) was established at ORNL in 2004 with the mission of standing up a supercomputer 100 times more powerful than the leading

    19. in High Performance Computing Computer System, Cluster, and Networking...

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      iSSH v. Auditd: Intrusion Detection in High Performance Computing Computer System, Cluster, and Networking Summer Institute David Karns, New Mexico State University Katy Protin,...

    20. Extensible Computational Chemistry Environment

      Energy Science and Technology Software Center (OSTI)

      2012-08-09

      ECCE provides a sophisticated graphical user interface, scientific visualization tools, and the underlying data management framework enabling scientists to efficiently set up calculations and store, retrieve, and analyze the rapidly growing volumes of data produced by computational chemistry studies. ECCE was conceived as part of the Environmental Molecular Sciences Laboratory construction to solve the problem of researchers being able to effectively utilize complex computational chemistry codes and massively parallel high performance compute resources. Bringing themore » power of these codes and resources to the desktops of researcher and thus enabling world class research without users needing a detailed understanding of the inner workings of either the theoretical codes or the supercomputers needed to run them was a grand challenge problem in the original version of the EMSL. ECCE allows collaboration among researchers using a web-based data repository where the inputs and results for all calculations done within ECCE are organized. ECCE is a first of kind end-to-end problem solving environment for all phases of computational chemistry research: setting up calculations with sophisticated GUI and direct manipulation visualization tools, submitting and monitoring calculations on remote high performance supercomputers without having to be familiar with the details of using these compute resources, and performing results visualization and analysis including creating publication quality images. ECCE is a suite of tightly integrated applications that are employed as the user moves through the modeling process.« less

    1. Information Science, Computing, Applied Math

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Capabilities » Information Science, Computing, Applied Math /science-innovation/_assets/images/icon-science.jpg Information Science, Computing, Applied Math National security depends on science and technology. The United States relies on Los Alamos National Laboratory for the best of both. No place on Earth pursues a broader array of world-class scientific endeavors. Computer, Computational, and Statistical Sciences (CCS)» High Performance Computing (HPC)» Extreme Scale Computing, Co-design»

    2. Super recycled water: quenching computers

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Super recycled water: quenching computers Super recycled water: quenching computers New facility and methods support conserving water and creating recycled products. Using reverse...

    3. Computer simulation | Open Energy Information

      Open Energy Info (EERE)

      Computer simulation Jump to: navigation, search OpenEI Reference LibraryAdd to library Web Site: Computer simulation Author wikipedia Published wikipedia, 2013 DOI Not Provided...

    4. NREL: Computational Science Home Page

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      high-performance computing, computational science, applied mathematics, scientific data management, visualization, and informatics. NREL is home to the largest high performance...

    5. SCC: The Strategic Computing Complex

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      computer room, which is an open room about three-fourths the size of a football field. The Strategic Computing Complex (SCC) at the Los Alamos National Laboratory...

    6. Human-computer interface

      DOE Patents [OSTI]

      Anderson, Thomas G.

      2004-12-21

      The present invention provides a method of human-computer interfacing. Force feedback allows intuitive navigation and control near a boundary between regions in a computer-represented space. For example, the method allows a user to interact with a virtual craft, then push through the windshield of the craft to interact with the virtual world surrounding the craft. As another example, the method allows a user to feel transitions between different control domains of a computer representation of a space. The method can provide for force feedback that increases as a user's locus of interaction moves near a boundary, then perceptibly changes (e.g., abruptly drops or changes direction) when the boundary is traversed.

    7. Synchronizing compute node time bases in a parallel computer

      DOE Patents [OSTI]

      Chen, Dong; Faraj, Daniel A; Gooding, Thomas M; Heidelberger, Philip

      2014-12-30

      Synchronizing time bases in a parallel computer that includes compute nodes organized for data communications in a tree network, where one compute node is designated as a root, and, for each compute node: calculating data transmission latency from the root to the compute node; configuring a thread as a pulse waiter; initializing a wakeup unit; and performing a local barrier operation; upon each node completing the local barrier operation, entering, by all compute nodes, a global barrier operation; upon all nodes entering the global barrier operation, sending, to all the compute nodes, a pulse signal; and for each compute node upon receiving the pulse signal: waking, by the wakeup unit, the pulse waiter; setting a time base for the compute node equal to the data transmission latency between the root node and the compute node; and exiting the global barrier operation.

    8. Synchronizing compute node time bases in a parallel computer

      DOE Patents [OSTI]

      Chen, Dong; Faraj, Daniel A; Gooding, Thomas M; Heidelberger, Philip

      2015-01-27

      Synchronizing time bases in a parallel computer that includes compute nodes organized for data communications in a tree network, where one compute node is designated as a root, and, for each compute node: calculating data transmission latency from the root to the compute node; configuring a thread as a pulse waiter; initializing a wakeup unit; and performing a local barrier operation; upon each node completing the local barrier operation, entering, by all compute nodes, a global barrier operation; upon all nodes entering the global barrier operation, sending, to all the compute nodes, a pulse signal; and for each compute node upon receiving the pulse signal: waking, by the wakeup unit, the pulse waiter; setting a time base for the compute node equal to the data transmission latency between the root node and the compute node; and exiting the global barrier operation.

    9. Computer Security Risk Assessment

      Energy Science and Technology Software Center (OSTI)

      1992-02-11

      LAVA/CS (LAVA for Computer Security) is an application of the Los Alamos Vulnerability Assessment (LAVA) methodology specific to computer and information security. The software serves as a generic tool for identifying vulnerabilities in computer and information security safeguards systems. Although it does not perform a full risk assessment, the results from its analysis may provide valuable insights into security problems. LAVA/CS assumes that the system is exposed to both natural and environmental hazards and tomore » deliberate malevolent actions by either insiders or outsiders. The user in the process of answering the LAVA/CS questionnaire identifies missing safeguards in 34 areas ranging from password management to personnel security and internal audit practices. Specific safeguards protecting a generic set of assets (or targets) from a generic set of threats (or adversaries) are considered. There are four generic assets: the facility, the organization''s environment; the hardware, all computer-related hardware; the software, the information in machine-readable form stored both on-line or on transportable media; and the documents and displays, the information in human-readable form stored as hard-copy materials (manuals, reports, listings in full-size or microform), film, and screen displays. Two generic threats are considered: natural and environmental hazards, storms, fires, power abnormalities, water and accidental maintenance damage; and on-site human threats, both intentional and accidental acts attributable to a perpetrator on the facility''s premises.« less

    10. MHD computations for stellarators

      SciTech Connect (OSTI)

      Johnson, J.L.

      1985-12-01

      Considerable progress has been made in the development of computational techniques for studying the magnetohydrodynamic equilibrium and stability properties of three-dimensional configurations. Several different approaches have evolved to the point where comparison of results determined with different techniques shows good agreement. 55 refs., 7 figs.

    11. Sandia National Laboratories: Advanced Simulation and Computing...

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      ASC Advanced Simulation and Computing Computational Systems & Software Environment Crack Modeling The Computational Systems & Software Environment program builds integrated,...

    12. Extreme Scale Computing, Co-design

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Information Science, Computing, Applied Math Extreme Scale Computing, Co-design Extreme Scale Computing, Co-design Computational co-design may facilitate revolutionary designs ...

    13. Software and High Performance Computing

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Software and High Performance Computing Software and High Performance Computing Providing world-class high performance computing capability that enables unsurpassed solutions to complex problems of strategic national interest Contact thumbnail of Kathleen McDonald Head of Intellectual Property, Business Development Executive Kathleen McDonald Richard P. Feynman Center for Innovation (505) 667-5844 Email Software Computational physics, computer science, applied mathematics, statistics and the

    14. Magellan: A Cloud Computing Testbed

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Magellan News & Announcements Archive Petascale Initiative Exascale Computing APEX Home » R & D » Archive » Magellan: A Cloud Computing Testbed Magellan: A Cloud Computing Testbed Cloud computing is gaining a foothold in the business world, but can clouds meet the specialized needs of scientists? That was one of the questions NERSC's Magellan cloud computing testbed explored between 2009 and 2011. The goal of Magellan, a project funded through the U.S. Department of Energy (DOE) Oce

    15. Computer Algebra System

      Energy Science and Technology Software Center (OSTI)

      1992-05-04

      DOE-MACSYMA (Project MAC''s SYmbolic MAnipulation system) is a large computer programming system written in LISP. With DOE-MACSYMA the user can differentiate, integrate, take limits, solve systems of linear or polynomial equations, factor polynomials, expand functions in Laurent or Taylor series, solve differential equations (using direct or transform methods), compute Poisson series, plot curves, and manipulate matrices and tensors. A language similar to ALGOL-60 permits users to write their own programs for transforming symbolic expressions. Franzmore » Lisp OPUS 38 provides the environment for the Encore, Celerity, and DEC VAX11 UNIX,SUN(OPUS) versions under UNIX and the Alliant version under Concentrix. Kyoto Common Lisp (KCL) provides the environment for the SUN(KCL),Convex, and IBM PC under UNIX and Data General under AOS/VS.« less

    16. Exploratory Experimentation and Computation

      SciTech Connect (OSTI)

      Bailey, David H.; Borwein, Jonathan M.

      2010-02-25

      We believe the mathematical research community is facing a great challenge to re-evaluate the role of proof in light of recent developments. On one hand, the growing power of current computer systems, of modern mathematical computing packages, and of the growing capacity to data-mine on the Internet, has provided marvelous resources to the research mathematician. On the other hand, the enormous complexity of many modern capstone results such as the Poincare conjecture, Fermat's last theorem, and the classification of finite simple groups has raised questions as to how we can better ensure the integrity of modern mathematics. Yet as the need and prospects for inductive mathematics blossom, the requirement to ensure the role of proof is properly founded remains undiminished.

    17. GPU Computational Screening

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      GPU Computational Screening of Carbon Capture Materials J. Kim 1 , A Koniges 1 , R. Martin 1 , M. Haranczyk 1 , J. Swisher 2 , and B. Smit 1,2 1 Lawrence Berkeley National Laboratory, Berkeley, CA 94720 2 Department of Chemical Engineering, University of California, Berkeley, Berkeley, CA 94720 E-mail: jihankim@lbl.gov Abstract. In order to reduce the current costs associated with carbon capture technologies, novel materials such as zeolites and metal-organic frameworks that are based on

    18. Cloud Computing Services

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Computing Services - Sandia Energy Energy Search Icon Sandia Home Locations Contact Us Employee Locator Energy & Climate Secure & Sustainable Energy Future Stationary Power Energy Conversion Efficiency Solar Energy Wind Energy Water Power Supercritical CO2 Geothermal Natural Gas Safety, Security & Resilience of the Energy Infrastructure Energy Storage Nuclear Power & Engineering Grid Modernization Battery Testing Nuclear Fuel Cycle Defense Waste Management Programs Advanced

    19. High Performance Computing

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Performance Computing - Sandia Energy Energy Search Icon Sandia Home Locations Contact Us Employee Locator Energy & Climate Secure & Sustainable Energy Future Stationary Power Energy Conversion Efficiency Solar Energy Wind Energy Water Power Supercritical CO2 Geothermal Natural Gas Safety, Security & Resilience of the Energy Infrastructure Energy Storage Nuclear Power & Engineering Grid Modernization Battery Testing Nuclear Fuel Cycle Defense Waste Management Programs Advanced

    20. Argonne Leadership Computing Facility

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Anti-HIV antibody Software optimized on Mira advances design of mini-proteins for medicines, materials Scientists at the University of Washington are using Mira to virtually design unique, artificial peptides, or short proteins. Read More Celebrating 10 years 10 science highlights celebrating 10 years of Argonne Leadership Computing Facility To celebrate our 10th anniversary, we're highlighting 10 science accomplishments since we opened our doors. Read More Bill Gropp works with students during

    1. Applied & Computational Math

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      & Computational Math - Sandia Energy Energy Search Icon Sandia Home Locations Contact Us Employee Locator Energy & Climate Secure & Sustainable Energy Future Stationary Power Energy Conversion Efficiency Solar Energy Wind Energy Water Power Supercritical CO2 Geothermal Natural Gas Safety, Security & Resilience of the Energy Infrastructure Energy Storage Nuclear Power & Engineering Grid Modernization Battery Testing Nuclear Fuel Cycle Defense Waste Management Programs Advanced

    2. From Federal Computer Week:

      National Nuclear Security Administration (NNSA)

      Federal Computer Week: Energy agency launches performance-based pay system By Richard W. Walker Published on March 27, 2008 The Energy Department's National Nuclear Security Administration has launched a new performance- based pay system involving about 2,000 of its 2,500 employees. NNSA officials described the effort as a pilot project that will test the feasibility of the new system, which collapses the traditional 15 General Schedule pay bands into broader pay bands. The new structure

    3. Computed Tomography Status

      DOE R&D Accomplishments [OSTI]

      Hansche, B. D.

      1983-01-01

      Computed tomography (CT) is a relatively new radiographic technique which has become widely used in the medical field, where it is better known as computerized axial tomographic (CAT) scanning. This technique is also being adopted by the industrial radiographic community, although the greater range of densities, variation in samples sizes, plus possible requirement for finer resolution make it difficult to duplicate the excellent results that the medical scanners have achieved.

    4. Development of computer graphics

      SciTech Connect (OSTI)

      Nuttall, H.E.

      1989-07-01

      The purpose of this project was to screen and evaluate three graphics packages as to their suitability for displaying concentration contour graphs. The information to be displayed is from computer code simulations describing air-born contaminant transport. The three evaluation programs were MONGO (John Tonry, MIT, Cambridge, MA, 02139), Mathematica (Wolfram Research Inc.), and NCSA Image (National Center for Supercomputing Applications at the University of Illinois at Urbana-Champaign). After a preliminary investigation of each package, NCSA Image appeared to be significantly superior for generating the desired concentration contour graphs. Hence subsequent work and this report describes the implementation and testing of NCSA Image on both an Apple MacII and Sun 4 computers. NCSA Image includes several utilities (Layout, DataScope, HDF, and PalEdit) which were used in this study and installed on Dr. Ted Yamada`s Mac II computer. Dr. Yamada provided two sets of air pollution plume data which were displayed using NCSA Image. Both sets were animated into a sequential expanding plume series.

    5. Computing for Finance

      ScienceCinema (OSTI)

      None

      2011-10-06

      The finance sector is one of the driving forces for the use of distributed or Grid computing for business purposes. The speakers will review the state-of-the-art of high performance computing in the financial sector, and provide insight into how different types of Grid computing ? from local clusters to global networks - are being applied to financial applications. They will also describe the use of software and techniques from physics, such as Monte Carlo simulations, in the financial world. There will be four talks of 20min each. The talk abstracts and speaker bios are listed below. This will be followed by a Q&A; panel session with the speakers. From 19:00 onwards there will be a networking cocktail for audience and speakers. This is an EGEE / CERN openlab event organized in collaboration with the regional business network rezonance.ch. A webcast of the event will be made available for subsequent viewing, along with powerpoint material presented by the speakers. Attendance is free and open to all. Registration is mandatory via www.rezonance.ch, including for CERN staff. 1. Overview of High Performance Computing in the Financial Industry Michael Yoo, Managing Director, Head of the Technical Council, UBS Presentation will describe the key business challenges driving the need for HPC solutions, describe the means in which those challenges are being addressed within UBS (such as GRID) as well as the limitations of some of these solutions, and assess some of the newer HPC technologies which may also play a role in the Financial Industry in the future. Speaker Bio: Michael originally joined the former Swiss Bank Corporation in 1994 in New York as a developer on a large data warehouse project. In 1996 he left SBC and took a role with Fidelity Investments in Boston. Unable to stay away for long, he returned to SBC in 1997 while working for Perot Systems in Singapore. Finally, in 1998 he formally returned to UBS in Stamford following the merger with SBC and has remained with UBS for the past 9 years. During his tenure at UBS, he has had a number of leadership roles within IT in development, support and architecture. In 2006 Michael relocated to Switzerland to take up his current role as head of the UBS IB Technical Council, responsible for the overall technology strategy and vision of the Investment Bank. One of Michael's key responsibilities is to manage the UBS High Performance Computing Research Lab and he has been involved in a number of initiatives in the HPC space. 2. Grid in the Commercial WorldFred Gedling, Chief Technology Officer EMEA and Senior Vice President Global Services, DataSynapse Grid computing gets mentions in the press for community programs starting last decade with "Seti@Home". Government, national and supranational initiatives in grid receive some press. One of the IT-industries' best-kept secrets is the use of grid computing by commercial organizations with spectacular results. Grid Computing and its evolution into Application Virtualization is discussed and how this is key to the next generation data center. Speaker Bio: Fred Gedling holds the joint roles of Chief Technology Officer for EMEA and Senior Vice President of Global Services at DataSynapse, a global provider of application virtualisation software. Based in London and working closely with organisations seeking to optimise their IT infrastructures, Fred offers unique insights into the technology of virtualisation as well as the methodology of establishing ROI and rapid deployment to the immediate advantage of the business. Fred has more than fifteen years experience of enterprise middleware and high-performance infrastructures. Prior to DataSynapse he worked in high performance CRM middleware and was the CTO EMEA for New Era of Networks (NEON) during the rapid growth of Enterprise Application Integration. His 25-year career in technology also includes management positions at Goldman Sachs and Stratus Computer. Fred holds a First Class Bsc (Hons) degree in Physics with Astrophysics from the University of Leeds and had the privilege of being a summer student at CERN.3. Opportunities for gLite in finance and related industriesAdam Vile, Head of Grid, HPC and Technical Computing, Excelian Ltd.gLite, the Grid software developed by the EGEE project, has been exceedingly successful as an enabling infrastructure, and has been a massive success in bringing together scientific and technical communities to provide the compute power to address previously incomputable problems. Not so in the finance industry. In its current form gLite would be a business disabler. There are other middleware tools that solve the finance communities compute problems much better. Things are moving on, however. There are moves afoot in the open source community to evolve the technology to address other, more sophisticated needs such as utility and interactive computing. In this talk, I will describe how Excelian is providing Grid consultancy services for the finance community and how, through its relationship to the EGEE project, Excelian is helping to identify and exploit opportunities as the research and business worlds converge. Because of the strong third party presence in the finance industry, such opportunities are few and far between, but they are there, especially as we expand sideways into related verticals such as the smaller hedge funds and energy companies. This talk will give an overview of the barriers to adoption of gLite in the finance industry and highlight some of the opportunities offered in this and related industries as the ideas around Grid mature. Speaker Bio: Dr Adam Vile is a senior consultant and head of the Grid and HPC practice at Excelian, a consultancy that focuses on financial markets professional services. He has spent many years in investment banking, as a developer, project manager and architect in both front and back office. Before joining Excelian he was senior Grid and HPC architect at Barclays Capital. Prior to joining investment banking, Adam spent a number of years lecturing in IT and mathematics at a UK University and maintains links with academia through lectures, research and through validation and steering of postgraduate courses. He is a chartered mathematician and was the conference chair of the Institute of Mathematics and its Applications first conference in computational Finance.4. From Monte Carlo to Wall Street Daniel Egloff, Head of Financial Engineering Computing Unit, Zrich Cantonal Bank High performance computing techniques provide new means to solve computationally hard problems in the financial service industry. First I consider Monte Carlo simulation and illustrate how it can be used to implement a sophisticated credit risk management and economic capital framework. From a HPC perspective, basic Monte Carlo simulation is embarrassingly parallel and can be implemented efficiently on distributed memory clusters. Additional difficulties arise for adaptive variance reduction schemes, if the information content in a sample is very small, and if the amount of simulated date becomes huge such that incremental processing algorithms are indispensable. We discuss the business value of an advanced credit risk quantification which is particularly compelling in these days. While Monte Carlo simulation is a very versatile tool it is not always the preferred solution for the pricing of complex products like multi asset options, structured products, or credit derivatives. As a second application I show how operator methods can be used to develop a pricing framework. The scalability of operator methods relies heavily on optimized dense matrix-matrix multiplications and requires specialized BLAS level-3 implementations provided by specialized FPGA or GPU boards. Speaker Bio: Daniel Egloff studied mathematics, theoretical physics, and computer science at the University of Zurich and the ETH Zurich. He holds a PhD in Mathematics from University of Fribourg, Switzerland. After his PhD he started to work for a large Swiss insurance company in the area of asset and liability management. He continued his professional career in the consulting industry. At KPMG and Arthur Andersen he consulted international clients and implemented quantitative risk management solutions for financial institutions and insurance companies. In 2002 he joined Zurich Cantonal Bank. He was assigned to develop and implement credit portfolio risk and economic capital methodologies. He built up a competence center for high performance and cluster computing. Currently, Daniel Egloff is heading the Financial Computing unit in the ZKB Financial Engineering division. He and his team is engineering and operating high performance cluster applications for computationally intensive problems in financial risk management.

    6. Computing for Finance

      ScienceCinema (OSTI)

      None

      2011-10-06

      The finance sector is one of the driving forces for the use of distributed or Grid computing for business purposes. The speakers will review the state-of-the-art of high performance computing in the financial sector, and provide insight into how different types of Grid computing ? from local clusters to global networks - are being applied to financial applications. They will also describe the use of software and techniques from physics, such as Monte Carlo simulations, in the financial world. There will be four talks of 20min each. The talk abstracts and speaker bios are listed below. This will be followed by a Q&A; panel session with the speakers. From 19:00 onwards there will be a networking cocktail for audience and speakers. This is an EGEE / CERN openlab event organized in collaboration with the regional business network rezonance.ch. A webcast of the event will be made available for subsequent viewing, along with powerpoint material presented by the speakers. Attendance is free and open to all. Registration is mandatory via www.rezonance.ch, including for CERN staff. 1. Overview of High Performance Computing in the Financial Industry Michael Yoo, Managing Director, Head of the Technical Council, UBS Presentation will describe the key business challenges driving the need for HPC solutions, describe the means in which those challenges are being addressed within UBS (such as GRID) as well as the limitations of some of these solutions, and assess some of the newer HPC technologies which may also play a role in the Financial Industry in the future. Speaker Bio: Michael originally joined the former Swiss Bank Corporation in 1994 in New York as a developer on a large data warehouse project. In 1996 he left SBC and took a role with Fidelity Investments in Boston. Unable to stay away for long, he returned to SBC in 1997 while working for Perot Systems in Singapore. Finally, in 1998 he formally returned to UBS in Stamford following the merger with SBC and has remained with UBS for the past 9 years. During his tenure at UBS, he has had a number of leadership roles within IT in development, support and architecture. In 2006 Michael relocated to Switzerland to take up his current role as head of the UBS IB Technical Council, responsible for the overall technology strategy and vision of the Investment Bank. One of Michael's key responsibilities is to manage the UBS High Performance Computing Research Lab and he has been involved in a number of initiatives in the HPC space. 2. Grid in the Commercial WorldFred Gedling, Chief Technology Officer EMEA and Senior Vice President Global Services, DataSynapse Grid computing gets mentions in the press for community programs starting last decade with "Seti@Home". Government, national and supranational initiatives in grid receive some press. One of the IT-industries' best-kept secrets is the use of grid computing by commercial organizations with spectacular results. Grid Computing and its evolution into Application Virtualization is discussed and how this is key to the next generation data center. Speaker Bio: Fred Gedling holds the joint roles of Chief Technology Officer for EMEA and Senior Vice President of Global Services at DataSynapse, a global provider of application virtualisation software. Based in London and working closely with organisations seeking to optimise their IT infrastructures, Fred offers unique insights into the technology of virtualisation as well as the methodology of establishing ROI and rapid deployment to the immediate advantage of the business. Fred has more than fifteen years experience of enterprise middleware and high-performance infrastructures. Prior to DataSynapse he worked in high performance CRM middleware and was the CTO EMEA for New Era of Networks (NEON) during the rapid growth of Enterprise Application Integration. His 25-year career in technology also includes management positions at Goldman Sachs and Stratus Computer. Fred holds a First Class Bsc (Hons) degree in Physics with Astrophysics from the University of Leeds and had the privilege of being a summer student at CERN.3. Opportunities for gLite in finance and related industriesAdam Vile, Head of Grid, HPC and Technical Computing, Excelian Ltd.gLite, the Grid software developed by the EGEE project, has been exceedingly successful as an enabling infrastructure, and has been a massive success in bringing together scientific and technical communities to provide the compute power to address previously incomputable problems. Not so in the finance industry. In its current form gLite would be a business disabler. There are other middleware tools that solve the finance communities compute problems much better. Things are moving on, however. There are moves afoot in the open source community to evolve the technology to address other, more sophisticated needs such as utility and interactive computing. In this talk, I will describe how Excelian is providing Grid consultancy services for the finance community and how, through its relationship to the EGEE project, Excelian is helping to identify and exploit opportunities as the research and business worlds converge. Because of the strong third party presence in the finance industry, such opportunities are few and far between, but they are there, especially as we expand sideways into related verticals such as the smaller hedge funds and energy companies. This talk will give an overview of the barriers to adoption of gLite in the finance industry and highlight some of the opportunities offered in this and related industries as the ideas around Grid mature. Speaker Bio: Dr Adam Vile is a senior consultant and head of the Grid and HPC practice at Excelian, a consultancy that focuses on financial markets professional services. He has spent many years in investment banking, as a developer, project manager and architect in both front and back office. Before joining Excelian he was senior Grid and HPC architect at Barclays Capital. Prior to joining investment banking, Adam spent a number of years lecturing in IT and mathematics at a UK University and maintains links with academia through lectures, research and through validation and steering of postgraduate courses. He is a chartered mathematician and was the conference chair of the Institute of Mathematics and its Applications first conference in computational Finance.4. From Monte Carlo to Wall Street Daniel Egloff, Head of Financial Engineering Computing Unit, Zürich Cantonal Bank High performance computing techniques provide new means to solve computationally hard problems in the financial service industry. First I consider Monte Carlo simulation and illustrate how it can be used to implement a sophisticated credit risk management and economic capital framework. From a HPC perspective, basic Monte Carlo simulation is embarrassingly parallel and can be implemented efficiently on distributed memory clusters. Additional difficulties arise for adaptive variance reduction schemes, if the information content in a sample is very small, and if the amount of simulated date becomes huge such that incremental processing algorithms are indispensable. We discuss the business value of an advanced credit risk quantification which is particularly compelling in these days. While Monte Carlo simulation is a very versatile tool it is not always the preferred solution for the pricing of complex products like multi asset options, structured products, or credit derivatives. As a second application I show how operator methods can be used to develop a pricing framework. The scalability of operator methods relies heavily on optimized dense matrix-matrix multiplications and requires specialized BLAS level-3 implementations provided by specialized FPGA or GPU boards. Speaker Bio: Daniel Egloff studied mathematics, theoretical physics, and computer science at the University of Zurich and the ETH Zurich. He holds a PhD in Mathematics from University of Fribourg, Switzerland. After his PhD he started to work for a large Swiss insurance company in the area of asset and liability management. He continued his professional career in the consulting industry. At KPMG and Arthur Andersen he consulted international clients and implemented quantitative risk management solutions for financial institutions and insurance companies. In 2002 he joined Zurich Cantonal Bank. He was assigned to develop and implement credit portfolio risk and economic capital methodologies. He built up a competence center for high performance and cluster computing. Currently, Daniel Egloff is heading the Financial Computing unit in the ZKB Financial Engineering division. He and his team is engineering and operating high performance cluster applications for computationally intensive problems in financial risk management.

    7. Computing for Finance

      SciTech Connect (OSTI)

      2010-03-24

      The finance sector is one of the driving forces for the use of distributed or Grid computing for business purposes. The speakers will review the state-of-the-art of high performance computing in the financial sector, and provide insight into how different types of Grid computing – from local clusters to global networks - are being applied to financial applications. They will also describe the use of software and techniques from physics, such as Monte Carlo simulations, in the financial world. There will be four talks of 20min each. The talk abstracts and speaker bios are listed below. This will be followed by a Q&A; panel session with the speakers. From 19:00 onwards there will be a networking cocktail for audience and speakers. This is an EGEE / CERN openlab event organized in collaboration with the regional business network rezonance.ch. A webcast of the event will be made available for subsequent viewing, along with powerpoint material presented by the speakers. Attendance is free and open to all. Registration is mandatory via www.rezonance.ch, including for CERN staff. 1. Overview of High Performance Computing in the Financial Industry Michael Yoo, Managing Director, Head of the Technical Council, UBS Presentation will describe the key business challenges driving the need for HPC solutions, describe the means in which those challenges are being addressed within UBS (such as GRID) as well as the limitations of some of these solutions, and assess some of the newer HPC technologies which may also play a role in the Financial Industry in the future. Speaker Bio: Michael originally joined the former Swiss Bank Corporation in 1994 in New York as a developer on a large data warehouse project. In 1996 he left SBC and took a role with Fidelity Investments in Boston. Unable to stay away for long, he returned to SBC in 1997 while working for Perot Systems in Singapore. Finally, in 1998 he formally returned to UBS in Stamford following the merger with SBC and has remained with UBS for the past 9 years. During his tenure at UBS, he has had a number of leadership roles within IT in development, support and architecture. In 2006 Michael relocated to Switzerland to take up his current role as head of the UBS IB Technical Council, responsible for the overall technology strategy and vision of the Investment Bank. One of Michael's key responsibilities is to manage the UBS High Performance Computing Research Lab and he has been involved in a number of initiatives in the HPC space. 2. Grid in the Commercial WorldFred Gedling, Chief Technology Officer EMEA and Senior Vice President Global Services, DataSynapse Grid computing gets mentions in the press for community programs starting last decade with "Seti@Home". Government, national and supranational initiatives in grid receive some press. One of the IT-industries' best-kept secrets is the use of grid computing by commercial organizations with spectacular results. Grid Computing and its evolution into Application Virtualization is discussed and how this is key to the next generation data center. Speaker Bio: Fred Gedling holds the joint roles of Chief Technology Officer for EMEA and Senior Vice President of Global Services at DataSynapse, a global provider of application virtualisation software. Based in London and working closely with organisations seeking to optimise their IT infrastructures, Fred offers unique insights into the technology of virtualisation as well as the methodology of establishing ROI and rapid deployment to the immediate advantage of the business. Fred has more than fifteen years experience of enterprise middleware and high-performance infrastructures. Prior to DataSynapse he worked in high performance CRM middleware and was the CTO EMEA for New Era of Networks (NEON) during the rapid growth of Enterprise Application Integration. His 25-year career in technology also includes management positions at Goldman Sachs and Stratus Computer. Fred holds a First Class Bsc (Hons) degree in Physics with Astrophysics from the University of Leeds and had the privilege of being a summer student at CERN.3. Opportunities for gLite in finance and related industriesAdam Vile, Head of Grid, HPC and Technical Computing, Excelian Ltd.gLite, the Grid software developed by the EGEE project, has been exceedingly successful as an enabling infrastructure, and has been a massive success in bringing together scientific and technical communities to provide the compute power to address previously incomputable problems. Not so in the finance industry. In its current form gLite would be a business disabler. There are other middleware tools that solve the finance communities compute problems much better. Things are moving on, however. There are moves afoot in the open source community to evolve the technology to address other, more sophisticated needs such as utility and interactive computing. In this talk, I will describe how Excelian is providing Grid consultancy services for the finance community and how, through its relationship to the EGEE project, Excelian is helping to identify and exploit opportunities as the research and business worlds converge. Because of the strong third party presence in the finance industry, such opportunities are few and far between, but they are there, especially as we expand sideways into related verticals such as the smaller hedge funds and energy companies. This talk will give an overview of the barriers to adoption of gLite in the finance industry and highlight some of the opportunities offered in this and related industries as the ideas around Grid mature. Speaker Bio: Dr Adam Vile is a senior consultant and head of the Grid and HPC practice at Excelian, a consultancy that focuses on financial markets professional services. He has spent many years in investment banking, as a developer, project manager and architect in both front and back office. Before joining Excelian he was senior Grid and HPC architect at Barclays Capital. Prior to joining investment banking, Adam spent a number of years lecturing in IT and mathematics at a UK University and maintains links with academia through lectures, research and through validation and steering of postgraduate courses. He is a chartered mathematician and was the conference chair of the Institute of Mathematics and its Applications first conference in computational Finance.4. From Monte Carlo to Wall Street Daniel Egloff, Head of Financial Engineering Computing Unit, Zürich Cantonal Bank High performance computing techniques provide new means to solve computationally hard problems in the financial service industry. First I consider Monte Carlo simulation and illustrate how it can be used to implement a sophisticated credit risk management and economic capital framework. From a HPC perspective, basic Monte Carlo simulation is embarrassingly parallel and can be implemented efficiently on distributed memory clusters. Additional difficulties arise for adaptive variance reduction schemes, if the information content in a sample is very small, and if the amount of simulated date becomes huge such that incremental processing algorithms are indispensable. We discuss the business value of an advanced credit risk quantification which is particularly compelling in these days. While Monte Carlo simulation is a very versatile tool it is not always the preferred solution for the pricing of complex products like multi asset options, structured products, or credit derivatives. As a second application I show how operator methods can be used to develop a pricing framework. The scalability of operator methods relies heavily on optimized dense matrix-matrix multiplications and requires specialized BLAS level-3 implementations provided by specialized FPGA or GPU boards. Speaker Bio: Daniel Egloff studied mathematics, theoretical physics, and computer science at the University of Zurich and the ETH Zurich. He holds a PhD in Mathematics from University of Fribourg, Switzerland. After his PhD he started to work for a large Swiss insurance company in the area of asset and liability management. He continued his professional career in the consulting industry. At KPMG and Arthur Andersen he consulted international clients and implemented quantitative risk management solutions for financial institutions and insurance companies. In 2002 he joined Zurich Cantonal Bank. He was assigned to develop and implement credit portfolio risk and economic capital methodologies. He built up a competence center for high performance and cluster computing. Currently, Daniel Egloff is heading the Financial Computing unit in the ZKB Financial Engineering division. He and his team is engineering and operating high performance cluster applications for computationally intensive problems in financial risk management.

    8. High Performance Computing at the Oak Ridge Leadership Computing Facility

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      High Performance Computing at the Oak Ridge Leadership Computing Facility Go to Menu Page 2 Outline * Our Mission * Computer Systems: Present, Past, Future * Challenges Along the Way * Resources for Users Go to Menu Page 3 Our Mission Go to Menu Page 4 * World's most powerful computing facility * Nation's largest concentration of open source materials research * $1.3B budget * 4,250 employees * 3,900 research guests annually * $350 million invested in modernization * Nation's most diverse energy

    9. Multiprocessor computing for images

      SciTech Connect (OSTI)

      Cantoni, V. ); Levialdi, S. )

      1988-08-01

      A review of image processing systems developed until now is given, highlighting the weak points of such systems and the trends that have dictated their evolution through the years producing different generations of machines. Each generation may be characterized by the hardware architecture, the programmability features and the relative application areas. The need for multiprocessing hierarchical systems is discussed focusing on pyramidal architectures. Their computational paradigms, their virtual and physical implementation, their programming and software requirements, and capabilities by means of suitable languages, are discussed.

    10. computational-hydraulics

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      and Aerodynamics using STAR-CCM+ for CFD Analysis March 21-22, 2012 Argonne, Illinois Dr. Steven Lottes This email address is being protected from spambots. You need JavaScript enabled to view it. A training course in the use of computational hydraulics and aerodynamics CFD software using CD-adapco's STAR-CCM+ for analysis will be held at TRACC from March 21-22, 2012. The course assumes a basic knowledge of fluid mechanics and will make extensive use of hands on tutorials. CD-adapco will issue

    11. developing-compute-efficient

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Developing Compute-efficient, Quality Models with LS-PrePost® 3 on the TRACC Cluster Oct. 21-22, 2010 Argonne TRACC Dr. Cezary Bojanowski Dr. Ronald F. Kulak This email address is being protected from spambots. You need JavaScript enabled to view it. Announcement pdficon small The LS-PrePost Introductory Course was held October 21-22, 2010 at TRACC in West Chicago with interactive participation on-site as well as remotely via the Internet. Intended primarily for finite element analysts with

    12. Computing for Finance

      ScienceCinema (OSTI)

      None

      2011-10-06

      The finance sector is one of the driving forces for the use of distributed or Grid computing for business purposes. The speakers will review the state-of-the-art of high performance computing in the financial sector, and provide insight into how different types of Grid computing ? from local clusters to global networks - are being applied to financial applications. They will also describe the use of software and techniques from physics, such as Monte Carlo simulations, in the financial world. There will be four talks of 20min each. The talk abstracts and speaker bios are listed below. This will be followed by a Q&A; panel session with the speakers. From 19:00 onwards there will be a networking cocktail for audience and speakers. This is an EGEE / CERN openlab event organized in collaboration with the regional business network rezonance.ch. A webcast of the event will be made available for subsequent viewing, along with powerpoint material presented by the speakers. Attendance is free and open to all. Registration is mandatory via www.rezonance.ch, including for CERN staff. 1. Overview of High Performance Computing in the Financial Industry Michael Yoo, Managing Director, Head of the Technical Council, UBS Presentation will describe the key business challenges driving the need for HPC solutions, describe the means in which those challenges are being addressed within UBS (such as GRID) as well as the limitations of some of these solutions, and assess some of the newer HPC technologies which may also play a role in the Financial Industry in the future. Speaker Bio: Michael originally joined the former Swiss Bank Corporation in 1994 in New York as a developer on a large data warehouse project. In 1996 he left SBC and took a role with Fidelity Investments in Boston. Unable to stay away for long, he returned to SBC in 1997 while working for Perot Systems in Singapore. Finally, in 1998 he formally returned to UBS in Stamford following the merger with SBC and has remained with UBS for the past 9 years. During his tenure at UBS, he has had a number of leadership roles within IT in development, support and architecture. In 2006 Michael relocated to Switzerland to take up his current role as head of the UBS IB Technical Council, responsible for the overall technology strategy and vision of the Investment Bank. One of Michael's key responsibilities is to manage the UBS High Performance Computing Research Lab and he has been involved in a number of initiatives in the HPC space. 2. Grid in the Commercial WorldFred Gedling, Chief Technology Officer EMEA and Senior Vice President Global Services, DataSynapse Grid computing gets mentions in the press for community programs starting last decade with "Seti@Home". Government, national and supranational initiatives in grid receive some press. One of the IT-industries' best-kept secrets is the use of grid computing by commercial organizations with spectacular results. Grid Computing and its evolution into Application Virtualization is discussed and how this is key to the next generation data center. Speaker Bio: Fred Gedling holds the joint roles of Chief Technology Officer for EMEA and Senior Vice President of Global Services at DataSynapse, a global provider of application virtualisation software. Based in London and working closely with organisations seeking to optimise their IT infrastructures, Fred offers unique insights into the technology of virtualisation as well as the methodology of establishing ROI and rapid deployment to the immediate advantage of the business. Fred has more than fifteen years experience of enterprise middleware and high-performance infrastructures. Prior to DataSynapse he worked in high performance CRM middleware and was the CTO EMEA for New Era of Networks (NEON) during the rapid growth of Enterprise Application Integration. His 25-year career in technology also includes management positions at Goldman Sachs and Stratus Computer. Fred holds a First Class Bsc (Hons) degree in Physics with Astrophysics from the University of Leeds and had the privilege o

    13. Computer generated holographic microtags

      DOE Patents [OSTI]

      Sweatt, William C.

      1998-01-01

      A microlithographic tag comprising an array of individual computer generated holographic patches having feature sizes between 250 and 75 nanometers. The tag is a composite hologram made up of the individual holographic patches and contains identifying information when read out with a laser of the proper wavelength and at the proper angles of probing and reading. The patches are fabricated in a steep angle Littrow readout geometry to maximize returns in the -1 diffracted order. The tags are useful as anti-counterfeiting markers because of the extreme difficulty in reproducing them.

    14. Computer generated holographic microtags

      DOE Patents [OSTI]

      Sweatt, W.C.

      1998-03-17

      A microlithographic tag comprising an array of individual computer generated holographic patches having feature sizes between 250 and 75 nanometers is disclosed. The tag is a composite hologram made up of the individual holographic patches and contains identifying information when read out with a laser of the proper wavelength and at the proper angles of probing and reading. The patches are fabricated in a steep angle Littrow readout geometry to maximize returns in the -1 diffracted order. The tags are useful as anti-counterfeiting markers because of the extreme difficulty in reproducing them. 5 figs.

    15. Scanning computed confocal imager

      DOE Patents [OSTI]

      George, John S. (Los Alamos, NM)

      2000-03-14

      There is provided a confocal imager comprising a light source emitting a light, with a light modulator in optical communication with the light source for varying the spatial and temporal pattern of the light. A beam splitter receives the scanned light and direct the scanned light onto a target and pass light reflected from the target to a video capturing device for receiving the reflected light and transferring a digital image of the reflected light to a computer for creating a virtual aperture and outputting the digital image. In a transmissive mode of operation the invention omits the beam splitter means and captures light passed through the target.

    16. Introduction to High Performance Computing

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Introduction to High Performance Computing Introduction to High Performance Computing June 10, 2013 Photo on 7 30 12 at 7.10 AM Downloads Download File Gerber-HPC-2.pdf...

    17. Computer Wallpaper | The Ames Laboratory

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Computer Wallpaper We've incorporated the tagline, Creating Materials and Energy Solutions, into a computer wallpaper so you can display it on your desktop as a constant reminder....

    18. W7 printing from VISITOR network

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      2. Click on "Add a printer" and choose "Add a local printer". 3 . Choose "create a new port" and pick "Standard TCPIP Port" from the drop down menu. 4. Type the following for the...

    19. Jefferson Lab Visitor's Center - Driving in Virginia

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Driving in virginia Effective January 1, 2004, all applicants for a driver's license or photo ID card must show proof of U.S. Citizenship or Legal Presence in the United States. A U.S. birth certificate or unexpired U.S. passport will serve as both proof of identity and proof of legal presence. It can also be proved using a variety of other documents, such as a Certificate of Citizenship, Resident Alien Card or a valid foreign passport with a visa, I-94 or I-94W from a participating country. If

    20. Jefferson Lab Visitor's Center - Travel Accommodations

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Travel Accommodations While visiting or working at JLab there are many options for living accommodations. These options vary depending upon the expected length of stay, transportation available and money allotted for housing expenses. The following information has been compiled to assist in the search for living accommodations near the lab. For further assistance e-mail User Liaison or call 757-269-6388. On-Site Accommodations SURA Residence Facility Off-Site Accommodations Negotiated Lodging

    1. Super recycled water: quenching computers

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Super recycled water: quenching computers Super recycled water: quenching computers New facility and methods support conserving water and creating recycled products. Using reverse osmosis to "super purify" water allows the system to reuse water and cool down our powerful yet thirsty computers. January 30, 2014 Super recycled water: quenching computers LANL's Sanitary Effluent Reclamation Facility, key to reducing the Lab's discharge of liquid. Millions of gallons of industrial

    2. Fermilab | Science at Fermilab | Computing

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Computing Computing is indispensable to science at Fermilab. High-energy physics experiments generate an astounding amount of data that physicists need to store, analyze and communicate with others. Cutting-edge technology allows scientists to work quickly and efficiently to advance our understanding of the world . Fermilab's Computing Division is recognized for its expertise in handling huge amounts of data, its success in high-speed parallel computing and its willingness to take its craft in

    3. History | Argonne Leadership Computing Facility

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Leadership Computing The Argonne Leadership Computing Facility (ALCF) was established at Argonne National Laboratory in 2004 as part of a U.S. Department of Energy (DOE) initiative dedicated to enabling leading-edge computational capabilities to advance fundamental discovery and understanding in a broad range of scientific and engineering disciplines. Supported by the Advanced Scientific Computing Research (ASCR) program within DOE's Office of Science, the ALCF is one half of the DOE Leadership

    4. Computing architecture for autonomous microgrids

      DOE Patents [OSTI]

      Goldsmith, Steven Y.

      2015-09-29

      A computing architecture that facilitates autonomously controlling operations of a microgrid is described herein. A microgrid network includes numerous computing devices that execute intelligent agents, each of which is assigned to a particular entity (load, source, storage device, or switch) in the microgrid. The intelligent agents can execute in accordance with predefined protocols to collectively perform computations that facilitate uninterrupted control of the microgrid.

    5. Noise tolerant spatiotemporal chaos computing

      SciTech Connect (OSTI)

      Kia, Behnam; Kia, Sarvenaz; Ditto, William L.; Lindner, John F.; Sinha, Sudeshna

      2014-12-01

      We introduce and design a noise tolerant chaos computing system based on a coupled map lattice (CML) and the noise reduction capabilities inherent in coupled dynamical systems. The resulting spatiotemporal chaos computing system is more robust to noise than a single map chaos computing system. In this CML based approach to computing, under the coupled dynamics, the local noise from different nodes of the lattice diffuses across the lattice, and it attenuates each other's effects, resulting in a system with less noise content and a more robust chaos computing architecture.

    6. AMRITA -- A computational facility

      SciTech Connect (OSTI)

      Shepherd, J.E.; Quirk, J.J.

      1998-02-23

      Amrita is a software system for automating numerical investigations. The system is driven using its own powerful scripting language, Amrita, which facilitates both the composition and archiving of complete numerical investigations, as distinct from isolated computations. Once archived, an Amrita investigation can later be reproduced by any interested party, and not just the original investigator, for no cost other than the raw CPU time needed to parse the archived script. In fact, this entire lecture can be reconstructed in such a fashion. To do this, the script: constructs a number of shock-capturing schemes; runs a series of test problems, generates the plots shown; outputs the LATEX to typeset the notes; performs a myriad of behind-the-scenes tasks to glue everything together. Thus Amrita has all the characteristics of an operating system and should not be mistaken for a common-or-garden code.

    7. Computer memory management system

      DOE Patents [OSTI]

      Kirk, III, Whitson John

      2002-01-01

      A computer memory management system utilizing a memory structure system of "intelligent" pointers in which information related to the use status of the memory structure is designed into the pointer. Through this pointer system, The present invention provides essentially automatic memory management (often referred to as garbage collection) by allowing relationships between objects to have definite memory management behavior by use of coding protocol which describes when relationships should be maintained and when the relationships should be broken. In one aspect, the present invention system allows automatic breaking of strong links to facilitate object garbage collection, coupled with relationship adjectives which define deletion of associated objects. In another aspect, The present invention includes simple-to-use infinite undo/redo functionality in that it has the capability, through a simple function call, to undo all of the changes made to a data model since the previous `valid state` was noted.

    8. Intro to computer programming, no computer required! | Argonne Leadership

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Computing Facility Intro to computer programming, no computer required! Author: Laura Wolf January 6, 2016 Facebook Twitter LinkedIn Google E-mail Printer-friendly version Pairing the volunteers with interested schools was the easy part. School administrators and teachers alike were delighted to have Argonne National Laboratory volunteers visit and help guide their Hour of Code activities last December. In all, Argonne's Educational Programs department helped place 44 volunteers in Chicago

    9. Other World Computing | Open Energy Information

      Open Energy Info (EERE)

      World Computing Jump to: navigation, search Name Other World Computing Facility Other World Computing Sector Wind energy Facility Type Community Wind Facility Status In Service...

    10. CLAMR (Compute Language Adaptive Mesh Refinement)

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      CLAMR (Compute Language Adaptive Mesh Refinement) CLAMR (Compute Language Adaptive Mesh Refinement) CLAMR (Compute Language Adaptive Mesh Refinement) is being developed as a DOE...

    11. Computer_Vision

      Energy Science and Technology Software Center (OSTI)

      2002-10-04

      The Computer_Vision software performs object recognition using a novel multi-scale characterization and matching algorithm. To understand the multi-scale characterization and matching software, it is first necessary to understand some details of the Computer Vision (CV) Project. This project has focused on providing algorithms and software that provide an end-to-end toolset for image processing applications. At a high-level, this end-to-end toolset focuses on 7 coy steps. The first steps are geometric transformations. 1) Image Segmentation. Thismore » step essentially classifies pixels in foe input image as either being of interest or not of interest. We have also used GENIE segmentation output for this Image Segmentation step. 2 Contour Extraction (patent submitted). This takes the output of Step I and extracts contours for the blobs consisting of pixels of interest. 3) Constrained Delaunay Triangulation. This is a well-known geometric transformation that creates triangles inside the contours. 4 Chordal Axis Transform (CAT) . This patented geometric transformation takes the triangulation output from Step 3 and creates a concise and accurate structural representation of a contour. From the CAT, we create a linguistic string, with associated metrical information, that provides a detailed structural representation of a contour. 5.) Normalization. This takes an attributed linguistic string output from Step 4 and balances it. This ensures that the linguistic representation accurately represents the major sections of the contour. Steps 6 and 7 are implemented by the multi-scale characterization and matching software. 6) Multi scale Characterization. This takes as input the attributed linguistic string output from Normalization. Rules from a context free grammar are applied in reverse to create a tree-like representation for each contour. For example, one of the grammar’s rules is L -> (LL ). When an (LL) is seen in a string, a parent node is created that points to the four child symbols ‘(‘ , ‘L’ , ‘L’, and ‘)‘ . Levels in the tree can then be thought of as coarser (towards the root) or finer (towards the leaves) representations of the same contours. 7.) Multi scale Matching. Having a multi-scale characterization allows us to compare objects at a coarser level before matching at finer levels of detail. Matching at a coarser level not only increases the speed of the matching process (you’re comparing fewer symbols) , but also increases accuracy since small variations along contours do not significantly detract from two objects’ similarity.« less

    12. Computational Fluid Dynamics Library

      Energy Science and Technology Software Center (OSTI)

      2005-03-04

      CFDLib05 is the Los Alamos Computational Fluid Dynamics LIBrary. This is a collection of hydrocodes using a common data structure and a common numerical method, for problems ranging from single-field, incompressible flow, to multi-species, multi-field, compressible flow. The data structure is multi-block, with a so-called structured grid in each block. The numerical method is a Finite-Volume scheme employing a state vector that is fully cell-centered. This means that the integral form of the conservation lawsmore » is solved on the physical domain that is represented by a mesh of control volumes. The typical control volume is an arbitrary quadrilateral in 2D and an arbitrary hexahedron in 3D. The Finite-Volume scheme is for time-unsteady flow and remains well coupled by means of time and space centered fluxes; if a steady state solution is required, the problem is integrated forward in time until the user is satisfied that the state is stationary.« less

    13. Bioinformatics Computing Consultant Position Available

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Bioinformatics Computing Consultant Position Available Bioinformatics Computing Consultant Position Available October 31, 2011 by Katie Antypas NERSC and the Joint Genome Institute (JGI) are searching for two individuals who can help biologists exploit advanced computing platforms. JGI provides production sequencing and genomics for the Department of Energy. These activities are critical to the DOE missions in areas related to clean energy generation and environmental characterization and

    14. Parallel Computing Summer Research Internship

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Parallel Computing Parallel Computing Summer Research Internship Creates next-generation leaders in HPC research and applications development Contacts Program Co-Lead Robert (Bob) Robey Email Program Co-Lead Gabriel Rockefeller Email Program Co-Lead Hai Ah Nam Email Professional Staff Assistant Nickole Aguilar Garcia (505) 665-3048 Email The Parallel Computing Summer Research Internship is an intense 10 week program aimed at providing students with a solid foundation in modern high performance

    15. computational-fluid-dynamics-training

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Table of Contents Date Location Advanced Hydraulic and Aerodynamic Analysis Using CFD March 27-28, 2013 Argonne TRACC Argonne, IL Computational Hydraulics and Aerodynamics using STAR-CCM+ for CFD Analysis March 21-22, 2012 Argonne TRACC Argonne, IL Computational Hydraulics and Aerodynamics using STAR-CCM+ for CFD Analysis March 30-31, 2011 Argonne TRACC Argonne, IL Computational Hydraulics for Transportation Workshop September 23-24, 2009 Argonne TRACC West Chicago, IL

    16. Careers | Argonne Leadership Computing Facility

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Careers at Argonne Looking for a unique opportunity to work at the forefront of high-performance computing? At the Argonne Leadership Computing Facility, we are helping to redefine what's possible in computational science. With some of the most powerful supercomputers in the world and a talented and diverse team of experts, we enable researchers to pursue groundbreaking discoveries that would otherwise not be possible. Check out our open positions below. For the most current listing of

    17. Bioinformatics Computing Consultant Position Available

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      You can read more about the positions and apply at jobs.lbl.gov: Bioinformatics High Performance Computing Consultant (job number: 73194) and Software Developer for High...

    18. Parallel Computing Summer Research Internship

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      should have basic experience with a scientific computing language, such as C, C++, Fortran and with the LINUX operating system. Duration & Location The program will last ten...

    19. Tukey | Argonne Leadership Computing Facility

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Documentation Feedback Please provide feedback to help guide us as we continue to build documentation for our new computing resource. Feedback Form Tukey The primary purpose of...

    20. QBox | Argonne Leadership Computing Facility

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      computers. Obtaining Qbox http:eslab.ucdavis.edusoftwareqbox Building Qbox for Blue GeneQ Qbox requires the standard math libraries plus the Xerces-C http:...

    1. Thrusts in High Performance Computing

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Exascale computers (1000x Hopper) in next decade: - Manycore processors using graphics, games, embedded cores, or other low power designs offer 100x in power efficiency -...

    2. Advanced Simulation and Computing Program

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      The SSP mission is to analyze and predict the performance, safety, and reliability of nuclear weapons and certify their functionality. ASC works in partnership with computer ...

    3. Institutional computing (IC) information session

      SciTech Connect (OSTI)

      Koch, Kenneth R; Lally, Bryan R

      2011-01-19

      The LANL Institutional Computing Program (IC) will host an information session about the current state of unclassified Institutional Computing at Los Alamos, exciting plans for the future, and the current call for proposals for science and engineering projects requiring computing. Program representatives will give short presentations and field questions about the call for proposals and future planned machines, and discuss technical support available to existing and future projects. Los Alamos has started making a serious institutional investment in open computing available to our science projects, and that investment is expected to increase even more.

    4. Manufacturing Energy and Carbon Footprint - Sector: Computer...

      Office of Energy Efficiency and Renewable Energy (EERE) Indexed Site

      Computers, Electronics and Electrical Equipment (NAICS 334, 335) Process Energy ... Carbon Footprint Sector: Computers, Electronics and Electrical Equipment (NAICS 334, ...

    5. Computation Directorate 2008 Annual Report

      SciTech Connect (OSTI)

      Crawford, D L

      2009-03-25

      Whether a computer is simulating the aging and performance of a nuclear weapon, the folding of a protein, or the probability of rainfall over a particular mountain range, the necessary calculations can be enormous. Our computers help researchers answer these and other complex problems, and each new generation of system hardware and software widens the realm of possibilities. Building on Livermore's historical excellence and leadership in high-performance computing, Computation added more than 331 trillion floating-point operations per second (teraFLOPS) of power to LLNL's computer room floors in 2008. In addition, Livermore's next big supercomputer, Sequoia, advanced ever closer to its 2011-2012 delivery date, as architecture plans and the procurement contract were finalized. Hyperion, an advanced technology cluster test bed that teams Livermore with 10 industry leaders, made a big splash when it was announced during Michael Dell's keynote speech at the 2008 Supercomputing Conference. The Wall Street Journal touted Hyperion as a 'bright spot amid turmoil' in the computer industry. Computation continues to measure and improve the costs of operating LLNL's high-performance computing systems by moving hardware support in-house, by measuring causes of outages to apply resources asymmetrically, and by automating most of the account and access authorization and management processes. These improvements enable more dollars to go toward fielding the best supercomputers for science, while operating them at less cost and greater responsiveness to the customers.

    6. Oak Ridge National Laboratory - Computing and Computational Sciences

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Directorate Oak Ridge to acquire next generation supercomputer Oak Ridge to acquire next generation supercomputer The U.S. Department of Energy's (DOE) Oak Ridge Leadership Computing Facility (OLCF) has signed a contract with IBM to bring a next-generation supercomputer to Oak Ridge National Laboratory (ORNL). The OLCF's new hybrid CPU/GPU computing system, Summit, will be delivered in 2017. (more) Links Department of Energy Consortium for Advanced Simulation of Light Water Reactors Extreme

    7. Introduction to computers: Reference guide

      SciTech Connect (OSTI)

      Ligon, F.V.

      1995-04-01

      The ``Introduction to Computers`` program establishes formal partnerships with local school districts and community-based organizations, introduces computer literacy to precollege students and their parents, and encourages students to pursue Scientific, Mathematical, Engineering, and Technical careers (SET). Hands-on assignments are given in each class, reinforcing the lesson taught. In addition, the program is designed to broaden the knowledge base of teachers in scientific/technical concepts, and Brookhaven National Laboratory continues to act as a liaison, offering educational outreach to diverse community organizations and groups. This manual contains the teacher`s lesson plans and the student documentation to this introduction to computer course.

    8. Power throttling of collections of computing elements

      DOE Patents [OSTI]

      Bellofatto, Ralph E. (Ridgefield, CT); Coteus, Paul W. (Yorktown Heights, NY); Crumley, Paul G. (Yorktown Heights, NY); Gara, Alan G. (Mount Kidsco, NY); Giampapa, Mark E. (Irvington, NY); Gooding; Thomas M. (Rochester, MN); Haring, Rudolf A. (Cortlandt Manor, NY); Megerian, Mark G. (Rochester, MN); Ohmacht, Martin (Yorktown Heights, NY); Reed, Don D. (Mantorville, MN); Swetz, Richard A. (Mahopac, NY); Takken, Todd (Brewster, NY)

      2011-08-16

      An apparatus and method for controlling power usage in a computer includes a plurality of computers communicating with a local control device, and a power source supplying power to the local control device and the computer. A plurality of sensors communicate with the computer for ascertaining power usage of the computer, and a system control device communicates with the computer for controlling power usage of the computer.

    9. Quantum Computing: Solving Complex Problems

      ScienceCinema (OSTI)

      DiVincenzo, David [IBM Watson Research Center

      2009-09-01

      One of the motivating ideas of quantum computation was that there could be a new kind of machine that would solve hard problems in quantum mechanics. There has been significant progress towards the experimental realization of these machines (which I will review), but there are still many questions about how such a machine could solve computational problems of interest in quantum physics. New categorizations of the complexity of computational problems have now been invented to describe quantum simulation. The bad news is that some of these problems are believed to be intractable even on a quantum computer, falling into a quantum analog of the NP class. The good news is that there are many other new classifications of tractability that may apply to several situations of physical interest.

    10. SSRL Computer Account Request Form

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      SSRL/LCLS Computer Account Request Form August 2009 Fill in this form and sign the security statement mentioned at the bottom of this page to obtain an account. Your Name: __________________________________________________________ Institution: ___________________________________________________________ Mailing Address: ______________________________________________________ Email Address: _______________________________________________________ Telephone:

    11. Computing at SSRL Home Page

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      contents you are looking for have moved. You will be redirected to the new location automatically in 5 seconds. Please bookmark the correct page at http://www-ssrl.slac.stanford.edu/content/staff-resources/computer-networking-group

    12. Events | Argonne Leadership Computing Facility

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      2:00 PM Finding Multiple Local Minima of Computationally Expensive Simulations Jeffery Larson Postdoctoral Appointee, MCS Building 240Room 4301 Pages 1 2 3 4 5 6 7 8 9 ... next...

    13. Present and Future Computing Requirements

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Cosmology SciDAC-3 Project Ann Almgren (LBNL) Nick Gnedin (FNAL) Dave Higdon (LANL) Rob Ross (ANL) Martin White (UC Berkeley LBNL) Large Scale Production Computing and Storage...

    14. SSRL Computer Account Request Form

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      SSRLLCLS Computer Account Request Form August 2009 Fill in this form and sign the security statement mentioned at the bottom of this page to obtain an account. Your Name:...

    15. Computer Assisted Virtual Environment - CAVE

      ScienceCinema (OSTI)

      Erickson, Phillip; Podgorney, Robert; Weingartner, Shawn; Whiting, Eric

      2014-06-09

      Research at the Center for Advanced Energy Studies is taking on another dimension with a 3-D device known as a Computer Assisted Virtual Environment. The CAVE uses projection to display high-end computer graphics on three walls and the floor. By wearing 3-D glasses to create depth perception and holding a wand to move and rotate images, users can delve into data.

    16. Secure computing for the 'Everyman'

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Secure computing for the 'Everyman' Secure computing for the 'Everyman' If implemented on a wide scale, quantum key distribution technology could ensure truly secure commerce, banking, communications and data transfer. September 2, 2014 This small device developed at Los Alamos National Laboratory uses the truly random spin of light particles as defined by laws of quantum mechanics to generate a random number for use in a cryptographic key that can be used to securely transmit information

    17. computational-hydraulics-for-transportation

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Transportation Workshop Sept. 23-24, 2009 Argonne TRACC Dr. Steven Lottes This email address is being protected from spambots. You need JavaScript enabled to view it. Announcement pdficon small The Transportation Research and Analysis Computing Center at Argonne National Laboratory will hold a workshop on the use of computational hydraulics for transportation applications. The goals of the workshop are: Bring together people who are using or would benefit from the use of high performance cluster

    18. Computational Sciences and Engineering Division

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      If you have questions or comments regarding any of our research and development activities, how to work with ORNL and the Computational Sciences and Engineering (CSE) Division, or the content of this website please contact one of the following people: If you have questions regarding CSE technologies and capabilities, job opportunities, working with ORNL and the CSE Division, intellectual property, etc., contact, Shaun S. Gleason, Ph.D. Division Director, Computational Sciences and Engineering

    19. Computational Sciences and Engineering Division

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      The Computational Sciences and Engineering Division is a major research division at the Department of Energy's Oak Ridge National Laboratory. CSED develops and applies creative information technology and modeling and simulation research solutions for National Security and National Energy Infrastructure needs. The mission of the Computational Sciences and Engineering Division is to enhance the country's capabilities in achieving important objectives in the areas of national defense, homeland

    20. Mira | Argonne Leadership Computing Facility

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Computing Resources Mira Cetus and Vesta Visualization Cluster Data and Networking Software JLSE Featured Videos Mira: Argonne's 10-Petaflop Supercomputer Mira's Dedication Ceremony Introducing Mira: Our Next-Generation Supercomputer Mira Mira Ushers in a New Era of Scientific Supercomputing As one of the fastest supercomputers, Mira, our 10-petaflops IBM Blue Gene/Q system, is capable of 10 quadrillion calculations per second. With this computing power, Mira can do in one day what it would take

    1. Cooley | Argonne Leadership Computing Facility

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Changes from Tukey to Cooley Compiling and Linking Using Cobalt on Cooley Visit on Cooley Paraview on Cooley ParaView Tutorial VNC on Cooley Policies Documentation Feedback Please provide feedback to help guide us as we continue to build documentation for our new computing resource. [Feedback Form] Cooley The primary purpose of Cooley is to analyze and visualize data produced on Mira. Equipped with state-of-the-art graphics processing units (GPUs), Cooley converts computational data from Mira

    2. LAMMPS | Argonne Leadership Computing Facility

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Software & Libraries Boost CPMD Code_Saturne GAMESS GPAW GROMACS LAMMPS MADNESS QBox IBM References Cooley Policies Documentation Feedback Please provide feedback to help guide us as we continue to build documentation for our new computing resource. [Feedback Form] LAMMPS Overview LAMMPS is a general-purpose molecular dynamics software package for massively parallel computers. It is written in an exceptionally clean style that makes it one of the most popular codes for users to extend and

    3. Automatic computation of transfer functions

      DOE Patents [OSTI]

      Atcitty, Stanley; Watson, Luke Dale

      2015-04-14

      Technologies pertaining to the automatic computation of transfer functions for a physical system are described herein. The physical system is one of an electrical system, a mechanical system, an electromechanical system, an electrochemical system, or an electromagnetic system. A netlist in the form of a matrix comprises data that is indicative of elements in the physical system, values for the elements in the physical system, and structure of the physical system. Transfer functions for the physical system are computed based upon the netlist.

    4. Proposal for grid computing for nuclear applications

      SciTech Connect (OSTI)

      Idris, Faridah Mohamad; Ismail, Saaidi; Haris, Mohd Fauzi B.; Sulaiman, Mohamad Safuan B.; Aslan, Mohd Dzul Aiman Bin.; Samsudin, Nursuliza Bt.; Ibrahim, Maizura Bt.; Ahmad, Megat Harun Al Rashid B. Megat; Yazid, Hafizal B.; Jamro, Rafhayudi B.; Azman, Azraf B.; Rahman, Anwar B. Abdul; Ibrahim, Mohd Rizal B. Mamat; Muhamad, Shalina Bt. Sheik; Hassan, Hasni; Abdullah, Wan Ahmad Tajuddin Wan; Ibrahim, Zainol Abidin; Zolkapli, Zukhaimira; Anuar, Afiq Aizuddin; Norjoharuddeen, Nurfikri; and others

      2014-02-12

      The use of computer clusters for computational sciences including computational physics is vital as it provides computing power to crunch big numbers at a faster rate. In compute intensive applications that requires high resolution such as Monte Carlo simulation, the use of computer clusters in a grid form that supplies computational power to any nodes within the grid that needs computing power, has now become a necessity. In this paper, we described how the clusters running on a specific application could use resources within the grid, to run the applications to speed up the computing process.

    5. Computation Directorate 2007 Annual Report

      SciTech Connect (OSTI)

      Henson, V E; Guse, J A

      2008-03-06

      If there is a single word that both characterized 2007 and dominated the thoughts and actions of many Laboratory employees throughout the year, it is transition. Transition refers to the major shift that took place on October 1, when the University of California relinquished management responsibility for Lawrence Livermore National Laboratory (LLNL), and Lawrence Livermore National Security, LLC (LLNS), became the new Laboratory management contractor for the Department of Energy's (DOE's) National Nuclear Security Administration (NNSA). In the 55 years under the University of California, LLNL amassed an extraordinary record of significant accomplishments, clever inventions, and momentous contributions in the service of protecting the nation. This legacy provides the new organization with a built-in history, a tradition of excellence, and a solid set of core competencies from which to build the future. I am proud to note that in the nearly seven years I have had the privilege of leading the Computation Directorate, our talented and dedicated staff has made far-reaching contributions to the legacy and tradition we passed on to LLNS. Our place among the world's leaders in high-performance computing, algorithmic research and development, applications, and information technology (IT) services and support is solid. I am especially gratified to report that through all the transition turmoil, and it has been considerable, the Computation Directorate continues to produce remarkable achievements. Our most important asset--the talented, skilled, and creative people who work in Computation--has continued a long-standing Laboratory tradition of delivering cutting-edge science even in the face of adversity. The scope of those achievements is breathtaking, and in 2007, our accomplishments span an amazing range of topics. From making an important contribution to a Nobel Prize-winning effort to creating tools that can detect malicious codes embedded in commercial software; from expanding BlueGene/L, the world's most powerful computer, by 60% and using it to capture the most prestigious prize in the field of computing, to helping create an automated control system for the National Ignition Facility (NIF) that monitors and adjusts more than 60,000 control and diagnostic points; from creating a microarray probe that rapidly detects virulent high-threat organisms, natural or bioterrorist in origin, to replacing large numbers of physical computer servers with small numbers of virtual servers, reducing operating expense by 60%, the people in Computation have been at the center of weighty projects whose impacts are felt across the Laboratory and the DOE community. The accomplishments I just mentioned, and another two dozen or so, make up the stories contained in this report. While they form an exceptionally diverse set of projects and topics, it is what they have in common that excites me. They share the characteristic of being central, often crucial, to the mission-driven business of the Laboratory. Computational science has become fundamental to nearly every aspect of the Laboratory's approach to science and even to the conduct of administration. It is difficult to consider how we would proceed without computing, which occurs at all scales, from handheld and desktop computing to the systems controlling the instruments and mechanisms in the laboratories to the massively parallel supercomputers. The reasons for the dramatic increase in the importance of computing are manifest. Practical, fiscal, or political realities make the traditional approach to science, the cycle of theoretical analysis leading to experimental testing, leading to adjustment of theory, and so on, impossible, impractical, or forbidden. How, for example, can we understand the intricate relationship between human activity and weather and climate? We cannot test our hypotheses by experiment, which would require controlled use of the entire earth over centuries. It is only through extremely intricate, detailed computational simulation that we can test our theories, and simulati

    6. Snowmass Computing Frontier I2: Distributed

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Snowmass Computing Frontier I2: Distributed Computing and Facility Infrastructures Ken Bloom Richard Gerber July 31, 2013 Thursday, October 10, 13 Computing Frontier I2: Distributed Computing and Facility Infrastructures 7/31/13 Who we are ‣ Ken Bloom, Associate Professor, Department of Physics and Astronomy, University of Nebraska-Lincoln ‣ Co-PI for the Nebraska CMS Tier-2 computing facility ‣ Tier-2 program manager and Deputy Manager of Software and Computing for US CMS ‣ Tier-2

    7. Yuri Alexeev | Argonne Leadership Computing Facility

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Yuri Alexeev Assistant Computational Scientist Yury Alekseev Argonne National Laboratory 9700 South Cass Avenue Building 240 - Rm. 1126 Argonne IL, 60439 630-252-0157 yuri@alcf.anl.gov Yuri Alexeev is an Assistant Computational Scientist at the Argonne Leadership Computing Facility where he applies his skills, knowledge and experience for using and enabling computational methods in chemistry and biology for high-performance computing on next-generation high-performance computers. Yuri is

    8. Computer Science and Information Technology Student Pipeline

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Divisions recruit and hire promising undergraduate and graduate students in the areas of Computer Science, Information Technology, Management Information Systems, Computer...

    9. Applications for Postdoctoral Fellowship in Computational Science...

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      at Berkeley Lab due November 26 October 15, 2012 by Francesca Verdier Researchers in computer science, applied mathematics or any computational science discipline who have...

    10. Sandia National Laboratories: Careers: Computer Science

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Computer Science Red Storm photo Sandia's supercomputing research is reaching for tomorrow's exascale performance while solving real-world problems today. Computer scientists and...

    11. Personal Computing Equipment | Open Energy Information

      Open Energy Info (EERE)

      Computing Equipment Jump to: navigation, search TODO: Add description List of Personal Computing Equipment Incentives Retrieved from "http:en.openei.orgwindex.php?titlePersona...

    12. Advanced Materials Development through Computational Design ...

      Office of Energy Efficiency and Renewable Energy (EERE) Indexed Site

      Development through Computational Design Advanced Materials Development through Computational Design Presentation given at the 2007 Diesel Engine-Efficiency & Emissions Research ...

    13. Thermoelectric Materials by Design, Computational Theory and...

      Office of Energy Efficiency and Renewable Energy (EERE) Indexed Site

      by Design, Computational Theory and Structure Thermoelectric Materials by Design, Computational Theory and Structure 2009 DOE Hydrogen Program and Vehicle Technologies Program...

    14. Extreme Scale Computing, Co-Design

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Information Science, Computing, Applied Math Extreme Scale Computing, Co-design Publications Publications Ramon Ravelo, Qi An, Timothy C. Germann, and Brad Lee Holian, ...

    15. Energy Storage Computational Tool | Open Energy Information

      Open Energy Info (EERE)

      Energy Storage Computational Tool Jump to: navigation, search Tool Summary LAUNCH TOOL Name: Energy Storage Computational Tool AgencyCompany Organization: Navigant Consulting...

    16. Integrated Computational Materials Engineering (ICME) for Mg...

      Office of Energy Efficiency and Renewable Energy (EERE) Indexed Site

      lm012li2012o.pdf More Documents & Publications Integrated Computational Materials Engineering (ICME) for Mg: International Pilot Project Integrated Computational Materials...

    17. Integrated Computational Materials Engineering (ICME) for Mg...

      Office of Energy Efficiency and Renewable Energy (EERE) Indexed Site

      lm012li2011o.pdf More Documents & Publications Integrated Computational Materials Engineering (ICME) for Mg: International Pilot Project Integrated Computational Materials...

    18. Hybrid Rotaxanes: Interlocked Structures for Quantum Computing...

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      based on molecular magnets that may make them suitable as qubits for quantum computers. Chemistry Aids Quantum Computing Quantum bits or qubits are the fundamental...

    19. Compare Activities by Number of Computers

      U.S. Energy Information Administration (EIA) Indexed Site

      of Computers Office buildings contained the most computers per square foot, followed by education and outpatient health care buildings. Education buildings were the only type...

    20. Hybrid Rotaxanes: Interlocked Structures for Quantum Computing...

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Hybrid Rotaxanes: Interlocked Structures for Quantum Computing? Hybrid Rotaxanes: Interlocked Structures for Quantum Computing? Print Wednesday, 26 August 2009 00:00 Rotaxanes are...

    1. NERSC Enhances PDSF, Genepool Computing Capabilities

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Computing Capabilities NERSC Enhances PDSF, Genepool Computing Capabilities Linux cluster expansion speeds data access and analysis January 3, 2014 Christmas came early for...

    2. Solvate Structures and Computational/Spectroscopic Characterization...

      Office of Scientific and Technical Information (OSTI)

      Solvate Structures and ComputationalSpectroscopic Characterization of LiPF6 Electrolytes Citation Details In-Document Search Title: Solvate Structures and Computational...

    3. Solvate Structures and Computational/Spectroscopic Characterization...

      Office of Scientific and Technical Information (OSTI)

      Solvate Structures and ComputationalSpectroscopic Characterization of LiBF4 Electrolytes Citation Details In-Document Search Title: Solvate Structures and Computational...

    4. Improved computer models support genetics research

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      February Simple computer models unravel genetic stress reactions in cells Simple computer models unravel genetic stress reactions in cells Integrated biological and...

    5. Computer Accounts | Stanford Synchrotron Radiation Lightsource

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Computer Accounts Each user group must have a computer account. Additionally, all persons using these accounts are responsible for understanding and complying with the terms...

    6. LANL computer model boosts engine efficiency

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      LANL computer model boosts engine efficiency LANL computer model boosts engine efficiency The KIVA model has been instrumental in helping researchers and manufacturers understand...

    7. Sandia National Laboratories: Advanced Simulation Computing:...

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      These collaborations help solve the challenges of developing computing platforms and simulation tools across a number of disciplines. Computer Science Research Institute The...

    8. Significant Enhancement of Computational Efficiency in Nonlinear Multiscale Battery Model for Computer Aided Engineering

      SciTech Connect (OSTI)

      Smith, Kandler; Graf, Peter; Jun, Myungsoo; Yang, Chuanbo; Li, Genong; Li, Shaoping; Hochman, Amit; Tselepidakis, Dimitrios

      2015-06-09

      This presentation provides an update on improvements in computational efficiency in a nonlinear multiscale battery model for computer aided engineering.

    9. Argonne Lea Computing F A

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Lea Computing F A r g o n n e L e a d e r s h i p C o m p u t i n g FA c i l i t y 2 0 1 3 S c i e n c e H i g H l i g H t S Argonne leadership computing Facility C O N T E N T S About ALCF ...............................................................................................................................2 MirA...............................................................................................................................................3 SCienCe DireCtor'S MeSSAge

    10. Supporting collaborative computing and interaction

      SciTech Connect (OSTI)

      Agarwal, Deborah; McParland, Charles; Perry, Marcia

      2002-05-22

      To enable collaboration on the daily tasks involved in scientific research, collaborative frameworks should provide lightweight and ubiquitous components that support a wide variety of interaction modes. We envision a collaborative environment as one that provides a persistent space within which participants can locate each other, exchange synchronous and asynchronous messages, share documents and applications, share workflow, and hold videoconferences. We are developing the Pervasive Collaborative Computing Environment (PCCE) as such an environment. The PCCE will provide integrated tools to support shared computing and task control and monitoring. This paper describes the PCCE and the rationale for its design.

    11. Computing

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      ... The Challenge is project-based learning geared to teaching a wide range of skills - 41615 A simulation of vortex induced motion shows how ocean currents affect offshore oil rigs. ...

    12. Computing

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      This turbulent transport is caused by drift-wave instabilities, driven by free energy in plasma temperature and density gradients. * Unavoidable: These instabilities will persist ...

    13. PERTURBATION APPROACH FOR QUANTUM COMPUTATION

      SciTech Connect (OSTI)

      G. P. BERMAN; D. I. KAMENEV; V. I. TSIFRINOVICH

      2001-04-01

      We discuss how to simulate errors in the implementation of simple quantum logic operations in a nuclear spin quantum computer with many qubits, using radio-frequency pulses. We verify our perturbation approach using the exact solutions for relatively small (L = 10) number of qubits.

    14. New challenges in computational biochemistry

      SciTech Connect (OSTI)

      Honig, B.

      1996-12-31

      The new challenges in computational biochemistry to which the title refers include the prediction of the relative binding free energy of different substrates to the same protein, conformational sampling, and other examples of theoretical predictions matching known protein structure and behavior.

    15. Radiological Worker Computer Based Training

      Energy Science and Technology Software Center (OSTI)

      2003-02-06

      Argonne National Laboratory has developed an interactive computer based training (CBT) version of the standardized DOE Radiological Worker training program. This CD-ROM based program utilizes graphics, animation, photographs, sound and video to train users in ten topical areas: radiological fundamentals, biological effects, dose limits, ALARA, personnel monitoring, controls and postings, emergency response, contamination controls, high radiation areas, and lessons learned.

    16. Experimental Mathematics and Computational Statistics

      SciTech Connect (OSTI)

      Bailey, David H.; Borwein, Jonathan M.

      2009-04-30

      The field of statistics has long been noted for techniques to detect patterns and regularities in numerical data. In this article we explore connections between statistics and the emerging field of 'experimental mathematics'. These includes both applications of experimental mathematics in statistics, as well as statistical methods applied to computational mathematics.

    17. Computer Science and Information Technology Student Pipeline

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Science and Information Technology Student Pipeline Program Description Los Alamos National Laboratory's High Performance Computing and Information Technology Divisions recruit and hire promising undergraduate and graduate students in the areas of Computer Science, Information Technology, Management Information Systems, Computer Security, Software Engineering, Computer Engineering, and Electrical Engineering. Students are provided a mentor and challenging projects to demonstrate their

    18. Internode data communications in a parallel computer

      DOE Patents [OSTI]

      Archer, Charles J.; Blocksome, Michael A.; Miller, Douglas R.; Parker, Jeffrey J.; Ratterman, Joseph D.; Smith, Brian E.

      2013-09-03

      Internode data communications in a parallel computer that includes compute nodes that each include main memory and a messaging unit, the messaging unit including computer memory and coupling compute nodes for data communications, in which, for each compute node at compute node boot time: a messaging unit allocates, in the messaging unit's computer memory, a predefined number of message buffers, each message buffer associated with a process to be initialized on the compute node; receives, prior to initialization of a particular process on the compute node, a data communications message intended for the particular process; and stores the data communications message in the message buffer associated with the particular process. Upon initialization of the particular process, the process establishes a messaging buffer in main memory of the compute node and copies the data communications message from the message buffer of the messaging unit into the message buffer of main memory.

    19. Internode data communications in a parallel computer

      DOE Patents [OSTI]

      Archer, Charles J; Blocksome, Michael A; Miller, Douglas R; Parker, Jeffrey J; Ratterman, Joseph D; Smith, Brian E

      2014-02-11

      Internode data communications in a parallel computer that includes compute nodes that each include main memory and a messaging unit, the messaging unit including computer memory and coupling compute nodes for data communications, in which, for each compute node at compute node boot time: a messaging unit allocates, in the messaging unit's computer memory, a predefined number of message buffers, each message buffer associated with a process to be initialized on the compute node; receives, prior to initialization of a particular process on the compute node, a data communications message intended for the particular process; and stores the data communications message in the message buffer associated with the particular process. Upon initialization of the particular process, the process establishes a messaging buffer in main memory of the compute node and copies the data communications message from the message buffer of the messaging unit into the message buffer of main memory.

    20. Link failure detection in a parallel computer

      DOE Patents [OSTI]

      Archer, Charles J. (Rochester, MN); Blocksome, Michael A. (Rochester, MN); Megerian, Mark G. (Rochester, MN); Smith, Brian E. (Rochester, MN)

      2010-11-09

      Methods, apparatus, and products are disclosed for link failure detection in a parallel computer including compute nodes connected in a rectangular mesh network, each pair of adjacent compute nodes in the rectangular mesh network connected together using a pair of links, that includes: assigning each compute node to either a first group or a second group such that adjacent compute nodes in the rectangular mesh network are assigned to different groups; sending, by each of the compute nodes assigned to the first group, a first test message to each adjacent compute node assigned to the second group; determining, by each of the compute nodes assigned to the second group, whether the first test message was received from each adjacent compute node assigned to the first group; and notifying a user, by each of the compute nodes assigned to the second group, whether the first test message was received.

    1. Broadcasting a message in a parallel computer

      DOE Patents [OSTI]

      Berg, Jeremy E. (Rochester, MN); Faraj, Ahmad A. (Rochester, MN)

      2011-08-02

      Methods, systems, and products are disclosed for broadcasting a message in a parallel computer. The parallel computer includes a plurality of compute nodes connected together using a data communications network. The data communications network optimized for point to point data communications and is characterized by at least two dimensions. The compute nodes are organized into at least one operational group of compute nodes for collective parallel operations of the parallel computer. One compute node of the operational group assigned to be a logical root. Broadcasting a message in a parallel computer includes: establishing a Hamiltonian path along all of the compute nodes in at least one plane of the data communications network and in the operational group; and broadcasting, by the logical root to the remaining compute nodes, the logical root's message along the established Hamiltonian path.

    2. Newsletter Southern Great Plains

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      facility use by total visitor days and to track actual visitors and active user research computer accounts. Historical data show that the total number of users is related to the...

    3. 2011 Computation Directorate Annual Report

      SciTech Connect (OSTI)

      Crawford, D L

      2012-04-11

      From its founding in 1952 until today, Lawrence Livermore National Laboratory (LLNL) has made significant strategic investments to develop high performance computing (HPC) and its application to national security and basic science. Now, 60 years later, the Computation Directorate and its myriad resources and capabilities have become a key enabler for LLNL programs and an integral part of the effort to support our nation's nuclear deterrent and, more broadly, national security. In addition, the technological innovation HPC makes possible is seen as vital to the nation's economic vitality. LLNL, along with other national laboratories, is working to make supercomputing capabilities and expertise available to industry to boost the nation's global competitiveness. LLNL is on the brink of an exciting milestone with the 2012 deployment of Sequoia, the National Nuclear Security Administration's (NNSA's) 20-petaFLOP/s resource that will apply uncertainty quantification to weapons science. Sequoia will bring LLNL's total computing power to more than 23 petaFLOP/s-all brought to bear on basic science and national security needs. The computing systems at LLNL provide game-changing capabilities. Sequoia and other next-generation platforms will enable predictive simulation in the coming decade and leverage industry trends, such as massively parallel and multicore processors, to run petascale applications. Efficient petascale computing necessitates refining accuracy in materials property data, improving models for known physical processes, identifying and then modeling for missing physics, quantifying uncertainty, and enhancing the performance of complex models and algorithms in macroscale simulation codes. Nearly 15 years ago, NNSA's Accelerated Strategic Computing Initiative (ASCI), now called the Advanced Simulation and Computing (ASC) Program, was the critical element needed to shift from test-based confidence to science-based confidence. Specifically, ASCI/ASC accelerated the development of simulation capabilities necessary to ensure confidence in the nuclear stockpile-far exceeding what might have been achieved in the absence of a focused initiative. While stockpile stewardship research pushed LLNL scientists to develop new computer codes, better simulation methods, and improved visualization technologies, this work also stimulated the exploration of HPC applications beyond the standard sponsor base. As LLNL advances to a petascale platform and pursues exascale computing (1,000 times faster than Sequoia), ASC will be paramount to achieving predictive simulation and uncertainty quantification. Predictive simulation and quantifying the uncertainty of numerical predictions where little-to-no data exists demands exascale computing and represents an expanding area of scientific research important not only to nuclear weapons, but to nuclear attribution, nuclear reactor design, and understanding global climate issues, among other fields. Aside from these lofty goals and challenges, computing at LLNL is anything but 'business as usual.' International competition in supercomputing is nothing new, but the HPC community is now operating in an expanded, more aggressive climate of global competitiveness. More countries understand how science and technology research and development are inextricably linked to economic prosperity, and they are aggressively pursuing ways to integrate HPC technologies into their native industrial and consumer products. In the interest of the nation's economic security and the science and technology that underpins it, LLNL is expanding its portfolio and forging new collaborations. We must ensure that HPC remains an asymmetric engine of innovation for the Laboratory and for the U.S. and, in doing so, protect our research and development dynamism and the prosperity it makes possible. One untapped area of opportunity LLNL is pursuing is to help U.S. industry understand how supercomputing can benefit their business. Industrial investment in HPC applications has historically been limited by the prohibitive cost of entry, the inaccessibility of software to run the powerful systems, and the years it takes to grow the expertise to develop codes and run them in an optimal way. LLNL is helping industry better compete in the global market place by providing access to some of the world's most powerful computing systems, the tools to run them, and the experts who are adept at using them. Our scientists are collaborating side by side with industrial partners to develop solutions to some of industry's toughest problems. The goal of the Livermore Valley Open Campus High Performance Computing Innovation Center is to allow American industry the opportunity to harness the power of supercomputing by leveraging the scientific and computational expertise at LLNL in order to gain a competitive advantage in the global economy.

    4. NERSC Enhances PDSF, Genepool Computing Capabilities

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Enhances PDSF, Genepool Computing Capabilities NERSC Enhances PDSF, Genepool Computing Capabilities Linux cluster expansion speeds data access and analysis January 3, 2014 Christmas came early for users of the Parallel Distributed Systems Facility (PDSF) and Genepool systems at Department of Energy's National Energy Research Scientific Computer Center (NERSC). Throughout November members of NERSC's Computational Systems Group were busy expanding the Linux computing resources that support PDSF's

    5. Extreme Scale Computing, Co-design

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Information Science, Computing, Applied Math » Extreme Scale Computing, Co-design Extreme Scale Computing, Co-design Computational co-design may facilitate revolutionary designs in the next generation of supercomputers. Get Expertise Tim Germann Physics and Chemistry of Materials Email Allen McPherson Energy and Infrastructure Analysis Email Turab Lookman Physics and Condensed Matter and Complex Systems Email Computational co-design involves developing the interacting components of a

    6. NERSC seeks Computational Systems Group Lead

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      seeks Computational Systems Group Lead NERSC seeks Computational Systems Group Lead January 6, 2011 by Katie Antypas Note: This position is now closed. The Computational Systems Group provides production support and advanced development for the supercomputer systems at NERSC. Manage the Computational Systems Group (CSG) which provides production support and advanced development for the supercomputer systems at NERSC (National Energy Research Scientific Computing Center). These systems, which

    7. Presentation: High Performance Computing Applications | Department of

      Energy Savers [EERE]

      Energy High Performance Computing Applications Presentation: High Performance Computing Applications A briefing to the Secretary's Energy Advisory Board on High Performance Computing Applications delivered by Frederick H. Streitz, Lawrence Livermore National Laboratory. PDF icon High Performance Computing More Documents & Publications Presentation: QER Energy Topics 2011_INCITE_Fact_Sheets.pdf DOEs Effort to Reduce Truck Aerodynamic Drag through Joint Experiments and Computations

    8. Advanced Computing Tech Team | Department of Energy

      Office of Energy Efficiency and Renewable Energy (EERE) Indexed Site

      Advanced Computing Tech Team Advanced Computing Tech Team Advanced Computing Tech Team The Advanced Computing Tech Team is working with the DOE Energy Technology Offices, the Office of Science, and the National Nuclear Security Administration to deliver technologies that will be used to create new scientific insights into complex physical systems. Advanced computing technologies have been used for decades to provide better understanding of the performance and reliability of the nuclear stockpile

    9. Energy Efficient Computer Use | Department of Energy

      Office of Energy Efficiency and Renewable Energy (EERE) Indexed Site

      Electricity & Fuel » Appliances & Electronics » Energy Efficient Computer Use Energy Efficient Computer Use Use sleep mode and power management features on your computer to save money on your energy bill. Use sleep mode and power management features on your computer to save money on your energy bill. If you wonder when you should turn off your personal computer for energy savings, here are some general guidelines to help you make that decision. Though there is a small surge in energy

    10. Numerical computation of Pop plot

      SciTech Connect (OSTI)

      Menikoff, Ralph

      2015-03-23

      The Pop plot — distance-of-run to detonation versus initial shock pressure — is a key characterization of shock initiation in a heterogeneous explosive. Reactive burn models for high explosives (HE) must reproduce the experimental Pop plot to have any chance of accurately predicting shock initiation phenomena. This report describes a methodology for automating the computation of a Pop plot for a specific explosive with a given HE model. Illustrative examples of the computation are shown for PBX 9502 with three burn models (SURF, WSD and Forest Fire) utilizing the xRage code, which is the Eulerian ASC hydrocode at LANL. Comparison of the numerical and experimental Pop plot can be the basis for a validation test or as an aid in calibrating the burn rate of an HE model. Issues with calibration are discussed.

    11. Addressing failures in exascale computing

      SciTech Connect (OSTI)

      Snir, Marc; Wisniewski, Robert W.; Abraham, Jacob A.; Adve, Sarita; Bagchi, Saurabh; Balaji, Pavan; Belak, Jim; Bose, Pradip; Cappello, Franck; Carlson, William; Chien, Andrew A.; Coteus, Paul; Debardeleben, Nathan A.; Diniz, Pedro; Engelmann, Christian; Erez, Mattan; Saverio, Fazzari; Geist, Al; Gupta, Rinku; Johnson, Fred; Krishnamoorthy, Sriram; Leyffer, Sven; Liberty, Dean; Mitra, Subhasish; Munson, Todd; Schreiber, Robert; Stearly, Jon; Van Hensbergen, Eric

      2014-05-01

      We present here a report produced by a workshop on Addressing Failures in Exascale Computing held in Park City, Utah, August 411, 2012. The charter of this workshop was to establish a common taxonomy about resilience across all the levels in a computing system; discuss existing knowledge on resilience across the various hardware and software layers of an exascale system; and build on those results, examining potential solutions from both a hardware and software perspective and focusing on a combined approach. The workshop brought together participants with expertise in applications, system software, and hardware; they came from industry, government, and academia; and their interests ranged from theory to implementation. The combination allowed broad and comprehensive discussions and led to this document, which summarizes and builds on those discussions.

    12. gdb | Argonne Leadership Computing Facility

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Allinea DDT Core File Settings Determining Memory Use Using VNC with a Debugger bgq_stack gdb Coreprocessor Runjob termination TotalView Performance Tools & APIs Software & Libraries IBM References Cooley Policies Documentation Feedback Please provide feedback to help guide us as we continue to build documentation for our new computing resource. [Feedback Form] gdb Using gdb Preliminaries You should prepare a debug version of your code: Compile using -O0 -g If you are using the XL

    13. Darshan | Argonne Leadership Computing Facility

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Performance Tools & APIs Tuning MPI on BG/Q Tuning and Analysis Utilities (TAU) HPCToolkit HPCTW mpiP gprof Profiling Tools Darshan PAPI BG/Q Performance Counters BGPM Openspeedshop Scalasca BG/Q DGEMM Performance Automatic Performance Collection (AutoPerf) Software & Libraries IBM References Cooley Policies Documentation Feedback Please provide feedback to help guide us as we continue to build documentation for our new computing resource. [Feedback Form] Darshan References Darshan

    14. GAMESS | Argonne Leadership Computing Facility

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Performance Tools & APIs Software & Libraries Boost CPMD Code_Saturne GAMESS GPAW GROMACS LAMMPS MADNESS QBox IBM References Cooley Policies Documentation Feedback Please provide feedback to help guide us as we continue to build documentation for our new computing resource. [Feedback Form] GAMESS What Is GAMESS? The General Atomic and Molecular Electronic Structure System (GAMESS) is a general ab initio quantum chemistry package. For more information on GAMESS, see the Gordon research

    15. GROMACS | Argonne Leadership Computing Facility

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Software & Libraries Boost CPMD Code_Saturne GAMESS GPAW GROMACS LAMMPS MADNESS QBox IBM References Cooley Policies Documentation Feedback Please provide feedback to help guide us as we continue to build documentation for our new computing resource. [Feedback Form] GROMACS Building and Running GROMACS on Vesta/Mira The Gromacs Molecular Dynamics package has a large number of executables. Some of them, such as luck, are just utilities that do not need to be built for the back end. Begin by

    16. Policies | Argonne Leadership Computing Facility

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Accounts Policy Account Sponsorship & Retention Policy ALCC Quarterly Report Policy ALCF Acknowledgment Policy Data Policy INCITE Quarterly Report Policy Job Scheduling Policy on BG/Q Job Scheduling Policies on Cooley Pullback Policy Refund Policy Software Policy User Authentication Policy Documentation Feedback Please provide feedback to help guide us as we continue to build documentation for our new computing resource. [Feedback Form] Policies Official policies and procedures of the ALCF.

    17. Molecular Science Computing: 2010 Greenbook

      SciTech Connect (OSTI)

      De Jong, Wibe A.; Cowley, David E.; Dunning, Thom H.; Vorpagel, Erich R.

      2010-04-02

      This 2010 Greenbook outlines the science drivers for performing integrated computational environmental molecular research at EMSL and defines the next-generation HPC capabilities that must be developed at the MSC to address this critical research. The EMSL MSC Science Panel used EMSL’s vision and science focus and white papers from current and potential future EMSL scientific user communities to define the scientific direction and resulting HPC resource requirements presented in this 2010 Greenbook.

    18. Programs | Argonne Leadership Computing Facility

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      INCITE Program ALCC Program Director's Discretionary (DD) Program Early Science Program INCITE 2016 Projects ALCC 2015 Projects ESP Projects View All Projects Publications ALCF Tech Reports Industry Collaborations Featured Science Snapshot of the global structure of a radiation-dominated accretion flow around a black hole computed using the Athena++ code Magnetohydrodynamic Models of Accretion Including Radiation Transport James Stone Allocation Program: INCITE Allocation Hours: 47 Million

    19. Projects | Argonne Leadership Computing Facility

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Projects bgclang Compiler Hal Finkel Cobalt Scheduler Bill Allcock, Paul Rich, Brian Toonen, Tom Uram GLEAN: Scalable In Situ Analysis and I/O Acceleration on Leadership Computing Systems Michael E. Papka, Venkat Vishwanath, Mark Hereld, Preeti Malakar, Joe Insley, Silvio Rizzi, Tom Uram Petrel: Data Management and Sharing Pilot Ian Foster, Michael E. Papka, Bill Allcock, Ben Allen, Rachana Ananthakrishnan, Lukasz Lacinski The Swift Parallel Scripting Language for ALCF Systems Michael Wilde,

    20. MADNESS | Argonne Leadership Computing Facility

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Software & Libraries Boost CPMD Code_Saturne GAMESS GPAW GROMACS LAMMPS MADNESS QBox IBM References Cooley Policies Documentation Feedback Please provide feedback to help guide us as we continue to build documentation for our new computing resource. [Feedback Form] MADNESS Overview MADNESS is a numerical tool kit used to solve integral differential equations using multi-resolution analysis and a low-rank separation representation. MADNESS can solve multi-dimensional equations, currently up

    1. Michael Levitt and Computational Biology

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Michael Levitt and Computational Biology Resources with Additional Information * Publications Michael Levitt Courtesy of Linda A. Cicero / Stanford News Service Michael Levitt, PhD, professor of structural biology at the Stanford University School of Medicine, has won the 2013 Nobel Prize in Chemistry. ... Levitt ... shares the ... prize with Martin Karplus ... and Arieh Warshel ... "for the development of multiscale models for complex chemical systems." Levitt's work focuses on

    2. Allocations | Argonne Leadership Computing Facility

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Allocation Management Determining Allocation Requirements Querying Allocations Using cbank Mira/Cetus/Vesta Cooley Policies Documentation Feedback Please provide feedback to help guide us as we continue to build documentation for our new computing resource. [Feedback Form] Allocations ALCF resources are primarily used for DOE INCITE and ALCC awarded projects. Additional information on the INCITE program can be found on the DOE INCITE website and the ALCC program can be found on the Office of

    3. computation | National Nuclear Security Administration

      National Nuclear Security Administration (NNSA)

      computation | National Nuclear Security Administration Facebook Twitter Youtube Flickr RSS People Mission Managing the Stockpile Preventing Proliferation Powering the Nuclear Navy Emergency Response Recapitalizing Our Infrastructure Countering Nuclear Terrorism About Our Programs Our History Who We Are Our Leadership Our Locations Budget Our Operations Library Bios Congressional Testimony Fact Sheets Newsletters Press Releases Photo Gallery Jobs Apply for Our Jobs Our Jobs Working at NNSA Blog

    4. computers | National Nuclear Security Administration

      National Nuclear Security Administration (NNSA)

      computers | National Nuclear Security Administration Facebook Twitter Youtube Flickr RSS People Mission Managing the Stockpile Preventing Proliferation Powering the Nuclear Navy Emergency Response Recapitalizing Our Infrastructure Countering Nuclear Terrorism About Our Programs Our History Who We Are Our Leadership Our Locations Budget Our Operations Library Bios Congressional Testimony Fact Sheets Newsletters Press Releases Photo Gallery Jobs Apply for Our Jobs Our Jobs Working at NNSA Blog

    5. computing | National Nuclear Security Administration

      National Nuclear Security Administration (NNSA)

      computing | National Nuclear Security Administration Facebook Twitter Youtube Flickr RSS People Mission Managing the Stockpile Preventing Proliferation Powering the Nuclear Navy Emergency Response Recapitalizing Our Infrastructure Countering Nuclear Terrorism About Our Programs Our History Who We Are Our Leadership Our Locations Budget Our Operations Library Bios Congressional Testimony Fact Sheets Newsletters Press Releases Photo Gallery Jobs Apply for Our Jobs Our Jobs Working at NNSA Blog

    6. TRIDAC host computer functional specification

      SciTech Connect (OSTI)

      Hilbert, S.M.; Hunter, S.L.

      1983-08-23

      The purpose of this document is to outline the baseline functional requirements for the Triton Data Acquisition and Control (TRIDAC) Host Computer Subsystem. The requirements presented in this document are based upon systems that currently support both the SIS and the Uranium Separator Technology Groups in the AVLIS Program at the Lawrence Livermore National Laboratory and upon the specific demands associated with the extended safe operation of the SIS Triton Facility.

    7. Towards Energy-Centric Computing and Computer Architecture

      SciTech Connect (OSTI)

      2011-02-09

      Technology forecasts indicate that device scaling will continue well into the next decade. Unfortunately, it is becoming extremely difficult to harness this increase in the number of transistorsinto performance due to a number of technological, circuit, architectural, methodological and programming challenges.In this talk, I will argue that the key emerging showstopper is power. Voltage scaling as a means to maintain a constant power envelope with an increase in transistor numbers is hitting diminishing returns. As such, to continue riding the Moore's law we need to look for drastic measures to cut power. This is definitely the case for server chips in future datacenters,where abundant server parallelism, redundancy and 3D chip integration are likely to remove programming, reliability and bandwidth hurdles, leaving power as the only true limiter.I will present results backing this argument based on validated models for future server chips and parameters extracted from real commercial workloads. Then I use these results to project future research directions for datacenter hardware and software.About the speakerBabak Falsafi is a Professor in the School of Computer and Communication Sciences at EPFL, and an Adjunct Professor of Electrical and Computer Engineering and Computer Science at Carnegie Mellon. He is thefounder and the director ofthe Parallel Systems Architecture Laboratory (PARSA) at EPFL where he conducts research onarchitectural support for parallel programming, resilient systems, architectures to break the memory wall, and analytic and simulation tools for computer system performance evaluation.In 1999, in collaboration with T. N. Vijaykumar he showed for the first time that, contrary to conventional wisdom,multiprocessors do not needrelaxed memory consistency models (and the resulting convoluted programming interfaces found and used in modern systems) to achieve high performance. He is a recipient of an NSF CAREER award in 2000, IBM Faculty Partnership Awards between 2001 and 2004, and an Alfred P. Sloan Research Fellowship in 2004. He is a senior member of IEEE and ACM.

    8. Directions | Argonne National Laboratory

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Getting to Argonne Drive south on Cass Avenue Just south of I-55 turn right onto Northgate Road (just south of the Frontage Road) As you approach the main Argonne entrance gate you will see the Visitor Reception Building (VRC) on your right - turn into the VRC parking lot Everyone in your bus over the age of 16 must enter the VRC and obtain a pass at the reception desk Visitor Check-in Procedure After everyone over the age of 16 has obtained a gate pass and has re-boarded the bus pull out of

    9. Toward Molecular Catalysts by Computer

      SciTech Connect (OSTI)

      Raugei, Simone; DuBois, Daniel L.; Rousseau, Roger J.; Chen, Shentan; Ho, Ming-Hsun; Bullock, R. Morris; Dupuis, Michel

      2015-02-17

      Rational design of molecular catalysts requires a systematic approach to designing ligands with specific functionality and precisely tailored electronic and steric properties. It then becomes possible to devise computer protocols to predict accurately the required properties and ultimately to design catalysts by computer. In this account we first review how thermodynamic properties such as oxidation-reduction potentials (E0), acidities (pKa), and hydride donor abilities (ΔGH-) form the basis for a systematic design of molecular catalysts for reactions that are critical for a secure energy future (hydrogen evolution and oxidation, oxygen and nitrogen reduction, and carbon dioxide reduction). We highlight how density functional theory allows us to determine and predict these properties within “chemical” accuracy (~ 0.06 eV for redox potentials, ~ 1 pKa unit for pKa values, and ~ 1.5 kcal/mol for hydricities). These quantities determine free energy maps and profiles associated with catalytic cycles, i.e. the relative energies of intermediates, and help us distinguish between desirable and high-energy pathways and mechanisms. Good catalysts have flat profiles that avoid high activation barriers due to low and high energy intermediates. We illustrate how the criterion of a flat energy profile lends itself to the prediction of design points by computer for optimum catalysts. This research was carried out in the Center for Molecular Electro-catalysis, an Energy Frontier Research Center funded by the U.S. Department of Energy (DOE), Office of Science, Office of Basic Energy Sciences. Pacific Northwest National Laboratory (PNNL) is operated for the DOE by Battelle.

    10. GPU COMPUTING FOR PARTICLE TRACKING

      SciTech Connect (OSTI)

      Nishimura, Hiroshi; Song, Kai; Muriki, Krishna; Sun, Changchun; James, Susan; Qin, Yong

      2011-03-25

      This is a feasibility study of using a modern Graphics Processing Unit (GPU) to parallelize the accelerator particle tracking code. To demonstrate the massive parallelization features provided by GPU computing, a simplified TracyGPU program is developed for dynamic aperture calculation. Performances, issues, and challenges from introducing GPU are also discussed. General purpose Computation on Graphics Processing Units (GPGPU) bring massive parallel computing capabilities to numerical calculation. However, the unique architecture of GPU requires a comprehensive understanding of the hardware and programming model to be able to well optimize existing applications. In the field of accelerator physics, the dynamic aperture calculation of a storage ring, which is often the most time consuming part of the accelerator modeling and simulation, can benefit from GPU due to its embarrassingly parallel feature, which fits well with the GPU programming model. In this paper, we use the Tesla C2050 GPU which consists of 14 multi-processois (MP) with 32 cores on each MP, therefore a total of 448 cores, to host thousands ot threads dynamically. Thread is a logical execution unit of the program on GPU. In the GPU programming model, threads are grouped into a collection of blocks Within each block, multiple threads share the same code, and up to 48 KB of shared memory. Multiple thread blocks form a grid, which is executed as a GPU kernel. A simplified code that is a subset of Tracy++ [2] is developed to demonstrate the possibility of using GPU to speed up the dynamic aperture calculation by having each thread track a particle.

    11. Controlling data transfers from an origin compute node to a target compute node

      DOE Patents [OSTI]

      Archer, Charles J. (Rochester, MN); Blocksome, Michael A. (Rochester, MN); Ratterman, Joseph D. (Rochester, MN); Smith, Brian E. (Rochester, MN)

      2011-06-21

      Methods, apparatus, and products are disclosed for controlling data transfers from an origin compute node to a target compute node that include: receiving, by an application messaging module on the target compute node, an indication of a data transfer from an origin compute node to the target compute node; and administering, by the application messaging module on the target compute node, the data transfer using one or more messaging primitives of a system messaging module in dependence upon the indication.

    12. Scott Runnels of Computational Physics

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Scott Runnels of Computational Physics to teach at West Point March 19, 2013 LOS ALAMOS, N. M., March 19, 2013- Under an agreement between Los Alamos National Laboratory and the U.S. Military Academy, Scott Runnels has been selected for a two-year faculty post in the Department of Physics and Nuclear Engineering at West Point. The teaching position is intended to strengthen the ties between the U.S. national laboratories and the U.S. military academies by bringing in a top scientist to teach at

    13. computational-hydaulics-march-30

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      and Aerodynamics using STAR-CCM+ for CFD Analysis March 30-31, 2011 Argonne, Illinois Dr. Steven Lottes This email address is being protected from spambots. You need JavaScript enabled to view it. Announcement pdficon small A training course in the use of computational hydraulics and aerodynamics CFD software using CD-adapco's STAR-CCM+ for analysis was held at TRACC from March 30-31, 2011. The course assumes a basic knowledge of fluid mechanics and made extensive use of hands on tutorials.

    14. computational-structural-mechanics-training

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Table of Contents Date Location Training Course: HyperMesh and HyperView April 12-14, 2011 Argonne TRACC Argonne, IL Introductory Course: Developing Compute-efficient, Quality Models with LS-PrePost® 3 on the TRACC Cluster October 21-22, 2010 Argonne TRACC West Chicago, IL Modeling and Simulation with LS-DYNA®: Insights into Modeling with a Goal of Providing Credible Predictive Simulations February 11-12, 2010 Argonne TRACC West Chicago, IL Introductory Course: Using LS-OPT® on the TRACC

    15. An Arbitrary Precision Computation Package

      Energy Science and Technology Software Center (OSTI)

      2003-06-14

      This package permits a scientist to perform computations using an arbitrarily high level of numeric precision (the equivalent of hundreds or even thousands of digits), by making only minor changes to conventional C++ or Fortran-90 soruce code. This software takes advantage of certain properties of IEEE floating-point arithmetic, together with advanced numeric algorithms, custom data types and operator overloading. Also included in this package is the "Experimental Mathematician's Toolkit", which incorporates many of these facilitiesmore » into an easy-to-use interactive program.« less

    16. Computer network control plane tampering monitor

      DOE Patents [OSTI]

      Michalski, John T.; Tarman, Thomas D.; Black, Stephen P.; Torgerson, Mark D.

      2010-06-08

      A computer network control plane tampering monitor that detects unauthorized alteration of a label-switched path setup for an information packet intended for transmission through a computer network.

    17. Fermilab | Science | Particle Physics | Scientific Computing

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Today Fermilab serves as one of two US computing centers that processes and analyzes data from experiments at the Large Hadron Collider. The worldwide LHC computing project is one ...

    18. NERSC Intern Wins Award for Computing Achievement

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      (NCWIT) Aspirations in Computing award on Saturday, March 16, 2013 in a ceremony in San Jose, CA. The award honors young women at the high school level for their computing-related...

    19. The Brain: Key To a Better Computer

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Brain: Key To a Better Computer - Sandia Energy Energy Search Icon Sandia Home Locations ... Twitter Google + Vimeo GovDelivery SlideShare The Brain: Key To a Better Computer Home...

    20. High Performance Computational Biology: A Distributed computing Perspective (2010 JGI/ANL HPC Workshop)

      ScienceCinema (OSTI)

      Konerding, David [Google, Inc

      2011-06-08

      David Konerding from Google, Inc. gives a presentation on "High Performance Computational Biology: A Distributed Computing Perspective" at the JGI/Argonne HPC Workshop on January 26, 2010.

    1. Significant Enhancement of Computational Efficiency in Nonlinear Multiscale Battery Model for Computer Aided Engineering (Presentation)

      SciTech Connect (OSTI)

      Kim, G.; Pesaran, A.; Smith, K.; Graf, P.; Jun, M.; Yang, C.; Li, G.; Li, S.; Hochman, A.; Tselepidakis, D.; White, J.

      2014-06-01

      This presentation discusses the significant enhancement of computational efficiency in nonlinear multiscale battery model for computer aided engineering in current research at NREL.

    2. NERSC Intern Wins Award for Computing Achievement

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Intern Wins Award for Computing Achievement NERSC Intern Wins Award for Computing Achievement March 27, 2013 Linda Vu, lvu@lbl.gov, +1 510 495 2402 ncwit1 Stephanie Cabanela, a student intern in the National Energy Research Scientific Computing Center's (NERSC) Operation Technologies Group was honored with the Bay Area Affiliate National Center for Women and Information Technology (NCWIT) Aspirations in Computing award on Saturday, March 16, 2013 in a ceremony in San Jose, CA. The award honors

    3. Validating Computer-Designed Proteins for Vaccines

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Validating Computer-Designed Proteins for Vaccines Validating Computer-Designed Proteins for Vaccines Print Thursday, 21 August 2014 12:05 In the struggle to keep up with microbes whose rapid mutations outpace our ability to produce vaccines, the human race has a powerful ally: computers. Researchers have now figured out a way to use computational protein design to generate small, stable proteins that accurately mimic key viral structures; these can then be used in vaccines to induce potent

    4. Future Computing Needs for Innovative Confinement Concepts

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      of Plasma Science and Innovation Center Current Computing Utilization and Resources Near Term Needs Concluding Comments Future Computing Needs for Innovative Confinement Concepts Charlson C. Kim charlson@aa.washington.edu Plasma Science and Innovation Center University of Washington, Seattle August 3, 2010 Large Scale Computing Needs for Fusion Energy Science Workshop Rockville, MD Charlson C. Kim, PSI-Center Future Computing Needs of ICC's Introduction of Plasma Science and Innovation Center

    5. Computers and Monitors | The Ames Laboratory

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Computers and Monitors Buying a Computer or Monitor If you have a need to purchase a computer and/or monitor, follow this How To Guide to search the registry for EPEAT products. On your purchase requisition, indicate whether or not the item is EPEAT registered. *Acceptable Justifications/exceptions will be rare for computer or monitor through the Ames Laboratory storeroom. Both items offered through the storeroom are registered as EPEAT Gold* Office Electronics - look for ENERGY STAR and

    6. NERSC National Energy Research Scientific Computing Center

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      National Energy Research Scientific Computing Center 2007 Annual Report National Energy Research Scientific Computing Center 2007 Annual Report Ernest Orlando Lawrence Berkeley National Laboratory 1 Cyclotron Road, Berkeley, CA 94720-8148 This work was supported by the Director, Office of Science, Office of Ad- vanced Scientific Computing Research of the U.S. Department of Energy under Contract No. DE-AC02-05CH11231. LBNL-1143E, October 2008 iii National Energy Research Scientific Computing

    7. National Energ y Research Scientific Computing Center

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Annual Report This work was supported by the Director, Office of Science, Office of Advanced Scientific Computing Research of the U.S. Department of Energy under Contract No. DE-AC 03-76SF00098. LBNL-49186, December 2001 National Energ y Research Scientific Computing Center 2001 Annual Report NERSC aspires to be a world leader in accelerating scientific discovery through computation. Our vision is to provide high- performance computing tools to tackle science's biggest and most challenging

    8. Computing contingency statistics in parallel.

      SciTech Connect (OSTI)

      Bennett, Janine Camille; Thompson, David; Pebay, Philippe Pierre

      2010-09-01

      Statistical analysis is typically used to reduce the dimensionality of and infer meaning from data. A key challenge of any statistical analysis package aimed at large-scale, distributed data is to address the orthogonal issues of parallel scalability and numerical stability. Many statistical techniques, e.g., descriptive statistics or principal component analysis, are based on moments and co-moments and, using robust online update formulas, can be computed in an embarrassingly parallel manner, amenable to a map-reduce style implementation. In this paper we focus on contingency tables, through which numerous derived statistics such as joint and marginal probability, point-wise mutual information, information entropy, and {chi}{sup 2} independence statistics can be directly obtained. However, contingency tables can become large as data size increases, requiring a correspondingly large amount of communication between processors. This potential increase in communication prevents optimal parallel speedup and is the main difference with moment-based statistics where the amount of inter-processor communication is independent of data size. Here we present the design trade-offs which we made to implement the computation of contingency tables in parallel.We also study the parallel speedup and scalability properties of our open source implementation. In particular, we observe optimal speed-up and scalability when the contingency statistics are used in their appropriate context, namely, when the data input is not quasi-diffuse.

    9. Argonne's Laboratory computing center - 2007 annual report.

      SciTech Connect (OSTI)

      Bair, R.; Pieper, G. W.

      2008-05-28

      Argonne National Laboratory founded the Laboratory Computing Resource Center (LCRC) in the spring of 2002 to help meet pressing program needs for computational modeling, simulation, and analysis. The guiding mission is to provide critical computing resources that accelerate the development of high-performance computing expertise, applications, and computations to meet the Laboratory's challenging science and engineering missions. In September 2002 the LCRC deployed a 350-node computing cluster from Linux NetworX to address Laboratory needs for mid-range supercomputing. This cluster, named 'Jazz', achieved over a teraflop of computing power (1012 floating-point calculations per second) on standard tests, making it the Laboratory's first terascale computing system and one of the 50 fastest computers in the world at the time. Jazz was made available to early users in November 2002 while the system was undergoing development and configuration. In April 2003, Jazz was officially made available for production operation. Since then, the Jazz user community has grown steadily. By the end of fiscal year 2007, there were over 60 active projects representing a wide cross-section of Laboratory expertise, including work in biosciences, chemistry, climate, computer science, engineering applications, environmental science, geoscience, information science, materials science, mathematics, nanoscience, nuclear engineering, and physics. Most important, many projects have achieved results that would have been unobtainable without such a computing resource. The LCRC continues to foster growth in the computational science and engineering capability and quality at the Laboratory. Specific goals include expansion of the use of Jazz to new disciplines and Laboratory initiatives, teaming with Laboratory infrastructure providers to offer more scientific data management capabilities, expanding Argonne staff use of national computing facilities, and improving the scientific reach and performance of Argonne's computational applications. Furthermore, recognizing that Jazz is fully subscribed, with considerable unmet demand, the LCRC has framed a 'path forward' for additional computing resources.

    10. Radiological Safety Analysis Computer Program

      Energy Science and Technology Software Center (OSTI)

      2001-08-28

      RSAC-6 is the latest version of the RSAC program. It calculates the consequences of a release of radionuclides to the atmosphere. Using a personal computer, a user can generate a fission product inventory; decay and in-grow the inventory during transport through processes, facilities, and the environment; model the downwind dispersion of the activity; and calculate doses to downwind individuals. Internal dose from the inhalation and ingestion pathways is calculated. External dose from ground surface andmore » plume gamma pathways is calculated. New and exciting updates to the program include the ability to evaluate a release to an enclosed room, resuspension of deposited activity and evaluation of a release up to 1 meter from the release point. Enhanced tools are included for dry deposition, building wake, occupancy factors, respirable fraction, AMAD adjustment, updated and enhanced radionuclide inventory and inclusion of the dose-conversion factors from FOR 11 and 12.« less

    11. Collective network for computer structures

      DOE Patents [OSTI]

      Blumrich, Matthias A. (Ridgefield, CT); Coteus, Paul W. (Yorktown Heights, NY); Chen, Dong (Croton On Hudson, NY); Gara, Alan (Mount Kisco, NY); Giampapa, Mark E. (Irvington, NY); Heidelberger, Philip (Cortlandt Manor, NY); Hoenicke, Dirk (Ossining, NY); Takken, Todd E. (Brewster, NY); Steinmacher-Burow, Burkhard D. (Wernau, DE); Vranas, Pavlos M. (Bedford Hills, NY)

      2011-08-16

      A system and method for enabling high-speed, low-latency global collective communications among interconnected processing nodes. The global collective network optimally enables collective reduction operations to be performed during parallel algorithm operations executing in a computer structure having a plurality of the interconnected processing nodes. Router devices ate included that interconnect the nodes of the network via links to facilitate performance of low-latency global processing operations at nodes of the virtual network and class structures. The global collective network may be configured to provide global barrier and interrupt functionality in asynchronous or synchronized manner. When implemented in a massively-parallel supercomputing structure, the global collective network is physically and logically partitionable according to needs of a processing algorithm.

    12. Foundational Tools for Petascale Computing

      SciTech Connect (OSTI)

      Miller, Barton

      2014-05-19

      The Paradyn project has a history of developing algorithms, techniques, and software that push the cutting edge of tool technology for high-end computing systems. Under this funding, we are working on a three-year agenda to make substantial new advances in support of new and emerging Petascale systems. The overall goal for this work is to address the steady increase in complexity of these petascale systems. Our work covers two key areas: (1) The analysis, instrumentation and control of binary programs. Work in this area falls under the general framework of the Dyninst API tool kits. (2) Infrastructure for building tools and applications at extreme scale. Work in this area falls under the general framework of the MRNet scalability framework. Note that work done under this funding is closely related to work done under a contemporaneous grant, High-Performance Energy Applications and Systems, SC0004061/FG02-10ER25972, UW PRJ36WV.

    13. Collective network for computer structures

      DOE Patents [OSTI]

      Blumrich, Matthias A; Coteus, Paul W; Chen, Dong; Gara, Alan; Giampapa, Mark E; Heidelberger, Philip; Hoenicke, Dirk; Takken, Todd E; Steinmacher-Burow, Burkhard D; Vranas, Pavlos M

      2014-01-07

      A system and method for enabling high-speed, low-latency global collective communications among interconnected processing nodes. The global collective network optimally enables collective reduction operations to be performed during parallel algorithm operations executing in a computer structure having a plurality of the interconnected processing nodes. Router devices are included that interconnect the nodes of the network via links to facilitate performance of low-latency global processing operations at nodes of the virtual network. The global collective network may be configured to provide global barrier and interrupt functionality in asynchronous or synchronized manner. When implemented in a massively-parallel supercomputing structure, the global collective network is physically and logically partitionable according to the needs of a processing algorithm.

    14. ASCR Workshop on Quantum Computing for Science

      SciTech Connect (OSTI)

      Aspuru-Guzik, Alan; Van Dam, Wim; Farhi, Edward; Gaitan, Frank; Humble, Travis; Jordan, Stephen; Landahl, Andrew J; Love, Peter; Lucas, Robert; Preskill, John; Muller, Richard P.; Svore, Krysta; Wiebe, Nathan; Williams, Carl

      2015-06-01

      This report details the findings of the DOE ASCR Workshop on Quantum Computing for Science that was organized to assess the viability of quantum computing technologies to meet the computational requirements of the DOE’s science and energy mission, and to identify the potential impact of quantum technologies. The workshop was held on February 17-18, 2015, in Bethesda, MD, to solicit input from members of the quantum computing community. The workshop considered models of quantum computation and programming environments, physical science applications relevant to DOE's science mission as well as quantum simulation, and applied mathematics topics including potential quantum algorithms for linear algebra, graph theory, and machine learning. This report summarizes these perspectives into an outlook on the opportunities for quantum computing to impact problems relevant to the DOE’s mission as well as the additional research required to bring quantum computing to the point where it can have such impact.

    15. Cupola Furnace Computer Process Model

      SciTech Connect (OSTI)

      Seymour Katz

      2004-12-31

      The cupola furnace generates more than 50% of the liquid iron used to produce the 9+ million tons of castings annually. The cupola converts iron and steel into cast iron. The main advantages of the cupola furnace are lower energy costs than those of competing furnaces (electric) and the ability to melt less expensive metallic scrap than the competing furnaces. However the chemical and physical processes that take place in the cupola furnace are highly complex making it difficult to operate the furnace in optimal fashion. The results are low energy efficiency and poor recovery of important and expensive alloy elements due to oxidation. Between 1990 and 2004 under the auspices of the Department of Energy, the American Foundry Society and General Motors Corp. a computer simulation of the cupola furnace was developed that accurately describes the complex behavior of the furnace. When provided with the furnace input conditions the model provides accurate values of the output conditions in a matter of seconds. It also provides key diagnostics. Using clues from the diagnostics a trained specialist can infer changes in the operation that will move the system toward higher efficiency. Repeating the process in an iterative fashion leads to near optimum operating conditions with just a few iterations. More advanced uses of the program have been examined. The program is currently being combined with an ''Expert System'' to permit optimization in real time. The program has been combined with ''neural network'' programs to affect very easy scanning of a wide range of furnace operation. Rudimentary efforts were successfully made to operate the furnace using a computer. References to these more advanced systems will be found in the ''Cupola Handbook''. Chapter 27, American Foundry Society, Des Plaines, IL (1999).

    16. Chameleon: A Computer Science Testbed as Application of Cloud...

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Chameleon: A Computer Science Testbed as Application of Cloud Computing Event Sponsor: Mathematics and Computing Science Brownbag Lunch Start Date: Dec 15 2015 - 12:00pm Building...

    17. Large Scale Computing and Storage Requirements for Advanced Scientific...

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Large Scale Computing and Storage Requirements for Advanced Scientific Computing Research: Target 2014 ASCRFrontcover.png Large Scale Computing and Storage Requirements for ...

    18. Previous Computer Science Award Announcements | U.S. DOE Office...

      Office of Science (SC) Website

      Previous Computer Science Award Announcements Advanced Scientific Computing Research (ASCR) ASCR Home About Research Applied Mathematics Computer Science Exascale Tools Workshop...

    19. Technical Standards, Guidance on MELCOR computer code - May 3...

      Office of Environmental Management (EM)

      Standards, Guidance on MELCOR computer code - May 3, 2004 Technical Standards, Guidance on MELCOR computer code - May 3, 2004 May 3, 2004 MELCOR Computer Code Application Guidance...

    20. High-performance computer system installed at Los Alamos National...

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      High-performance computer system installed at Lab High-performance computer system installed at Los Alamos National Laboratory New high-performance computer system, called Wolf,...

    1. Locating hardware faults in a parallel computer

      DOE Patents [OSTI]

      Archer, Charles J.; Megerian, Mark G.; Ratterman, Joseph D.; Smith, Brian E.

      2010-04-13

      Locating hardware faults in a parallel computer, including defining within a tree network of the parallel computer two or more sets of non-overlapping test levels of compute nodes of the network that together include all the data communications links of the network, each non-overlapping test level comprising two or more adjacent tiers of the tree; defining test cells within each non-overlapping test level, each test cell comprising a subtree of the tree including a subtree root compute node and all descendant compute nodes of the subtree root compute node within a non-overlapping test level; performing, separately on each set of non-overlapping test levels, an uplink test on all test cells in a set of non-overlapping test levels; and performing, separately from the uplink tests and separately on each set of non-overlapping test levels, a downlink test on all test cells in a set of non-overlapping test levels.

    2. Impact analysis on a massively parallel computer

      SciTech Connect (OSTI)

      Zacharia, T.; Aramayo, G.A.

      1994-06-01

      Advanced mathematical techniques and computer simulation play a major role in evaluating and enhancing the design of beverage cans, industrial, and transportation containers for improved performance. Numerical models are used to evaluate the impact requirements of containers used by the Department of Energy (DOE) for transporting radioactive materials. Many of these models are highly compute-intensive. An analysis may require several hours of computational time on current supercomputers despite the simplicity of the models being studied. As computer simulations and materials databases grow in complexity, massively parallel computers have become important tools. Massively parallel computational research at the Oak Ridge National Laboratory (ORNL) and its application to the impact analysis of shipping containers is briefly described in this paper.

    3. Performing an allreduce operation on a plurality of compute nodes of a parallel computer

      DOE Patents [OSTI]

      Faraj, Ahmad (Rochester, MN)

      2012-04-17

      Methods, apparatus, and products are disclosed for performing an allreduce operation on a plurality of compute nodes of a parallel computer. Each compute node includes at least two processing cores. Each processing core has contribution data for the allreduce operation. Performing an allreduce operation on a plurality of compute nodes of a parallel computer includes: establishing one or more logical rings among the compute nodes, each logical ring including at least one processing core from each compute node; performing, for each logical ring, a global allreduce operation using the contribution data for the processing cores included in that logical ring, yielding a global allreduce result for each processing core included in that logical ring; and performing, for each compute node, a local allreduce operation using the global allreduce results for each processing core on that compute node.

    4. Multicore: Fallout from a Computing Evolution

      ScienceCinema (OSTI)

      Yelick, Kathy [Director, NERSC

      2009-09-01

      July 22, 2008 Berkeley Lab lecture: Parallel computing used to be reserved for big science and engineering projects, but in two years that's all changed. Even laptops and hand-helds use parallel processors. Unfortunately, the software hasn't kept pace. Kathy Yelick, Director of the National Energy Research Scientific Computing Center at Berkeley Lab, describes the resulting chaos and the computing community's efforts to develop exciting applications that take advantage of tens or hundreds of processors on a single chip.

    5. Validating Computer-Designed Proteins for Vaccines

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Validating Computer-Designed Proteins for Vaccines Print In the struggle to keep up with microbes whose rapid mutations outpace our ability to produce vaccines, the human race has a powerful ally: computers. Researchers have now figured out a way to use computational protein design to generate small, stable proteins that accurately mimic key viral structures; these can then be used in vaccines to induce potent neutralizing antibodies. The results were validated in part using protein structures

    6. Shaping Future Supercomputing Argonne Leadership Computing Facility

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      0 1 1 a n n u a l r e p o r t Shaping Future Supercomputing Argonne Leadership Computing Facility ANL-12/22 Argonne Leadership Computing Facility 2 0 1 1 a l c f a n n u a l r e p o r t w w w . a l c f . a n l . g o v Contents Overview .......................................2 Mira ..............................................4 Science Highlights ...........................8 Computing Resources ..................... 26 2011 ALCF Publications .................. 28 2012 INCITE Projects

    7. Present and Future Computing Requirements for PETSc

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      and Future Computing Requirements for PETSc Jed Brown jedbrown@mcs.anl.gov Mathematics and Computer Science Division, Argonne National Laboratory Department of Computer Science, University of Colorado Boulder NERSC ASCR Requirements for 2017 2014-01-15 Extending PETSc's Hierarchically Nested Solvers ANL Lois C. McInnes, Barry Smith, Jed Brown, Satish Balay UChicago Matt Knepley IIT Hong Zhang LBL Mark Adams Linear solvers, nonlinear solvers, time integrators, optimization methods (merged TAO)

    8. High Performance Computing Student Career Resources

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      HPC » Students High Performance Computing Student Career Resources Explore the multiple dimensions of a career at Los Alamos Lab: work with the best minds on the planet in an inclusive environment that is rich in intellectual vitality and opportunities for growth. Contact Us Student Liaison Josephine Kilde (505) 667-5086 Email High Performance Computing Capabilities The High Performance Computing (HPC) Division supports the Laboratory mission by managing world-class Supercomputing Centers. Our

    9. Validating Computer-Designed Proteins for Vaccines

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Validating Computer-Designed Proteins for Vaccines Print In the struggle to keep up with microbes whose rapid mutations outpace our ability to produce vaccines, the human race has a powerful ally: computers. Researchers have now figured out a way to use computational protein design to generate small, stable proteins that accurately mimic key viral structures; these can then be used in vaccines to induce potent neutralizing antibodies. The results were validated in part using protein structures

    10. Validating Computer-Designed Proteins for Vaccines

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Validating Computer-Designed Proteins for Vaccines Print In the struggle to keep up with microbes whose rapid mutations outpace our ability to produce vaccines, the human race has a powerful ally: computers. Researchers have now figured out a way to use computational protein design to generate small, stable proteins that accurately mimic key viral structures; these can then be used in vaccines to induce potent neutralizing antibodies. The results were validated in part using protein structures

    11. Validating Computer-Designed Proteins for Vaccines

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Validating Computer-Designed Proteins for Vaccines Print In the struggle to keep up with microbes whose rapid mutations outpace our ability to produce vaccines, the human race has a powerful ally: computers. Researchers have now figured out a way to use computational protein design to generate small, stable proteins that accurately mimic key viral structures; these can then be used in vaccines to induce potent neutralizing antibodies. The results were validated in part using protein structures

    12. INCITE Program | Argonne Leadership Computing Facility

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Science at ALCF Allocation Programs INCITE Program 5 Checks & 5 Tips for INCITE Mira Computational Readiness Assessment ALCC Program Director's Discretionary (DD) Program Early Science Program INCITE 2016 Projects ALCC 2015 Projects ESP Projects View All Projects Publications ALCF Tech Reports Industry Collaborations INCITE Program Innovative and Novel Computational Impact on Theory and Experiment (INCITE) Program The INCITE program provides allocations to computationally intensive,

    13. Improved computer models support genetics research

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      February » Simple computer models unravel genetic stress reactions in cells Simple computer models unravel genetic stress reactions in cells Integrated biological and computational methods provide insight into why genes are activated. February 8, 2013 When complete, these barriers will be a portion of the NMSSUP upgrade. This molecular structure depicts a yeast transfer ribonucleic acid (tRNA), which carries a single amino acid to the ribosome during protein construction. A combined

    14. Fermilab | Science at Fermilab | Computing | Networking

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Detectors and Computing Detectors and Computing Computing Networking Physicists are constantly exchanging information, within Fermilab and between Fermilab and collaborating institutions. They do this from the design phase of an experiment to long after they have finished collecting data. To move huge amounts of data from one place to another, Fermilab needs high-performance networking. For years, Fermilab has been the largest user of Energy Services Network, or ESnet, a network the Department

    15. Accounts Policy | Argonne Leadership Computing Facility

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Accounts Policy All holders of user accounts must abide by all appropriate Argonne Leadership Computing Facility and Argonne National Laboratory computing usage policies. These are described at the time of the account request and include requirements such as using a sufficiently strong password, appropriate use of the system, and so on. Any user not following these requirements will have their account disabled. Furthermore, ALCF resources are intended to be used as a computing resource for

    16. Introduction to High Performance Computing Using GPUs

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      HPC Using GPUs Introduction to High Performance Computing Using GPUs July 11, 2013 NERSC, NVIDIA, and The Portland Group presented a one-day workshop "Introduction to High Performance Computing Using GPUs" on July 11, 2013 in Room 250 of Sutardja Dai Hall on the University of California, Berkeley, campus. Registration was free and open to all NERSC users; Berkeley Lab Researchers; UC students, faculty, and staff; and users of the Oak Ridge Leadership Computing Facility. This workshop

    17. James Osborn | Argonne Leadership Computing Facility

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Osborn Computational Scientist / Catalyst Team Lead James Osborn Argonne National Laboratory 9700 South Cass Avenue Building 240 - Rm. 2126 Argonne, IL 60439 630-252-6327 osborn@alcf.anl.gov James Osborn is a Computational Scientist at the ALCF and a Fellow of the Computation Institute at The University of Chicago and Argonne. He specializes in the application of Lattice Field Theory, Random Matrix Theory, and cluster algorithms to strongly interacting systems. He is also interested in

    18. Marta Garcia Martinez | Argonne Leadership Computing Facility

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Marta Garcia Martinez Assistant Computational Scientist Marta Garcia Martinez Argonne National Laboratory 9700 South Cass Avenue Building 240 - Rm. 1132 Argonne IL, 60439 630-252-0091 mgarcia@alcf.anl.gov http://web.alcf.anl.gov/~mgarcia/ Marta García is an Assistant Computational Scientist at the ALCF. She is part of the Catalyst Team, where she focuses on assisting Computational Fluid Dynamics projects to maximize and accelerate their research on ALCF resources. She obtained a degree in

    19. Computer Networking Group | Stanford Synchrotron Radiation Lightsource

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Computer Networking Group Do you need help? For assistance please submit a CNG Help Request ticket. CNG Logo Chris Ramirez SSRL Computer and Networking Group (650) 926-2901 | email Jerry Camuso SSRL Computer and Networking Group (650) 926-2994 | email Networking Support The Networking group provides connectivity and communications services for SSRL. The services provided by the Networking Support Group include: Local Area Network support for cable and wireless connectivity. Installation and

    20. Digital computer operation of a nuclear reactor

      DOE Patents [OSTI]

      Colley, Robert W. (Richland, WA)

      1984-01-01

      A method is described for the safe operation of a complex system such as a nuclear reactor using a digital computer. The computer is supplied with a data base containing a list of the safe state of the reactor and a list of operating instructions for achieving a safe state when the actual state of the reactor does not correspond to a listed safe state, the computer selects operating instructions to return the reactor to a safe state.

    1. User Advisory Council | Argonne Leadership Computing Facility

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      About Overview History Staff Directory Our Teams User Advisory Council Careers Margaret Butler Fellowship Visiting Us Contact Us User Advisory Council The User Advisory Council meets regularly to review major policies and to provide user feedback to the facility leadership. All council members are active Principal Investigators or users of ALCF computational resources through one or more of the allocation programs. Martin Berzins Professor Department of Computer Science Scientific Computing and

    2. Improved computer models support genetics research

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Simple computer models unravel genetic stress reactions in cells Simple computer models unravel genetic stress reactions in cells Integrated biological and computational methods provide insight into why genes are activated. February 8, 2013 When complete, these barriers will be a portion of the NMSSUP upgrade. This molecular structure depicts a yeast transfer ribonucleic acid (tRNA), which carries a single amino acid to the ribosome during protein construction. A combined experimental and

    3. Digital computer operation of a nuclear reactor

      DOE Patents [OSTI]

      Colley, R.W.

      1982-06-29

      A method is described for the safe operation of a complex system such as a nuclear reactor using a digital computer. The computer is supplied with a data base containing a list of the safe state of the reactor and a list of operating instructions for achieving a safe state when the actual state of the reactor does not correspond to a listed safe state, the computer selects operating instructions to return the reactor to a safe state.

    4. About ALCF | Argonne Leadership Computing Facility

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      HPC at Argonne About ALCF The Argonne Leadership Computing Facility's (ALCF) mission is to accelerate major scientific discoveries and engineering breakthroughs for humanity by designing and providing world-leading computing facilities in partnership with the computational science community. We help researchers solve some of the world's largest and most complex problems with our unique combination of supercomputing resources and expertise. ALCF projects cover many scientific disciplines, ranging

    5. Covered Product Category: Computers | Department of Energy

      Office of Environmental Management (EM)

      Computers Covered Product Category: Computers The Federal Energy Management Program (FEMP) provides acquisition guidance for computers, a product category covered by the ENERGY STAR program. Federal laws and requirements mandate that agencies buy ENERGY STAR-qualified products in all product categories covered by this program and any acquisition actions that are not specifically exempted by law. MEETING EFFICIENCY REQUIREMENTS FOR FEDERAL PURCHASES The U.S. Environmental Protection Agency (EPA)

    6. Extreme Scale Computing, Co-design

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      ... software, and hardware in an integrated computational co-design process. * Designed Cruft, a suite of molecular dynamics proxy applications (software) developed to explore ...

    7. Computer virus information update CIAC-2301

      SciTech Connect (OSTI)

      Orvis, W.J.

      1994-01-15

      While CIAC periodically issues bulletins about specific computer viruses, these bulletins do not cover all the computer viruses that affect desktop computers. The purpose of this document is to identify most of the known viruses for the MS-DOS and Macintosh platforms and give an overview of the effects of each virus. The authors also include information on some windows, Atari, and Amiga viruses. This document is revised periodically as new virus information becomes available. This document replaces all earlier versions of the CIAC Computer virus Information Update. The date on the front cover indicates date on which the information in this document was extracted from CIAC`s Virus database.

    8. Bayesian approaches for combining computational model output...

      Office of Scientific and Technical Information (OSTI)

      for combining computational model output and physical observations Authors: Higdon, David M 1 ; Lawrence, Earl 1 ; Heitmann, Katrin 2 ; Habib, Salman 2 + Show Author...

    9. Climate Modeling using High-Performance Computing

      SciTech Connect (OSTI)

      Mirin, A A

      2007-02-05

      The Center for Applied Scientific Computing (CASC) and the LLNL Climate and Carbon Science Group of Energy and Environment (E and E) are working together to improve predictions of future climate by applying the best available computational methods and computer resources to this problem. Over the last decade, researchers at the Lawrence Livermore National Laboratory (LLNL) have developed a number of climate models that provide state-of-the-art simulations on a wide variety of massively parallel computers. We are now developing and applying a second generation of high-performance climate models. Through the addition of relevant physical processes, we are developing an earth systems modeling capability as well.

    10. Computable General Equilibrium Models for Sustainability Impact...

      Open Energy Info (EERE)

      Publications, Softwaremodeling tools User Interface: Other Website: iatools.jrc.ec.europa.eudocsecolecon2006.pdf Computable General Equilibrium Models for Sustainability...

    11. Computers in Commercial Buildings - Table 3

      U.S. Energy Information Administration (EIA) Indexed Site

      Back to Computers in Commercial Buildings Specific questions may be directed to: Jay Olsen jay.olsen@eia.doe.gov http:www.eia.govconsumptioncommercialdataarchive...

    12. Computers in Commercial Buildings - Table 2

      U.S. Energy Information Administration (EIA) Indexed Site

      Back to Computers in Commercial Buildings Specific questions may be directed to: Jay Olsen jay.olsen@eia.doe.gov http:www.eia.govconsumptioncommercialdataarchive...

    13. Computers in Commercial Buildings - Table 4

      U.S. Energy Information Administration (EIA) Indexed Site

      Back to Computers in Commercial Buildings Specific questions may be directed to: Jay Olsen jay.olsen@eia.doe.gov http:www.eia.govconsumptioncommercialdataarchive...

    14. Computers in Commercial Buildings - Table 1

      U.S. Energy Information Administration (EIA) Indexed Site

      Back to Computers in Commercial Buildings Specific questions may be directed to: Jay Olsen jay.olsen@eia.doe.gov http:www.eia.govconsumptioncommercialdataarchive...

    15. Computational Scientist | Princeton Plasma Physics Lab

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Department, with interest in leadership class computing of gyrokinetic fusion edge plasma research. A candidate who has knowledge in hybrid parallel programming with MPI, OpenMP,...

    16. Computational Design of Interfaces for Photovoltaics | Argonne...

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Computational Design of Interfaces for Photovoltaics PI Name: Noa Marom PI Email: nmarom@tulane.edu Institution: Tulane University Allocation Program: ALCC Allocation Hours at...

    17. Computational Challenges for Nanostructure Solar Cells

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Challenges for Nanostructure Solar Cells Computational Challenges for Nanostructure Solar Cells ZZ2.jpg Key Challenges: Current nanostructure solar cells often have energy...

    18. Integrated Computational Materials Engineering (ICME) for Mg...

      Office of Energy Efficiency and Renewable Energy (EERE) Indexed Site

      Project (Part 1) Integrated Computational Materials Engineering (ICME) for Mg: International Pilot Project (Part 1) 2010 DOE Vehicle Technologies and Hydrogen Programs Annual Merit...

    19. DOE Advanced Scientific Computing Advisory Subcommittee (ASCAC...

      Office of Scientific and Technical Information (OSTI)

      existing computer designs to reach exascale. The technical challenges revolve around energy consumption, memory performance, resilience, extreme concurrency, and big data....

    20. BETO Webinar: Computational Studies of Lignocellulose Deconstruction

      Office of Energy Efficiency and Renewable Energy (EERE) Indexed Site

      ... - lower temperature CH2-CH2 cleavage - facilitated by CaO Dehydrogenation Electron transfer > Exploiting quantum and classical computational methods to study the atomistic ...

    1. Hybrid Rotaxanes: Interlocked Structures for Quantum Computing...

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Hybrid Rotaxanes: Interlocked Structures for Quantum Computing? Print Rotaxanes are mechanically interlocked molecular architectures consisting of a dumbbell-shaped molecule, the...

    2. Hal Finkel | Argonne Leadership Computing Facility

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Hal Finkel Assistant Computational Scientist Hal Finkel Argonne National Laboratory 9700 South Cass Avenue Building 240 - Rm. 2126 Argonne, IL 60439 630-252-0023 hfinkel

    3. Anouar Benali | Argonne Leadership Computing Facility

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Anouar Benali Assistant Computational Scientist Anouar Benali Argonne National Laboratory 9700 South Cass Avenue Building 240 - Rm. 2127 Argonne, IL 60439 630-252-0058 benali@anl

    4. Adrian C. Pope | Argonne Leadership Computing Facility

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Adrian C. Pope Assistant Computational Scientist Argonne National Laboratory 9700 S. Cass Avenue Building 240 - Rm. 11213 Argonne, IL 60439 630-252-8745 apope

    5. Innovative Computational Tools for Reducing Exploration Risk...

      Office of Energy Efficiency and Renewable Energy (EERE) Indexed Site

      of Water-Rock Interactions and Magnetotelluric Surveys Innovative Computational Tools for Reducing Exploration Risk Through Integration of Water-Rock Interactions and ...

    6. Inexpensive computer data-acquisition system

      SciTech Connect (OSTI)

      Galvin, J.E.; Brown, I.G.

      1985-10-01

      A system based on an Apple II+ personal computer is used for on-line monitoring of ion-beam characteristics in accelerator ion source development.

    7. Ni Clusterbank Replacement Project | Argonne Leadership Computing...

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Ni Clusterbank Replacement Project Event Sponsor: Argonne Leadership Computing Facility Seminar Start Date: Oct 20 2015 - 12:00pm BuildingRoom: Building 241Room D173...

    8. Computer System Retirement Guidelines | Department of Energy

      Office of Energy Efficiency and Renewable Energy (EERE) Indexed Site

      contains DOE headquarters-specific information that may be adapted for use by any site or organization PDF icon Computer System Retirement Guidelines More Documents & Publications...

    9. Nichols A. Romero | Argonne Leadership Computing Facility

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Nichols A. Romero Principal Project Specialist - Computational Science Catalyst Team Lead Nichols Romero Argonne National Laboratory 9700 South Cass Avenue Building 240 - Rm....

    10. The Magellan Final Report on Cloud Computing

      SciTech Connect (OSTI)

      ,; Coghlan, Susan; Yelick, Katherine

      2011-12-21

      The goal of Magellan, a project funded through the U.S. Department of Energy (DOE) Office of Advanced Scientific Computing Research (ASCR), was to investigate the potential role of cloud computing in addressing the computing needs for the DOE Office of Science (SC), particularly related to serving the needs of mid- range computing and future data-intensive computing workloads. A set of research questions was formed to probe various aspects of cloud computing from performance, usability, and cost. To address these questions, a distributed testbed infrastructure was deployed at the Argonne Leadership Computing Facility (ALCF) and the National Energy Research Scientific Computing Center (NERSC). The testbed was designed to be flexible and capable enough to explore a variety of computing models and hardware design points in order to understand the impact for various scientific applications. During the project, the testbed also served as a valuable resource to application scientists. Applications from a diverse set of projects such as MG-RAST (a metagenomics analysis server), the Joint Genome Institute, the STAR experiment at the Relativistic Heavy Ion Collider, and the Laser Interferometer Gravitational Wave Observatory (LIGO), were used by the Magellan project for benchmarking within the cloud, but the project teams were also able to accomplish important production science utilizing the Magellan cloud resources.

    11. Cobalt Scheduler | Argonne Leadership Computing Facility

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Narayan Desai, "Bandwidth-Aware Resource Management for Extreme Scale Systems", IEEEACM International Conference for High Performance Computing, Networking, Storage, and...

    12. User Guides | Argonne Leadership Computing Facility

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      MiraCetusVesta Cooley Policies Documentation Feedback Please provide feedback to help guide us as we continue to build documentation for our new computing resource. Feedback...

    13. New User Guide | Argonne Leadership Computing Facility

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Documentation Feedback Please provide feedback to help guide us as we continue to build documentation for our new computing resource. Feedback Form New User Guide USER TIP:...

    14. Our Teams | Argonne Leadership Computing Facility

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      work directly with project teams to maximize and accelerate their research efforts. With multidisciplinary domain expertise, a deep knowledge of the ALCF computing environment,...

    15. Collaboration to advance high-performance computing

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      cyber security, data sharing and mobility, cloud computing, large-scale analytics, and materials science. This first Project Task Statement (PTS) under the Umbrella CRADA is...

    16. High Performance Computing Facility Operational Assessment, CY...

      Office of Scientific and Technical Information (OSTI)

      around the world for computational simulations relevant to national and energy security; advancing the frontiers of knowledge in physical sciences and areas of ...

    17. Smart Grid Computational Tool | Open Energy Information

      Open Energy Info (EERE)

      project benefits. The Smart Grid Computational Tool employs the benefit analysis methodology that DOE uses to evaluate the Recovery Act smart grid projects. How it works: The...

    18. Argonne's Super Computer Mira | Argonne National Laboratory

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Argonne's Super Computer Mira Share Browse By - Any - Energy -Energy efficiency --Vehicles ---Alternative fuels ---Diesel ---Electric drive technology ---Hybrid & electric vehicles...

    19. Modeling of Geothermal Reservoirs: Fundamental Processes, Computer...

      Open Energy Info (EERE)

      of Geothermal Reservoirs: Fundamental Processes, Computer Simulation and Field Applications Jump to: navigation, search OpenEI Reference LibraryAdd to library Journal Article:...

    20. Section 23: Models and Computer Codes

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Application-2014 for the Waste Isolation Pilot Plant Models and Computer Codes (40 CFR 194.23) United States Department of Energy Waste Isolation Pilot Plant Carlsbad Field...

    1. SciTech Connect: "high performance computing"

      Office of Scientific and Technical Information (OSTI)

      Advanced Search Term Search Semantic Search Advanced Search All Fields: "high performance computing" Semantic Semantic Term Title: Full Text: Bibliographic Data: Creator ...

    2. Radiation Detection Computational Benchmark Scenarios

      SciTech Connect (OSTI)

      Shaver, Mark W.; Casella, Andrew M.; Wittman, Richard S.; McDonald, Ben S.

      2013-09-24

      Modeling forms an important component of radiation detection development, allowing for testing of new detector designs, evaluation of existing equipment against a wide variety of potential threat sources, and assessing operation performance of radiation detection systems. This can, however, result in large and complex scenarios which are time consuming to model. A variety of approaches to radiation transport modeling exist with complementary strengths and weaknesses for different problems. This variety of approaches, and the development of promising new tools (such as ORNL’s ADVANTG) which combine benefits of multiple approaches, illustrates the need for a means of evaluating or comparing different techniques for radiation detection problems. This report presents a set of 9 benchmark problems for comparing different types of radiation transport calculations, identifying appropriate tools for classes of problems, and testing and guiding the development of new methods. The benchmarks were drawn primarily from existing or previous calculations with a preference for scenarios which include experimental data, or otherwise have results with a high level of confidence, are non-sensitive, and represent problem sets of interest to NA-22. From a technical perspective, the benchmarks were chosen to span a range of difficulty and to include gamma transport, neutron transport, or both and represent different important physical processes and a range of sensitivity to angular or energy fidelity. Following benchmark identification, existing information about geometry, measurements, and previous calculations were assembled. Monte Carlo results (MCNP decks) were reviewed or created and re-run in order to attain accurate computational times and to verify agreement with experimental data, when present. Benchmark information was then conveyed to ORNL in order to guide testing and development of hybrid calculations. The results of those ADVANTG calculations were then sent to PNNL for compilation. This is a report describing the details of the selected Benchmarks and results from various transport codes.

    3. In the OSTI Collections: High-Performance Computing | OSTI, US...

      Office of Scientific and Technical Information (OSTI)

      Performance Computing Computing efficiently Programming efficiently Correcting mistakes, avoiding failures Projections References Research Organizations Reports Available through ...

    4. Extreme Scale Computing to Secure the Nation

      SciTech Connect (OSTI)

      Brown, D L; McGraw, J R; Johnson, J R; Frincke, D

      2009-11-10

      Since the dawn of modern electronic computing in the mid 1940's, U.S. national security programs have been dominant users of every new generation of high-performance computer. Indeed, the first general-purpose electronic computer, ENIAC (the Electronic Numerical Integrator and Computer), was used to calculate the expected explosive yield of early thermonuclear weapons designs. Even the U. S. numerical weather prediction program, another early application for high-performance computing, was initially funded jointly by sponsors that included the U.S. Air Force and Navy, agencies interested in accurate weather predictions to support U.S. military operations. For the decades of the cold war, national security requirements continued to drive the development of high performance computing (HPC), including advancement of the computing hardware and development of sophisticated simulation codes to support weapons and military aircraft design, numerical weather prediction as well as data-intensive applications such as cryptography and cybersecurity U.S. national security concerns continue to drive the development of high-performance computers and software in the U.S. and in fact, events following the end of the cold war have driven an increase in the growth rate of computer performance at the high-end of the market. This mainly derives from our nation's observance of a moratorium on underground nuclear testing beginning in 1992, followed by our voluntary adherence to the Comprehensive Test Ban Treaty (CTBT) beginning in 1995. The CTBT prohibits further underground nuclear tests, which in the past had been a key component of the nation's science-based program for assuring the reliability, performance and safety of U.S. nuclear weapons. In response to this change, the U.S. Department of Energy (DOE) initiated the Science-Based Stockpile Stewardship (SBSS) program in response to the Fiscal Year 1994 National Defense Authorization Act, which requires, 'in the absence of nuclear testing, a progam to: (1) Support a focused, multifaceted program to increase the understanding of the enduring stockpile; (2) Predict, detect, and evaluate potential problems of the aging of the stockpile; (3) Refurbish and re-manufacture weapons and components, as required; and (4) Maintain the science and engineering institutions needed to support the nation's nuclear deterrent, now and in the future'. This program continues to fulfill its national security mission by adding significant new capabilities for producing scientific results through large-scale computational simulation coupled with careful experimentation, including sub-critical nuclear experiments permitted under the CTBT. To develop the computational science and the computational horsepower needed to support its mission, SBSS initiated the Accelerated Strategic Computing Initiative, later renamed the Advanced Simulation & Computing (ASC) program (sidebar: 'History of ASC Computing Program Computing Capability'). The modern 3D computational simulation capability of the ASC program supports the assessment and certification of the current nuclear stockpile through calibration with past underground test (UGT) data. While an impressive accomplishment, continued evolution of national security mission requirements will demand computing resources at a significantly greater scale than we have today. In particular, continued observance and potential Senate confirmation of the Comprehensive Test Ban Treaty (CTBT) together with the U.S administration's promise for a significant reduction in the size of the stockpile and the inexorable aging and consequent refurbishment of the stockpile all demand increasing refinement of our computational simulation capabilities. Assessment of the present and future stockpile with increased confidence of the safety and reliability without reliance upon calibration with past or future test data is a long-term goal of the ASC program. This will be accomplished through significant increases in the scientific bases that underlie the computational tools. Computer codes must be de

    5. Guide to Preventing Computer Software Piracy

      Broader source: Directives, Delegations, and Requirements [Office of Management (MA)]

      2001-07-12

      Guide to Preventing Computer Software Piracy It is the intent of the Department of Energy (DOE) to issue guidance in accordance with Federal CIO Council recommendations and in compliance with Executive Order 13103. The guidance in this document is based on the CIO Council's recommendations in reference to computer software piracy, and applies to all DOE elements. Canceled by DOE N 205.18

    6. Debugging a high performance computing program

      DOE Patents [OSTI]

      Gooding, Thomas M.

      2014-08-19

      Methods, apparatus, and computer program products are disclosed for debugging a high performance computing program by gathering lists of addresses of calling instructions for a plurality of threads of execution of the program, assigning the threads to groups in dependence upon the addresses, and displaying the groups to identify defective threads.

    7. Debugging a high performance computing program

      DOE Patents [OSTI]

      Gooding, Thomas M.

      2013-08-20

      Methods, apparatus, and computer program products are disclosed for debugging a high performance computing program by gathering lists of addresses of calling instructions for a plurality of threads of execution of the program, assigning the threads to groups in dependence upon the addresses, and displaying the groups to identify defective threads.

    8. Argonne's Magellan Cloud Computing Research Project

      ScienceCinema (OSTI)

      Beckman, Pete

      2013-04-19

      Pete Beckman, head of Argonne's Leadership Computing Facility (ALCF), discusses the Department of Energy's new $32-million Magellan project, which designed to test how cloud computing can be used for scientific research. More information: http://www.anl.gov/Media_Center/News/2009/news091014a.html

    9. Multiprocessor computer overset grid method and apparatus

      DOE Patents [OSTI]

      Barnette, Daniel W.; Ober, Curtis C.

      2003-01-01

      A multiprocessor computer overset grid method and apparatus comprises associating points in each overset grid with processors and using mapped interpolation transformations to communicate intermediate values between processors assigned base and target points of the interpolation transformations. The method allows a multiprocessor computer to operate with effective load balance on overset grid applications.

    10. Low Mach Number Models in Computational Astrophysics

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Ann Almgren Low Mach Number Models in Computational Astrophysics February 4, 2014 Ann Almgren. Berkeley Lab Downloads Almgren-nug2014.pdf | Adobe Acrobat PDF file Low Mach Number Models in Computational Astrophysics - Ann Almgren, Berkeley Lab Last edited: 2016-02-01 08:06:52

    11. Introduction to the session: Computing with biomolecules

      SciTech Connect (OSTI)

      Head, T.; Yokomori, Takashi

      1996-12-31

      This session presents three distinct presentations: the first presentation deals with the design of a photosensitive receptor from the protein bacteriorhodopsin; the second presentation deals with biocomputing applications using DNA molecules and enzymes to implement algorithmic computations. The third presentation discusses the computational generative capabilities of DNA subsegments.

    12. Pacing a data transfer operation between compute nodes on a parallel computer

      DOE Patents [OSTI]

      Blocksome, Michael A. (Rochester, MN)

      2011-09-13

      Methods, systems, and products are disclosed for pacing a data transfer between compute nodes on a parallel computer that include: transferring, by an origin compute node, a chunk of an application message to a target compute node; sending, by the origin compute node, a pacing request to a target direct memory access (`DMA`) engine on the target compute node using a remote get DMA operation; determining, by the origin compute node, whether a pacing response to the pacing request has been received from the target DMA engine; and transferring, by the origin compute node, a next chunk of the application message if the pacing response to the pacing request has been received from the target DMA engine.

    13. Intranode data communications in a parallel computer

      DOE Patents [OSTI]

      Archer, Charles J; Blocksome, Michael A; Miller, Douglas R; Ratterman, Joseph D; Smith, Brian E

      2013-07-23

      Intranode data communications in a parallel computer that includes compute nodes configured to execute processes, where the data communications include: allocating, upon initialization of a first process of a compute node, a region of shared memory; establishing, by the first process, a predefined number of message buffers, each message buffer associated with a process to be initialized on the compute node; sending, to a second process on the same compute node, a data communications message without determining whether the second process has been initialized, including storing the data communications message in the message buffer of the second process; and upon initialization of the second process: retrieving, by the second process, a pointer to the second process's message buffer; and retrieving, by the second process from the second process's message buffer in dependence upon the pointer, the data communications message sent by the first process.

    14. Intranode data communications in a parallel computer

      DOE Patents [OSTI]

      Archer, Charles J; Blocksome, Michael A; Miller, Douglas R; Ratterman, Joseph D; Smith, Brian E

      2014-01-07

      Intranode data communications in a parallel computer that includes compute nodes configured to execute processes, where the data communications include: allocating, upon initialization of a first process of a computer node, a region of shared memory; establishing, by the first process, a predefined number of message buffers, each message buffer associated with a process to be initialized on the compute node; sending, to a second process on the same compute node, a data communications message without determining whether the second process has been initialized, including storing the data communications message in the message buffer of the second process; and upon initialization of the second process: retrieving, by the second process, a pointer to the second process's message buffer; and retrieving, by the second process from the second process's message buffer in dependence upon the pointer, the data communications message sent by the first process.

    15. Predictive Dynamic Security Assessment through Advanced Computing

      SciTech Connect (OSTI)

      Huang, Zhenyu; Diao, Ruisheng; Jin, Shuangshuang; Chen, Yousu

      2014-11-30

      Abstract— Traditional dynamic security assessment is limited by several factors and thus falls short in providing real-time information to be predictive for power system operation. These factors include the steady-state assumption of current operating points, static transfer limits, and low computational speed. This addresses these factors and frames predictive dynamic security assessment. The primary objective of predictive dynamic security assessment is to enhance the functionality and computational process of dynamic security assessment through the use of high-speed phasor measurements and the application of advanced computing technologies for faster-than-real-time simulation. This paper presents algorithms, computing platforms, and simulation frameworks that constitute the predictive dynamic security assessment capability. Examples of phasor application and fast computation for dynamic security assessment are included to demonstrate the feasibility and speed enhancement for real-time applications.

    16. Low latency, high bandwidth data communications between compute nodes in a parallel computer

      DOE Patents [OSTI]

      Archer, Charles J. (Rochester, MN); Blocksome, Michael A. (Rochester, MN); Ratterman, Joseph D. (Rochester, MN); Smith, Brian E. (Rochester, MN)

      2010-11-02

      Methods, parallel computers, and computer program products are disclosed for low latency, high bandwidth data communications between compute nodes in a parallel computer. Embodiments include receiving, by an origin direct memory access (`DMA`) engine of an origin compute node, data for transfer to a target compute node; sending, by the origin DMA engine of the origin compute node to a target DMA engine on the target compute node, a request to send (`RTS`) message; transferring, by the origin DMA engine, a predetermined portion of the data to the target compute node using memory FIFO operation; determining, by the origin DMA engine whether an acknowledgement of the RTS message has been received from the target DMA engine; if the an acknowledgement of the RTS message has not been received, transferring, by the origin DMA engine, another predetermined portion of the data to the target compute node using a memory FIFO operation; and if the acknowledgement of the RTS message has been received by the origin DMA engine, transferring, by the origin DMA engine, any remaining portion of the data to the target compute node using a direct put operation.

    17. Method for transferring data from an unsecured computer to a secured computer

      DOE Patents [OSTI]

      Nilsen, Curt A. (Castro Valley, CA)

      1997-01-01

      A method is described for transferring data from an unsecured computer to a secured computer. The method includes transmitting the data and then receiving the data. Next, the data is retransmitted and rereceived. Then, it is determined if errors were introduced when the data was transmitted by the unsecured computer or received by the secured computer. Similarly, it is determined if errors were introduced when the data was retransmitted by the unsecured computer or rereceived by the secured computer. A warning signal is emitted from a warning device coupled to the secured computer if (i) an error was introduced when the data was transmitted or received, and (ii) an error was introduced when the data was retransmitted or rereceived.

    18. Method and system for benchmarking computers

      DOE Patents [OSTI]

      Gustafson, John L.

      1993-09-14

      A testing system and method for benchmarking computer systems. The system includes a store containing a scalable set of tasks to be performed to produce a solution in ever-increasing degrees of resolution as a larger number of the tasks are performed. A timing and control module allots to each computer a fixed benchmarking interval in which to perform the stored tasks. Means are provided for determining, after completion of the benchmarking interval, the degree of progress through the scalable set of tasks and for producing a benchmarking rating relating to the degree of progress for each computer.

    19. Traffic information computing platform for big data

      SciTech Connect (OSTI)

      Duan, Zongtao Li, Ying Zheng, Xibin Liu, Yan Dai, Jiting Kang, Jun

      2014-10-06

      Big data environment create data conditions for improving the quality of traffic information service. The target of this article is to construct a traffic information computing platform for big data environment. Through in-depth analysis the connotation and technology characteristics of big data and traffic information service, a distributed traffic atomic information computing platform architecture is proposed. Under the big data environment, this type of traffic atomic information computing architecture helps to guarantee the traffic safety and efficient operation, more intelligent and personalized traffic information service can be used for the traffic information users.

    20. Mobile Visitors in Particular to Note Web Upgrade

      Broader source: Energy.gov [DOE]

      WASHINGTON, D.C. – DOE recently launched changes to Energy.gov, which includes the public website for EM.

    1. Visitor Info | NEES - EFRC | University of Maryland Energy Frontier

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Research Center Visiting the University of Maryland Air Travel The University is served by three major airports. All have transportation to the College Park area. Baltimore/Washington International Thurgood Marshall Airport (BWI) provides regular shuttle service to the College Park area or a convenient drive by car. Reagan National Airport is not only an easy car ride from College Park, it is also located right on the green Metrorail line. Passengers can take public transportation directly

    2. NREL: Technology Transfer - Kuwait Visitors Interested in NREL...

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      to improve energy efficiency in their refining operations. KOC may also apply concentrated solar power technology to produce some of the steam needed in the company's operations...

    3. Fernald Preserve Visitors Center Grand Opening and LEED Platinum...

      Office of Energy Efficiency and Renewable Energy (EERE) Indexed Site

      ... URS, Southwest Services, and Water Quality Systems Incorporated for their efforts. ... site will rapidly become a resource for teachers throughout the tri-state area of ...

    4. Large Scale Computing and Storage Requirements for Advanced Scientific

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Computing Research: Target 2014 Large Scale Computing and Storage Requirements for Advanced Scientific Computing Research: Target 2014 ASCRFrontcover.png Large Scale Computing and Storage Requirements for Advanced Scientific Computing Research An ASCR / NERSC Review January 5-6, 2011 Final Report Large Scale Computing and Storage Requirements for Advanced Scientific Computing Research, Report of the Joint ASCR / NERSC Workshop conducted January 5-6, 2011 Goals This workshop is being

    5. About the Advanced Computing Tech Team | Department of Energy

      Energy Savers [EERE]

      Advanced Computing Tech Team About the Advanced Computing Tech Team The Advanced Computing Tech Team is made up of representatives from DOE and its national laboratories who are involved with developing and using advanced computing tools. The following is a list of some of those programs and what how they are currently using advanced computing in pursuit of their respective missions. Advanced Science Computing Research (ASCR) The mission of the Advanced Scientific Computing Research (ASCR)

    6. Computer Model Buildings Contaminated with Radioactive Material

      Energy Science and Technology Software Center (OSTI)

      1998-05-19

      The RESRAD-BUILD computer code is a pathway analysis model designed to evaluate the potential radiological dose incurred by an individual who works or lives in a building contaminated with radioactive material.

    7. Venkatram Vishwanath | Argonne Leadership Computing Facility

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      leadership-class computers, including IO forwarding and power consumption on the Blue GenP and Blue GeneQ systems. Vishwanath won a Department of Energy SciDAC Scientific...

    8. Quantum Process Matrix Computation by Monte Carlo

      Energy Science and Technology Software Center (OSTI)

      2012-09-11

      The software package, processMC, is a python script that allows for the rapid modeling of small , noisy quantum systems and the computation of the averaged quantum evolution map.

    9. Bridging the Gap to 64-bit Computing

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Opteron and AMD64 A Commodity 64 bit x86 SOC Fred Weber Vice President and CTO Computation Products Group Advanced Micro Devices 22 April 2003 AMD - Salishan HPC 2003 2 Opteron...

    10. Validating Computer-Designed Proteins for Vaccines

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      apply to a variety of other vaccine targets, such as human immunodeficiency virus and influenza. Wanted: Dead or Computed As strange as it sounds, most vaccines are composed of...

    11. Validating Computer-Designed Proteins for Vaccines

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      keep up with microbes whose rapid mutations outpace our ability to produce vaccines, the human race has a powerful ally: computers. Researchers have now figured out a way to use...

    12. Data triage enables extreme-scale computing

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      ranking large-scale data. The researchers presented a whitepaper on the subject for the Big Data Exascale Computing workshop in Japan. Significance of the research The main focus...

    13. High Performance Computing Data Center (Fact Sheet)

      SciTech Connect (OSTI)

      Not Available

      2012-08-01

      This two-page fact sheet describes the new High Performance Computing Data Center being built in the ESIF and talks about some of the capabilities and unique features of the center.

    14. Bringing Advanced Computational Techniques to Energy Research

      SciTech Connect (OSTI)

      Mitchell, Julie C

      2012-11-17

      Please find attached our final technical report for the BACTER Institute award. BACTER was created as a graduate and postdoctoral training program for the advancement of computational biology applied to questions of relevance to bioenergy research.

    15. THE CENTER FOR DATA INTENSIVE COMPUTING

      SciTech Connect (OSTI)

      GLIMM,J.

      2002-11-01

      CDIC will provide state-of-the-art computational and computer science for the Laboratory and for the broader DOE and scientific community. We achieve this goal by performing advanced scientific computing research in the Laboratory's mission areas of High Energy and Nuclear Physics, Biological and Environmental Research, and Basic Energy Sciences. We also assist other groups at the Laboratory to reach new levels of achievement in computing. We are ''data intensive'' because the production and manipulation of large quantities of data are hallmarks of scientific research in the 21st century and are intrinsic features of major programs at Brookhaven. An integral part of our activity to accomplish this mission will be a close collaboration with the University at Stony Brook.

    16. THE CENTER FOR DATA INTENSIVE COMPUTING

      SciTech Connect (OSTI)

      GLIMM,J.

      2001-11-01

      CDIC will provide state-of-the-art computational and computer science for the Laboratory and for the broader DOE and scientific community. We achieve this goal by performing advanced scientific computing research in the Laboratory's mission areas of High Energy and Nuclear Physics, Biological and Environmental Research, and Basic Energy Sciences. We also assist other groups at the Laboratory to reach new levels of achievement in computing. We are ''data intensive'' because the production and manipulation of large quantities of data are hallmarks of scientific research in the 21st century and are intrinsic features of major programs at Brookhaven. An integral part of our activity to accomplish this mission will be a close collaboration with the University at Stony Brook.

    17. THE CENTER FOR DATA INTENSIVE COMPUTING

      SciTech Connect (OSTI)

      GLIMM,J.

      2003-11-01

      CDIC will provide state-of-the-art computational and computer science for the Laboratory and for the broader DOE and scientific community. We achieve this goal by performing advanced scientific computing research in the Laboratory's mission areas of High Energy and Nuclear Physics, Biological and Environmental Research, and Basic Energy Sciences. We also assist other groups at the Laboratory to reach new levels of achievement in computing. We are ''data intensive'' because the production and manipulation of large quantities of data are hallmarks of scientific research in the 21st century and are intrinsic features of major programs at Brookhaven. An integral part of our activity to accomplish this mission will be a close collaboration with the University at Stony Brook.

    18. Wei Jiang | Argonne Leadership Computing Facility

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Wei Jiang Assistant Computational Scientist Wei Jiang Argonne National Laboratory 9700 S. Cass Avenue Building 240 - Rm. 2128 Argonne IL, 60439 630-252-8688 wjiang@alcf.anl.gov...

    19. High Performance Computing Data Center (Fact Sheet)

      SciTech Connect (OSTI)

      Not Available

      2014-08-01

      This two-page fact sheet describes the new High Performance Computing Data Center in the ESIF and talks about some of the capabilities and unique features of the center.

    20. Christopher J. Knight | Argonne Leadership Computing Facility

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Christopher J. Knight Assistant Computational Scientist Christopher Knight Argonne National Laboratory 9700 S. Cass Avenue Building 240 - Rm. 1132 Argonne, IL 60439 630-252-9793 knightc@anl

    1. Validating Computer-Designed Proteins for Vaccines

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Computed As strange as it sounds, most vaccines are composed of actual dead viruses and bacteria. The idea is that presenting a dead form of the pathogen will fake your body into...

    2. PCs and Computer Terminals in Commercial Buildings

      U.S. Energy Information Administration (EIA) Indexed Site

      thousand users would be much lower. The more PCs and computer terminals used in a given building, the greater the impact on the building's energy consumption. By this measure,...

    3. Applications in Data-Intensive Computing

      SciTech Connect (OSTI)

      Shah, Anuj R.; Adkins, Joshua N.; Baxter, Douglas J.; Cannon, William R.; Chavarra-Miranda, Daniel; Choudhury, Sutanay; Gorton, Ian; Gracio, Deborah K.; Halter, Todd D.; Jaitly, Navdeep; Johnson, John R.; Kouzes, Richard T.; Macduff, Matt C.; Marquez, Andres; Monroe, Matthew E.; Oehmen, Christopher S.; Pike, William A.; Scherrer, Chad; Villa, Oreste; Webb-Robertson, Bobbie-Jo M.; Whitney, Paul D.; Zuljevic, Nino

      2010-04-01

      This book chapter, to be published in Advances in Computers, Volume 78, in 2010 describes applications of data intensive computing (DIC). This is an invited chapter resulting from a previous publication on DIC. This work summarizes efforts coming out of the PNNL's Data Intensive Computing Initiative. Advances in technology have empowered individuals with the ability to generate digital content with mouse clicks and voice commands. Digital pictures, emails, text messages, home videos, audio, and webpages are common examples of digital content that are generated on a regular basis. Data intensive computing facilitates human understanding of complex problems. Data-intensive applications provide timely and meaningful analytical results in response to exponentially growing data complexity and associated analysis requirements through the development of new classes of software, algorithms, and hardware.

    4. Innovative & Novel Computational Impact on Theory & Experiement...

      Office of Science (SC) Website

      DOE Leadership Computing Web Site External link Last modified: 352016 7:57:43 PM Share Page Share with Facebook Facebook External link Share with Twitter Twitter External link ...

    5. Computes Generalized Electromagnetic Interactions Between Structures

      Energy Science and Technology Software Center (OSTI)

      1999-02-20

      Object oriented software for computing generalized electromagnetic interactions between structures in the frequency domains. The software is based on integral equations. There is also a static integral equation capability.

    6. Session on computation in biological pathways

      SciTech Connect (OSTI)

      Karp, P.D.; Riley, M.

      1996-12-31

      The papers in this session focus on the development of pathway databases and computational tools for pathway analysis. The discussion involves existing databases of sequenced genomes, as well as techniques for studying regulatory pathways.

    7. Validating Computer-Designed Proteins for Vaccines

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Validating Computer-Designed Proteins for Vaccines Print In the struggle to keep up with microbes whose rapid mutations outpace our ability to produce vaccines, the human race has...

    8. Visualization Clusters | Argonne Leadership Computing Facility

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Cooley has a total of 126 compute nodes; each node has 12 CPU cores and one NVIDIA Tesla K80 dual-GPU card. Aggregate GPU peak performance is over 293 teraflops double...

    9. Human Brain vs. Computer | GE Global Research

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Computer Processors Beat the Human Mind in the Future? Click to email this to a friend (Opens in new window) Share on Facebook (Opens in new window) Click to share (Opens in new ...

    10. ACM TOMS replicated computational results initiative

      DOE Public Access Gateway for Energy & Science Beta (PAGES Beta)

      Heroux, Michael Allen

      2015-06-03

      In this study, the scientific community relies on the peer review process for assuring the quality of published material, the goal of which is to build a body of work we can trust. Computational journals such as The ACM Transactions on Mathematical Software (TOMS) use this process for rigorously promoting the clarity and completeness of content, and citation of prior work. At the same time, it is unusual to independently confirm computational results.

    11. ACM TOMS replicated computational results initiative

      SciTech Connect (OSTI)

      Heroux, Michael Allen

      2015-06-03

      In this study, the scientific community relies on the peer review process for assuring the quality of published material, the goal of which is to build a body of work we can trust. Computational journals such as The ACM Transactions on Mathematical Software (TOMS) use this process for rigorously promoting the clarity and completeness of content, and citation of prior work. At the same time, it is unusual to independently confirm computational results.

    12. Sandia National Laboratories: Advances toward quantum computing

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      By Neal Singer Photography By Randy Montoya Thursday, March 03, 2016 Precise atom implants in silicon provide first step toward quantum computers Precise atom implants in silicon provide first step toward quantum computers Post-doc fellows Meenakshi Singh and Jose Pacheco stand in front of Sandia's ion beam generator. Project PI Meenakshi holds a sample qubit structure embedded in silicon. (Photo by Randy Montoya) Sandia researchers have supported post-doctoral fellow Meenakshi Singh (1132) in

    13. computational-fluid-dynamics-student-thesis

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Fluid Dynamics Student Thesis Abstract DEVELOPMENT OF A THREE-DIMENSIONAL SCOURING METHODOLOGY AND ITS IMPLEMENTATION IN A COMMERCIAL CFD CODE FOR OPEN CHANNEL FLOW OVER A FLOODED BRIDGE DECK The Computational Fluid Dynamics staff at TRACC is supporting three students from Northern Illinois University who are working for a Masters degree. The CFD staff is directing the thesis research and working with them on three projects: (1) a three-dimensional scour computation methodology for pressure flow

    14. High Performance Computing Data Center Metering Protocol

      Office of Energy Efficiency and Renewable Energy (EERE) Indexed Site

      High Performance Computing Data Center Metering Protocol Prepared for: U.S. Department of Energy Office of Energy Efficiency and Renewable Energy Federal Energy Management Program Prepared by: Thomas Wenning Michael MacDonald Oak Ridge National Laboratory September 2010 ii Introduction Data centers in general are continually using more compact and energy intensive central processing units, but the total number and size of data centers continues to increase to meet progressive computing

    15. Applying computationally efficient schemes for biogeochemical cycles

      Office of Scientific and Technical Information (OSTI)

      (ACES4BGC) (Technical Report) | SciTech Connect Applying computationally efficient schemes for biogeochemical cycles (ACES4BGC) Citation Details In-Document Search Title: Applying computationally efficient schemes for biogeochemical cycles (ACES4BGC) NCAR contributed to the ACES4BGC project through software engineering work on aerosol model implementation, build system and script changes, coupler enhancements for biogeochemical tracers, improvements to the Community Land Model (CLM) code and

    16. Data triage enables extreme-scale computing

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Data triage enables extreme-scale computing Data triage enables extreme-scale computing Data selection and triage are important techniques for large-scale data, which can drastically reduce the amount of data written to disk or transmitted over a network. August 1, 2014 Spatial partitioning for the ocean simulation data set. Spatial partitioning for the ocean simulation data set. The main focus for ADR is to prioritize data primarily generated by large-scale scientific simulations run on

    17. Graham Fletcher | Argonne Leadership Computing Facility

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Graham Fletcher Principal Project Specialist in Computational Science Graham Fletcher Argonne National Laboratory 9700 South Cass Avenue Building 240 - Rm. 1123 Argonne, IL 60439 630-252-0755 fletcher@alcf.anl.gov Graham Fletcher is a Principal Project Specialist in Computational Science at the ALCF with a background in quantum chemistry and supercomputing. His research interests focus on the development of highly scalable methods and algorithms for the accurate and reliable prediction of

    18. Extreme Scale Computing, Co-Design

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Information Science, Computing, Applied Math » Extreme Scale Computing, Co-design » Publications Publications Ramon Ravelo, Qi An, Timothy C. Germann, and Brad Lee Holian, "Large-scale molecular dynamics simulations of shock induced plasticity in tantalum single crystals," AIP Conference Proceedings 1426, 1263-1266 (2012). Frank J. Cherne, Guy Dimonte, and Timothy C. Germann, "Richtymer-Meshkov instability examined with large-scale molecular dynamics simulations," AIP

    19. Extreme Scale Computing, Co-design

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Extreme Scale Computing, Co-design Informing system design, ensuring productive and efficient code Project Description To address the increasingly complex problems of the modern world, scientists at Los Alamos are pushing the scale of computing to the extreme, forming partnerships with other national laboratories and industry to develop supercomputers that can achieve "exaflop" speeds-that is, a quintillion (a million trillion) calculations per second. To put such speed in perspective,

    20. Account Information | Argonne Leadership Computing Facility

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Account Information Accounts and Access FAQ Connect & Log In Using CRYPTOCards Disk Space Quota Management Allocations Mira/Cetus/Vesta Cooley Policies Documentation Feedback Please provide feedback to help guide us as we continue to build documentation for our new computing resource. [Feedback Form] Account Information All computing carried out on the ALCF systems is associated with a user "account." This account is used to log onto the login servers and run jobs on the

    1. LANL computer model boosts engine efficiency

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      LANL computer model boosts engine efficiency LANL computer model boosts engine efficiency The KIVA model has been instrumental in helping researchers and manufacturers understand combustion processes, accelerate engine development and improve engine design and efficiency. September 25, 2012 KIVA simulation of an experimental engine with DOHC quasi-symmetric pent-roof combustion chamber and 4 valves. KIVA simulation of an experimental engine with DOHC quasi-symmetric pent-roof combustion chamber

    2. Computational Research and Theory (CRT) Facility

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Computational Research and Theory (CRT) Facility Community Berkeley Global Campus Environmental Documents Tours Community Programs Friends of Berkeley Lab ⇒ Navigate Section Community Berkeley Global Campus Environmental Documents Tours Community Programs Friends of Berkeley Lab Project Description The Computational Research and Theory (CRT) Facility will be on the forefront of high-performance supercomputing research and will be DOE's most efficient facility of its kind. Designed to take

    3. Mark Hereld | Argonne Leadership Computing Facility

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Hereld Manager, Visualization and Data Analysis Mark Hereld Argonne National Laboratory 9700 South Cass Avenue Building 240 - Rm. 4139 Argonne, IL 60439 630-252-4170 hereld@mcs.anl.gov Mark Hereld is the ALCF's Visualization and Data Analysis Manager. He is also a member of the research staff in Argonne's Mathematics and Computer Science Division and a Senior Fellow of the Computation Institute with a joint appointment at the University of Chicago. His work in understanding simulation on future

    4. Michael Papka | Argonne Leadership Computing Facility

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Papka Division Director Michael Papka Argonne National Laboratory 9700 South Cass Avenue Building 240 - Rm. 4134 Argonne, IL 60439 630-252-1556 papka@anl.gov http://papka.alcf.anl.gov Michael E. Papka is the Director of the ALCF. He is also Argonne's Deputy Associate Laboratory Director for Computing, Environment and Life Sciences. Both his laboratory leadership roles and his research interests relate to high-performance computing in support of scientific discovery. Dr. Papka holds a Senior

    5. Introducing Aurora | Argonne Leadership Computing Facility

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Aurora Announcement Press Conference: DOE announces next-gen supercomputer Aurora to be built at Argonne Introducing Aurora Author: ALCF staff April 9, 2015 Facebook Twitter LinkedIn Google E-mail Printer-friendly version Today, U.S. Department of Energy Under Secretary for Science and Energy Lynn Orr announced two new High Performance Computing (HPC) awards that continue to advance U.S. leadership in developing exascale computing. The announcement was made alongside leaders from Argonne

    6. Computational Spectroscopy of Heterogeneous Interfaces | Argonne Leadership

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Computing Facility Complex interfaces between nanoparticles and a solvent Complex interfaces between nanoparticles and a solvent. N. Brawand, University of Chicago Computational Spectroscopy of Heterogeneous Interfaces PI Name: Giulia Galli PI Email: gagalli@uchicago.edu Institution: University of Chicago Allocation Program: INCITE Allocation Hours at ALCF: 150 Million Year: 2016 Research Domain: Materials Science The interfaces between solids, nanoparticles and liquids play a fundamental

    7. ALCF Acknowledgment Policy | Argonne Leadership Computing Facility

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      ALCF Acknowledgment Policy As a U.S. Department of Energy user facility dedicated to the advancement of scientific discoveries, the Argonne Leadership Computing Facility (ALCF) provides unique computing resources and expertise to a user community that is bound by certain policies designed to acknowledge and promote the work of others as well as the resources used to accomplish this work. The ALCF requests your continued compliance with the terms of your program or discretionary award,

    8. Low Mach Number Models in Computational Astrophysics

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      In memoriam: Michael Welcome 1957 - 2014 RIP Almgren CCSE Low Mach Number Models in Computational Astrophysics Ann Almgren Center for Computational Sciences and Engineering Lawrence Berkeley National Laboratory NUG 2014: NERSC@40 February 4, 2014 Collaborators: John Bell, Chris Malone, Andy Nonaka, Stan Woosley, Michael Zingale Almgren CCSE Introduction We often associate astrophysics with explosive phenomena: novae supernovae gamma-ray bursts X-ray bursts Type Ia Supernovae Largest

    9. Staff Directory | Argonne Leadership Computing Facility

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      About Overview History Staff Directory Our Teams User Advisory Council Careers Margaret Butler Fellowship Visiting Us Contact Us Staff Directory Yury Alekseev Yuri Alexeev Assistant Computational Scientist Catalyst 630-252-0157 yuri@alcf.anl.gov Bill Allcock Bill Allcock Manager, Advanced Integration Group Leadership, AIG 630-252-7573 allcock@anl.gov Ben Allen HPC Systems Administration Specialist Systems 630-252-0554 allen@alcf.anl.gov Ramesh Balakrishnan Computational Scientist Catalyst

    10. Hybrid Rotaxanes: Interlocked Structures for Quantum Computing?

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Hybrid Rotaxanes: Interlocked Structures for Quantum Computing? Hybrid Rotaxanes: Interlocked Structures for Quantum Computing? Print Wednesday, 26 August 2009 00:00 Rotaxanes are mechanically interlocked molecular architectures consisting of a dumbbell-shaped molecule, the "axle," that threads through a ring called a macrocycle. Because the rings can spin around and slide along the axle, rotaxanes are promising components of molecular machines. While most rotaxanes have been entirely

    11. Advances and Challenges in Computational Plasma Science

      SciTech Connect (OSTI)

      W.M. Tang; V.S. Chan

      2005-01-03

      Scientific simulation, which provides a natural bridge between theory and experiment, is an essential tool for understanding complex plasma behavior. Recent advances in simulations of magnetically-confined plasmas are reviewed in this paper with illustrative examples chosen from associated research areas such as microturbulence, magnetohydrodynamics, and other topics. Progress has been stimulated in particular by the exponential growth of computer speed along with significant improvements in computer technology.

    12. Accounts & Access | Argonne Leadership Computing Facility

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      Account Information Accounts and Access FAQ Connect & Log In Using CRYPTOCards Disk Space Quota Management Allocations Mira/Cetus/Vesta Cooley Policies Documentation Feedback Please provide feedback to help guide us as we continue to build documentation for our new computing resource. [Feedback Form] Accounts & Access Account Information Account Information: All computing carried out on the ALCF systems is associated with a user "account." This account is used to log onto the

    13. Avanced Large-scale Integrated Computational Environment

      Energy Science and Technology Software Center (OSTI)

      1998-10-27

      The ALICE Memory Snooper is a software applications programming interface (API) and library for use in implementing computational steering systems. It allows distributed memory parallel programs to publish variables in the computation that may be accessed over the Internet. In this way, users can examine and even change the variables in their running application remotely. The API and library ensure the consistency of the variables across the distributed memory system.

    14. BETO Webinar: Computational Studies of Lignocellulose Deconstruction |

      Office of Energy Efficiency and Renewable Energy (EERE) Indexed Site

      Department of Energy Webinar: Computational Studies of Lignocellulose Deconstruction BETO Webinar: Computational Studies of Lignocellulose Deconstruction Dr. Gnanakaran of Los Alamos National Laboratory (LANL) presented LANL's molecular research on lignocellulose on April 15, 2013, at the Bioenergy Technologies Office's webinar series. PDF icon april2013_lanl_webinar.pdf More Documents & Publications Process Design and Economics for the Conversion of Lignocellulosic Biomass to

    15. High-performance computing for airborne applications

      SciTech Connect (OSTI)

      Quinn, Heather M; Manuzzato, Andrea; Fairbanks, Tom; Dallmann, Nicholas; Desgeorges, Rose

      2010-06-28

      Recently, there has been attempts to move common satellite tasks to unmanned aerial vehicles (UAVs). UAVs are significantly cheaper to buy than satellites and easier to deploy on an as-needed basis. The more benign radiation environment also allows for an aggressive adoption of state-of-the-art commercial computational devices, which increases the amount of data that can be collected. There are a number of commercial computing devices currently available that are well-suited to high-performance computing. These devices range from specialized computational devices, such as field-programmable gate arrays (FPGAs) and digital signal processors (DSPs), to traditional computing platforms, such as microprocessors. Even though the radiation environment is relatively benign, these devices could be susceptible to single-event effects. In this paper, we will present radiation data for high-performance computing devices in a accelerated neutron environment. These devices include a multi-core digital signal processor, two field-programmable gate arrays, and a microprocessor. From these results, we found that all of these devices are suitable for many airplane environments without reliability problems.

    16. Computer System, Cluster, and Networking Summer Institute Program Description

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      System, Cluster, and Networking Summer Institute Program Description The Computer System, Cluster, and Networking Summer Institute (CSCNSI) is a focused technical enrichment program targeting third-year college undergraduate students currently engaged in a computer science, computer engineering, or similar major. The program emphasizes practical skill development in setting up, configuring, administering, testing, monitoring, and scheduling computer systems, supercomputer clusters, and computer

    17. ASCR Budget | U.S. DOE Office of Science (SC)

      Office of Science (SC) Website

      Budget Advanced Scientific Computing Research (ASCR) ASCR Home About Organization Chart .pdf file (40KB) Staff ASCR Budget ASCR Committees of Visitors Directions Jobs Research...

    18. Mobile computing device configured to compute irradiance, glint, and glare of the sun

      DOE Patents [OSTI]

      Gupta, Vipin P; Ho, Clifford K; Khalsa, Siri Sahib

      2014-03-11

      Described herein are technologies pertaining to computing the solar irradiance distribution on a surface of a receiver in a concentrating solar power system or glint/glare emitted from a reflective entity. A mobile computing device includes at least one camera that captures images of the Sun and the entity of interest, wherein the images have pluralities of pixels having respective pluralities of intensity values. Based upon the intensity values of the pixels in the respective images, the solar irradiance distribution on the surface of the entity or glint/glare corresponding to the entity is computed by the mobile computing device.

    19. Demystifying computer code for northern New Mexico students

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      school students in northern New Mexico try their hands at computer programming during the Computer Science Education Week's Hour of Code event, a one-hour introduction to computer...

    20. ALCF summer students gain experience with high-performance computing...

      Broader source: All U.S. Department of Energy (DOE) Office Webpages (Extended Search)

      of computing that my textbooks couldn't keep up with," said Brown, who is majoring in computer science and computer game design. "Getting exposed to many-core machines and...