skip to main content
OSTI.GOV title logo U.S. Department of Energy
Office of Scientific and Technical Information
  1. Polydots, soft nanoparticles, at membrane interfaces

    Soft nanoparticles (NPs) are emerging candidates for nano medicine, particularly for intercellular imaging and targeted drug delivery. Their soft nature, manifested in their dynamics, allows translocation into organisms without damaging their membranes. A crucial step towards incorporating soft dynamic NPs in nano medicine, is to resolve their interrelation with membranes. Here using atomistic molecular dynamics (MD) simulations we probe the interaction of soft NPs formed by conjugated polymers with a model membrane. These NPs, often termed polydots, are confined to their nano dimensions without any chemical tethers, forming dynamic long lived nano structures. Specifically, polydots formed by dialkyl para polymore » phenylene ethylene (PPE), with a varying number of carboxylate groups tethered to the alkyl chains to tune the interfacial charge of the surface of the NP are investigated at the interface with a model membrane that consists of di-palmitoyl phosphatidylcholine (DPPC). We find that even though polydots are controlled only by physical forces, they retain their NP configuration as they transcend the membrane. Regardless of their size, neutral polydots spontaneously penetrate the membrane whereas carboxylated polydots must be driven in, with a force that depends on the charge at their interface, all without significant disruption to the membrane. These fundamental results provide a means to control the position of the nanoparticles with respect to the membrane interfaces, which is key to their therapeutic use.« less
  2. Code modernization strategies for short-range non-bonded molecular dynamics simulations

    Modern HPC systems are increasingly relying on greater core counts and wider vector registers. Thus, applications need to be adapted to fully utilize these hardware capabilities. One class of applications that can benefit from this increase in parallelism are molecular dynamics simulations. In this paper, we describe our efforts at modernizing the ESPResSo++ simulation package for molecular dynamics by restructuring its particle data layout for efficient memory accesses and applying vectorization techniques to benefit the calculation of short-range non-bonded forces, which results in an overall three times speedup and serves as a baseline for further optimizations. We also implement fine-grainedmore » parallelism for multi-core CPUs through HPX, a C++ runtime system which uses lightweight threads and an asynchronous many-task approach to maximize concurrency. Our goal is to evaluate the performance of an HPX-based approach compared to the bulk-synchronous MPI-based implementation. This requires the introduction of an additional layer to the domain decomposition scheme that defines the task granularity. On spatially inhomogeneous systems, which impose a corresponding load-imbalance in traditional MPI-based approaches, we demonstrate that by choosing an optimal task size, the efficient work-stealing mechanisms of HPX can overcome the overhead of communication resulting in an overall 1.4 times speedup compared to the baseline MPI version.« less
  3. GLUE Code: A framework handling communication and interfaces between scales

    Many scientific applications are inherently multiscale in nature. Such complex physical phenomena often require simultaneous execution and coordination of simulations spanning multiple time and length scales. This is possible by combining expensive small-scale simulations (such as molecular dynamics simulations) with larger scale simulations (such continuum limit/hydro solvers) to allow for considerably larger systems using task and data parallelism. However, the granularity of the tasks can be very large and often leads to load imbalance. Traditionally, we use approximations to streamline the computation of the more costly interactions and this introduces trade-offs between simulation cost and accuracy. In recent years, themore » available computational power and the advances in machine learning have made computing these scale-bridging interactions and multiscale simulations more feasible. One driving application has been plasma modeling in inertial confinement fusion (ICF), which is fundamentally multiscale in nature. This requires deep understanding of how to extrapolate microscopic information into macroscopically relevant scales. For example, in ICF one needs an accurate understanding of the connection between experimental observables and the underlying microphysics. The properties of the larger scales are often affected by the microscale behavior incorporated usually into the equations of state and ionic and electronic transport coefficients (Liboff, 1959; Rinderknecht et al., 2014; Rosenberg et al., 2015; Ross et al., 2017). Instead of incorporating this information using reliable molecular dynamics (MD) simulations, one often needs to use theoretical models, due to the inability of MD to reach engineering scales (Glosli et al., 2007; Marinak et al., 1998). One approach to resolve this issue is by coupling two MD simulations of different scales via force interpolation, e.g., the AdResS method (Krekeler et al., 2018; Nagarajan et al., 2013). Another approach, which we will pursue in the scope of this work, is by enabling scale bridging between MD simulations and meso/macro-scale models through the development and support of application programming interfaces that these different applications can interact through.« less
  4. Parthenon—a performance portable block-structured adaptive mesh refinement framework

    On the path to exascale the landscape of computer device architectures and corresponding programming models has become much more diverse. While various low-level performance portable programming models are available, support at the application level lacks behind. To address this issue, we present the performance portable block-structured adaptive mesh refinement (AMR) framework Parthenon, derived from the well-tested and widely used Athena++ astrophysical magnetohydrodynamics code, but generalized to serve as the foundation for a variety of downstream multi-physics codes. Parthenon adopts the Kokkos programming model, and provides various levels of abstractions from multidimensional variables, to packages defining and separating components, to launchingmore » of parallel compute kernels. Parthenon allocates all data in device memory to reduce data movement, supports the logical packing of variables and mesh blocks to reduce kernel launch overhead, and employs one-sided, asynchronous MPI calls to reduce communication overhead in multi-node simulations. Using a hydrodynamics miniapp, we demonstrate weak and strong scaling on various architectures including AMD and NVIDIA GPUs, Intel and AMD x86 CPUs, IBM Power9 CPUs, as well as Fujitsu A64FX CPUs. At the largest scale on Frontier (the first TOP500 exascale machine), the miniapp reaches a total of 1.7 × 10 13 zone-cycles/s on 9216 nodes (73,728 logical GPUs) at [Formula: see text] weak scaling parallel efficiency (starting from a single node). In combination with being an open, collaborative project, this makes Parthenon an ideal framework to target exascale simulations in which the downstream developers can focus on their specific application rather than on the complexity of handling massively-parallel, device-accelerated AMR.« less
  5. Cabana: A Performance Portable Library for Particle-Based Simulations

    Particle-based simulations are ubiquitous throughout many fields of computational science and engineering, spanning the atomistic level with molecular dynamics (MD), to mesoscale particle-in-cell (PIC) simulations for solid mechanics, device-scale modeling with PIC methods for plasma physics, and massive N-body cosmology simulations of galaxy structures, with many other methods in between (Hockney & Eastwood, 1989). While these methods use particles to represent significantly different entities with completely different physical models, many low-level details are shared including performant algorithms for short- and/or long-range particle interactions, multi-node particle communication patterns, and other data management tasks such as particle sorting and neighbor list construction.more » Cabana is a performance portable library for particle-based simulations, developed as part of the Co-Design Center for Particle Applications (CoPA) within the Exascale Computing Project (ECP) (Alexander et al., 2020). The CoPA project and its full development scope, including ECP partner applications, algorithm development, and similar software libraries for quantum MD, is described in (Mniszewski et al., 2021). Cabana uses the Kokkos library for on-node parallelism (Edwards et al., 2014; Trott et al., 2022), enabling simulation on multi-core CPU and GPU architectures, and MPI for GPU-aware, multi-node communication. Cabana provides particle simulation capabilities on almost all current Kokkos backends, including serial execution, OpenMP (including OpenMP-Target for GPUs), CUDA (NVIDIA GPUs), HIP (AMD GPUs), and SYCL (Intel GPUs), providing a clear path for the coming generation of accelerator-based exascale hardware. Cabana builds on Kokkos by providing new particle data structures and particle algorithms resulting in a similar execution policy-based, node-level programming model that is intended to be used in addition to the core Kokkos library within an application. Cabana is designed as an application and physics agnostic, but particle-specific toolkit which can either be used to generate a new application, or to be used as needed in existing applications at various levels of invasiveness including through interfaces that wrap user memory in existing data structures.« less
  6. Asynchronous distributed-memory task-parallel algorithm for compressible flows on unstructured 3D Eulerian grids

    Here, we discuss the implementation of a finite element method, used to numerically solve the Euler equations of compressible flows, using an asynchronous runtime system (RTS). The algorithm is implemented for distributed-memory machines, using stationary unstructured 3D meshes, combining data-, and task-parallelism on top of the Charm++ RTS. Charm++’s execution model is asynchronous by default, allowing arbitrary overlap of computation and communication. Task-parallelism allows scheduling parts of an algorithm independently of, or dependent on, each other. Built-in automatic load balancing enables continuous redistribution of computational load by migration of work units based on real-time CPU load measurement. The RTS alsomore » features automatic checkpointing, fault tolerance, resilience against hardware failure, and supports power-, and energy-aware computation. We demonstrate scalability up to 25 x 109 cells at $$\mathscr{O}$$104 compute cores and the benefits of automatic load balancing for irregular workloads. The full source code with documentation is available at https://quinoacomputing.org.« less
  7. Proxy App Suite Release (FY2020)

    Version 4.0 of the ECP Proxy App Suite is practically unchanged from the previous release. The current set of proxies has proven useful for many aspects of benchmarking and co-design and we see little reason to alter the suite. Although there have been few changes to the ECP suite, the team has been hard at work in other areas. In the area of Machine Learning (ML) we have now created a separate proxy suite dedicated to this scientific applications of ML.
  8. Excited-state electronic structure of molecules using many-body Green’s functions: Quasiparticles and electron–hole excitations with VOTCA-XTP

    Here, we present the open-source VOTCA-XTP software for the calculation of the excited-state electronic structure of molecules using many-body Green’s function theory in the GW approximation with the Bethe–Salpeter equation (BSE). This work provides a summary of the underlying theory and discusses the details of its implementation based on Gaussian orbitals, including resolution-of-identity techniques and different approaches to the frequency integration of the self-energy or acceleration by offloading compute-intensive matrix operations using graphics processing units in a hybrid OpenMP/Cuda scheme. A distinctive feature of VOTCA–XTP is the capability to couple the calculation of electronic excitations to a classical polarizable environmentmore » on an atomistic level in a coupled quantum- and molecular-mechanics (QM/MM) scheme, where a complex morphology can be imported from Molecular Dynamics simulations. The capabilities and limitations of the GW–BSE implementation are illustrated with two examples. First, we study the dependence of optically active electron–hole excitations in a series of diketopyrrolopyrrole-based oligomers on molecular-architecture modifications and the number of repeat units. Second, we use the GW–BSE/MM setup to investigate the effect of polarization on localized and intermolecular charge-transfer excited states in morphologies of low-donor content rubrene–fullerene mixtures. These showcases demonstrate that our implementation currently allows us to treat systems with up to 2500 basis functions on regular shared-memory workstations, providing accurate descriptions of quasiparticle and coupled electron–hole excited states of various characters on an equal footing.« less
  9. The Ristra Project: FY20/21 Milestone Report

    The ASC Advanced Technology Development and Mitigation (ATDM) sub-program was established in 2014 to develop new simulation tools operating on exascale-class computers to serve NNSA (see Appendix B). Over the course of ATDM, LANL management have set a strategy for exascale-class application codes that follows two supportive and mutually risk-mitigating paths: evolution for established production integrated design codes (IDCs) – with a strong pedigree within the user community – based upon existing programming paradigms(MPI+X); and a new start ATDM project, Ristra, a high-risk/high-reward push for a next-generation multi-physics, multi-scale simulation toolkit based on emerging advanced programming systems(with an initial focusmore » on data-flow task-based models exemplified by Legion). The role of Ristra as the high-risk/high-reward path for LANL’s codes was fully consistent with the goals of ATDM as described in Appendix B, in particular its emphasis on evolving ASC capabilities through novel computing programming models and computing technologies.« less
  10. Asynchronous Navier-Stokes Solver on 3D Unstructured Grids for the Exascale Era

    This project has developed multiple fluid dynamics solvers for complex 3D flows using fully asynchronous distributed-memory task-parallel algorithms on top of the Charm++ runtime system. The algorithms solve the Euler or Navier-Stokes equations of compressible flows using unstructured tetrahedron meshes with optional solution-adaptive mesh-, and polynomial-order refinement. We have demonstrated excellent strong scaling up 50K compute cores and the benefits of Charm++’s automatic load balancing.
...

Search for:
All Records
Author / Contributor
0000000309251458

Refine by:
Resource Type
Availability
Publication Date
Author / Contributor
Research Organization