HPX V0.9.6 Released

The STE||AR Group is proud to announce the sixth formal release of HPX (V0.9.6).  We would like to thank everyone for their time and dedication which continues to push HPX to the edge of parallel computation.

HPX (High Performance ParalleX) provides a unified programming model for parallel and distributed applications of any scale. It is the first freely available, open source, feature-complete, modular, and performance oriented implementation of the ParalleX execution model. HPX is a general purpose C++ runtime system for applications targeted at conventional, widely available architectures.

With the changes below, HPX is leading the charge of a whole new era of computation.  By intrinsically breaking down and synchronizing the work to be done, HPX insures that application developers will no longer have to fret about where a segment of code executes.  HPX allows coders to focus their time and energy to understanding the data dependencies of their algorithms and thereby the core obstacles to an efficient code.  Here are some of the advantages of using HPX:

  • HPX exposes an API equivalent to the facilities as standardized by C++11/14 extended to distributed computing. Everything programmers know about primitives in the standard C++ library is still valid in the context of HPX.
  • There is no need for the programmer to worry about lower level parallelization paradigms like threads or message passing; no need to understand pthreads, MPI, OpenMP, or Windows threads, etc.
  • There is no need to think about different types of parallelism such as tasks, pipelines, or fork-join, task or data parallelism.
  • The same source of your program compiles and runs on Linux, MacOS, Windows, and Android.
  • The same code runs on shared memory multi-core systems and supercomputers, on handheld devices and Xeon-Phi accelerators, or a heterogeneous mix of those.

In this release we have made several significant changes:

  • Consolidated API to be aligned with the C++11 (and the future C++14) Standard
  • Implemented a distributed version of our Active Global Address Space (AGAS)
  • Ported HPX to the Xeon-Phi device
  • Added support for the SLURM scheduling system
  • Improved the performance counter framework
  • Added parcel (message) compression and parcel coalescing systems
  • Allow different scheduling polices for different parts of code with experimental executors API
  • Added experimental security support on the locality level
  • Created a native transport layer on top of Infiniband networks
  • Created a native transport layer on top of low level MPI functions
  • Added an experimental tuple-space object

We hope you will try out V0.9.6 and begin to contemplate where HPX can take your applications to the next level.

You can download the release here, or get HPX directly from GitHub.  If you have suggestions, questions, or ideas we would love to hear from you. You can find us at our website, reach us at hpx-users@stellar.cct.lsu.edu, or chat with us live on IRC in the #ste||ar chat room on Freenode.

GD Star Rating
loading...
HPX V0.9.6 Released, 3.5 out of 5 based on 2 ratings
    This entry was posted in General, Release and tagged by Hartmut Kaiser. Bookmark the permalink.

    About Hartmut Kaiser

    Hartmut is an Adjunct Professor of Computer Science at Louisiana State University. At the same time, he holds the position of a senior scientist at the Center for Computation and Technology (LSU). He received his doctorate from the Technical University of Chemnitz (Germany) in 1988. He is probably best known through his involvement in open source software projects, mainly as the author of several C++ libraries he has contributed to Boost, which are in use by thousands of developers worldwide. His current research is focused on leading the STE||AR group at CCT working on the practical design and implementation of the ParalleX execution model and related programming methods. In addition, he architected and developed the core library modules of SAGA for C++, a Simple API for Grid Applications.

    Leave a Reply

    Your email address will not be published. Required fields are marked *