Recently released Clang 3.8 fully supports OpenMP 3.1 and some features of OpenMP 4.0 and 4.5.
For details, see the Clang 3.8 Release Notes.
Pre-compiled packages can be downloaded from http://llvm.org/releases/download.html#3.8.0
IWOMP 2016 - 12th International Workshop on OpenMP
October 5-7, 2016
For many years, OpenMP has provided a very rich and flexible programming model for shared memory architectures. OpenMP 4.0 was a major advance that added two new forms of parallelism and OpenMP 4.5 added substantial improvements for them and other aspects of the specification. OpenMP now includes a rich set of device constructs to support heterogeneous systems like GPUs,FPGAs, DSPs, and SIMD constructs to support vector units in addition to the support for task-based parallelism that has been present since OpenMP 3.0.
The International Workshop on OpenMP (IWOMP) is an annual workshop dedicated to the promotion and advancement of all aspects of parallel programming with OpenMP. It is the premier forum to present and discuss issues, trends, recent research ideas, and results related to parallel programming with OpenMP. We solicit submissions of unpublished technical papers detailing innovative, original research and development related to OpenMP.
All topics related to OpenMP are of interest, including OpenMP performance analysis and modeling, OpenMP performance and correctness tools, proposed OpenMP extensions, and OpenMP applications in any domain (e.g., scientific and numerical computation, video games, computer graphics, multimedia, information retrieval, optimization, text processing, data mining, finance, signal and image processing and machine learning).
Advances in technologies, such as multi-core processors and OpenMP devices (accelerators such as GPGPUs, DSPs or FPGAs), Multiprocessor Systems on a Chip (MPSoCs), and recent developments in OpenMP itself (e.g., devices) present new opportunities and challenges for software and hardware developers. Recent advances in the C, C++ and Fortran base languages also offer interesting opportunities and challenges to the OpenMP programming model. IWOMP 2016 particularly solicits submissions in these areas as well as ones that discuss how to apply OpenMP to additional models of parallelism such as event loops.
Paper Submission and Registration
Submitted papers for review should be limited to 12 pages and follow LNCS guidelines. Submission deadline is May 18th, 2016. Submit your paper to: https://www.easychair.org/conferences/?conf=iwomp2016. Authors of accepted papers will be asked to prepare a final paper of up to 15 pages. As in previous years, IWOMP 2016 will publish formal proceedings of the accepted papers in Springer Verlag’s LNCS series.
Paper Submission Deadline: May 18th, 2016
Notification of acceptance: June 18th, 2016
Deadline for final version: July 8th, 2016
- Naoya Maruyama, RIKEN Advanced Institute for Computational Science, Japan
Program Committee Co-chairs:
- Naoya Maruyama, RIKEN Advanced Institute for Computational Science, Japan
- Bronis R. de Supinski, Lawrence Livermore National Laboratory, USA
Videos and slides of most of the presentations given at our first OpenMPCon developers conference, held September 29-30 at Aachen University, are now available on our YouTube channel as a playlist.
Here are the individual videos:
Five short presentations given by members of the OpenMP team in the OpenMP booth at SC15 are now available on YouTube (playlist):
- OpenMP Tasking (Slides)
- Getting OpenMP Up to Speed (Slides)
- OpenMP for Embedded Systems (Slides)
- OpenMP and Parallware (Slides)
- SC 15 OpenMP - Performance and Energy Optimization with OpenMP (Slides)
The OpenMP ARB is pleased to announce OpenMP 4.5, a major upgrade of the OpenMP standard language specifications. This release provides a substantial improvement on the support for programming of accelerator and GPU devices, and supports now also the parallelization of loops with well-structured dependencies. Implementation is underway in GCC and Clang. The new specification can be downloaded from here.
Standard for parallel programming extends its reach
With this release, OpenMP, the de-facto standard for parallel programming on shared memory systems, continues to extend its reach beyond pure HPC to include DSPs, real time systems, and accelerators. OpenMP aims to provide high-level parallel language support for a wide range of applications, from biotech and automotive to aeronautics, automation, robotics and financial analysis.
“OpenMP 4.5 is a significant achievement that demonstrates the industry-wide collaboration and the hard work and dedication within the OpenMP community,” says Michael Wong, CEO of the OpenMP ARB. “It is more than a minor release, representing the road towards OpenMP 5.0 while we continue on a cadence that delivers Technical Reports and/or Ratified Specifications annually, in keeping pace with the marketplace.”
Many new features
- Significantly improved support for devices. OpenMP now provides mechanisms for unstructured data mapping and asynchronous execution and also runtime routines for device memory management. These routines allow for allocating, copying and freeing.
- Support for doacross loops. A natural mechanism to parallelize loops with well-structured dependences is provided.
- New taskloop construct. Support to divide loops into tasks, avoiding the requirement that all threads execute the loop.
- Reductions for C/C++ arrays. This often requested feature is now available by building on support for array sections.
- New hint mechanisms. Hint mechanisms can provide guidance on the relative priority of tasks and on preferred synchronization implementations.
- Thread affinity support. It is now possible to use runtime functions to determine the effect of thread affinity clauses.
- Improved support for Fortran 2003. Users can now parallelize many Fortran 2003 programs.
- SIMD extensions. These extensions include the ability to specify exact SIMD width and additional data-sharing attributes.
Implementation is already almost complete in GCC version 6.0. It is starting in the current trunk of Clang 4.8. Other vendor compilers are following.
OpenMP 4.5 Specifications
- OpenMP 4.5 Complete Specifications (November 2015) PDF
- OpenMP 4.5 Summary Card - C/C++ (November 2015) PDF
- OpenMP 4.5 Summary Card - Fortran (November 2015) PDF
OpenMP: Where Are We and What’s Next? [Room 18AB]
We will showcase modern OpenMP (“The language that lets you use all your compute resources”) by presenting OpenMP 4.1 and the vision for OpenMP 5.0. Experts will give short presentations on key OpenMP 4.1 features and answer your technical questions. OpenMP CEO Michael Wong will present our vision for OpenMP 5.0 and we’ll finish with an audience-led discussion with a panel including members of the OpenMP ARB.
- Sunday, Nov 15th, 8:30am - 5:00pm
Hands-on Introduction to OpenMP [Room 17B]
Bring your laptop with installed OpenMP compiler for a hands-on introduction.
- Monday, Nov 16th, 8:30am - 5:00pm
Advanced OpenMP Tutorial: Performance and 4.1 Features [Room 13AB]
OpenMP performance, parallelization strategies, advanced features, and more.
- Monday, Nov 16th, 1:30pm - 5:00pm
Debugging and Performance Tools for MPI and OpenMP 4.0 APIs [Room 14]
Parallel debugging and optimization focused on techniques used with accelerators and coprocessors.
RELAX WITH A COLD BEER:
Attend a short talk in our booth [Booth# 2036, meet some of the pros behind the API, and participate in drawings after each talk for the latest book by James Reinders and Jim Jeffers.
- Tuesday, Nov 18th, 11:15 - 11:40am
Power Optimization Strategies in OpenMP – Abid Malik, University of Houston
- Tuesday, Nov 18th, 2:15 - 2:40pm
OpenMP and the Multicore Associates (MCA) – Sunita Chandrasekaran, (University of Delaware)
- Wednesday, Nov 19th, 11:15 - 11:40am
Getting Open Up to Speed – Ruud van der Pas, Oracle
- Wednesday, Nov 19th, 2:15 - 2:40pm
OpenMP Tasking – Michael Klemm, Intel and Christian Terboven, RWTH-Aachen
- Thursday, Nov 20th, 11:15 - 11:30am
OpenMP and Parallware – Manuel Arenaz, Appentra
- Overcoming Distributed Debugging Challenges in MPI + OpenMP Programming Model
- Directive-Based Pipelining Extension for OpenMP
- A Standard Debug Interface for OpenMP Target Regions
- A Deadlock Detection Concept for OpenMP Tasks and Fully Hybrid MPI-OpenMP Applications
Cray: Booth 1833
EPCC: Booth 2503
Fujitsu: Booth 1827
HP: Booth 603
IBM: Booth 522
Intel: Booths 1333, 1533
NASA: Booth 333
NEC: Booth 313
Nvidia: Booth 1021
Oracle: Booth 1327
RedHat: Booth 3112
TACC: Booth 548
Texas Instruments (in the Rapid IO Trade Assoc. Booth): Booth 2917
We managed to interview some of the participants, and those short interview videos are now up on our ≫OpenMP YouTube channel
Topics ranged from OpenMP SIMD to OpenACC and OpenCL, OpenMP usage in the national labs and in financial applications, automatic parallelization tools from Appentra and Silexica, and what’s in store for the next OpenMPCon in Kobe, Japan.
Videos of the actual presentations and the slideware are currently being prepared and will be made available very soon.
And thanks to everyone who attended this very first OpenMPCon, and the staff at SuperC, RWTH Aachen University!
OpenMPCon, the first OpenMP Conference is happening now in Aachen, Germany. Watch this space for links to video interviews and papers.