The Board of Directors of the OpenMP Architecture Review Board (ARB) is pleased to announce the appointment of Dr. Michael Klemm as the new CEO of the OpenMP organization, a group of vendors and research organizations creating the standard for one of the most popular shared-memory and embedded parallel programming models in use today. Dr. Klemm succeeds Mr. Michael Wong, who served as the OpenMP CEO for almost 5 years. The Board of Directors and the members of the OpenMP ARB would like to thank Mr. Wong for his dedication and leadership over the past years. During his tenure, the OpenMP language standard added support for coprocessors and accelerators, became more agile by adding a procedure for publishing Technical Reports and yearly ratifications to the standard, and significantly grew in membership.
“We are delighted to welcome Dr. Klemm as our CEO,” says Sanjiv Shah, Chairman of the Board of Directors. “His experience leading many organizational activities as well as several technical features of OpenMP make him well qualified to lead the OpenMP ARB.”
“I look forward to working with Dr. Klemm to ensure the continued success and modernization of OpenMP,” says Dr. Bronis de Supinski, Chair of the OpenMP Language Committee. “Very close collaboration between Bronis de Supinski as the Chair of the OpenMP Language Committee and myself will be key to maintain OpenMP the programming model of choice, not only in HPC but for parallel programming in general,” says Michael Klemm.
Dr. Michael Klemm obtained an M.Sc. in Computer Science in 2003 and received a Doctor of Engineering degree (Dr.-Ing.) from the Friedrich-Alexander-University Erlangen-Nuremberg, Germany, in 2008. Michael Klemm works in the Developer Relations Division at Intel in Germany and his areas of interest include compiler construction, design of programming languages, parallel programming, and performance analysis and tuning. Michael Klemm joined the OpenMP organization in 2009.
The OpenMP ARB announces the twelfth International Workshop on OpenMP (IWOMP), which is the premier forum to present and to discuss issues, trends, recent research ideas and results related to programming with OpenMP. The international workshop affords an opportunity for OpenMP users as well as developers to meet to discuss new ideas and information on this topic.
Submissions of unpublished technical papers detailing innovative, original research and development related to OpenMP are solicited. All topics related to OpenMP are of interest, including OpenMP performance analysis and modeling, OpenMP performance and correctness tools, proposed OpenMP extensions, and OpenMP applications in any domain (e.g., scientific and numerical computation, video games, computer graphics, multimedia, information retrieval, optimization, text processing, data mining, finance, signal and image processing and machine learning).
Advances in technologies, such as multicore processors and OpenMP devices (accelerators such as GPGPUs, DSPs or FPGAs), Multiprocessor Systems on a Chip (MPSoCs), and recent developments in OpenMP itself (e.g., tasking) present new opportunities and challenges for software and hardware developers. Recent advances in the C, C++ and Fortran base languages also offer interesting opportunities and challenges to the OpenMP programming model. IWOMP 2016 particularly solicits submissions in these areas as well as ones that discuss how to apply OpenMP to additional models of parallelism such as event loops.
Recently released Clang 3.8 fully supports OpenMP 3.1 and some features of OpenMP 4.0 and 4.5.
For details, see the Clang 3.8 Release Notes.
Pre-compiled packages can be downloaded from http://llvm.org/releases/download.html#3.8.0
IWOMP 2016 - 12th International Workshop on OpenMP
October 5-7, 2016
For many years, OpenMP has provided a very rich and flexible programming model for shared memory architectures. OpenMP 4.0 was a major advance that added two new forms of parallelism and OpenMP 4.5 added substantial improvements for them and other aspects of the specification. OpenMP now includes a rich set of device constructs to support heterogeneous systems like GPUs,FPGAs, DSPs, and SIMD constructs to support vector units in addition to the support for task-based parallelism that has been present since OpenMP 3.0.
The International Workshop on OpenMP (IWOMP) is an annual workshop dedicated to the promotion and advancement of all aspects of parallel programming with OpenMP. It is the premier forum to present and discuss issues, trends, recent research ideas, and results related to parallel programming with OpenMP. We solicit submissions of unpublished technical papers detailing innovative, original research and development related to OpenMP.
All topics related to OpenMP are of interest, including OpenMP performance analysis and modeling, OpenMP performance and correctness tools, proposed OpenMP extensions, and OpenMP applications in any domain (e.g., scientific and numerical computation, video games, computer graphics, multimedia, information retrieval, optimization, text processing, data mining, finance, signal and image processing and machine learning).
Advances in technologies, such as multi-core processors and OpenMP devices (accelerators such as GPGPUs, DSPs or FPGAs), Multiprocessor Systems on a Chip (MPSoCs), and recent developments in OpenMP itself (e.g., devices) present new opportunities and challenges for software and hardware developers. Recent advances in the C, C++ and Fortran base languages also offer interesting opportunities and challenges to the OpenMP programming model. IWOMP 2016 particularly solicits submissions in these areas as well as ones that discuss how to apply OpenMP to additional models of parallelism such as event loops.
Paper Submission and Registration
Submitted papers for review should be limited to 12 pages and follow LNCS guidelines. Submission deadline is May 18th, 2016. Submit your paper to: https://www.easychair.org/conferences/?conf=iwomp2016. Authors of accepted papers will be asked to prepare a final paper of up to 15 pages. As in previous years, IWOMP 2016 will publish formal proceedings of the accepted papers in Springer Verlag’s LNCS series.
Paper Submission Deadline: May 18th, 2016
Notification of acceptance: June 18th, 2016
Deadline for final version: July 8th, 2016
- Naoya Maruyama, RIKEN Advanced Institute for Computational Science, Japan
Program Committee Co-chairs:
- Naoya Maruyama, RIKEN Advanced Institute for Computational Science, Japan
- Bronis R. de Supinski, Lawrence Livermore National Laboratory, USA
Videos and slides of most of the presentations given at our first OpenMPCon developers conference, held September 29-30 at Aachen University, are now available on our YouTube channel as a playlist.
Here are the individual videos:
Five short presentations given by members of the OpenMP team in the OpenMP booth at SC15 are now available on YouTube (playlist):
- OpenMP Tasking (Slides)
- Getting OpenMP Up to Speed (Slides)
- OpenMP for Embedded Systems (Slides)
- OpenMP and Parallware (Slides)
- SC 15 OpenMP - Performance and Energy Optimization with OpenMP (Slides)
The OpenMP ARB is pleased to announce OpenMP 4.5, a major upgrade of the OpenMP standard language specifications. This release provides a substantial improvement on the support for programming of accelerator and GPU devices, and supports now also the parallelization of loops with well-structured dependencies. Implementation is underway in GCC and Clang. The new specification can be downloaded from here.
Standard for parallel programming extends its reach
With this release, OpenMP, the de-facto standard for parallel programming on shared memory systems, continues to extend its reach beyond pure HPC to include DSPs, real time systems, and accelerators. OpenMP aims to provide high-level parallel language support for a wide range of applications, from biotech and automotive to aeronautics, automation, robotics and financial analysis.
“OpenMP 4.5 is a significant achievement that demonstrates the industry-wide collaboration and the hard work and dedication within the OpenMP community,” says Michael Wong, CEO of the OpenMP ARB. “It is more than a minor release, representing the road towards OpenMP 5.0 while we continue on a cadence that delivers Technical Reports and/or Ratified Specifications annually, in keeping pace with the marketplace.”
Many new features
- Significantly improved support for devices. OpenMP now provides mechanisms for unstructured data mapping and asynchronous execution and also runtime routines for device memory management. These routines allow for allocating, copying and freeing.
- Support for doacross loops. A natural mechanism to parallelize loops with well-structured dependences is provided.
- New taskloop construct. Support to divide loops into tasks, avoiding the requirement that all threads execute the loop.
- Reductions for C/C++ arrays. This often requested feature is now available by building on support for array sections.
- New hint mechanisms. Hint mechanisms can provide guidance on the relative priority of tasks and on preferred synchronization implementations.
- Thread affinity support. It is now possible to use runtime functions to determine the effect of thread affinity clauses.
- Improved support for Fortran 2003. Users can now parallelize many Fortran 2003 programs.
- SIMD extensions. These extensions include the ability to specify exact SIMD width and additional data-sharing attributes.
Implementation is already almost complete in GCC version 6.0. It is starting in the current trunk of Clang 4.8. Other vendor compilers are following.
OpenMP 4.5 Specifications
- OpenMP 4.5 Complete Specifications (November 2015) PDF
- OpenMP 4.5 Summary Card - C/C++ (November 2015) PDF
- OpenMP 4.5 Summary Card - Fortran (November 2015) PDF
OpenMP: Where Are We and What’s Next? [Room 18AB]
We will showcase modern OpenMP (“The language that lets you use all your compute resources”) by presenting OpenMP 4.1 and the vision for OpenMP 5.0. Experts will give short presentations on key OpenMP 4.1 features and answer your technical questions. OpenMP CEO Michael Wong will present our vision for OpenMP 5.0 and we’ll finish with an audience-led discussion with a panel including members of the OpenMP ARB.
- Sunday, Nov 15th, 8:30am - 5:00pm
Hands-on Introduction to OpenMP [Room 17B]
Bring your laptop with installed OpenMP compiler for a hands-on introduction.
- Monday, Nov 16th, 8:30am - 5:00pm
Advanced OpenMP Tutorial: Performance and 4.1 Features [Room 13AB]
OpenMP performance, parallelization strategies, advanced features, and more.
- Monday, Nov 16th, 1:30pm - 5:00pm
Debugging and Performance Tools for MPI and OpenMP 4.0 APIs [Room 14]
Parallel debugging and optimization focused on techniques used with accelerators and coprocessors.
RELAX WITH A COLD BEER:
Attend a short talk in our booth [Booth# 2036, meet some of the pros behind the API, and participate in drawings after each talk for the latest book by James Reinders and Jim Jeffers.
- Tuesday, Nov 18th, 11:15 - 11:40am
Power Optimization Strategies in OpenMP – Abid Malik, University of Houston
- Tuesday, Nov 18th, 2:15 - 2:40pm
OpenMP and the Multicore Associates (MCA) – Sunita Chandrasekaran, (University of Delaware)
- Wednesday, Nov 19th, 11:15 - 11:40am
Getting Open Up to Speed – Ruud van der Pas, Oracle
- Wednesday, Nov 19th, 2:15 - 2:40pm
OpenMP Tasking – Michael Klemm, Intel and Christian Terboven, RWTH-Aachen
- Thursday, Nov 20th, 11:15 - 11:30am
OpenMP and Parallware – Manuel Arenaz, Appentra
- Overcoming Distributed Debugging Challenges in MPI + OpenMP Programming Model
- Directive-Based Pipelining Extension for OpenMP
- A Standard Debug Interface for OpenMP Target Regions
- A Deadlock Detection Concept for OpenMP Tasks and Fully Hybrid MPI-OpenMP Applications
Cray: Booth 1833
EPCC: Booth 2503
Fujitsu: Booth 1827
HP: Booth 603
IBM: Booth 522
Intel: Booths 1333, 1533
NASA: Booth 333
NEC: Booth 313
Nvidia: Booth 1021
Oracle: Booth 1327
RedHat: Booth 3112
TACC: Booth 548
Texas Instruments (in the Rapid IO Trade Assoc. Booth): Booth 2917