Welcome to the 11th CHPC National Conference, to be held at the picturesque Velmoré Hotel Estate, Pretoria.
The theme for this year's gathering is: HPC convergence with novel applications models for better service to research and industry. The theme comes as HPC is gaining more interest in application to non-traditional markets, and largely due to the convergence of HPC to Cloud and the re-emergence of Artificial Intelligence and strengthening of Machine Learning in broader applications.
The comprehensive programme will include national and international contributions as well as contributions from our colleagues in the cyberinfrastructure system: the South African National Research Network (SANReN) and the Data Intensive Research Initiative of South Africa (DIRISA). The 3rd and 7th December will be tutorials and workshops, while the main conference will be 4-6 December. Once again, the South African Development Community (SADC) HPC Collaboration Forum will form part of the conference to discuss the SADC HPC framework and implementation plans.
Our expo zone will be showcasing solutions by leading technology companies and will be the student competition battleground where almost 20 teams from universities across the country will be seeking national honours in the Student Cluster Competition and in the Student Cyber Security Competition - a new addition to the conference. We trust you will find an exciting programme and we look forward to meeting our regular and new delegates.
We look forward to seeing you in Pretoria!
Dr Happy Sithole, Director: CHPC
The convergence of HPC, Big Data, and Deep Learning is the next game-changing business opportunity. Apache Hadoop, Spark, gRPC/TensorFlow, and Memcached are becoming standard building blocks in handling Big Data oriented processing and mining. Recent studies have shown that default designs of these components can not efficiently leverage the features of modern HPC clusters, like Remote Direct Memory Access (RDMA) enabled high-performance interconnects, high-throughput parallel storage systems (e.g. Lustre), Non-Volatile Memory (NVM). In this tutorial, we will provide an in-depth overview of the architecture of Hadoop, Spark, gRPC/TensorFlow, and Memcached.
We will examine the challenges in re-designing networking and I/O components of these middleware with modern interconnects, protocols (such as InfiniBand, RoCE) and storage architectures. Using the publicly available software packages in the High-Performance Big Data project (HiBD, http://hibd.cse.ohio-state.edu), we will provide case studies of the new designs for several Hadoop/Spark/gRPC/TensorFlow/Memcached components and their associated benefits. Through these, we will also examine the interplay between high-performance interconnects, storage (HDD, NVM, and SSD), and multi-core platforms to achieve the best solutions for these components and applications on modern HPC clusters.
We also present in-depth case-studies with modern Deep Learning tools (e.g., Caffe, TensorFlow, BigDL) with RDMA-enabled Hadoop, Spark, and gRPC.
Join us for a full day of technical sessions dedicated to Artificial Intelligence / Deep Learning showing how these technologies can be implemented on existing and upcoming CPUs and how performance can be scaled to any workload!
This tutorial will provide an entry-level presentation of the basic concepts, knowledge, and skills associated with a strong foundation in modern supercomputing. This introductory treatment is based on a first-year graduate course taught jointly at several universities and currently at Indiana University. Variations of this tutorial have been given at SC11, SC12, ISC13, ISC14, ISC15, ISC16, CHPC16, and ISC17. The tutorial will describe supercomputer architecture including multi-core organization and GPUs, parallel programming models like OpenMP and MPI, simple parallel algorithms, system software and tools for scheduling, debugging, performance monitoring and tuning, and underlying technology trends and future directions of this rapidly changing field. Emphasis will be given to practical information including current generation systems, sources of available software, and links for further reading. This tutorial provides the broadest outreach to people new to the field, students, managers, policy makers, and those needing a refresh in this rapidly advancing domain. Live demonstrations will be presented throughout the tutorial on the Big Red II+ Petaflops computer at Indiana University. This full-day tutorial will include hands-on use for those attendees wishing to engage the learning experience at this depth. Questions will be welcomed throughout the presentation.
Schrödinger is a leading provider of scientific software in the drug design industry. The workshop will give an overview of the drug discovery tools available from Schrödinger. Participants will be introduced to our new graphical interface, Maestro 11, and will work through hands-on examples for tasks like docking, pharmacophore based virtual screening and building homology models. The workshop will also touch on recent improvements on FEP+, our framework to run MD based free energy calculations. It will conclude with a general Q&A session were participants can discuss their own scientific projects. A trial version of our software will be available to participants prior to the workshop. Please bring your own laptop!
HPC services continue to provide ever increasing resources and many academic disciplines that formerly did not rely on computational resources are becoming interested in utilizing computational resources to analyze their problems. This gives us a strong incentive to provide more assistance to users over a wider range of disciplines which in turn taxes resources at HPC facilities. This leads us to investigate and/or develop the necessary tools to support easy-use and performant environments when running computations on HPC systems. TACC has developed a large number of tools that we can categorize as “user-facing software” that our users run (sometimes without their knowledge) to ensure that their jobs perform well, and give assistance to debugging efforts when jobs fail. We have selected four important tools, and plan to conduct a hands-on set up for administrators and/or provide details on configuration and management of them. They are primarily focused on environment control and performance analysis: Lmod, containerization (Singularity), Tau and REMORA.
One of TACC’s most successful open-source software packages is Lmod – the Lua based version of the “Environment Modules”. It is the first thing all TACC users encounter and learn to use when they login to a TACC system. We discuss and deploy this tool, and demonstrate what system administrators and package maintainers need to know to build/install/maintain an environment module tool. We also explain how users benefit from Lmod, and the simple and advanced use cases.
We often see communities in specific domain sciences build out similar environments that they want imported into HPC systems. They invest many hours in building these environments (same as we do at TACC for the environment that we present to users) and these environments usually end up in containers. Containerization substantially impacts system administrators because they are environments within the main environment, and many of them have substantial security implications. We demonstrate how to make containers available and investigate the typical usage.
High Performance in HPC systems is something administrators, developers and users should view as paramount in the design of applications, and setting up the optimal environment for production runs. The TAU profiler and analysis utility provides data for evaluating performance, similar to the GNU profiler, but has an advanced GUI interface to make analysis easier (and for viewing events). Another package, REMORA, is a resource monitoring tool, also developed at TACC, which allows users to easily collect execution data (memory, numa, i/o, networking and cpu load). The management, best usage practices and synergy between the TAU and REMORA tools will be presented.
The Student Cyber Security Challenge is an information security student competition with an emphasis on network security.
Students will have to solve network security problems and identify security issues such as Decrypting Passwords, Geo-locating pictures, Securing Websites, Finding information from TCP traffic and Extracting weak security keys.
Finally, the students will compete in on a live hacking scenario where they have to defend their network infrastructure while attacking their competitors' infrastructure.
This workshop will touch on the following topics:
Overview of the 3D visualization codes Paraview and VisIt
Similarities and differences between Paraview & VisIt
Limitations of single core processing of large data sets
OpenGL, VirtualGL and Mesa
Different approaches to remote visualization - VNC or X-forwarding
Different approaches to parallelisation
Decomposed datasets
Tutorial illustrating remote parallel visualization with Paraview
Tutorial illustrating remote parallel visualization with VisIt
Introductory remarks on the concept of in-situ visualization
Concluding remarks
End of the first day of Workshops
Refreshment break
SADC Cyberinfrastructure.
Understanding and Customizing Storage Systems:
General purpose storage systems aim to offer a least common denominator that works well enough for a variety of workloads. With inevitable failures and potential data loss at scale, it is important to understand what the chances are for data loss to minimize the probability. With this kind of knowledge, robust storage with interfaces customized to particular application requirements can be deployed. Through programmable storage systems interfaces, custom service features can be combined limiting overheads from unnecessary services. Using these tools, new kinds of storage services focused on data contents and quality of service requirements can be addressed. This session presents talks investigating understanding storage system component failures, programmable storage, and user-defined metadata services.
Organiser: Jay Lofstead (SANDIA)
SADC Cyberinfrastructure.
Refreshment break
Special Interest Group Meeting for Computational Chemistry and Materials Science
Students will be available to answer questions about their posters.
SADC Cyberinfrastructure.
SADC Cyberinfrastructure.
Students will be available to answer questions about their posters.
Birds of a feather session for all CHPC principal investigators and users.
Birds of a feather session for the HPC Ecosystems sites.
SADC Cyberinfrastructure.
Students will be available to answer questions about their posters.
Academics from relevant disciplines (Computer Science, Electronic and Computer Engineering, Computational Physics, Applied Mathematics, Data Science, etc.) and others interested in, or involved with, teaching and curriculum development are invited to discuss how core HPC concepts can be integrated into the undergraduate and postgraduate curricula at South African and African universities.
A broad range of topics related to HPC, including use cases, distributed computing, architectures and interconnects, operating systems, parallel programming and algorithms, file systems, debugging, performance scaling, accelerator technologies, benchmarking and visualization will be considered.
This session will focus on the unique challenges facing academics in South Africa and Africa, either integrating HPC topics into existing courses and curricula or developing and implementing new courses and specializations, and the ways in which these challenges can be addressed. The provision of access to high-end HPC compute infrastructure for student practical assignments and projects, along with teaching resources and course content, will be discussed with the goal of significantly lowering the barrier to entry for academics wishing to teach HPC topics. The alignment of the HPC education and human capital development mandates of the CHPC with other strategist initiatives such as the HPC Ecosystems Project (which has placed, and is continuing to place, donated HPC infrastructure at universities throughout Africa), will be investigated, along with other ways in which the CHPC can support academics in this critical endeavor.
Academics from relevant disciplines (Computer Science, Electronic and Computer Engineering, Computational Physics, Applied Mathematics, Data Science, etc.) and others interested in, or involved with, teaching and curriculum development are invited to discuss how core HPC concepts can be integrated into the undergraduate and postgraduate curricula at South African and African universities.
A broad range of topics related to HPC, including use cases, distributed computing, architectures and interconnects, operating systems, parallel programming and algorithms, file systems, debugging, performance scaling, accelerator technologies, benchmarking and visualization will be considered.
This session will focus on the unique challenges facing academics in South Africa and Africa, either integrating HPC topics into existing courses and curricula or developing and implementing new courses and specializations, and the ways in which these challenges can be addressed. The provision of access to high-end HPC compute infrastructure for student practical assignments and projects, along with teaching resources and course content, will be discussed with the goal of significantly lowering the barrier to entry for academics wishing to teach HPC topics. The alignment of the HPC education and human capital development mandates of the CHPC with other strategist initiatives such as the HPC Ecosystems Project (which has placed, and is continuing to place, donated HPC infrastructure at universities throughout Africa), will be investigated, along with other ways in which the CHPC can support academics in this critical endeavor.
Students will be available to answer questions about their posters.
Birds of a feather session dedicated to women in HPC.
The CHPC's non-academic users are invited to the annual industry engagement meeting,to be held on the morning of 7 December, with informal discussions to be continued during lunch. The purpose of the meeting is to touch base with existing non-academic users, discuss common points of interest and to inform prospective new commercial users. The meeting will consist of a presentation by CHPC staff, followed by brief presentations by users and a group discussion. The intention is for the CHPC to learn from the users' experience, and to inform the users of future plans.
his workshop will explore acceleration opportunities in existing and upcoming workflows. We will also explore the creation of a community & ecosystem around POWER9 acceleration technology for academics and industry practitioners.
In this workshop we will install and configure Puppet 5, a configuration management system, to manage our compute cluster.
We will install and configure the Puppet server, PuppetDB database, a Puppet agent and multiple modules from the Puppet Forge. Best practises will be implemented, including using Hiera and saving configurations to Git.
The workshop will be hands-on. Please ensure you have a working installation of Hyper-V, VMware (Workstation or Player) or VirtualBox. Copies of the VMDK/VHDX files will be provided.
This course will present an introduction to foundational concepts of parallel programming, and will specifically train attendees in the basic use of OpenMP for C/C++ and Fortran programmers. The fundamentals of creating parallel execution blocks, forking and synchronizing threads, and employing worksharing loops will be featured in a lecture and explored in a hands-on lab. Additionally, an introduction to new features in OpenMP (SIMD, Affinity) and a high-level introduction to tasking will be presented.
This workshop is for everyone who wants to learn, develop, or super-charge their research with AI and machine learning. You’ll discover Microsoft AI and Machine learning platform offerings . Come along if you want to experience hands-on deep-learning for image recognition, natural language processing, speech recognition, deploying machine learning apps using R and Python, mining academic publications, analyzing IoT data, or want to ask questions about how AI can help your research.
This one day workshop is intended for undergraduate project students, postgraduate students, postdoctoral researchers and researchers who are familiar with the field of Molecular Dynamics (MD) and want to employ state-of-the art methodology based on the density functional theory to understand bulk materials properties, surface science and heterogeneous catalysis phenomena. Molecular dynamics is a computer simulation method for studying the physical movements of atoms and molecules. The MD method can assist one in obtaining the static quantities and dynamic quantities.
The induction course covers the use of the CHPC systems from the user's point of view.