Astronomy Data And Computing Services - Upskilling the Australian astronomy community
The Astronomy Data And Computing Services (ADACS) initiative has been working with the Australian astronomy community for just over 3 years now. Our vision is to deliver astronomy-focused training, support and expertise to maximise the scientific return on investments in astronomical data &...
Keywords: astronomy, data skills, eresearch skills, skills, computational skills, training, skills gaps, astronomy-focused training, training material
Astronomy Data And Computing Services - Upskilling the Australian astronomy community
https://zenodo.org/records/4287748
https://dresa.org.au/materials/astronomy-data-and-computing-services-upskilling-the-australian-astronomy-community-57afa0b9-77da-4dc1-ad29-25089f19363d
The Astronomy Data And Computing Services (ADACS) initiative has been working with the Australian astronomy community for just over 3 years now. Our vision is to deliver astronomy-focused training, support and expertise to maximise the scientific return on investments in astronomical data & computing infrastructure.
During these last 3 years, we have delivered dozens of face-to-face, hands-on workshops and created several hours worth of online tutorial materials. This talk will focus on our journey to deliver this computational skills training to the community, exploring how we chose different delivery pathways and content, based both on community input as well as our professional expertise and understanding of existing skill gaps. Most importantly we will discuss our plans for the future and how we are working on actively including the community in developing new training material beyond the usual skills survey.
Come along to this talk if you would like to hear about a national effort to deliver computational skills training and would like to know more about potential new avenues to provide just-in-time training and how to collaborate with ADACS.
contact@ardc.edu.au
Lange, Rebecca (orcid: 0000-0002-9449-4384)
astronomy, data skills, eresearch skills, skills, computational skills, training, skills gaps, astronomy-focused training, training material
Successful data training stories from NCI
NCI Australia manages a multi-petabyte sized data repository, collocated with its HPC systems and data services, which allows high performance access to many scientific research datasets across many earth science domains.
An important aspect is to provide training materials that proactively...
Keywords: skills, training, eresearch skills, HPC training, domain-specific training, reproducible workflows, training material
Successful data training stories from NCI
https://zenodo.org/records/4287750
https://dresa.org.au/materials/successful-data-training-stories-from-nci-33f110e3-0c06-492e-9cc5-fa0f886ca1b8
NCI Australia manages a multi-petabyte sized data repository, collocated with its HPC systems and data services, which allows high performance access to many scientific research datasets across many earth science domains.
An important aspect is to provide training materials that proactively engages with the research community to improve their understanding of the data available, and to share knowledge and best practices in the use of tools and other software. We have developed multiple levels of training modules (introductory, intermediate and advanced) to cater for users with different levels of experience and interest. We have also tailored courses for each scientific domain, so that the use-cases and software will be most relevant to their interests and needs.
For our training, we combine brief lectures followed by hands-on training examples on how to use datasets, using working examples of well-known tools and software that people can use as a template and modify to fit their needs. For example, we take representative use-cases from some scientific activities, from our collaborations and from user support issues, and convert to Jupyter notebook examples so that people can repeat the workfIow and reproduce the results. We also use the training as an opportunity to raise awareness of growing issues in resource management. Some examples include a familiarity of the FAIR data principles, licensing, citation, data management and trusted digital repositories. This approach to both our online training materials and workshops has been well-received by PhD students, early careers, and cross disciplinary users.
contact@ardc.edu.au
Wang, Jingbo
skills, training, eresearch skills, HPC training, domain-specific training, reproducible workflows, training material
Accelerating skills development in Data science and AI at scale
At the Monash Data Science and AI platform, we believe that upskilling our research community and building a workforce with data science skills are key to accelerating the application of data science in research. To achieve this, we create and leverage new and existing training capabilities...
Keywords: AI, machine learning, eresearch skills, training, train the trainer, volunteer instructors, training partnerships, training material
Accelerating skills development in Data science and AI at scale
https://zenodo.org/records/4287746
https://dresa.org.au/materials/accelerating-skills-development-in-data-science-and-ai-at-scale-2d8a65fa-f96e-44ad-a026-cfae3f38d128
At the Monash Data Science and AI platform, we believe that upskilling our research community and building a workforce with data science skills are key to accelerating the application of data science in research. To achieve this, we create and leverage new and existing training capabilities within and outside Monash University. In this talk, we will discuss the principles and purpose of establishing collaborative models to accelerate skills development at scale. We will talk about our approach to identifying gaps in the existing skills and training available in data science, key areas of interest as identified by the research community and various sources of training available in the marketplace. We will provide insights into the collaborations we currently have and intend to develop in the future within the university and also nationally.
The talk will also cover our approach as outlined below
• Combined survey of gaps in skills and trainings for Data science and AI
• Provide seats to partners
• Share associate instructors/helpers/volunteers
• Develop combined training materials
• Publish a repository of open source trainings
• Train the trainer activities
• Establish a network of volunteers to deliver trainings at their local regions
Industry plays a significant role in making some invaluable training available to the research community either through self learning platforms like AWS Machine Learning University or Instructor led courses like NVIDIA Deep Learning Institute. We will discuss how we leverage our partnerships with Industry to bring these trainings to our research community.
Finally, we will discuss how we map our training to the ARDC skills roadmap and how the ARDC platforms project “Environments to accelerate Machine Learning based Discovery” has enabled collaboration between Monash University and University of Queensland to develop and deliver training together.
contact@ardc.edu.au
Tang, Titus
AI, machine learning, eresearch skills, training, train the trainer, volunteer instructors, training partnerships, training material
Data Fluency: a community of practice supporting a digitally skilled workforce
This presentation showcases the impact of the Monash Data Fluency Community of Practice upon digitally skilled Graduate Research students involved as learners and instructors in the program. The strong focus on building community to complement training, has fostered an environment of learning,...
Keywords: skills, training, eresearch skills, data skills, online learning, pedagogy, train the trainer, digitally skilled workforce, training material
Data Fluency: a community of practice supporting a digitally skilled workforce
https://zenodo.org/records/4287752
https://dresa.org.au/materials/data-fluency-a-community-of-practice-supporting-a-digitally-skilled-workforce-b911a1a8-0331-496e-95a6-0015a12acc34
This presentation showcases the impact of the Monash Data Fluency Community of Practice upon digitally skilled Graduate Research students involved as learners and instructors in the program. The strong focus on building community to complement training, has fostered an environment of learning, networking and sharing of expertise. Hear what the Graduate research students have to say about the value of skills training and how it has impacted their research; how the community has enabled them to network with a broad range of researchers and affiliate partner groups they would not ordinarily be in contact with; how their research journey has been enhanced by working as part of a multi-disciplinary team, as well as sharpening their teaching skills.
The rapid refocus from face - face to online delivery, as a result of the pandemic, highlights the importance of the multi-faceted online approach including workshops, drop-in sessions, SLACK chat and online learning resources. As a result of the shift to online, the range of strategic external partner/affiliate groups has extended and demand for workshops and drop-ins has increased. Learn how the instructors have altered their pedagogical approach to engage workshop and drop-in participants; how they have overcome some of the challenges of facilitating in an online environment; and how this is preparing them to become part of a digitally skilled workforce.
contact@ardc.edu.au
Groenewegen, David (orcid: 0000-0003-2523-1676)
skills, training, eresearch skills, data skills, online learning, pedagogy, train the trainer, digitally skilled workforce, training material
ARDC Skills Landscape
The Australian Research Data Commons is driving transformational change in the research data ecosystem, enabling researchers to conduct world class data-intensive research. One interconnected component of this ecosystem is skills development/uplift, which is critical to the Commons and its...
Keywords: skills, data skills, eresearch skills, community, skilled workforce, FAIR, research data management, data stewardship, data governance, data use, data generation, training material
ARDC Skills Landscape
https://zenodo.org/records/4287743
https://dresa.org.au/materials/ardc-skills-landscape-56b224ca-9e30-4771-8615-d028c7be86a6
The Australian Research Data Commons is driving transformational change in the research data ecosystem, enabling researchers to conduct world class data-intensive research. One interconnected component of this ecosystem is skills development/uplift, which is critical to the Commons and its purpose of providing Australian researchers with a competitive advantage through data.
In this presentation, Kathryn Unsworth introduces the ARDC Skills Landscape. The Landscape is a first step in developing a national skills framework to enable a coordinated and cohesive approach to skills development across the Australian eResearch sector. It is also a first step towards helping to analyse current approaches in data training to identify:
- Siloed skills initiatives, and finding ways to build partnerships and improve collaboration
- Skills deficits, and working to address the gaps in data skills
- Areas of skills development for investment by skills stakeholders like universities, research organisations, skills and training service providers, ARDC, etc.
contact@ardc.edu.au
Unsworth, Kathryn (orcid: 0000-0002-5407-9987)
skills, data skills, eresearch skills, community, skilled workforce, FAIR, research data management, data stewardship, data governance, data use, data generation, training material
PCon Preparing applications for El Capitan and beyond
As Lawrence Livermore National Laboratories (LLNL) prepares to stand up its next supercomputer, El Capitan, application teams prepare to pivot to another GPU architecture.
This talk presents how the LLNL application teams made the transition from distributed-memory, CPU-only architectures to...
Keywords: GPUs, supercomputing, HPC, PaCER
PCon Preparing applications for El Capitan and beyond
https://www.youtube.com/watch?v=cj7a7gWgt8o&list=PLmu61dgAX-aZ_aa6SmmExSJtXGS7L_BF9&index=4
https://dresa.org.au/materials/pcon-preparing-applications-for-el-capitan-and-beyond
As Lawrence Livermore National Laboratories (LLNL) prepares to stand up its next supercomputer, El Capitan, application teams prepare to pivot to another GPU architecture.
This talk presents how the LLNL application teams made the transition from distributed-memory, CPU-only architectures to GPUs. They share institutional best practices. They discuss new open-source software products as tools for porting and profiling applications and as avenues for collaboration across the computational science community.
Join LLNL's Erik Draeger and Jane Herriman, who presented this talk at Pawsey's PaCER Conference in September 2023.
training@pawsey.org.au
Erik Draeger
Jane Herriman
Pawsey Supercomputing Research Centre
GPUs, supercomputing, HPC, PaCER
masters
phd
researcher
ecr
support
professional
ugrad
OpenCL
Supercomputers make use of accelerators from a variety of different hardware vendors, using devices such as multi-core CPU’s, GPU’s and even FPGA’s. OpenCL is a way for your HPC application to make effective use of heterogeneous computing devices, and to avoid code refactoring for new HPC...
Keywords: supercomputing, Pawsey Supercomputing Centre, CPUs, GPUs, OpenCL, FPGAs
Resource type: activity
OpenCL
https://www.youtube.com/playlist?list=PLmu61dgAX-aa_lk5fby5PjuS49snHpyYL
https://dresa.org.au/materials/opencl
Supercomputers make use of accelerators from a variety of different hardware vendors, using devices such as multi-core CPU’s, GPU’s and even FPGA’s. OpenCL is a way for your HPC application to make effective use of heterogeneous computing devices, and to avoid code refactoring for new HPC infrastructure.
training@pawsey.org.au
Toby Potter
Pawsey Supercomputing Research Centre
Pelagos
Toby Potter
supercomputing, Pawsey Supercomputing Centre, CPUs, GPUs, OpenCL, FPGAs
masters
ecr
researcher
support
AMD Profiling
The AMD profiling workshop covers the AMD suite of tools for development of HPC applications on AMD GPUs.
You will learn how to use the rocprof profiler and trace visualization tool that has long been available as part of the ROCm software suite.
You will also learn how to use the new...
Keywords: supercomputing, performance, GPUs, CPUs, AMD, HPC, ROCm
Resource type: activity
AMD Profiling
https://www.youtube.com/playlist?list=PLmu61dgAX-aaQOCG5Jlw8oLBORJfoQC2o
https://dresa.org.au/materials/amd-profiling
The AMD profiling workshop covers the AMD suite of tools for development of HPC applications on AMD GPUs.
You will learn how to use the rocprof profiler and trace visualization tool that has long been available as part of the ROCm software suite.
You will also learn how to use the new Omnitools - Omnitrace and Omniperf - that were introduced at the end of 2022. Omnitrace is a powerful tracing profiler for both CPU and GPU. It can collect data from a much wider range of sources and includes hardware counters and sampling approaches. Omniperf is a performance analysis tool that can help you pinpoint how your application is performing with a visual view of the memory hierarchy on the GPU as well as reporting the percentage of peak for many different measurements.
training@pawsey.org.au
AMD
Pawsey Supercomputing Research Centre
supercomputing, performance, GPUs, CPUs, AMD, HPC, ROCm
HIP Workshop
The Heterogeneous Interface for Portability (HIP) provides a programming framework for harnessing the compute capabilities of multicore processors, such as the MI250X GPU’s on Setonix.
In this course we focus on the essentials of developing HIP applications with a focus on...
Keywords: HIP, supercomputing, Programming, GPUs, MPI, debugging
Resource type: full-course
HIP Workshop
https://support.pawsey.org.au/documentation/display/US/Pawsey+Training+Resources
https://dresa.org.au/materials/hip-workshop
The Heterogeneous Interface for Portability (HIP) provides a programming framework for harnessing the compute capabilities of multicore processors, such as the MI250X GPU’s on Setonix.
In this course we focus on the essentials of developing HIP applications with a focus on supercomputing.
Agenda
- Introduction to HIP and high level features
- How to build and run applications on Setonix with HIP and MPI
- A complete line-by-line walkthrough of a HIP-enabled application
- Tools and techniques for debugging and measuring the performance of HIP applications
training@pawsey.org.au
Pelagos
Pawsey Supercomputing Research Centre
HIP, supercomputing, Programming, GPUs, MPI, debugging
Porting the multi-GPU SELF-Fluids code to HIPFort
In this presentation by Dr. Joseph Schoonover of Fluid Numerics LLC, Joe shares their experience with the porting process for SELF-Fluids from multi-GPU CUDA-Fortran to multi-GPU HIPFort.
The presentation covers the design principles and roadmap for SELF and the strategy to port from...
Keywords: AMD, GPUs, supercomputer, supercomputing
Resource type: presentation
Porting the multi-GPU SELF-Fluids code to HIPFort
https://docs.google.com/presentation/d/1JUwFkrHLx5_hgjxsix8h498_YqvFkkcefNYbu-DsHio/edit#slide=id.g10626504d53_0_0
https://dresa.org.au/materials/porting-the-multi-gpu-self-fluids-code-to-hipfort
In this presentation by Dr. Joseph Schoonover of Fluid Numerics LLC, Joe shares their experience with the porting process for SELF-Fluids from multi-GPU CUDA-Fortran to multi-GPU HIPFort.
The presentation covers the design principles and roadmap for SELF and the strategy to port from Nvidia-only platforms to AMD & Nvidia GPUs. Also discussed are the hurdles encountered along the way and considerations for developing multi-GPU accelerated applications in Fortran.
SELF is an object-oriented Fortran library that supports the implementation of Spectral Element Methods for solving partial differential equations. SELF-Fluids is an implementation of SELF that solves the compressible Navier Stokes equations on CPU only and GPU accelerated compute platforms using the Discontinuous Galerkin Spectral Element Method. The SELF API is designed based on the assumption that SEM developers and researchers need to be able to implement derivatives in 1-D and divergence, gradient, and curl in 2-D and 3-D on scalar, vector, and tensor functions using spectral collocation, continuous Galerkin, and discontinuous Galerkin spectral element methods.
The presentation discussion is placed in context of the Exascale era, where we're faced with a zoo of available compute hardware. Because of this, SELF routines provide support for GPU acceleration through AMD’s HIP and support for multi-core, multi-node, and multi-GPU platforms with MPI.
training@pawsey.org.au
Joe Schoonover
AMD, GPUs, supercomputer, supercomputing
Embracing new solutions for in-situ visualisation
This PPT was used by Jean Favre, senior visualisation software engineer at CSCS, the Swiss National Supercomputing Centre during his presentation at P'Con '21 (Pawsey's first PaCER Conference).
This material discusses the upcoming release of ParaView v5.10, a leading scientific visualisation...
Keywords: ParaView, GPUs, supercomputer, supercomputing, visualisation, data visualisation
Resource type: presentation
Embracing new solutions for in-situ visualisation
https://github.com/jfavre/InSitu/blob/master/InSitu-Revisited.pdf
https://dresa.org.au/materials/embracing-new-solutions-for-in-situ-visualisation
This PPT was used by Jean Favre, senior visualisation software engineer at CSCS, the Swiss National Supercomputing Centre during his presentation at P'Con '21 (Pawsey's first PaCER Conference).
This material discusses the upcoming release of ParaView v5.10, a leading scientific visualisation application. In this release ParaView consolidates its implementation of the Catalyst API, a specification developed for simulations and scientific data producers to analyse and visualise data in situ.
The material reviews some of the terminology and issues of different in-situ visualisation scenarios, then reviews early Data Adaptors for tight-coupling of simulations and visualisation solutions. This is followed by an introduction of Conduit, an intuitive model for describing hierarchical scientific data. Both ParaView-Catalyst and Ascent use Conduit’s Mesh Blueprint, a set of conventions to describe computational simulation meshes.
Finally, the materials present CSCS’ early experience in adopting ParaView-Catalyst and Ascent via two concrete examples of instrumentation of some proxy numerical applications.
training@pawsey.org.au
Jean Favre
ParaView, GPUs, supercomputer, supercomputing, visualisation, data visualisation