Seminars
December 16, 2021
WIESŁAW NOWIŃSKI
Cardinal Stefan Wyszyński University in Warsaw
Human brain atlases and their applications
October 12, 2021
TOMASZ MALKIEWICZ
CSC – IT Center for Science and Nordic e-Infrastructure Collaboration (NeIC)
LUMI: the EuroHPC pre-exascale system of the North
May 27, 2021
David Winkler
Professor of Biochemistry and Genetics, La Trobe University
Professor of Medicinal Chemistry, Monash University
Visiting Professor in Pharmacy, University fo Nottingham
Fellow, Evolutionary Robotics, CSIRO Data61
Computational insights into the origin of SARS-CoV-2 and repurposing of drugs for COVID-19
Sven Dorkenwald
Princeton University, USA
Tentative topic:
Towards whole-brain Connectomes: Reconstructing all neurons in a fly brain at nanometer resolution
Abstract: Comprehensive neuronal wiring diagrams derived from Electron Microscopy images allow researchers to test models of how brain circuits give rise to neuronal activity and drive behavior. Due to advances in automated image acquisition and analysis, whole-brain connectomes with thousands of neurons are finally on the horizon. However, many person-years of manual proofreading are still required to correct errors in these automated reconstructions. We created FlyWire to facilitate the proofreading of neuronal circuits in an entire fly brain by a community of researchers distributed across the world. While FlyWire is dedicated to the fly brain, its methods will be generally applicable to whole-brain connectomics and are already in use to proofread multiple datasets. In this talk I will describe how FlyWire’s computational and social structures are organized to scale up to whole-brain connectomics and present on our progress towards the generation of a proofread whole-brain connectome of the fruit fly.
Biosketch: Sven Dorkenwald is currently a PhD student in the Seung Lab at Princeton University. In his PhD he is developing systems, infrastructure and machine learning methods to facilitate the analysis of large-scale connectomics datasets. Together with collaborators at the Allen Institute for Brain Science, he developed proofreading and annotation infrastructure that is used to host multiple large-scale connectomics datasets and runs FlyWire. FlyWire.ai is an online community for proofreading neural circuits in a whole fly brain based on the FAFB EM dataset.
Ivo F. Sbalzarini
Chair of Scientific Computing for Systems Biology, Faculty of Computer Science, TU Dresden;
MOSAIC Group, Center for Systems Biology Dresden;
Max Planck Institute of Molecular Cell Biology and Genetics, Dresden
Title: Computational Developmental Biology
Abstract: Our vision is to develop a computer simulation of a developing embryo, incorporating the known biochemistry and biophysics into a computational model in 3D-space and time, which predicts the emerging shape and function of the tissue. Development and morphogenesis of tissues, organs, and embryos emerges from the collective self-organization of cells that communicate though chemical and mechanical signals. Decisions about growth, division, and migration are taken locally by each cell based on the collective information. In this sense, a developing tissue is akin to a massively parallel computer system, where each cell computes robust local decisions, integrating communication with other cells. Mechanistically understanding and reprogramming this system is a grand challenge. While the “hardware” (proteins, lipids, etc.) and the “source code” (genome) are increasingly known, we known virtually nothing about the algorithms that this code implements on this hardware. Using examples from our work, I highlight computational challenges along the way. These range from content-adaptive data representations for machine learning, to novel languages for parallel high-performance computing, to virtual reality and real-time visualization for 3D microscopy and numerical simulations of nonlinear and non-equilibrium mechanical models. This cooperative interdisciplinary effort contributes to all involved disciplines.
Simon Mutch
Research Fellow & STA STEM Ambassador | ARC Centre of Excellence for All Sky Astrophysics in 3D | School of Physics Senior Research Data Specialist | Melbourne Data Analytics Platform & Petascale Campus Initiative | The University of Melbourne, Australia
Title: HPC Simulations of the early Universe
Abstract: Understanding the formation and evolution of the first galaxies in the Universe is a vital piece of the puzzle in understanding how all galaxies, including our own Milky Way, came to be. It is also a key aim of major forthcoming international facilities such as the Square Kilometre Array and James Webb Space Telescope. In order to maximise what we can learn from observations made by these facilities, we need to be able to accurately simulate the early Universe and model how galaxies affected and interacted with their environments.
Dr Simon Mutch a Postdoctoral Research Fellow in the Australian Research Council Centre of Excellence for All-Sky Astrophysics in 3-Dimensions (ASTRO 3D) and a Research Data Specialist in the Melbourne Data Analytics Platform (MDAP) based at the University of Melbourne. Dr Mutch received a Masters degree in Physics from the University of Edinburgh and was awarded his PhD in astrophysics from Swinburne University of Technology, Australia in 2013. His astronomy research is focused on the first galaxies and their impact on the evolution of the Universe, which he studies using a combination of supercomputer simulations and theoretical modelling. Through his role with MDAP, he is also currently collaborating with academics in areas such as climate science and ecology, to help uplift their digital research capabilities. Dr Mutch contributes to the Australian astronomical community through a number of national committees and is an inaugural Science and Technology Australia STEM Ambassador, a position which sees him meet with policy and decision makers to advocate for the importance of STEM and related fields.
Alan Edelman
Massachusetts Institute of Technology
Title: High Performance Computing: The Power of Language
Abstract: Julia is now being used for high performance computing for the important problems of today including climate change and Covid-19. We describe how language is making all the difference.
Alan Edelman is a professor of applied mathematics at MIT, is a member of MIT’s Computer Science & AI Laboratory, and is the leader of the JuliaLab and Applied Computing Group at MIT. He is also a cofounder of Julia Computing Inc. He works on numerical linear algebra, random matrix theory and parallel computing. He is a fellow of SIAM, IEEE, and the American Mathematial Society. He has won numerous prizes for his research, most recently the Fernbach Prize from IEEE for innovation in high performance computing.
STEPHEN WOLFRAM
Founder & CEO, Wolfram Research
Title: Emerging Surprises in Applying the Computational Paradigm (and the Deep Relations between Physics, Distributed Computing and AI)
Stephen Wolfram is the creator of Mathematica, Wolfram|Alpha and the Wolfram Language; the author of A New Kind of Science and other books; the originator of the Wolfram Physics Project; and the founder and CEO of Wolfram Research.
SARAH KENDERDINE
Digital Museology, Digital Humanities Institute | Lead: Laboratory for Experimental Museology (eM+) | Director: ArtLab | EPFL Lausanne Switzerland
Title: Cultural data sculpting
Abstract: In 1889 the curator G. B. Goode of the Smithsonian Institute delivered an anticipatory lecture entitled ‘The Future of the Museum’ in which he said this future museum would stand side by side with the library and the laboratory.’ Convergence in collecting organisations propelled by the liquidity of digital data now sees them reconciled as information providers in a networked world. The media theorist Lev Manovich described this world-order as “database logic,” whereby users transform the physical assets of cultural organisations into digital assets to be—uploaded, downloaded, visualized, shared, users who treat institutions not as storehouses of physical objects, but rather as datasets to be manipulated. This presentation explores how such a mechanistic description can replaced by ways in which computation has become ‘experiential, spatial and materialized; embedded and embodied’. It was at the birth of the Information Age in the 1950s that the prominent designer Gyorgy Kepes of MIT said “information abundance” should be a “landscapes of the senses” that organizes both perception and practice. “This ‘felt order’ he said should be “a source of beauty, data transformed from its measured quantities and recreated as sensed forms exhibiting properties of harmony, rhythm and proportion.”
Archives call for the creation of new prosthetic architectures for the production and sharing of archival resources. At the intersection of immersive visualisation technologies, visual analytics, aesthetics and cultural (big) data, this presentation explores diverse digital cultural heritage experiences of diverse archives from scientific, artistic and humanistic perspectives. Exploiting a series of experimental and embodied platforms, the discussion argues for a reformulation of engagement with digital archives at the intersection of the tangible and intangible and as a convergence across domains. The performative interfaces and repertoires described demonstrate opportunities to reformulate narrative in a digital context and they ways they support personal affective engagement with cultural memory.
Professor Sarah Kenderdine researches at the forefront of interactive and immersive experiences for galleries, libraries, archives and museums. In widely exhibited installation works, she has amalgamated cultural heritage with new media art practice, especially in the realms of interactive cinema, augmented reality and embodied narrative. In 2017, Sarah was appointed Professor of Digital Museology at the École polytechnique fédérale de Lausanne (EPFL), Switzerland where she has built a new laboratory for experimental museology (eM+), exploring the convergence of aesthetic practice, visual analytics and cultural data. She is also director and lead curator of EPFL’s new art/science initiative ArtLab.
FRANCISZEK RAKOWSKI
Interdisciplinary Centre for Mathematical and Computational Modelling, University of Warsaw | Samsung R&D, Poland
Title: Predicting the course of the COVID-19 epidemic in Poland
Abstract: One of the most promising approaches to predicting the possible scenarios of the epidemic is based on agent-based models. The idea of that model family is quite simple: reproduce the demographic and sociological structure of the society, and run the simulations of the disease spread throughout that structure. The direct reproduction of the contact structure allows investigating the consequences of various administrative measures applied, like school closure or travel restrictions. The model results can be visualised as a dynamic map of spreading disease, and enables the assessment of burden of disease factors locally. Our model, constructed more than 10 years ago for influenza epidemics, has been reanimated and tuned to COVID-19 parameters. It is now producing both scientific results as well as pragmatic reports, which are then being passed to Polish governmental authorities.
Affiliated Researcher at ICM University of Warsaw. Principal Data Scientist at the AI institute, Samsung R&D, Poland. Almost 10 years ago, when working at ICM UW as researcher, he led a project to construct a large-scale agent-based epidemic spread model for influenza disease.
Recently, during the times of COVID-19 pandemic, he became the project leader of development and adaptation of the ICM Epidemic Model to the coronavirus disease. His interests also cover cognitive science, neuroscience and computational biology.
ANETA AFELT
Interdisciplinary Centre for Mathematical and Computational Modelling, University of Warsaw | Espace-DEV, IRD – Institut de Recherche pour le Développement, Montpellier, France
Title: The promises of the One Health concept in the age of anthropocene
Abstract: In May 2019 an article was published: “Anthropocene now: influential panel votes to recognise Earth’s new epoch”, situating et the stratigraphy of Earth’s history a new geological epoch – the domination of human influence on shaping the Earth’s environment. When humansnas are a central figure in an ecological niche its results massive subordination and transformation of the environment for their needs.
Unfortunately, the consequence is robbery of natural resources.
The consequences are socially unexpected – a global epidemiological crisis. The current covid-19 pandemic is an excellent example. It seems that one of the most important questions of the anthropocene era is how to maintain stable epidemiological conditions for now and in the future. The One Health concept proposes a nwe paradigm – a deep look at the sources of our well-being: our relationship with the environment. Our health status is interdependent with the well-being of the environment. It is clear that the socio-ecological niche disturbance results in the spread of pathogens. Can sustainable development of socio-ecological niches help us? Let’s take a look at the results!
Aneta Afelt, PhD, is a geographer working in the area of health geography. Her interest in research is the One Health concept, where environment, epidemiology and epizootiology are considered as interconnected processes located in social-ecological niches. She shows in the research results that the destruction of ecosystems results in epidemiological consequences. She works at the Interdisciplinary Center for Mathematical and Computational Modeling of the University of Warsaw, Poland, and is currently a Guest Researcher at Espace-DEV, IRD – Institut de Recherche pour le Développement, Montpellier, France. She is also a member of the scientific committee for Covid-19 of the Ministry of Science in Poland and a scientific consultant of the European Research Agency for actions dedicated to Covid-19.
SELECTED PUBLICATIONS
Genetic diversity of coronaviruses in bats in Lao PDR and Cambodia. [Infection Genetics and Evolution, 48 (2017) 10–18]
Incidence of dengue and chikungunya viruses in mosquitoes and human patients in border provinces of Vietnam [BioMed Central, Springer Nature, 9 November 2017]
Japanese encephalitis in Indonesia: An update on epidemiology and transmission ecology [Acta Tropica, 187(2018), 240-247]
Distribution of bat-borne viruses and environment patterns [Infection Genetics and Evolution 58(2018), 181-191]
Bats, Coronaviruses, and Deforestation: Toward the Emergence of Novel Infectious Diseases? [Frontiers in Microbiology (11 April 2018)]
Bats, Bat-Borne Viruses, and Environmental Changes [in: Bats, IntechOpen (July 4th 2018)]
Hiroaki Kitano
President at The Systems Biology Institute, Tokyo;
Professor at Okinawa Institute of Science and Technology Graduate University, Okinawa;
President & CEO at Sony Computer Science Laboratories, Inc., Tokyo;
Executive Vice President at Sony Corporation, Tokyo
Title: Nobel Turing Challenge — Creating the Engine of Scientific Discovery
Abstract: A new grand challenge for AI: to develop an AI system that can make major scientific discoveries in biomedical sciences and that is worthy of a Nobel Prize. There are a series of human cognitive limitations that prevent us from making accelerated scientific discoveries, particularity in biomedical sciences. As a result, scientific discoveries are left at the level of a cottage industry. AI systems can transform scientific discoveries into highly efficient practices, thereby enabling us to expand knowledge in unprecedented ways. Such systems may outcompute all possible hypotheses and may redefine the nature of scientific intuition, hence the scientific discovery process.
Hiroaki Kitano received a PhD in computer science from Kyoto University in 1991 for the thesis in machine translation titled “Speech-to-speech translation: a massively parallel memory-based approach”. His work includes a broad spectrum of publications on artificial intelligence and systems biology.
Kitano is known for developing AIBO, Sony’s entertainment robot, and the grand challenge project known as RoboCup, that aims at developing a team of fully autonomous robot that can outperform the World Cup winning team in soccer. He is also a pioneer of systems biology, and has served as a scientific advisor for a number of research institutions and companies, including European Molecular Biology Laboratory (EMBL), ALSTOM, and others. He was awarded the IJCAI Computers and Thought Award in 1993 and the Nature Award for Creative Mentoring in Science in 2009.”
Ben Whitney
Postdoc in the Computer Science and Mathematics Division at Oak Ridge National Laboratory
Title: MGARD: Hierarchical Compression of Scientific Data
Abstract:The increasing scale of scientific datasets poses challenges to computational scientists needing to store, transfer, and understand their data. Lossy compression can help alleviate these problems, but practitioners are wary of incurring errors that may affect downstream analyses. In this talk, we present MGARD (MultiGrid Adaptive Reduction of Data), an algorithm for scientific data compression with guaranteed error bounds and the ability to preserve quantities of scientific interest. We begin by motivating the hierarchical decomposition underlying the algorithm. Next, we discuss quantization techniques for compression with guaranteed error bounds. Finally, we present numerical examples showcasing MGARD’s ability to preserve quantities of scientific interest.
Valerie E. Polichar
Sr. Director & SATO, Academic Technology Services
IT Services
University of California San Diego
Title: Creating a Technology Vision: Planning for the Next Generation of Research Support
Abstract:In 2009, the University of California San Diego released the Blueprint for a Digital University, a broad vision for research IT that set the direction for UC San Diego’s offerings for the next ten years and led to the creation of multiple campus services, as well as the university’s first Research IT Services and Research Data Curation teams. To refresh this vision for a new era, in late 2019, we embarked on a multi-year process to look to the next ten years and imagine a new range of research data and compute support services. We’ll discuss our process and some of the early themes emerging from the first year’s discussions – and how they differ from those of the past. Attendees will take away a how-to approach that can be applied at any research institution to build vision and develop their own strategic plan.
David Bader
Distinguished Professor in the Department of Computer Science;
Director of the Institute for Data Science at New Jersey Institute of Technology
Title: Solving Global Grand Challenges with High Performance Data Analytics
Abstract: Data science aims to solve grand global challenges such as: detecting and preventing disease in human populations; revealing community structure in large social networks; protecting our elections from cyber-threats, and improving the resilience of the electric power grid. Unlike traditional applications in computational science and engineering, solving these social problems at scale often raises new challenges because of the sparsity and lack of locality in the data, the need for research on scalable algorithms and architectures, and development of frameworks for solving these real-world problems on high performance computers, and for improved models that capture the noise and bias inherent in the torrential data streams. In this talk, Bader will discuss the opportunities and challenges in massive data science for applications in social sciences, physical sciences, and engineering.
David A. Bader is a Distinguished Professor in the Department of Computer Science and Director of the Institute for Data Science at New Jersey Institute of Technology. Prior to this, he served as founding Professor and Chair of the School of Computational Science and Engineering, College of Computing, at Georgia Institute of Technology. He is a Fellow of the IEEE, AAAS, and SIAM, and advises the White House, most recently on the National Strategic Computing Initiative (NSCI). Dr. Bader is a leading expert in solving global grand challenges in science, engineering, computing, and data science. His interests are at the intersection of high-performance computing and real-world applications, including cybersecurity, massive-scale analytics, and computational genomics, and he has co-authored over 250 articles in peer-reviewed journals and conferences. Dr. Bader has served as a lead scientist in several DARPA programs including High Productivity Computing Systems (HPCS) with IBM, Ubiquitous High Performance Computing (UHPC) with NVIDIA, Anomaly Detection at Multiple Scales (ADAMS), Power Efficiency Revolution For Embedded Computing Technologies (PERFECT), Hierarchical Identify Verify Exploit (HIVE), and Software-Defined Hardware (SDH). He has also served as Director of the Sony-Toshiba-IBM Center of Competence for the Cell Broadband Engine Processor. Bader is a cofounder of the Graph500 List for benchmarking “Big Data” computing platforms. Bader is recognized as a “RockStar” of High Performance Computing by InsideHPC and as HPCwire’s People to Watch in 2012 and 2014. In April 2019, Bader was awarded an NVIDIA AI Lab (NVAIL) award, and in July 2019, Bader received a Facebook Research AI Hardware/Software Co-Design award.
http://www.cs.njit.edu/~bader
Dr. Nicolas Weber, Senior Researcher, NEC Laboratories Europe
Dr. Erich Focht, Senior Manager R&D, NEC HPC Europe
Workshop title: EFFECTIVE NEURAL NETWORKS *WITHOUT* GPU
SOL: Transparent Neural Network Acceleration on NEC SX-Aurora TSUBASA
Date: Wed, Sept 30, 2020
Time: 09.00 am to 12.30 pm CEST
Limited access: 15 participants – first come first served
Payment: Free of charge
Note! Registration deadline, Sept 27, 2020
Participants will be informed about their qualification for the workshop by September 28th.
Introduction
In 2019, ICM University of Warsaw expanded its HPC infrastructure with a specialized vector computer, NEC SX Aurora TSUBASA [1], with eight vector processors. Aurora TSUBASA is used at ICM UW for calculations in physics, chemistry, AI, as well as development work intended to adapt and optimize the existing software to work on the new computer architecture.
Distinctive features of NEC SX-Aurora TSUBASA are:
- High memory bandwidth (48 GB, HBM2) of the Vector Engine (>1 TB/s) at < 300 W,
- 64 fully functional vector registers and 192 double precision FP operations per cycle,
- Works within the GNU/Linux environment – natively or in the accelerator mode.
The Workshop is intended as an introduction to two software frameworks designed specifically for NEC SX-Aurora TSUBASA
- NEC SOL – Transparent Neural Network Acceleration [2] – an AI acceleration middleware enabling wide range of optimizations for neural network workloads. It integrates with existing Machine Learning frameworks such as PyTorch and TensorFlow. SOL offers broad support for hardware architectures including CPUs (x86, arm64), GPUs (NVIDIA), and NEC SX-Aurora TSUBASA. It does not require modification of the original source code allowing the user to focus on solving the problem rather than on the specifics of the hardware architecture;
- Frovedis – FRamework Of VEctorized and DIStributed data analytics [3] – data analytics software primarily targeting the NEC SX-Aurora TSUBASA architecture.
Brief agenda
- SOL: Transparent Neural Network Acceleration:
- Introduction;
- Integration with PyTorch;
- Integration with ONNX;
- Deployment.
- Frovedis: FRamework Of VEctorized and DIStributed data analytics.
- Hands-on session: SOL at ICM infrastructure.
About SX-Aurora TSUBASA
NEC SX-Aurora Tsubasa, introduced to the market in 2018, is a vector processor (Vector Engine, VE) belonging to the SX architecture line which has been developed by NEC Corporation since mid-1980s [1]. Unlike its stand-alone predecessors, Tsubasa has been designed as a PCIe attached card working within and being operated by an x86_64 host server (Vector Host, VH) running a distribution of the GNU/Linux operating system. The latter provides a complete software development environment for the connected VEs and runs Vector Engine Operating System (VEOS) which, in turn, serves as operating system to the VE programs. [4].
[1] https://www.nec.com/en/global/solutions/hpc/sx/index.html
[2] http://sysml.neclab.eu/projects/sol/
[3] https://github.com/frovedis/frovedis
[4] https://github.com/veos-sxarr-NEC/veos
About presenters
Nicolas Weber is Senior Researcher at the NEC Laboratories Europe. He received his PhD for work on automated memory access optimizations for GPU in 2017 from TU Darmstadt. Since then he focusses on the efficient mapping of artificial intelligence workloads onto various accelerator processors such as NEC SX-Aurora or GPUs, to transparently increase performance and efficiency.
Erich Focht is Senior Manager of the Research and Development group at NEC HPC Europe. His work topics cover distributed systems software, parallel file systems, hybrid programming models, system software, tools and compilers for vector systems with the focus currently on applications, linear algebra, AI and cooperations on the SX-Aurora vector engine.
Nicolas Weber is Senior Researcher at the NEC Laboratories Europe. He received his PhD for work on automated memory access optimizations for GPU in 2017 from TU Darmstadt. Since then he focusses on the efficient mapping of artificial intelligence workloads onto various accelerator processors such as NEC SX-Aurora or GPUs, to transparently increase performance and efficiency.
Erich Focht is Senior Manager of the Research and Development group at NEC HPC Europe. His work topics cover distributed systems software, parallel file systems, hybrid programming models, system software, tools and compilers for vector systems with the focus currently on applications, linear algebra, AI and cooperations on the SX-Aurora vector engine.