• Aucun résultat trouvé

HPC challenges for the next years: the rising of heterogeneity and its impact on simulations

N/A
N/A
Protected

Academic year: 2021

Partager "HPC challenges for the next years: the rising of heterogeneity and its impact on simulations"

Copied!
24
0
0

Texte intégral

(1)cres%c.univ-reims.fr. HPC challenges for the next years The rising of heterogeneity and its impact on simula%ons. CECAM Workshop Microscopic simula%ons: forecas%ng the next two decades Toulouse, April 24-26 2019. Luiz Angelo STEFFENEL angelo.steffenel@univ-reims.fr.

(2) About my team • Luiz Angelo Steffenel • Associate Professor, CReSTIC Laboratory • CASH Team (HPC, Autonomous compuAng, Heterogeneity) • Our team has a long tradiAon on HPC • ROMEO supercompuAng center • Part of MASCa. 2.

(3) Maison de la Simula%on Champagne-Ardenne More than 15 years associaAng HPC and applied compuAng 1 3 Molecular and Mul,scale Modeling Pla1orm. Image Center. 2. 1 2. ROMEO HPC Center. 3. 3.

(4) 2013 - Biggest hybrid CPU/GPU cluster in France 270 TFlops 151th in Top500 5th in Green500 2018 – Biggest academic cluster in France 1022 Tflops 249th in Top500 20th in Green500. 4.

(5) Top 500 ranking over the time We are in a "calm" period. Besides mulAcore, what is the biggest "innovaAon" since 2008? 2008. 2018 5.

(6) Hybrid architectures • Mix of CPUs and accelerators • GPUs (mostly NVIDIA) • Other accelerators (Xeon Phi). 6.

(7) TOP500 – Which is the impact of accelerators? GPUs as a way to reduce overall costs and look nice at TOP500 rank. Accelerators can deliver extra FLOPS but they add an extra heterogeneity layer à harder to explore 7.

(8) How to extract more from the hardware? • GPUs are good tools • Useful with specific code parts • Some problems are intrinsically hard • Hardware evoluAon helps doing faster, but does not reduce complexity • Beder results only come with addiAonal soeware development • Extra hardware = extra complexity. 8.

(9) Example: n-body problems • In 30 years • 107 hardware • 1010 soeware • Our problem now is that hardware is much more complex • Soeware has to struggle to control it. 9.

(10) The cost of Heterogeneity • Most of our programming models are 20+ years old (MPI, OpenMP, etc.) • Designed for homogeneous environments • Node-node, CPU-CPU, CPU-memory. • Current HPC has several layers • GPUs Good GPU Programming is not • Cores in a CPU trivial • MulA CPUs • MulAple layers of memory (cache, RAM, etc.) • InterconnecAons 10.

(11) So what are GPUs good for? • As a "piece of hardware", GPUs are no more special than co-processors for i386/i486 • Early HPC developments with GPUs started by exploring their parallel processing capabiliAes (SIMT) • GROMACS ✅ • Fluent ✅ • OpenFOAM ✅ • Autodock with GPU ❎ • Performance gains limited by memory size and latency constraints • Hard to code (CUDA, OpenCL, ...) 11.

(12) The revival of Neural Networks • GPUs are well-suited for the matrix/vector math involved in machine learning • Especially the famous Deep Learning • Data is oeen provided as a matrix of pixels • Or matrices of n-dimensions called "tensors" • The work can be split in several parallel tasks • Data is kept in the GPU memory for a long Ame. 12.

(13) Is AI the future of HPC?. nel1, Gustavo Rasera1, Nelson Begue3, Damaris Kirsch Pinheiro2, and Hassan Bencherif3 1 Université. de Reims Champagne-Ardenne, CReSTIC Laboratory EA 3804, Reims, France 2 Universidade Federal de Santa Maria, Santa Maria RS, Brazil 3 Université de La Réunion, LACy Laboratory UMR 8105, Reunion Island, France. • Once again, it's a good tool, not the answer • AI can help us to speed up simulaAons • What AI can do for us? • Unveil correla%ons. e scientific by episodes During these wards mide Antarctic last several potentially. tratospheric ) and 4,500 nd flora are. ly observed deed, recent he south of % in densely elow, which mid-latitudes ction in the ial vorticity e movement ca.. he numeric atic models er forecast), e layer.. METHODOLOGY. Input Sequence. t=1. Predicting the shape and movement of atmospheric events is a real challenge of predictive learning. Contrarily to some datasets studied by [7], atmospheric phenomena are not rigid, their speeds and trajectories vary, and their shapes may accumulate, dissipate or change rapidly due to the complex atmospheric environment. Furthermore, Ozone datasets are usually provided in a daily basis, making the tracking of the air masses even more complex. Hence, being able to model the spatial deformation of the Ozone concentration is a key aspect for the prediction of OSE events.. t=7. t=8. t=9. Ground Truth and Prediction. t=10. …. t=11. t=12. t=13. t=14. t=15. t=16. t=11. t=12. t=13. t=14. t=15. t=16. t=11. t=12. t=13. t=14. t=15. t=16. ->. April 2010. t=1. In this work we used part of the Multi Sensor Re-analysis (MSR2) [3] dataset corresponding to the OMI sensor, and covering the period between Jan 1, 2004 to Dec 31, 2018. This dataset is composed of daily measures of the Total Column Ozone (TCO) in Dobson units.. t=7. t=8. t=9. t=10. …. ->. October 2012. We preprocessed this dataset in order to highlight only the OSE occurrences, using these steps:. • Help improve the simulaAon models • Ex: meteorological models. 1. For a given day k, compute !"#$% , the average TCO for the precedent 15 days. 2. Compute the variation between mean of the previous days and the observation for day k ∆% = − *+,% − !"#$% ×100/!"#$% 3. Retain only the measures that show more than 10% of reduction ∆ IJ ≥ 10% ,FG = H % 0 LMℎ"OPIQ". t=1. t=7. t=8. …. These values were mapped as pixels in a 32 × 32 gray-scale images, each pixel with a value between 0 and 1 (0 to 100% variation). For each day in the dataset, we define a sequence consisting of 20 frames, 10 for the input (the previous 10 days), and 10 for the output (forecasting). The total 5,075 sequences are split into a training set of 3,998 samples and a test set of 1,077 samples. We trained the PredRNN++ model [7] over 10,000 iterations. After prediction, we transform the resulted intensities into colored maps. These predictions are compared with the real observations (ground truth).. ->. CONCLUSIONS. This preliminary work gives us some insights on the efficiency and precision of Deep Learning forecasts for Ozone Secondary Effects. The obtained results are encouraging as we were able to obtain some predictions with similar forms and intensities than the observed events. However, the results are not always precise, and the quality of the forecast easily degrade for long term predictions, requiring therefore an additional work on the learning algorithms and datasets.. • Fill the gap between simulaAons steps • Ex: molecular docking. Due to our previous works on OSE, we focus the experiment in the zone surrounding the Southern Spatial Observatory in Brazil (Observatório Espacial do Sul – OES/CRS/INPE – MCTI), located at 29.4°S; 53.8° W; 488.7 m). Hence, we selected a 32° × 32° grid centered at these coordinates.. t=10. October 2016. • Iden%fy/reproduce paVerns. We chose to observe the mean of a moving time window as it is more adapted to capture the seasonal fluctuations than the simple historical month mean. Also, we chose to represent the TCO variation instead of the raw Dobson values to highlight the events, simplifying the learning task. In addition, the main harm of OSE happens during the days following the sudden drop of the Ozone concentration. After a few days, even if the TCO concentration remains low, the population will probably be warned by the governmental health agencies to take the required actions (use sunscreen, hats, avoid exposition at given hours, etc.).. t=9. This work also helped us to improve data extraction and feature selection techniques, and to understand the impact of different parameters such as the number and size of hidden convolution layers. Hence, we shall continue to investigate how to better tune the Deep Learning network, as well as trying other algorithms adapted from the video frame.. REFERENCES. 1. Bencherif H., et al.: Examination of the 2002 major warming in the southern hemisphere using ground-based and Odin/SMR assimilated data: stratospheric ozone distributions and tropic/mid-latitude exchange, Can. J. Phys., 85, 12871300, 2007 2. Casiccia C., et al..: Erythemal irradiance at the Magellan’s region and Antarctic ozone hole 1999-2005, Atmosfera, 21, 1-12, 2008 3. Eskes, H. et al.: Assimilation of GOME total ozone satellite observations in a three-dimensional tracer transport model, Q.J.R.Meteorol.Soc. 129, 1663-1681, 2003. 4. Hauchecorne A., et al.: Quantification of the transport of chemical constituents from the polar vortex to midlatitudes in the lower stratosphere using the high-resolution advection model MIMOSA and effective diffusivity, J.Geophys.. 13.

(14) Ex: quantum many-body problem • Microsoe and ETH project • Use neural networks to represent the wave funcAon and reduce the compuAng complexity • AI does not replace the simulaAon models, just accelerate some steps •. hdps://science.sciencemag.org/content/355/6325/602. 14.

(15) BY COMBINING SIMULATION RESULTS FROM DIFFERENT MODELS TO PRODUCE A SUPERIOR MODEL.. AI + Simula%on = Synthesis Models. AI+HPC WORKFLOW FOR ENHANCEMENT MODELING Using Simulation Data To Train AI- Fermilab NOvA DL Training. 4.7M simulated events. 3.7M events for training. 1M events for Testing. CVN-Trained for 1 week. F. DLT FrameworksIGURE. CRY. Cosmic-Ray. 3:. Neutrino detection efficiency increasedCONDENSATE ACHIEVED CONVERGENCE BOSE-EINSTEIN. AFTER ONLY 10-12 EXPERIMENTS USING MACHINE LEARNING, by 33% COMPARED TO 140 EXPERIMENTS USING THE TRADITIONAL APPROACH. 1. Source: NVIDIA. 2. Modulation. AI+HPC WORKFLOW FOR MODULATION. AI-led Experiment To Converge Faster-bose Einstien Condensate AI Synthesis Modeling. Conventional Approach Nelder-Mead Convergence. ML can also be used to steer simulation or experiments between successive iterations, accelerating convergence to a stable, reliable model. In this case, simulation is used to train the neural network, which then refines the model to create input for the next simulation run or experiment. Researchers at the University of adjusts New each South Wales 1. Slowly parameter experimenting with Bose-Einstein Condensates (BEC) have been able to achieve the BEC state in only 10-12 experiments using synthesis modeling, compared to the 140 2. BEC reached in 140 experiments that are typically required using traditional models. experiments Page 3. Synthesis Modeling: The Intersection of HPC and Machine Learning November 2017 Copyright © 2017 Moor Insights & Strategy. ML Training. AI-Led Convergence. 1. Trained on 10 random experiments. 2. AI-Led experiments to converge on BEC. 3. BEC reached in 10-12. experiments. 2. Source: NVIDIA. 15.

(16) WARNING • Artificial Intelligence is handy, but has a (hidden) cost • Most AI models are developed for the same kind of problems • Ex: image recognition • Adapting our data is an expensive task • And no one dares to tell you this • Good training requires sufficient train data • What to do if we lack data? • Worst: AI is a black box • No explanation on the reasoning • Reproducibility is not a priority 16.

(17) Can we rely on GPUs for general compu%ng? • Trends for NVIDIA/AMD • 7nm or less, energy constraints, interconnecAon speed, but... • More and more dedicated for AI • Ex: TPUs from Google • Autonomous cars (Tesla, etc) • WARNING: all GPU development points towards mixed precision • Faster, acceptable precision • Not adapted for all problems 17.

(18) Even the CPUs are changing. • Arrival of ARM processors on the HPC market • Just an European dream? 18.

(19) European processor Ini%a%ve (EPI). Sovereignty. Independence from US components. Cost savings. Supercomputer. CPU + Memory = 65% of BOM. +70%. 3rd Parties. Blad es. supercomputer. 19. Developing a pan-European supercompu%ng infrastructure Public Members: 1 B€ ; EU Financial: 486 M€; Private partners 400 M€ 19.

(20) Europe can provide most of the elements. Own processor?. Interconnect. EPI (ARM). BXI. Scale-up system. HPC system. AI system. Consulting & services. Aries OEM by Atos. Power. CAPI. Intel. OPA. Huawei (ARM) Sunway. Post-K (ARM). Tofu OEM by Atos. IB. 20. available. planned. OEM. unavailable. Lacks only a good GPU # 20.

(21) And what about Quantum Computing • •. • •. PotenAal to solve difficult problems • Classical bit VS Qubits Only a few "real" quantum computers • Mostly simulators • Ex: QLM (ATOS + partners). Develop new algorithms • The "logic" is not the same Designing compuAng architectures • Many challenges on memory access, interconnecAon 21.

(22) This imply that a measurement will have equal pr become p 1 or 0. It represents a rotation of ⇡ about ẑ)/ 2. This is the combination of two rotations: ⇡ axis followed by ⇡/2 about the Y-axis. Its matrix r Quantum gates canisspeed up some complex  operaAons 1 1 1 p H ⌘ Quantum CompuAng 2 1 1 • Ex: Hadamard Gate. Quantum as accelerator. •. 5 to 10 years, minimum. The Hadamard gate is the one-qubit • Equivalent to a Discrete Fourrier Transform. •. uare • OT N OT ). version of fourier transform [42]. This is extremely useful fo Can replace part of the complex operaAons first computation in anyand quantum program bec initialized qubits back into their natural fluid outperform classicalforms algorithms to leverage their full quantum powers [32]. • hdps://www.ncbi.nlm.nih.gov/pmc/arAcles/PMC2596249/ - n-body • hdps://pubs.acs.org/doi/10.1021/acscentsci.8b00788 – Schrodinger. SAll, ofgate qubits! Rootrequires dozens This maps the basis states as follow Gate. (1 + i) |0i + (1 i) |1i |0i ! 2 (1 i) |0i + (1 + i) |1i |1i ! 2 The matrix representation of this gate is as follow 22.

(23) Some Conclusions and Forethoughts • Aeer a calm period, HPC is facing a new "Cambrian explosion" due to hardware heterogeneity • HPC soeware is sAll bound to 2000's methods à not enough!!! • GPUs have driven developers towards a risky path • Architecture-dependent • Low-level programming • This has a price • AI is not the "holy grail" • Neither Quantum CompuAng • The next years will be agitated! 23.

(24) Thanks!. 24.

(25)

Références

Documents relatifs

We can increase the flux by adding a series of fins, but there is an optimum length L m for fins depending on their thickness e and the heat transfer coefficient h. L m ~ (e

P 2 is easily extended to infer principal typings for recursive denitions under the new rule, resulting in a type system with decidable type inference that can type many

Dr Shadd is Assistant Professor, Dr Terry is Assistant Professor, Ms Chevendra is IT Consultant, and Dr Thind is Associate Professor, all at the Centre for Studies in Family

• le conseil d’établissement et l’ensemble des services de la commission scolaire; • les autres partenaires qui œuvrent en réussite éducative (certains ministères, les

سفن َيبط وىو ، تجلاعمل وئاقدصأ دقأ ىلإ يذخأ ررق اذل » ( 4 ) بمقف ، دجي مل دلاولا .ونبا ىمع نانئمطلااب وحور نكست ىتح ةحارلا ضفرف.. ريظت بقرتو فوخ

At our school -Holy Redeemer Upper School in Belize City- the struggle to keep the children healthy is un- ceasing.. We may lack the finance and technology needed to combat all

When accessing a database through SQL from an ordinary programming language, such as C# or Java, there is typically a strict segregation between data under the control of the DB

Les différentes étapes d’élaboration des films minces d’oxyde d’étain SnO2 par la technique de spray pyrolyse, l’effet des paramètres de dépôt tels que la