Meeting Notes 09-26-2022

Meeting Notes from 09-26-2022

Minutes of SBI-FAIR September 26, 2022, Meeting

Present: Kamil Iskra, Xiaodong Yu, Deborah Penchoff, Shantenu Jha, Geoffrey Fox, Piotr Luszczek, Baixi Sun. Vikram Jadhao, Gregor von Laszewski

Updates

Virginia

Geoffrey discussed

  • The transfer of the DOE grant is still making progress
  • He noted two nearly completed new surrogates
    • paper on Tsunami simulation surrogates entitled “Forecasting tsunami inundation with convolutional neural networks for a potential Cascadia Subduction Zone rupture”
    • Rough draft of the diffusion model for cell simulations GENERALIZATION AND TRANSFER LEARNING IN A DEEP DIFFUSION SURROGATE FOR MECHANISTIC REAL-WORLD SIMULATIONS. Interesting is the study of dataset sizes 5000-400,000 and the importance of dealing with the large numeric range in computed values
  • He summarized the MLCommons status with the move to continuous (rolling) submissions rather than fixed date submissions

Indiana University

  • Vikram presented some of his recent work
  • He studied sensitivity to input training set showing some dramatic effects from seemingly small changes – removing one value of electrolyte concentration c

Tennessee

Piotr reported

  • There was a Data Challenge at Smoky Mountain meeting with a smaller version of the Cloudmask dataset from MLCommons 2022 Challenge 6: SMCEFR: Sentinel-3 Satellite Dataset « SMC Data Challange 2021
  • Two Submitted papers: one on Performance Surrogate and the other a SABATH paper at HPEC Conference IEEE HPEC 26th Annual 2022 IEEE High Performance Extreme Computing Virtual Conference 19 - 23 September 2022
    • paper and presentation Deep Gaussian process with multitask and transfer learning for performance optimization
  • Questions included reproducibility and overheads from using FAIR metadata
  • It was asked if SABATH recorded training time; it does record loss versus epoch number.
  • Tennessee will give a detailed presentation on SABATH next time.

Rutgers

Shantenu reported

  • Drug and Quantum surrogates
  • He noted a new DOE $25M award for climate surrogates revisiting the startling Oxford paper https://iopscience.iop.org/article/10.1088/2632-2153/ac3ffa/meta and https://arxiv.org/pdf/2001.08055v1
  • Work with Indiana University was continuing with efforts to get system running on Summit
  • There was a discussion of Large Language models LLM and DOE interest in using them on scientific literature. There is a challenge with the current $10-100 million computing training cost possibly reaching a billion dollars.

Argonne

  • Xiaodong Yu discussed the ASPLOS paper which was unfortunately rejected
  • Baixi presented their results commenting on referee remarks
  • One question prompted observation that surrogate MODEL sizes are comparatively small
  • Another question was answered by noting that scheduling was a one-time cost
  • In some cases their custom training order outperformed the baseline training
Last modified January 26, 2024: add notes (fa4a2ea)