Sam Foreman’s personal website
Author

Sam Foreman

Published

January 9, 2026

Modified

January 7, 2026

Sam Foreman

👋 Hi, I’m Sam!

I’m a Computational Scientist in the AI / ML group at the Argonne Leadership Computing Facility (ALCF).

I’m generally interested in the large scale distributed training of AI models for scientific applications, and am the co-lead of the Models / Pre-Training group for the AuroraGPT project.

Prior to this, I received my PhD in Physics from the University of Iowa in 2019, where I used ML to build better Markov Chain Monte Carlo sampling techniques for Lattice Quantum Chromodynamics (l2hmc-qcd).

Tip✨ New!
Tip✏️ Last Updated
Updated: 2026-01-09 @ 15:40:03

My current research focuses on using deep generative modeling to help build better sampling algorithms in lattice gauge theory. In particular, I’m interested in building gauge equivariant neural network architectures and using inductive priors to incorporate physical symmetries into machine learning models.


I received my PhD in Physics from the University of Iowa in 2019 and my thesis was on Learning Better Physics: A Machine Learning Approach to Lattice Gauge Theory.


Prior to this, I completed two bachelors degrees (Engineering Physics and Applied Mathematics, 2015) at The University of Illinois at Urbana-Champaign. My undergraduate dissertation was titled Energy Storage in Quantum Resonators and was supervised by Professor Alfred Hübler within the Center for Complex Systems Research at UIUC.

This work ultimately resulted in a patent !!

hits

© Copyright 2025 Sam Foreman

🏷️ ✏️
🎉 Happy New Year! 2026-01-07 2026-01-09
🧊 Cooling Down Checkpoints: Best Practices for Model Evaluation 2025-11-12 2025-12-11
🎨 Mixing Between Distributions While Training 2025-10-06 2025-12-14
📊 pbs-tui: TUI for PBS Job Scheduler Monitoring 2025-09-17 2026-01-09
🍹 BlendCorpus + TorchTitan @ ALCF 2025-09-12 2025-10-07
🏗️ Building PyTorch 2.8 from Source on Aurora 2025-06-14 2025-12-10
📆 2025 2025-06-14 2025-10-06
🧜‍♀️ Mermaid 2025-06-02 2025-09-14
📰 Nice Headings 2025-06-01 2025-09-14
06 2025-06-01 2025-06-14
🚧 Frameworks Issue with numpy > 2 2025-05-03 2025-06-07
🔥 Building PyTorch 2.6 from Source on Aurora 2025-04-28 2025-06-14
🧑🏻‍💻 Sam Foreman’s Résumé 2025-04-26 2025-11-09
🪛 Torchtune on Aurora 2025-03-23 2025-03-29
🚑 Torchtune Patch on Aurora 2025-03-23 2025-05-01
🫥 svgbob 2024-11-15 2025-05-01
💾 Converting Checkpoints 2024-10-17 2025-04-28
🏔️ Spike Skipper 2024-09-17 2025-03-29
🍋 ezpz @ ALCF 2024-08-23 2025-03-29
📝 ezpz-v1 2024-08-23 2025-07-04
💅 How to Make Dope Slides 2024-08-13 2025-12-31
🔳 l2hmc-qcd Example: 4D SU(3) 2024-07-24 2026-01-08
🎰 Deterministic flash-attn 2024-06-17 2025-04-26
📸 flash-attn on Sunspot 2024-06-17 2025-04-26
🏎️ Megatron-DeepSpeed on Intel XPU 2024-06-15 2025-04-26
🐛 mpi4py bug on Sunspot 2024-05-25 2025-03-29
🎲 MCMC + Diffusion Sampling 2024-04-15 2025-04-07
🐢 Starting Up Distributed Training on Aurora 2024-03-21 2025-04-07
🚂 Loooooooong Sequence Lengths 2024-02-12 2025-09-20
🏁 l2hmc Example: 2D U(1)U(1) 2024-02-12 2025-04-07
No matching items

See talks for a live view!

Tip[HTML ⇆ Reveal.js]

Convert from HTML to slideshow version of a page by appending /slides to the end of its URL, e.g.

📆 2025

📆 2024

📆 2023

📆 2022

::: ::: {.callout-tip icon="false" title="[[Scientific Data Science: An Emerging Symbiosis](https://saforem2.github.io/anl-job-talk/) @ ANL (05/2022)]{.dim-text}" collapse="true" style="width:100%;background-color: rgba(0,0,0,0.0)!important;"} ```{=html}
::: ## 📆 2021 ::: {.callout-tip icon="false" title="[[Accelerated Sampling Methods for LGT](https://saforem2.github.io/l2hmc-dwq25/), @ [DWQ @ 25 BNLBNL](https://indico.bnl.gov/event/13576/) 12/202112/2021]{.dim-text}" collapse="true" style="width:100%; background-color: rgba(0,0,0,0.0)!important;"} ```{=html}

l2hmc-qcd at the MIT Lattice Group Seminar, 2021

📆 2020

You can find a full list of my publications on my Google Scholar

  1. 🌎 AERIS: Argonne Earth Systems Model for Reliable and Skillful Predictions (Hatanpää et al. ())
  2. Aurora: Architecting Argonne’s First Exascale Supercomputer for Accelerated Scientific Discovery (Allen et al. ())
  3. HiPerRAG: High-Performance Retrieval Augmented Generation for Scientific Insights (Gokdemir et al. ())
  4. Automated Tuning for HMC Mass Ratios (Torsiello et al. ())
  5. MOFA: Discovering Materials for Carbon Capture with a GenAI and Simulation-Based Workflow (Yan et al. ())
  6. 🧪 MProt-DPO: Breaking the ExaFLOPS Barrier for Multimodal Protein Design with DPO (Dharuman et al. ())
  7. Intro to HPC Bootcamp: Engaging New Communities Through Energy Justice Projects (Leung et al. ())
  8. Thorough Characterization and Analysis of Large Transformer Model Training At-Scale (Cheng et al. ())
  9. MLMC: Machine Learning Monte Carlo for Lattice Gauge Theory (Sam Foreman, Jin, and Osborn ())
  10. Protein Generation via Genome-scale Language Models with Bio-physical Scoring (Dharuman et al. ())
  11. DeepSpeed4Science Initiative: Enabling Large-Scale Scientific Discovery (Song et al. ())
  12. Comprehensive Performance Study of LLMs on Novel AI Accelerators (Emani et al. ())
  13. Exploratory Analysis of Climate Data with ClimRR, Intro to HPC Bootcamp @ NERSC (Sam Foreman ())
  14. 🧬 GenSLMs: Genome-scale language models reveal SARS-Cov-2 evolutionary dynamics (Zvyagin et al. ())
  15. Lattice QCD and Particle Physics (Kronfeld et al. ())
  16. Applications of ML to Lattice QFT (Boyda et al. ())
  17. LeapFrogLayers: Trainable Framework for Effective Sampling (Sam Foreman et al. ())
  18. HMC with Normalizing Flows [slides] (Sam Foreman et al. ())
  19. Deep Learning Hamiltonian Monte Carlo [+ poster] (Sam Foreman, Jin, and C. ())
  20. Machine Learning and Neural Networks for Field Theory (Sam Foreman, Jin, and Osborn ())
  21. Examples of renormalization group transformations for image sets (Samuel Foreman et al. ())
  22. RG inspired Machine Learning for lattice field theory (Sam Foreman et al. ())
  23. Large Energy Density in Three-Plate Nanocapacitors due to Coulomb Blockade (Hubler et al. ())
  24. Superconductivity of In and Sn Samples (Deamont and Foreman ())
Allen, Benjamin S., James Anchell, Victor Anisimov, Thomas Applencourt, Abhishek Bagusetty, Ramesh Balakrishnan, Riccardo Balin, et al. 2025. “Aurora: Architecting Argonne’s First Exascale Supercomputer for Accelerated Scientific Discovery.” https://arxiv.org/abs/2509.08207.
Boyda, Denis, Salvatore Calı̀, Sam Foreman, Lena Funcke, Daniel C Hackett, Yin Lin, Gert Aarts, et al. 2022. “Applications of Machine Learning to Lattice Quantum Field Theory.” arXiv Preprint arXiv:2202.05838. https://arxiv.org/abs/2202.05838.
Cheng, Scott, Jun-Liang Lin, Murali Emani, Siddhisanket Raskar, Sam Foreman, Zhen Xie, Venkatram Vishwanath, and Mahmut Taylan Kandemir. 2024. “Thorough Characterization and Analysis of Large Transformer Model Training at-Scale.” Proc. ACM Meas. Anal. Comput. Syst. 8 (1). https://doi.org/10.1145/3639034.
Deamont, George, and Sam Foreman. 2014. “Superconductivity of in and Sn Samples.”
Dharuman, Gautham, Kyle Hippe, Alexander Brace, Sam Foreman, Väinö Hatanpää, Varuni K. Sastry, Huihuo Zheng, et al. 2024. “MProt-DPO: Breaking the ExaFLOPS Barrier for Multimodal Protein Design Workflows with Direct Preference Optimization.” In Proceedings of the International Conference for High Performance Computing, Networking, Storage, and Analysis. SC ’24. Atlanta, GA, USA: IEEE Press. https://doi.org/10.1109/SC41406.2024.00013.
Dharuman, Gautham, Logan Ward, Heng Ma, Priyanka V Setty, Ozan Gokdemir, Sam Foreman, Murali Emani, et al. 2023. “Protein Generation via Genome-Scale Language Models with Bio-Physical Scoring.” In Proceedings of the SC’23 Workshops of the International Conference on High Performance Computing, Network, Storage, and Analysis, 95–101.
Emani, Murali, Sam Foreman, Varuni Sastry, Zhen Xie, Siddhisanket Raskar, William Arnold, Rajeev Thakur, Venkatram Vishwanath, and Michael E Papka. 2023. “A Comprehensive Performance Study of Large Language Models on Novel AI Accelerators.” arXiv Preprint arXiv:2310.04607. https://arxiv.org/abs/2310.04607.
Foreman, Sam. 2023. “Energy Justice Analysis of Climate Data with ClimRR.” August 7, 2023. https://saforem2.github.io/climate-analysis.
Foreman, Sam, Joel Giedt, Yannick Meurice, and Judah Unmuth-Yockey. 2018. RG-inspired machine learning for lattice field theory.” In European Physical Journal Web of Conferences, 175:11025. European Physical Journal Web of Conferences. https://doi.org/10.1051/epjconf/201817511025.
Foreman, Sam, Taku Izubuchi, Luchang Jin, Xiao-Yong Jin, James C Osborn, and Akio Tomiya. 2021. “HMC with Normalizing Flows.” arXiv Preprint arXiv:2112.01586. https://arxiv.org/abs/2112.01586.
Foreman, Sam, Xiao-Yong Jin, and Osborn James C. 2021. “Deep Learning Hamiltonian Monte Carlo.” https://arxiv.org/abs/2105.03418.
Foreman, Sam, Xiao-Yong Jin, and James C Osborn. 2020. “Machine Learning and Neural Networks for Field Theory.”
Foreman, Sam, Xiao-Yong Jin, and James C. Osborn. 2023. “MLMC: Machine Learning Monte Carlo for Lattice Gauge Theory.” https://arxiv.org/abs/2312.08936.
Foreman, Samuel, Joel Giedt, Yannick Meurice, and Judah Unmuth-Yockey. 2018. “Examples of Renormalization Group Transformations for Image Sets.” Physical Review E 98 (5): 052129.
Gokdemir, Ozan, Carlo Siebenschuh, Alexander Brace, Azton Wells, Brian Hsu, Kyle Hippe, Priyanka V. Setty, et al. 2025. “HiPerRAG: High-Performance Retrieval Augmented Generation for Scientific Insights.” https://arxiv.org/abs/2505.04846.
Hatanpää, Väinö, Eugene Ku, Jason Stock, Murali Emani, Sam Foreman, Chunyong Jung, Sandeep Madireddy, et al. 2025. “AERIS: Argonne Earth Systems Model for Reliable and Skillful Predictions.” https://arxiv.org/abs/2509.13523.
Hubler, A, S Foreman, J Liu, and L Wortsmann. 2018. “Large Energy Density in Three-Plate Nanocapacitors Due to Coulomb Blockade.” Journal of Applied Physics 123 (10).
Kronfeld, Andreas S, Tanmoy Bhattacharya, Thomas Blum, Norman H Christ, Carleton DeTar, William Detmold, Robert Edwards, et al. 2022. “Lattice QCD and Particle Physics.” arXiv Preprint arXiv:2207.07641. https://arxiv.org/abs/2207.07641.
Leung, Mary Ann, Katharine Cahill, Rebecca Hartman-Baker, Paige Kinsley, Lois Curfman McInnes, Suzanne Parete-Koon, Sreeranjani Ramprakash, et al. 2024. “Intro to HPC Bootcamp: Engaging New Communities Through Energy Justice Projects.” Journal of Computational Science Education 15 (1). https://doi.org/10.22369/issn.2153-4136/15/1/10.
Song, Shuaiwen Leon, Bonnie Kruft, Minjia Zhang, Conglong Li, Shiyang Chen, Chengming Zhang, Masahiro Tanaka, et al. 2023. “DeepSpeed4Science Initiative: Enabling Large-Scale Scientific Discovery Through Sophisticated AI System Technologies.” arXiv Preprint arXiv:2310.04610. https://arxiv.org/abs/2310.04610.
Torsiello, J., G. T. Fleming, S. Foreman, X.-Y. Jin, and J. C. Osborn. 2025. “Automated Tuning for HMC Mass Ratios.” PoS. Argonne, ALCF; Argonne National Laboratory (ANL), Argonne, IL (United States); Temple U.; Fermi National Accelerator Laboratory (FNAL), Batavia, IL (United States). https://doi.org/10.22323/1.466.0052.
Yan, Xiaoli, Nathaniel Hudson, Hyun Park, Daniel Grzenda, J. Gregory Pauloski, Marcus Schwarting, Haochen Pan, et al. 2025. “MOFA: Discovering Materials for Carbon Capture with a GenAI- and Simulation-Based Workflow.” https://arxiv.org/abs/2501.10651.
Zvyagin, Maxim, Alexander Brace, Kyle Hippe, Yuntian Deng, Bin Zhang, Cindy Orozco Bohorquez, Austin Clyde, et al. 2023. “GenSLMs: Genome-Scale Language Models Reveal SARS-CoV-2 Evolutionary Dynamics.” The International Journal of High Performance Computing Applications 37 (6): 683–705.

saforem2s GitHub Repositories

ezpz

Public

Train across all your devices, ezpz 🍋

Python⭐️ 25Updated Jan 11, 2026

Github Stats

Python⭐️ 1Updated Jan 11, 2026

A curated list of my GitHub stars!

⭐️ 16Updated Jan 11, 2026

saforem2

Public

Profile README

⭐️ 5Updated Jan 11, 2026

My personal website

Python⭐️ 11Updated Jan 10, 2026

sam.onl

Public

No description provided.

Nix⭐️ 0Updated Jan 9, 2026

pbs-tui

Public

TUI for PBS Pro Scheduler

Python⭐️ 5Updated Jan 9, 2026

Modern parallelism techniques for training LLMs

SCSS⭐️ 12Updated Jan 8, 2026

Example repo demonstrating MLOps use cases on ALCF systems

Python⭐️ 4Updated Dec 30, 2025

amsc

Public

Python

Makefile⭐️ 0Updated Dec 16, 2025

Configuration for Kitty

⭐️ 0Updated Nov 22, 2025

Minimal, beautiful (+ highly-customizable) styles for Matplotlib.

Python⭐️ 19Updated Nov 21, 2025

diagrams

Public

Repo of various `draw.io` diagrams

⭐️ 7Updated Nov 15, 2025

Personal Website (using Quarto)

TeX⭐️ 3Updated Oct 10, 2025

ezpz-ai

Public

PyPi alias for https://github.com/saforem2/ezpz

⭐️ 0Updated Oct 9, 2025

dotfiles

Vim script⭐️ 1Updated Oct 6, 2025

chunkwm

Public

Tiling window manager for MacOS based on plugin architecture

C++⭐️ 49Updated Oct 4, 2025

Monte Carlo simulation of Z(N) models in lattice gauge theory.

Python⭐️ 4Updated Sep 24, 2025

lattice23

Public

Slides for Lattice 2023

TeX⭐️ 7Updated Sep 23, 2025

mmm

Public

Multi-Modal Modeling

Python⭐️ 6Updated Sep 8, 2025

mccl

Public

Collective communications using mpi4py

⭐️ 1Updated Aug 28, 2025

m

Public

monorepo

⭐️ 0Updated Aug 24, 2025

l2hmc-qcd

Public

Application of the L2HMC algorithm to simulations in lattice QCD.

Jupyter Notebook⭐️ 67Updated Aug 18, 2025

Intro to HPC Bootcamp 2025

HTML⭐️ 0Updated Aug 14, 2025

wordplay

Public

Playing with words

Python⭐️ 4Updated Aug 12, 2025

No description provided.

⭐️ 0Updated Jul 22, 2025

Quarto codespaces

Shell⭐️ 0Updated Jul 19, 2025

Worm algorithm implementation for 2D Ising model

Jupyter Notebook⭐️ 6Updated Jul 12, 2025

Demo Obsidian Vault

CSS⭐️ 4Updated Jul 8, 2025

sf

Public

sf: So Fast

⭐️ 0Updated Jul 7, 2025

blog-old

Public

New domain, new blog

⭐️ 0Updated Jul 3, 2025

No description provided.

⭐️ 0Updated Jun 9, 2025

orkz

Public

🎶 `orkz`: your devices, your symphony. Library for large scale orchestration of accelerators.

Python⭐️ 0Updated Mar 26, 2025

`pyorch`: PyTorch Orchestration. Your devices, your symphony 🎶

Python⭐️ 0Updated Mar 26, 2025

orch

Public

🎶 Orchestrator: Your devices, your symphony.

Python⭐️ 0Updated Mar 26, 2025

Personal Website (development version)

Python⭐️ 0Updated Feb 2, 2025

glam

Public

Neovim colorscheme that pops 💅

Lua⭐️ 0Updated Jan 28, 2025

Slides from Statistical Learning Talk @ ATPESC 2022

⭐️ 2Updated Jan 1, 2025

aoc24

Public

AOC 24

Python⭐️ 1Updated Dec 27, 2024

lazy-vim

Public

LazyVim starter config

Lua⭐️ 0Updated Dec 26, 2024

Playing with Large Vision Models and ViTs

Jupyter Notebook⭐️ 1Updated Dec 25, 2024

No description provided.

Lua⭐️ 0Updated Dec 25, 2024

No description provided.

CSS⭐️ 0Updated Dec 25, 2024

Simple tutorial on creating Small(-ish) LLMs (pt. 2 🎉!!)

Python⭐️ 3Updated Dec 13, 2024

LLMs at ALCF

TeX⭐️ 3Updated Dec 6, 2024

yap

Public

Learning to yap

⭐️ 0Updated Nov 7, 2024

Simple tutorial on creating Small(-ish) LLMs

TeX⭐️ 1Updated Oct 27, 2024

🎓 Education

👔 Professional Experience

  • Assistant Computational Scientist
    • Argonne National Laboratory, Leadership Computing Facility (ALCF) Lemont, IL | 2022–Present
      • Research lead on scaling large language models (LLMs) and generative AI for science on supercomputers (Aurora, Frontier, LUMI, Leonardo, …).
        • Co-lead the Models and Pretraining team of the AuroraGPT project
      • Optimize large-scale training of foundation models and language models for scientific applications.
      • Collaborate with interdisciplinary teams to enhance simulation efficiency and scalability
      • Focus on AI and HPC for scientific applications, including:
        • Training large language models on supercomputers
        • Genome scale language models (GenSLMs) for studying SARS-CoV-2 evolutionary dynamics
        • Direct Preference Optimization (DPO) for multimodal protein design workflows
        • Climate modeling and weather forecasting using foundation models
        • Developing improved sampling algorithms for lattice quantum chromodynamics (QCD)
      • https://www.alcf.anl.gov/about/people/sam-foreman
  • Postdoctoral Researcher
    • Argonne National Laboratory, Leadership Computing Facility (ALCF) Lemont, IL | 2019 – 2022
      • Applied deep learning to lattice gauge theory and quantum field simulations.
      • Developed ML-enhanced Monte Carlo methods for QCD (l2hmc-qcd).
      • Engaged in AI-for-Science collaborations with national labs and university partners.
  • Graduate Researcher (DOE SCGSR Fellowship)
    • Argonne National Laboratory, Mathematics and Computer Sciences Division (MCS)
      Lemont, IL | 2018 – 2019
      • Development of l2hmc-qcd in collaboration with ALCF for my PhD Thesis research

🏆 Awards and Honors

  • Member of the DeepSpeed Technical Steering Commiittee, 2025 – Present

    • Contributing to the development and direction of the DeepSpeed library for large-scale model training.
  • Nominated to serve on the US Coordinating Panel for Software and Computing by the Division of Particles and Fields of the American Physical Society (APS).

  • Finalist, ACM Gordon Bell Prize in Climate Modeling, 2025

    • Recognized for our work on
      🌎 AERIS (Hatanpää et al. ()): The first billion-parameter pixel-level diffusion model for global weather and subseasonal-to-seasonal forecasting. Trained efficiently at scales from 1.3–80B parameters with our sequence-window parallelism (SWiPe) strategy, we achieve a sustained mixed-precision performance of 10.21 ExaFLOPS and peak performance of 11.21 ExaFLOPS, scaling to 10,080 nodes (120,960 GPUs) on the Aurora supercomputer.
  • Finalist, ACM Gordon Bell Prize, 2024

  • ACM Gordon Bell Special Prize for High Performance Computing-Based COVID-19 Research, 2022

  • DOE Office of Science Graduate Student Research Fellow, 2018

    • Awarded by the Department of Energy for outstanding research contributions during graduate studies.

🎪 Events

My current research focuses on using deep generative modeling to help build better sampling algorithms in lattice gauge theory. In particular, I’m interested in building gauge equivariant neural network architectures and using inductive priors to incorporate physical symmetries into machine learning models.


I received my PhD in Physics from the University of Iowa in 2019 and my thesis was on Learning Better Physics: A Machine Learning Approach to Lattice Gauge Theory.


Prior to this, I completed two bachelors degrees (Engineering Physics and Applied Mathematics, 2015) at The University of Illinois at Urbana-Champaign. My undergraduate dissertation was titled Energy Storage in Quantum Resonators and was supervised by Professor Alfred Hübler within the Center for Complex Systems Research at UIUC.

This work ultimately resulted in a patent !!

hits

© Copyright 2025 Sam Foreman

Temporarily disabled while guesbooks gets their Azure issues worked out :(

Footnotes

  1. 🏅 Finalist for the Gordon Bell Prize in Climate Based Modeling at SC25!↩︎

Citation

BibTeX citation:
@online{foreman2026,
  author = {Foreman, Sam},
  date = {2026-01-09},
  url = {https://samforeman.me/},
  langid = {en}
}
For attribution, please cite this work as:
Foreman, Sam. 2026. January 9, 2026. https://samforeman.me/.