Sam Foreman
2024-10-30
AuroraGPT: General purpose scientific LLM
Broadly trained on a general corpora plus scientific {papers, texts, data}
Hannibal046/Awesome-LLM
argonne-lcf/Megatron-DeepSpeed
| Racks | 166 | 
| Nodes | 10,624 | 
| CPUs | 21,248 | 
| GPUs | 63,744 | 
| NICs | 84,992 | 
| HBM | 8 PB | 
| DDR5c | 10 PB | 
Up to 25× improvement for genomic foundation models with 6.5× energy efficiency
✅ Goals
❌ Challenges
Megatron-DeepSpeedezpz🙏 Acknowledgements
This research used resources of the Argonne Leadership Computing Facility, which is a DOE Office of Science User Facility supported under Contract DE-AC02-06CH11357.
SEQ_LEN for both 25B and 33B models (See: Song et al. (2023))
