共计 13 篇文章
2025
【论文阅读】ByteScale:Efficient Scaling of LLM Training with a 2048K Context Length on More Than 12,000
【论文阅读】ScheMoE:An Extensible Mixture-of-Experts Distributed Training System with Tasks Scheduling
【论文阅读】The Llama 3 Herd of Models(Section 3 Pre-Training)
【论文阅读】Reducing Activation Recomputation in Large Transformer Models
【论文阅读】Megatron-LM论文阅读
2024
【论文阅读】{MegaScale}:Scaling Large Language Model Training to More Than 10,000 {GPUs}
【论文阅读】Fluid:Dataset Abstraction and Elastic Acceleration for Cloud-native Deep Learning Training Jobs
【论文阅读】Gödel:Unified Large-Scale Resource Management and Scheduling at ByteDance
【论文阅读】Not All Resources are Visible:Exploiting Fragmented Shadow Resources in Shared-State Scheduler Architecture
2023
【论文阅读】In Search of an Understandable Consensus Algorithm