default search action
"Efficiently Training 7B LLM with 1 Million Sequence Length on 8 GPUs."
Pinxue Zhao et al. (2024)
- Pinxue Zhao, Hailin Zhang, Fangcheng Fu, Xiaonan Nie, Qibin Liu, Fang Yang, Yuanbo Peng, Dian Jiao, Shuaipeng Li, Jinbao Xue, Yangyu Tao, Bin Cui:
Efficiently Training 7B LLM with 1 Million Sequence Length on 8 GPUs. CoRR abs/2407.12117 (2024)
manage site settings
To protect your privacy, all features that rely on external API calls from your browser are turned off by default. You need to opt-in for them to become active. All settings here will be stored as cookies with your web browser. For more information see our F.A.Q.