1. AMD. 2023. AMD Infinity Architecture. https://www.amd.com/en/technologies/infinity-architecture
2. DeepSpeed- Inference: Enabling Efficient Inference of Transformer Models at Unprecedented Scale
3. Tom Brown Benjamin Mann Nick Ryder et al. 2020. Language models are few-shot learners. Advances in neural information processing systems 33 (2020) 1877--1901.
4. Shenggan Cheng, Ziming Liu, Jiangsu Du, and Yang You. 2023. ATP: Adaptive Tensor Parallelism for Foundation Models. arXiv preprint arXiv:2301.08658 (2023).
5. Shenggan Cheng Xuanlei Zhao Guangyang Lu Jiarui Fang Zhongming Yu Tian Zheng Ruidong Wu Xiwen Zhang Jian Peng and Yang You. 2023. FastFold: Reducing AlphaFold Training Time from 11 Days to 67 Hours. arXiv:2203.00854 [cs.LG]