Hello, I'm Zhenhuan Wei.
My current research focuses on AI Infra, training and inference optimization for LLM(large language models), with a particular emphasis on distributed systems.
I will soon join Alibaba T-Head, where I will work on framework adaptation and performance optimization for LLM on domestic PPU chips.
Previously, I interned at ByteDance in the Data-AML team, where I contributed to the research and development of a LLM framework serving ByteDance's internal search and advertising businesses. Before that, I interned at Tencent IEG, focusing on inference optimization for a TTS model.
Feel free to reach out: frankwei0109@gmail.com