Siyuan Feng

Assistant Professor at The Apache Software Foundation

Xuhui District, Shanghai, China
email-iconphone-icongithub-logolinkedin-logotwitter-logostackoverflow-logofacebook-logo
Join Prog.AI to see contacts
email-iconphone-icongithub-logolinkedin-logotwitter-logostackoverflow-logofacebook-logo
Join Prog.AI to see contacts

Summary

🤩
Rockstar
🎓
Top School
Siyuan Feng is an assistant professor at Shanghai Innovation Institute with a PhD in Computer Science from Shanghai Jiao Tong University, focused on distributed machine learning systems and machine learning compilers. With eight years of experience he combines academic rigor with hands-on open-source engineering as an ASF member and Apache TVM PMC member. His contributions include CUDA backend performance work, FP16/INT8 support, TVMScript parser/printer improvements and even a new "while" node, alongside compiler-centric enhancements for the MLC-LLM project and RWKV model deployment. Advised in APEX Lab by Weinan Zhang and Yong Yu and collaborating closely with Tianqi Chen via open source, he is adept at translating research ideas into production-grade compilation optimizations.
code9 years of coding experience
bookDoctor of Philosophy - PhD, Computer Science, Doctor of Philosophy - PhD, Computer Science at Shanghai Jiao Tong University
languagesChinese, English
github-logo-circle

Github Skills (11)

cuda10
tvm10
compiler10
code-generation10
language-model10
python10
llm10
cprogramming-language9
c-language9
webgpu8
vulkan5

Programming languages (11)

TypeScriptC++ShellCSSCScalaJavaScriptHTML

Github contributions (5)

github-logo-circle
mlc-ai/mlc-llm

May 2023 - Mar 2025

Universal LLM Deployment Engine with ML Compilation
Role in this project:
userML Engineer
Contributions:54 reviews, 79 PRs, 93 pushes in 1 year 10 months
Contributions summary:Siyuan's contributions primarily focused on modifications and improvements to the ML compilation aspects of the project. The commits show the user fixing issues related to GPU parameters, addressing typos, and rebasing code related to the RWKV model. The user also updated documentation concerning model compilation for the RWKV architecture. The most significant contribution appears to be related to the compilation and optimization of various models within the MLC-LLM framework.
language-modelllmmachine-learning-compilationtvm
apache/tvm

Mar 2019 - Jan 2023

Open deep learning compiler stack for cpu, gpu and specialized accelerators
Role in this project:
userBack-end Developer & Compiler Engineer
Contributions:4 releases, 912 reviews, 68 commits in 3 years 10 months
Contributions summary:Siyuan's commits primarily focus on enhancing the performance and functionality of a deep learning compiler stack. They are responsible for adding new features, providing support for the tensor core, and addressing compilation warnings in source files. They implemented and tested performance improvements for the CUDA backend, including support for FP16 and INT8, as well as modifications to the handling of memory allocations. Their work involves developing a new "while" node, enhancing the TVMScript parser and printer, and optimizing code for WebGPU.
metalvulkancompilertensoropencl
Find and Hire Top DevelopersWe’ve analyzed the programming source code of over 60 million software developers on GitHub and scored them by 50,000 skills. Sign-up on Prog,AI to search for software developers.
Request Free Trial