- 🔭 Interests focusing on: Computer Systems (LLMSys, GPUSys, StorageSys).
- 😄 Currently Permanent @Huawei Ascend NPU - MindIE AI Inference Team, Ex-Internship @Alibaba.
- 📬 How to reach me: huangxin.hust@gmail.com
GSoC 2022 - Final Project Report
Forked from sgl-project/sglang
SGLang is a high-performance serving framework for large language models and multimodal models.
Python
Forked from kvcache-ai/Mooncake
Mooncake is the serving platform for Kimi, a leading LLM service provided by Moonshot AI.
C++
Forked from tile-ai/tilelang
Domain-specific language designed to streamline the development of high-performance GPU/CPU/Accelerators kernels
Python 2
Forked from vllm-project/vllm
A high-throughput and memory-efficient inference and serving engine for LLMs
Python