- 👋 Hi, I’m @JerryGJX
-
Shanghai Jiao Tong University
- Shanghai , China
Highlights
- Pro
Popular repositories Loading
-
-
-
-
-
llm-awq-long-tinychat
llm-awq-long-tinychat PublicForked from mit-han-lab/llm-awq
[MLSys 2024 Best Paper Award] AWQ: Activation-aware Weight Quantization for LLM Compression and Acceleration
Python
-
Block-Sparse-Attention
Block-Sparse-Attention PublicForked from mit-han-lab/Block-Sparse-Attention
A sparse attention kernel supporting mix sparse patterns
C++
Something went wrong, please refresh the page to try again.
If the problem persists, check the GitHub status page or contact support.
If the problem persists, check the GitHub status page or contact support.