From 95ac0afcb01ff50d945d69e48ba66aea2fb1064a Mon Sep 17 00:00:00 2001 From: Steven Shimizu Date: Thu, 22 Aug 2024 22:03:21 -0700 Subject: [PATCH] Added metadata for PyPI and bumped version (#52) --- .gitignore | 1 + README.md | 4 ++-- setup.py | 21 ++++++++++++++++++++- 3 files changed, 23 insertions(+), 3 deletions(-) diff --git a/.gitignore b/.gitignore index b5fc2aefc..6b0bd9a0b 100644 --- a/.gitignore +++ b/.gitignore @@ -2,6 +2,7 @@ __pycache__/ *.egg-info/ site/ .cache/ +.venv/ # Misc .DS_Store diff --git a/README.md b/README.md index ea61decc9..7b0a9d21e 100644 --- a/README.md +++ b/README.md @@ -10,14 +10,14 @@ ## Supercharge Your Model with Liger Kernel -![Banner](/docs/images/banner.GIF) +![Banner](https://raw.githubusercontent.com/linkedin/Liger-Kernel/main/docs/images/banner.GIF) With one line of code, Liger Kernel can increase throughput by more than 20% and reduce memory usage by 60%, thereby enabling longer context lengths, larger batch sizes, and massive vocabularies. | Speed Up | Memory Reduction | |--------------------------|-------------------------| -| ![Speed up](docs/images/e2e-tps.png) | ![Memory](docs/images/e2e-memory.png) | +| ![Speed up](https://raw.githubusercontent.com/linkedin/Liger-Kernel/main/docs/images/e2e-tps.png) | ![Memory](https://raw.githubusercontent.com/linkedin/Liger-Kernel/main/docs/images/e2e-memory.png) | > **Note:** > - Benchmark conditions: LLaMA 3-8B, Batch Size = 8, Data Type = `bf16`, Optimizer = AdamW, Gradient Checkpointing = True, Distributed Strategy = FSDP1 on 8 A100s. diff --git a/setup.py b/setup.py index f0eab207a..434eea525 100644 --- a/setup.py +++ b/setup.py @@ -1,12 +1,31 @@ from setuptools import find_namespace_packages, setup -__version__ = "0.1.0" +__version__ = "0.1.1" setup( name="liger_kernel", version=__version__, + description="Efficient Triton kernels for LLM Training", + long_description=open("README.md").read(), + long_description_content_type="text/markdown", + license="BSD-2-Clause", + url="https://github.com/linkedin/Liger-Kernel", package_dir={"": "src"}, packages=find_namespace_packages(where="src"), + classifiers=[ + 'Development Status :: 4 - Beta', + 'Intended Audience :: Developers', + 'Intended Audience :: Science/Research', + 'Intended Audience :: Education', + 'License :: OSI Approved :: BSD License', + 'Programming Language :: Python :: 3', + 'Programming Language :: Python :: 3.8', + 'Programming Language :: Python :: 3.9', + 'Programming Language :: Python :: 3.10', + 'Topic :: Software Development :: Libraries', + 'Topic :: Scientific/Engineering :: Artificial Intelligence', + ], + keywords="triton,kernels,LLM training,deep learning,Hugging Face,PyTorch,GPU optimization", include_package_data=True, install_requires=[ "torch>=2.1.2",