These 5 commits are when the Protocol Buffers files have changed:
Commit: | 876b7b4 | |
---|---|---|
Author: | Yao Fu | |
Committer: | GitHub |
[FIX] Fix loading loop in `SllmLocalStore` (#165) ## Description This PR enhances the `SllmLocalStore` by addressing issues in the loading loop and adapting to the latest automatic hardware configuration detection feature. ## Motivation - Fixes the first issue mentioned in #131. - Prepares the codebase for the upcoming live-migration feature release. ## Type of Change - [x] Bug fix - [x] Enhancement ## Checklist - [x] I have read the [CONTRIBUTING](https://github.com/ServerlessLLM/ServerlessLLM/blob/main/CONTRIBUTING.md) guide. - [x] The changes are tested and verified to maintain compatibility with existing features.
The documentation is generated from this commit.
Commit: | f773b68 | |
---|---|---|
Author: | Yao Fu | |
Committer: | GitHub |
Prepare for 0.5.0 release (#112) ## Description This PR includes necessary modifications for the 0.5.0 release: 1. Refactored the repository to shorten file paths (e.g., `serverless_llm/` to `sllm` and `serverless_llm/store/serverless_llm_store` to `sllm_store/sllm_store`). 2. Updated the README to clarify that `sllm` is pronounced "slim." 3. Added `pip install` instructions in the documentation. ## Motivation Previously, using the full name `serverless_llm` made it cumbersome to reference the codebase during discussions. We have now abbreviated `serverless_llm` to `sllm` (pronounced "slim"), as well as `serverless_llm_store` to `sllm_store` (pronounced "slim store"). However, "sllm" is unavailable on PyPI due to its similarity to an existing project `slim`. As a result, we will retain the full name `serverless_llm` for the `pip install` command but use `sllm` for all other references within the codebase and documentation. ## Type of Change - [x] Breaking change - [x] Documentation update ## Checklist - [x] I have read the [CONTRIBUTION](https://github.com/ServerlessLLM/ServerlessLLM/blob/main/CONTRIBUTING.md) guide. - [x] I have updated the tests (if applicable). - [x] I have updated the documentation (if applicable).
Commit: | 9b6a4db | |
---|---|---|
Author: | Yao Fu | |
Committer: | GitHub |
Fy/unified model path (#82) ## Description This PR cleans up ServerlessLLM Store design by provide dedicated interface for PyTorch and Transformers. ServerlessLLM Store now uses `model_path` instead of `model_name` as the global ID for a model checkpoint. By providing a clean API, ServerlessLLM now supports storage-aware model loading scheduling for vLLM backend. ## Changes - Move `save_model` and `load_model` to `serverless_llm_store.transformers` - Move 'save_dict` and `load_dict` to `serverless_llm_store.torch` - Use `serverless_llm_store.torch.save_dict/load_dict` in vLLM - Support and require model registration for vLLM backend
Commit: | 8796d13 | |
---|---|---|
Author: | Yao Fu | |
Committer: | GitHub |
feat: storage aware scheduler (#22) This PR implements a storage manager and a storage-aware model loading scheduler.
Commit: | 6847927 | |
---|---|---|
Author: | future-xy |
Initial commit: Launching ServerlessLLM This release includes basic modules of ServerlessLLM. This project builds upon our research published in OSDI 2024. As we transitioned from an internal development environment to this open-source platform, the detailed commit history and authorship information could not be preserved. Hence, we are taking this opportunity to formally acknowledge all contributors. - [Yao Fu](https://github.com/future-xy) - [Leyang Xue](https://github.com/drunkcoding) - [Yeqi Huang](https://github.com/Chivier) - [Andrei-Octavian Brabete](https://github.com/andrei3131) - [Matej Sandor](https://github.com/MatejSandor) - [Ruiqi Lai](https://github.com/lrq619) - [Siyang Shao](https://github.com/SiyangShao) - [Xinyuan Tong](https://github.com/JustinTong0323) Thank you to everyone involved for your dedication and expertise.