Skip to content

Commit 105dead

Browse files
authored
[Migration] Sendnn inference rename (#946)
<!-- markdownlint-disable --> ## Description This PR updates the `vllm_spyre` package to `sendnn_inference`. Docs and publication workflows have been updated in preparation to publish this as `sendnn-inference==2.0.0` ❗❗❗ Breaking configuration changes! - The plugin name for `VLLM_PLUGINS` is now `sendnn_inference` - All config options are now `SENDNN_INFERENCE_*` - The precompiled model parser now expects `sendnn_inference_version` ## Related Issues Closes #939 ## Test Plan - CPU unit tests - Spyre unit tests - Spot checks with full model integration tests - Ensure Test PyPI deploy works ## Checklist - [ ] I have read the [contributing guidelines](https://docs.vllm.ai/projects/spyre/en/latest/contributing) - [ ] My code follows the project's code style (run `bash format.sh`) - [ ] I have added tests for my changes (if applicable) - [ ] I have updated the documentation (if applicable) - [ ] My commits include a `Signed-off-by:` line (DCO compliance) --------- Signed-off-by: Joe Runde <joe@joerun.de>
1 parent 726bbde commit 105dead

120 files changed

Lines changed: 718 additions & 646 deletions

File tree

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

.github/CODEOWNERS

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -6,12 +6,12 @@
66
# unless a later match takes precedence
77
* @joerunde @prashantgupta24
88

9-
# This lists cover the "core" components of vLLM-Spyre that require careful review
10-
/vllm_spyre @yannicks1 @tdoublep @nikolaospapandreou @sducouedic
9+
# This lists cover the "core" components of SenDNN-Inference that require careful review
10+
/sendnn_inference @yannicks1 @tdoublep @nikolaospapandreou @sducouedic
1111

1212
# TODO: separate code into different packages:
13-
# /vllm_spyre/v1/worker/continuous_batching @nikolaospapandreou
14-
# /vllm_spyre/v1/worker/static_batching ...
13+
# /sendnn_inference/v1/worker/continuous_batching @nikolaospapandreou
14+
# /sendnn_inference/v1/worker/static_batching ...
1515

1616
# Tests
1717
/tests @rafvasq @prashantgupta24 @sducouedic

.github/ISSUE_TEMPLATE/bug-report.yml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,13 +1,13 @@
11
name: 🐛 Bug report
2-
description: Report a bug in vLLM Spyre
2+
description: Report a bug in SenDNN Inference
33
title: "[Bug]: "
44
labels: ["bug"]
55

66
body:
77
- type: markdown
88
attributes:
99
value: |
10-
Before submitting, please search [existing issues](https://github.com/vllm-project/vllm-spyre/issues) to avoid duplicates.
10+
Before submitting, please search [existing issues](https://github.com/torch-spyre/sendnn-inference/issues) to avoid duplicates.
1111
1212
- type: textarea
1313
attributes:

.github/ISSUE_TEMPLATE/config.yml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -2,12 +2,12 @@ blank_issues_enabled: false
22
contact_links:
33
- name: 📖 Documentation
44
url: https://docs.vllm.ai/projects/spyre/
5-
about: Read the vLLM Spyre documentation
5+
about: Read the SenDNN-Inference documentation
66
- name: 💬 Discussions
77
url: https://inviter.co/vllm-slack
88
about: Ask questions and discuss in the `#sig-spyre` channel.
99
- name: 🤝 Contributing Guide
1010
url: https://docs.vllm.ai/projects/spyre/en/latest/contributing
11-
about: Learn how to contribute to vLLM Spyre
11+
about: Learn how to contribute to SenDNN-Inference
1212

1313
# Made with Bob

.github/ISSUE_TEMPLATE/feature-request.yml

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -1,13 +1,13 @@
11
name: 🚀 Feature request
2-
description: Suggest a new feature for vLLM Spyre
2+
description: Suggest a new feature for SenDNN Inference
33
title: "[Feature]: "
44
labels: ["enhancement"]
55

66
body:
77
- type: markdown
88
attributes:
99
value: |
10-
Before submitting, please search [existing issues](https://github.com/vllm-project/vllm-spyre/issues) to avoid duplicates.
10+
Before submitting, please search [existing issues](https://github.com/torch-spyre/sendnn-inference/issues) to avoid duplicates.
1111
1212
- type: textarea
1313
attributes:
@@ -22,7 +22,7 @@ body:
2222
attributes:
2323
label: Motivation and context
2424
description: |
25-
Why is this feature important? How would it benefit vLLM Spyre users?
25+
Why is this feature important? How would it benefit SenDNN Inference users?
2626
2727
Is this related to a specific problem or limitation?
2828
validations:

.github/ISSUE_TEMPLATE/rfc.yml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,13 +1,13 @@
11
name: 💭 Request for comments
2-
description: Start a discussion for larger changes related to vLLM Spyre
2+
description: Start a discussion for larger changes related to SenDNN Inference
33
title: "[RFC]: "
44
labels: []
55

66
body:
77
- type: markdown
88
attributes:
99
value: |
10-
Before submitting, please search [existing issues](https://github.com/vllm-project/vllm-spyre/issues) to avoid duplicates.
10+
Before submitting, please search [existing issues](https://github.com/torch-spyre/sendnn-inference/issues) to avoid duplicates.
1111
1212
- type: textarea
1313
attributes:

.github/workflows/build_and_publish.yaml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,12 +1,12 @@
1-
name: Build and Publish vllm-spyre Library
1+
name: Build and Publish sendnn-inference Library
22

33
on:
44
release:
55
types: [published]
66

77
jobs:
88
build:
9-
# Only run for vllm-spyre releases (v* tags).
9+
# Only run for sendnn-inference releases (v* tags).
1010
if: startsWith(github.ref, 'refs/tags/v')
1111
runs-on: ubuntu-latest
1212

.github/workflows/build_docker.yml

Lines changed: 4 additions & 19 deletions
Original file line numberDiff line numberDiff line change
@@ -1,23 +1,8 @@
11
name: "Build Docker"
22

3+
# Manual dispatch only for now
34
on:
45
workflow_dispatch:
5-
push:
6-
branches:
7-
- "main"
8-
paths-ignore:
9-
- "**.md"
10-
pull_request:
11-
branches:
12-
- "main"
13-
paths:
14-
- ".github/workflows/build_docker.yml"
15-
- "docker/**"
16-
- "vllm_spyre/**/*.py"
17-
- "pyproject.toml"
18-
- "uv.lock"
19-
release:
20-
types: [published]
216

227
defaults:
238
run:
@@ -27,8 +12,8 @@ jobs:
2712
build:
2813
runs-on: ubuntu-latest
2914
env:
30-
CACHE_IMAGE: "ghcr.io/vllm-project/vllm-spyre:build-cache"
31-
QUAY_REPOSITORY: "quay.io/ibm-aiu/vllm-spyre"
15+
CACHE_IMAGE: "ghcr.io/torch-spyre/sendnn-inference:build-cache"
16+
QUAY_REPOSITORY: "quay.io/ibm-aiu/sendnn-inference"
3217

3318
permissions:
3419
packages: write
@@ -38,7 +23,7 @@ jobs:
3823
- name: "Checkout"
3924
uses: actions/checkout@v4
4025
# We need to fetch history and tags so that we can correctly version the
41-
# vllm_spyre package, since it uses setuptools_scm to version based on
26+
# sendnn-inference package, since it uses setuptools_scm to version based on
4227
# git tags.
4328
with:
4429
fetch-depth: 0

.github/workflows/publish_to_test_pypi.yaml

Lines changed: 2 additions & 2 deletions
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,4 @@
1-
# This tests the publication process for vllm-spyre by pushing a pre-release to test.pypi.org
1+
# This tests the publication process for sendnn-inference by pushing a pre-release to test.pypi.org
22
# on all main commits
33
name: Test publication on test.pypi.org
44

@@ -13,7 +13,7 @@ jobs:
1313

1414
environment:
1515
name: testpypi
16-
url: https://test.pypi.org/p/vllm-spyre
16+
url: https://test.pypi.org/p/sendnn-inference
1717
permissions:
1818
id-token: write # IMPORTANT: this permission is mandatory for trusted publishing
1919

.github/workflows/test.yml

Lines changed: 7 additions & 7 deletions
Original file line numberDiff line numberDiff line change
@@ -18,7 +18,7 @@ env:
1818
FORCE_COLOR: "1"
1919
VLLM_CPU_DISABLE_AVX512: "true"
2020
VLLM_TARGET_DEVICE: "empty"
21-
VLLM_PLUGINS: "spyre"
21+
VLLM_PLUGINS: "sendnn_inference"
2222
HF_HUB_CACHE: "${{ github.workspace }}/.cache/huggingface/hub"
2323
DEFAULT_HF_MODEL: "ibm-ai-platform/micro-g3.3-8b-instruct-1b"
2424
DEFAULT_HF_MODEL_REV: "6e9c6465a9d7e5e9fa35004a29f0c90befa7d23f"
@@ -131,8 +131,8 @@ jobs:
131131
pyproject.toml
132132
uv.lock
133133
tests/**/*.py
134-
vllm_spyre/**/*.py
135-
vllm_spyre/config/model_configs.yaml
134+
sendnn_inference/**/*.py
135+
sendnn_inference/config/model_configs.yaml
136136
137137
- name: "Install uv"
138138
if: steps.changed-src-files.outputs.any_changed == 'true'
@@ -159,7 +159,7 @@ jobs:
159159
160160
# Syncs both the runtime and dev deps, based on the lockfile contents
161161
uv sync --frozen
162-
# Builds and installs the vllm_spyre wheel into .venv
162+
# Builds and installs the sendnn-inference wheel into .venv
163163
# This needs to be done after `uv sync`, or the wheel install will be
164164
# overwritten.
165165
uv pip install -v .
@@ -260,15 +260,15 @@ jobs:
260260
MASTER_PORT: 12355
261261
MASTER_ADDR: localhost
262262
DISTRIBUTED_STRATEGY_IGNORE_MODULES: WordEmbedding
263-
VLLM_SPYRE_TEST_MODEL_LIST: ""
263+
SENDNN_INFERENCE_TEST_MODEL_LIST: ""
264264
HF_HUB_OFFLINE: 1
265265
run: |
266266
# Delete the source code so we can ensure we're testing the installed
267267
# wheel
268-
rm -fr vllm_spyre
268+
rm -fr sendnn_inference
269269
# We activate .venv manually and run pytest directly instead of using
270270
# `uv run`, to avoid having `uv run` re-sync any dependencies or
271-
# re-install the vllm_sypre package from source
271+
# re-install the sendnn-inference package from source
272272
source .venv/bin/activate
273273
274274
${{matrix.test_suite.env_overrides}} python3 -m pytest ${{ matrix.test_suite.flags }} \

.pre-commit-config.yaml

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -37,7 +37,7 @@ repos:
3737
hooks:
3838
- id: ty
3939
name: ty check
40-
entry: uvx ty@0.0.16 check vllm_spyre
40+
entry: uvx ty@0.0.16 check sendnn_inference
4141
# `system` here instead of python so that we avoid running inside an isolated venv created by
4242
# prek, and instead have access to the .venv created by uv for the project
4343
language: system

0 commit comments

Comments
 (0)