Skip to content

Commit f42b34b

Browse files
authored
Removing leftovers fork from plugin (#525)
Signed-off-by: PatrykWo <patryk.wolsza@intel.com>
1 parent 82085eb commit f42b34b

File tree

5 files changed

+10
-12
lines changed

5 files changed

+10
-12
lines changed

.cd/Dockerfile.rhel.tenc.pytorch.vllm

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -12,7 +12,7 @@ ARG TORCH_TYPE_SUFFIX
1212

1313
FROM ${DOCKER_URL}/${VERSION}/${BASE_NAME}/${REPO_TYPE}/pytorch-${TORCH_TYPE_SUFFIX}installer-${PT_VERSION}:${REVISION}
1414

15-
# Parameterize commit/branch for vllm-fork checkout
15+
# Parameterize commit/branch for vllm-plugin checkout
1616
ARG VLLM_GAUDI_COMMIT=main
1717
# leave empty to use last-good-commit-for-vllm-gaudi
1818
ARG VLLM_PROJECT_COMMIT=

.cd/README.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -27,7 +27,7 @@ Supports a wide range of validated models including LLaMa, Mistral, and Qwen fam
2727

2828
### 0. Clone the Repository
2929

30-
Before proceeding with any of the steps below, make sure to clone the vLLM fork repository and navigate to the `.cd` directory. This ensures you have all necessary files and scripts for running the server or benchmarks.
30+
Before proceeding with any of the steps below, make sure to clone the vLLM plugin repository and navigate to the `.cd` directory. This ensures you have all necessary files and scripts for running the server or benchmarks.
3131

3232
```bash
3333
git clone https://github.com/vllm-project/vllm-gaudi.git

docs/features/compatibility_matrix.md

Lines changed: 4 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -7,7 +7,7 @@ title: Compatibility Matrix
77

88
The table below detail the configurations and versions supported for Intel® Gaudi® 2 AI accelerator, and Intel® Gaudi® 3 AI accelerator.
99

10-
| **Intel Gaudi Software** | **vLLM v0.10.0** | **vLLM v0.10.1** | **vLLM v0.10.2** | **vLLM v0.11.0** |
11-
| :---: | :---: | :---: | :---: | :---: |
12-
| 1.22.1 | ✅ alfa | ✅ beta |||
13-
| 1.23.0 |||| in development |
10+
| **Intel Gaudi Software** | **vLLM v0.10.0** | **vLLM v0.10.1** | **vLLM v0.10.2** | **vLLM v0.11.0** | **vLLM v0.11.1** |
11+
| :---: | :---: | :---: | :---: | :---: | :---: |
12+
| 1.22.1 | ✅ alfa | ✅ beta ||||
13+
| 1.23.0 |||| RC | in development |

docs/getting_started/quickstart.md

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -36,7 +36,7 @@ Guide](https://docs.habana.ai/en/latest/Installation_Guide/index.html) instructi
3636

3737
Follow these steps to run the vLLM server or launch benchmarks on Gaudi using Docker Compose.
3838

39-
1. Clone the vLLM fork repository to get the required files and Docker Compose configurations.
39+
1. Clone the vLLM plugin repository to get the required files and Docker Compose configurations.
4040

4141
```bash
4242
git clone https://github.com/vllm-project/vllm-gaudi.git

docs/user_guide/faq.md

Lines changed: 3 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -61,11 +61,9 @@ For more details, see [Quick Start Using Dockerfile](../getting_started/quicksta
6161
6262
- There are two different installation methods:
6363
64-
- (Recommended) Install the stable version from the HabanaAIvLLM-fork GitHub repo. This version is most suitable for production deployments.
64+
- (Recommended) Running vLLM Hardware Plugin for Intel® Gaudi® Using Dockerfile. This version is most suitable for production deployments.
6565
66-
- Install the latest version from the HabanaAI/vLLM-fork GitHub repo. This version is suitable for developers who would like to work on experimental code and new features that are still being tested.
67-
68-
- Install from the main vLLM source GitHub repo. This version is suitable for developers who would like to work with the official vLLM-project but may not have the latest Intel Gaudi features.
66+
- Building vLLM Hardware Plugin for Intel® Gaudi® from Source. This version is suitable for developers who would like to work on experimental code and new features that are still being tested.
6967
7068
## Examples and Model Support
7169
@@ -77,7 +75,7 @@ For more details, see [Quick Start Using Dockerfile](../getting_started/quicksta
7775
7876
- Different tensor parallelism configurations (single HPU, 2x, and 8x HPU).
7977
80-
- See [Supported Configurations](https://github.com/HabanaAI/vllm-fork/blob/v1.22.1/README_GAUDI.md#supported-configurations) for more details.
78+
- See [Validated Models](../models/validated_models.md) for more details.
8179
8280
## Features and Support
8381

0 commit comments

Comments
 (0)