Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Improve vLLM backend documentation #22

Merged
merged 9 commits into from
Nov 22, 2023
Merged

Improve vLLM backend documentation #22

merged 9 commits into from
Nov 22, 2023

Conversation

tanmayv25
Copy link
Contributor

Avoids a lot of confusion and improves user experience.

README.md Outdated Show resolved Hide resolved
README.md Outdated Show resolved Hide resolved
README.md Outdated Show resolved Hide resolved
README.md Outdated Show resolved Hide resolved
Copy link
Contributor

@oandreeva-nv oandreeva-nv left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM! Thanks for improving docs. Left some suggestions.

tanmayv25 and others added 3 commits November 22, 2023 11:10
oandreeva-nv
oandreeva-nv previously approved these changes Nov 22, 2023
README.md Outdated
registry. These are available starting in 23.10.
The tritonserver_vllm container has everything you need to run your vLLM model.
registry. \<xx.yy\> is the version of Triton that you want to use. Please note,
that Triton's vLLM container was first published in 23.10 release, so any prior
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I would get rid of the subordinate clause (, so any prior verson will not work.). It is redundant and can be unclear for future Triton versions.

README.md Outdated Show resolved Hide resolved
README.md Outdated
in [build.py](https://github.com/triton-inference-server/server/blob/main/build.py)
for the Triton version you are using.

If you would like to use a specific vLLM commit or the latest version of vLLM, you
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

I think this is no longer necessary.
To get the latest vLLM, you can do: pip install git+https://github.com/vllm-project/vllm.git
To get a specific vLLM commit, you can do something like: pip install git+https://github.com/vllm-project/vllm.git@7d761fe3c12e87df37383467c43c97dec2bb8470

Co-authored-by: dyastremsky <[email protected]>
Copy link
Contributor

@oandreeva-nv oandreeva-nv left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks!

@tanmayv25 tanmayv25 merged commit c736380 into main Nov 22, 2023
3 checks passed
@tanmayv25 tanmayv25 deleted the tanmayv-doc branch November 22, 2023 20:21
@dyastremsky
Copy link
Contributor

Great work!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

3 participants