Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

docs: Website documentation for vllm inferencing using rayserve on AWS Inferentia #637

Merged
merged 6 commits into from
Sep 19, 2024

Conversation

sindhupalakodety
Copy link
Contributor

…n AWS Inferentia

What does this PR do?

🛑 Please open an issue first to discuss any significant work and flesh out details/direction - we would hate for your time to be wasted.
Consult the CONTRIBUTING guide for submitting pull-requests.

Created a website documentation for vllm inferencing using rayserve on AWS Inferentia

Motivation

To have users use vllm in association with RayServe and Trainium and to use llmperf tool for benchmarking

More

  • Yes, I have tested the PR using my local account setup (Provide any test evidence report under Additional Notes)
  • Mandatory for new blueprints. Yes, I have added a example to support my blueprint PR
  • Mandatory for new blueprints. Yes, I have updated the website/docs or website/blog section for this feature
  • Yes, I ran pre-commit run -a with this PR. Link for installing pre-commit locally

For Moderators

  • E2E Test successfully complete before merge?

Additional Notes

Copy link
Collaborator

@askulkarni2 askulkarni2 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

@sindhupalakodety thanks for the PR! Minor comments. Also for next time, please open an issue in the repo for tracking.

website/docs/gen-ai/inference/Neuron/vLLM-rayserve.md Outdated Show resolved Hide resolved
website/docs/gen-ai/inference/Neuron/vLLM-rayserve.md Outdated Show resolved Hide resolved
@askulkarni2 askulkarni2 changed the title Created a website documentation for vllm inferencing using rayserve on AWS Inferentia feat: Website documentation for vllm inferencing using rayserve on AWS Inferentia Sep 5, 2024
@askulkarni2 askulkarni2 changed the title feat: Website documentation for vllm inferencing using rayserve on AWS Inferentia doc: Website documentation for vllm inferencing using rayserve on AWS Inferentia Sep 5, 2024
@askulkarni2 askulkarni2 changed the title doc: Website documentation for vllm inferencing using rayserve on AWS Inferentia docs: Website documentation for vllm inferencing using rayserve on AWS Inferentia Sep 5, 2024
Copy link
Collaborator

@vara-bonthu vara-bonthu left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks for the doc PR 👍🏼 Left few comments to be addressed.

Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

inf2.8x has only one neuron and we are showing 6 here. Please update accordingly

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Sure. Will work on it.

Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Don't delete this file as a part of this PR. We can deprecate this later on along with the code in favour of RayServe with vLLM

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Hi Vara, this is the file I changed. Changed the name of it to RayServe with vLLM .. do you want a copy of the original file to be in the repo.

---
title: RayServe with vLLM
sidebar_position: 2
description: Deploy Llama-3 models on AWS Inferentia accelerators for efficient inference using vLLM.
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Update the description to "Deploying Llama-3 Models on AWS Inferentia2 with Ray for Efficient Inference Using vLLM"

Comment on lines 134 to 138
kubectl get ds neuron-device-plugin-daemonset --namespace kube-system
```
```bash
NAME DESIRED CURRENT READY UP-TO-DATE AVAILABLE NODE SELECTOR AGE
neuron-device-plugin-daemonset 2 2 2 2 2 <none> 2d2h
Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Please change the daemonset name from neuron-device-plugin-daemonset to neuron-device-plugin.

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Sure will work on it.

Copy link
Collaborator

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Please remove this image.

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Sure. Will work on it.

@ratnopamc ratnopamc self-requested a review September 19, 2024 00:36
Copy link
Collaborator

@ratnopamc ratnopamc left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

lgtm.

@vara-bonthu vara-bonthu merged commit 5da52b2 into awslabs:main Sep 19, 2024
4 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

4 participants