Movatterモバイル変換


[0]ホーム

URL:


Skip to content

Navigation Menu

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Sign up
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to ourterms of service andprivacy statement. We’ll occasionally send you account related emails.

Already on GitHub?Sign in to your account

docs: update ray serve section#48770

Open
saihaj wants to merge1 commit intomaster
base:master
Choose a base branch
Loading
fromsaihaj/marketing-ray-serve

Conversation

saihaj
Copy link
Member

@saihajsaihaj commentedNov 16, 2024
edited by angelinalg
Loading

Signed-off-by: Saihajpreet Singhc-saihajpreet.singh@anyscale.com
Moving marketing type content to anyscale.com so we can keep user documentation concise. The removed content will now live here:https://www.anyscale.com/product/library/ray-serve
See this doc for more context:
ttps://docs.google.com/document/d/10xTHUhFDDD214xIeKPZK4Jrls_bdCvjR8phKQXygmgY/edit?tab=t.0

Signed-off-by: Saihajpreet Singh <c-saihajpreet.singh@anyscale.com>
@saihajsaihaj added the goadd ONLY when ready to merge, run all tests labelNov 18, 2024
Ray Serve is particularly well suited for [model composition](serve-model-composition) and many model serving, enabling you to build a complex inference service consisting of multiple ML models and business logic all in Python code.

Ray Serve is built on top of Ray, so it easily scales to many machines and offers flexible scheduling support such as fractional GPUs so you can share resources and serve many machine learning models at low cost.
Ray Serve is built on top of Ray, so it easily scales to many machines and offers flexible scheduling support such as fractional GPUs—so you can share resources and serve many machine learning models at low cost.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others.Learn more.

Suggested change
Ray Serve is built on top of Ray, so it easily scales to many machines and offers flexible scheduling support such as fractional GPUs—so you can share resources and serve many machine learning models at low cost.
Ray Serve is built on top of Ray, so it easily scales to many machines and offers flexible scheduling support such as fractional GPUs. You can share resources and serve many machine learning models at low cost.

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others.Learn more.

Just getting rid of the secondso.

@@ -35,12 +35,9 @@ api/index

(rayserve-overview)=

Ray Serve is a scalable model serving library for building online inference APIs.
Serve is framework-agnostic, so you can use a single toolkit to serve everything from deep learning models built with frameworks like PyTorch, TensorFlow, and Keras, to Scikit-Learn models, to arbitrary Python business logic. It has several features and performance optimizations for serving Large Language Models such as response streaming, dynamic request batching, multi-node/multi-GPU serving, etc.
Ray Serve is a scalable, framework-agnostic model serving library for building online inference APIs. Serve integrates with any ML framework including PyTorch, TensorFlow, Keras, Scikit-Learn, and more. It's particularly well suited for model composition and many model serving, and includes performance optimizations for serving LLMs such as response streaming, dynamic request batching, multi-mode/multi-GPU serving, and more.
Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others.Learn more.

Could you add a link to the first occurence ofRay Serve that goes to where the marketing content for Serve moved to?

Copy link
Contributor

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others.Learn more.

let's not forget this piece - "to arbitrary Python business logic" -> that's an important highlight. Also model composition and many model serving are more differentiated than the LLM serving features

@staleStale
Copy link

stalebot commentedFeb 1, 2025

This pull request has been automatically marked as stale because it has not had recent activity. It will be closed in 14 days if no further activity occurs. Thank you for your contributions.

  • If you'd like to keep this open, just leave any comment, and the stale label will be removed.

@stalestalebot added the staleThe issue is stale. It will be closed within 7 days unless there are further conversation labelFeb 1, 2025
Sign up for freeto join this conversation on GitHub. Already have an account?Sign in to comment
Reviewers

@angelinalgangelinalgangelinalg approved these changes

@edoakesedoakesAwaiting requested review from edoakesedoakes is a code owner

@zcinzcinAwaiting requested review from zcinzcin is a code owner

@GeneDerGeneDerAwaiting requested review from GeneDerGeneDer is a code owner

@akshay-anyscaleakshay-anyscaleAwaiting requested review from akshay-anyscaleakshay-anyscale is a code owner

Assignees
No one assigned
Labels
goadd ONLY when ready to merge, run all testsstaleThe issue is stale. It will be closed within 7 days unless there are further conversation
Projects
None yet
Milestone
No milestone
Development

Successfully merging this pull request may close these issues.

3 participants
@saihaj@akshay-anyscale@angelinalg

[8]ページ先頭

©2009-2025 Movatter.jp