Movatterモバイル変換


[0]ホーム

URL:


Skip to content

Navigation Menu

Sign in
Appearance settings

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Sign up
Appearance settings

Commit74ce6ae

Browse files
committed
add vllm docs
1 parentd017cd6 commit74ce6ae

File tree

2 files changed

+3
-1
lines changed

2 files changed

+3
-1
lines changed

‎pgml-extension/src/bindings/vllm/inference.rs‎

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -4,6 +4,8 @@ use serde_json::{json, Value};
44

55
usesuper::LLM;
66

7+
/// Cache a single model per client process. vLLM does not allow multiple, simultaneous models to be loaded.
8+
/// See GH issue, https://github.com/vllm-project/vllm/issues/565
79
staticMODEL:Mutex<Option<LLM>> =Mutex::new(None);
810

911
pubfnvllm_inference(task:&Value,inputs:&[&str]) ->PyResult<Value>{

‎pgml-extension/src/bindings/vllm/mod.rs‎

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,4 @@
1-
//! Rust bindings to the Python package`vllm`.
1+
//! Rust bindings to the Python package[vLLM](https://vllm.readthedocs.io/en/latest/)
22
33
mod inference;
44
mod llm;

0 commit comments

Comments
 (0)

[8]ページ先頭

©2009-2025 Movatter.jp