Movatterモバイル変換


[0]ホーム

URL:


Skip to content

Navigation Menu

Sign in
Appearance settings

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Sign up
Appearance settings

Add a llama-cpp Model? #1801

Open
Open
Assignees
Kludex
@joy-void-joy

Description

@joy-void-joy

Description

I know it is already possible to useOllama servers with pydantic using its OpenAI interface, however, there are a few reasons for why I would be interested in a directllama-cpp binding:

  • Tool-use and huggingface compatibility: When using Ollama on huggface models that support tool use (e.g.GLM-4-32B), Ollama says the model "does not support tool use", while llama-cpp works correctly
  • Standalone file: llama-cpp does not need to boot the server separately, which matters for ease of use and portability as everything can be included in a single python file
  • Better configurability/nativity? Unsure about this one, but it seems like llama-cpp-python might be faster and have overall more optimization builtin as compared to the ollama server

This is why I would like a llama-cpp Model to pydantic-ai.

Would there be interest for a PR regarding this? I already have started working on it for a personal project, so it is only a matter of packaging/adding tests etc...

References

No response

Metadata

Metadata

Assignees

Labels

No labels
No labels

Type

No type

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions


    [8]ページ先頭

    ©2009-2025 Movatter.jp