Movatterモバイル変換


[0]ホーム

URL:


Skip to content

Navigation Menu

Sign in
Appearance settings

Search code, repositories, users, issues, pull requests...

Provide feedback

We read every piece of feedback, and take your input very seriously.

Saved searches

Use saved searches to filter your results more quickly

Sign up
Appearance settings

Test your prompts, agents, and RAGs. AI Red teaming, pentesting, and vulnerability scanning for LLMs. Compare performance of GPT, Claude, Gemini, Llama, and more. Simple declarative configs with command line and CI/CD integration.

License

NotificationsYou must be signed in to change notification settings

promptfoo/promptfoo

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 

Repository files navigation

npmnpmGitHub Workflow StatusMIT licenseDiscord

promptfoo is a developer-friendly local tool for testing LLM applications. Stop the trial-and-error approach - start shipping secure, reliable AI apps.

Website ·Getting Started ·Red Teaming ·Documentation ·Discord

Quick Start

# Install and initialize projectnpx promptfoo@latest init# Run your first evaluationnpx promptfooeval

SeeGetting Started (evals) orRed Teaming (vulnerability scanning) for more.

What can you do with Promptfoo?

  • Test your prompts and models withautomated evaluations
  • Secure your LLM apps withred teaming and vulnerability scanning
  • Compare models side-by-side (OpenAI, Anthropic, Azure, Bedrock, Ollama, andmore)
  • Automate checks inCI/CD
  • Share results with your team

Here's what it looks like in action:

prompt evaluation matrix - web viewer

It works on the command line too:

prompt evaluation matrix - command line

It also can generatesecurity vulnerability reports:

gen ai red team

Why Promptfoo?

  • 🚀Developer-first: Fast, with features like live reload and caching
  • 🔒Private: LLM evals run 100% locally - your prompts never leave your machine
  • 🔧Flexible: Works with any LLM API or programming language
  • 💪Battle-tested: Powers LLM apps serving 10M+ users in production
  • 📊Data-driven: Make decisions based on metrics, not gut feel
  • 🤝Open source: MIT licensed, with an active community

Learn More

Contributing

We welcome contributions! Check out ourcontributing guide to get started.

Join ourDiscord community for help and discussion.

About

Test your prompts, agents, and RAGs. AI Red teaming, pentesting, and vulnerability scanning for LLMs. Compare performance of GPT, Claude, Gemini, Llama, and more. Simple declarative configs with command line and CI/CD integration.

Topics

Resources

License

Contributing

Security policy

Stars

Watchers

Forks

Sponsor this project

 

Packages

 
 
 

[8]ページ先頭

©2009-2025 Movatter.jp