Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We鈥檒l occasionally send you account related emails.

Already on GitHub? Sign in to your account

Add benchmarks and evals for jailbreaks #51

Open
steventkrawczyk opened this issue Aug 2, 2023 · 0 comments
Open

Add benchmarks and evals for jailbreaks #51

steventkrawczyk opened this issue Aug 2, 2023 · 0 comments
Labels
enhancement New feature or request help wanted Extra attention is needed

Comments

@steventkrawczyk
Copy link
Contributor

馃殌 The feature

As we add benchmarks, it would be good to cover common jailbreak scenarios. We should incorporate these benchmarks, and have auto-evals that check responses to see if they are "broken"

Motivation, pitch

https://github.com/llm-attacks/llm-attacks

Alternatives

No response

Additional context

No response

@steventkrawczyk steventkrawczyk added enhancement New feature or request help wanted Extra attention is needed labels Aug 2, 2023
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
enhancement New feature or request help wanted Extra attention is needed
Projects
None yet
Development

No branches or pull requests

1 participant