Skip to content
forked from Azure/PyRIT

The Python Risk Identification Tool for generative AI (PyRIT) is an open access automation framework to empower security professionals and machine learning engineers to proactively find risks in their generative AI systems.

License

Notifications You must be signed in to change notification settings

rdheekonda/PyRIT

This branch is 3 commits ahead of, 210 commits behind Azure/PyRIT:main.

Folders and files

NameName
Last commit message
Last commit date

Latest commit

author
rdheekonda
Oct 4, 2024
425c4c4 · Oct 4, 2024
Aug 13, 2024
May 16, 2024
Oct 1, 2024
Oct 4, 2024
Oct 3, 2024
Oct 2, 2024
Sep 23, 2024
Jul 25, 2024
Aug 14, 2024
Sep 23, 2024
Dec 12, 2023
Jan 26, 2024
Feb 24, 2024
Jan 29, 2024
Jan 30, 2024
Aug 8, 2024
Jan 29, 2024
Aug 8, 2024
Feb 6, 2024
Jun 28, 2024
Sep 23, 2024

Repository files navigation

Python Risk Identification Tool for generative AI (PyRIT)

The Python Risk Identification Tool for generative AI (PyRIT) is an open access automation framework to empower security professionals and ML engineers to red team foundation models and their applications.

Introduction

PyRIT is a library developed by the AI Red Team for researchers and engineers to help them assess the robustness of their LLM endpoints against different harm categories such as fabrication/ungrounded content (e.g., hallucination), misuse (e.g., bias), and prohibited content (e.g., harassment).

PyRIT automates AI Red Teaming tasks to allow operators to focus on more complicated and time-consuming tasks and can also identify security harms such as misuse (e.g., malware generation, jailbreaking), and privacy harms (e.g., identity theft).​

The goal is to allow researchers to have a baseline of how well their model and entire inference pipeline is doing against different harm categories and to be able to compare that baseline to future iterations of their model. This allows them to have empirical data on how well their model is doing today, and detect any degradation of performance based on future improvements.

Additionally, this tool allows researchers to iterate and improve their mitigations against different harms. For example, at Microsoft we are using this tool to iterate on different versions of a product (and its metaprompt) so that we can more effectively protect against prompt injection attacks.

PyRIT architecture

Where can I learn more?

Microsoft Learn has a dedicated page on AI Red Teaming.

Check out our docs for more information on how to install PyRIT, our How to Guide, and more, as well as our demos.

Trademarks

This project may contain trademarks or logos for projects, products, or services. Authorized use of Microsoft trademarks or logos is subject to and must follow Microsoft's Trademark & Brand Guidelines. Use of Microsoft trademarks or logos in modified versions of this project must not cause confusion or imply Microsoft sponsorship. Any use of third-party trademarks or logos are subject to those third-party's policies.

About

The Python Risk Identification Tool for generative AI (PyRIT) is an open access automation framework to empower security professionals and machine learning engineers to proactively find risks in their generative AI systems.

Resources

License

Code of conduct

Security policy

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 100.0%