Skip to content

AZ1688/SecurityofAI

Folders and files

NameName
Last commit message
Last commit date

Latest commit

 

History

9 Commits
 
 
 
 
 
 

Repository files navigation

Security of AI

Security Controls

Risk Managment Framework

Red Teaming for AI/ML

Risk areas & tools

  • Rebuff: A self-hardening prompt injection detector.
  • LLMFuzzer: A fuzzing framework for LLMs.
  • LLM Guard: A security toolkit for LLM interactions ,open source LLM Firewall
  • LLM Gateway: Helps ensure that data being fed into and output from the LLM tool is free of proprietary company data.
  • Data leakage, Inadequate sandboxing
  • LLM vulnerability scanner - garak https://github.com/leondz/garak
  • Arthur Shield: The First Firewall for LLMs
  • Lakera is an AI security company. They have specific products to protect against Prompt injection: Lakera Guard | Unlock LLMs for Production | Lakera – Protecting AI teams that disrupt the world.
  • Generative Discriminator Guided Sequence Generation (GeDi) or Plug and Play Language Models (PPLM)

AI Security for Enterprise

  • Enterprise architectures for LLMs (a16z)
  • Corporate GenAI policies gartner
  • Gandalf chatbot security game counters privacy fireballs
  • Should you buy or build: When it comes to large language models, should you build or buy? | TechCrunch
  • Companies blocking ChatGPT and other publicly trained chatbots: Employees are banned from using ChatGPT at these companies | Fortune
  • Google thinks open source LLMs will be as good as OpenAI soon: Google’s Leaked Document Reveals Open Source Threat: A New Era in Language Models | BigTechWire
  • Triveto language model whitepaper: https://www.truveta.com/wp-content/uploads/2023/04/Truveta-Language-Model.pdf

Playgrounds

  • Lope Kinz: GPT4 Free, PentestGPT, . ChatGPT client
  • Gelei Deng: Pentest GPT
  • GPT 3 Vulnerability Scanner
  • Orange Software for Customer ML models
  • AutocodePro
  • smol.ai - Promopt engineering
  • GreyDGL/PentestGPT
  • dorkgpt
  • GPT Integration - Nucliei, BurpGPT, HackGPT, CloudGPT,
  • Aviary Explorer: A way to compare results from open source LLMs: Aviary Explorer (anyscale.com)
  • A playground for prompt injection. Basically tricking LLMs in revealing secrets https://gandalf.lakera.ai/
  • Holistic evaluation of LLMs (HELM) from Stanford: https://crfm.stanford.edu/helm/latest/

About

No description, website, or topics provided.

Resources

License

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published