LLM Guard is a comprehensive tool designed to fortify the security of Large Language Models (LLMs) by offering sanitization, detection of harmful language, prevention of data leakage, and resistance against prompt injection attacks. It provides a range of features, including prompt scanners, anonymization, and output scanners, to ensure safe and secure interactions with LLMs. LLM Guard is an open-source solution that is easy to integrate and deploy in production environments, and is designed for easy integration and deployment in production environments. It is available for download and installation via pip, and is compatible with Python 3.9 or higher. The tool is constantly being improved and updated, and the community is encouraged to contribute to the package through bug reports, feature requests, and code contributions. For more information, please refer to the documentation and contribution guidelines.
FEATURES
SIMILAR TOOLS
Zania is an AI-driven platform that automates security and compliance tasks using autonomous agents for security inquiries, compliance assessments, and privacy regulation adherence.
Security platform that provides protection, monitoring and governance for enterprise generative AI applications and LLMs against various threats including prompt injection and data poisoning.
Unbound is a security platform that enables enterprises to control and protect the use of generative AI applications by employees while safeguarding sensitive information.
AI Access Security is a tool for managing and securing generative AI application usage in organizations, offering visibility, control, and protection features.
Mindgard is a continuous automated red teaming platform that enables security teams to identify and remediate vulnerabilities in AI systems, including generative AI and large language models.
Wald.ai is an AI security platform that provides enterprise access to multiple AI assistants while ensuring data protection and regulatory compliance.
VIDOC is an AI-powered security tool that automates code review, detects and fixes vulnerabilities, and monitors external security, ensuring the integrity of both human-written and AI-generated code in software development pipelines.
CalypsoAI is a platform that provides centralized security, observability, and control for deploying and scaling large language models and generative AI across an enterprise.
DIANNA is an AI-powered cybersecurity companion from Deep Instinct that analyzes and explains unknown threats, offering malware analysis and translating code intent into natural language.
PINNED

Checkmarx SCA
A software composition analysis tool that identifies vulnerabilities, malicious code, and license risks in open source dependencies throughout the software development lifecycle.

Orca Security
A cloud-native application protection platform that provides agentless security monitoring, vulnerability management, and compliance capabilities across multi-cloud environments.

DryRun
A GitHub application that performs automated security code reviews by analyzing contextual security aspects of code changes during pull requests.