𝗟𝗟𝗠 𝗚𝘂𝗮𝗿𝗱, 𝗮 𝗾𝘂𝗶𝗰𝗸 𝘄𝗮𝘆 𝘁𝗼 𝗿𝗲𝗺𝗼𝘃𝗲 𝗣𝗜𝗜 𝗮𝗻𝗱 𝗽𝗼𝘁𝗲𝗻𝘁𝗶𝗮𝗹𝗹𝘆 𝗵𝗮𝗿𝗺𝗳𝘂𝗹 𝗽𝗿𝗼𝗺𝗽𝘁𝘀 𝗪𝗵𝗮𝘁 𝗽𝗿𝗼𝗯𝗹𝗲𝗺 𝗱𝗼𝗲𝘀 𝗶𝘁 𝘀𝗼𝗹𝘃𝗲 ? • Blocks prompt injection before the request hits the model. • Scrubs output for PII, secrets, toxicity, bias. • Gives structured JSON verdicts (𝘢𝘭𝘭𝘰𝘸, 𝘴𝘢𝘯𝘪𝘵𝘪𝘻𝘦, 𝘣𝘭𝘰𝘤𝘬). 𝗪𝗵𝘆 𝗜 𝗹𝗶𝗸𝗲 𝗶𝘁 😇 1️⃣ Uses a stack of small, task‑specific BERT classifiers → runs on commodity CPUs, or GPUs if you have them. 2️⃣ Benchmarks show ~200 ms average latency on an AWS m5.xlarge CPU and single‑digit milliseconds on a G5 GPU with ONNX. 3️⃣ Specialized prompt‑injection model scores 𝟬.𝟵𝟳 𝗙𝟭 vs 0.81 for a compact LLM baseline -> understand better perf yet fast inferences ). 4️⃣ Open‑source and pip‑installable; the maintainers push frequent version bumps and new scanners, so we can just `pip install -U` in CI . ---- 𝗦𝗼𝘂𝗿𝗰𝗲𝘀 • LLM Guard docs – Index page (protectai.github.io (https://lnkd.in/eBXvWGxq)) • LLM Guard docs – Prompt Injection benchmarks (protectai.github.io (https://lnkd.in/e-j8eR67)) • Protect AI blog – 𝘚𝘱𝘦𝘤𝘪𝘢𝘭𝘪𝘻𝘦𝘥 𝘔𝘰𝘥𝘦𝘭𝘴 𝘉𝘦𝘢𝘵 𝘚𝘪𝘯𝘨𝘭𝘦 𝘓𝘓𝘔𝘴 𝘧𝘰𝘳 𝘈𝘐 𝘚𝘦𝘤𝘶𝘳𝘪𝘵𝘺 (protectai.com (https://lnkd.in/eSBuVMVc)) • GitHub – protectai/llm‑guard (github.com (https://lnkd.in/eFKtYVwR)) Have you already wrapped your LLMs with guardrails? What gaps are you still seeing/framework do you recommend? #LLMSecurity #GenAI #AIEngineering #MLOps #OpenSource
Love this! Your post is my first look at the project, and it immediately brought me back to a past role where I worked with similar HAP filters. Input filtering is always the smooth sailing part. However, having a streaming output scanner that is good at flagging unwanted outputs is challenging. Excited to see how this package pushes the capabilities further. Are there any real-world use cases you're most proud of so far?
How is multilingual performance?
It makes a lot of sense to use "small" classifiers. Text classification will never die.