Skip to content

Features Overview

SafeLLM provides a waterfall security pipeline (L0-L2) designed for maximum performance and security.

Feature Comparison: OSS vs. Enterprise (Paid)

Section titled “Feature Comparison: OSS vs. Enterprise (Paid)”
CategoryFeatureOSSEnterprise (Paid)Description
L0: PerformanceSmart CacheSemantic caching to reduce LLM costs and latency (<0.1ms).
Distributed CoalescerCross-pod request deduplication to prevent redundant LLM calls.
L1: Static GuardKeyword GuardUltra-fast (O(n)) deterministic guard for banned phrases and jailbreaks.
Regex PIIFast detection of basic PII (Emails, Phones, Credit Cards).
L1.5: AI GuardAI PII (GLiNER)Context-aware detection of 25+ PII types.
L2: Neural GuardPrompt InjectionNeural network (ONNX) analysis for sophisticated jailbreaks.
Data ProtectionDLP Output ScanAudit (log-only)Block/Anonymize/LogPreventing data leakage in LLM responses.
DLP Streaming ModeAudit (async)Zero-latency PII detection in output stream.
IntegrationMCP Server (stdio)JSON-RPC MCP tools for policy checks and guarded tool orchestration.
ObservabilityAudit LoggingLoki/S3Tamper-proof, persistent audit trails for compliance.
Admin DashboardReal-time security posture and rule management.

Every request is processed through layers, allowing for “short-circuit” rejection:

  1. L0 (Cache): If a similar prompt was processed recently, return cached response.
  2. L1 (Keywords): Instantly block known bad phrases or patterns.
  3. L1.5 (PII Shield): Detect and mask sensitive information.
  4. L2 (Neural Guard): Use AI to detect complex semantic attacks.

SafeLLM is built for high-throughput enterprise environments:

  • Targets below apply to Enterprise (Paid) deployments with AI layers enabled.
  • Accuracy: >95% (ONNX + GLiNER)
  • E2E Latency: <10ms sidecar overhead
  • Throughput: 1000+ RPS (Scalable with APISIX)
  • False Positives: <0.3%