Director-AI

The only LLM guardrail that halts hallucinations mid-stream. NLI + RAG dual scoring against your own knowledge base. Enterprise-ready. Open-source.

Why Director-AI

Token-Level Streaming Halt

Not post-hoc review. The safety kernel monitors coherence token-by-token and severs output the moment it degrades. No other guardrail does this.

🔒

Dual-Entropy Scoring

NLI contradiction detection (DeBERTa) + RAG fact-checking against your own knowledge base. Both must pass. One model catching what the other misses.

🏗

Your Data, Your Rules

Ingest PDFs, directories, or any text into a ChromaDB-backed knowledge base. The scorer checks LLM output against your ground truth, not a generic model.

Head-to-Head Comparison

FeatureDirector-AINeMo GuardrailsGuardrails AISelfCheckGPT
Streaming halt (token-level)YesNoNoNo
NLI contradiction detectionYesNoNoYes
Custom knowledge baseYes (ChromaDB)YesYesNo
LangChain integrationBuilt-inBuilt-inBuilt-inManual
LlamaIndex integrationBuilt-inNoBuilt-inNo
Policy engine (YAML)YesColangRAILNo
Multi-tenant isolationYesNoNoNo
Audit trailJSONLNoNoNo
Self-hosted / air-gappedYesYesYesYes
LicenseAGPL v3 + CommercialApache 2.0Apache 2.0MIT

Works in 4 Lines

LangChain
from director_ai.integrations.langchain import DirectorAIGuard

guard = DirectorAIGuard(
    facts={"refund": "Refunds within 30 days."},
    threshold=0.6,
)
chain = my_llm | guard  # halts hallucinations
LlamaIndex
from director_ai.integrations.llamaindex import DirectorAIPostprocessor

pp = DirectorAIPostprocessor(
    facts={"pricing": "Enterprise: $99/mo."},
    threshold=0.6,
)
engine = index.as_query_engine(
    node_postprocessors=[pp]
)  # filters hallucinated nodes

Benchmarks

Evaluated on LLM-AggreFact (29,320 samples across 11 datasets)

ToolBal. AccParamsLatencyStreaming
Bespoke-MiniCheck-7B77.4%7B~100 ms (GPU)No
Director-AI (FactCG)75.8%0.4B575 ms (CPU)Yes
MiniCheck-Flan-T5-L75.0%0.8B~120 msNo
MiniCheck-DeBERTa-L72.6%0.4B~120 msNo
HHEM-2.1-Open71.8%~0.4B~200 msNo

Honest assessment: 75.8% balanced accuracy ranks 4th on the LLM-AggreFact leaderboard — within 1.6pp of the top-scoring 7B model while using only 0.4B params. Director-AI's unique value is the system: NLI + KB facts + streaming token-level halt. No competitor offers real-time streaming gating. CPU latency (~575 ms with source chunking) drops to ~50-80 ms on GPU.

Pricing

AGPL v3 open-source free forever. Commercial license removes copyleft.

Hobbyist
$9/mo
$90/yr (save ~17%)
  • Students & side projects
  • 1 local deployment
  • Community support
  • Delayed updates
  • Non-production use
Indie
$49/mo
$490/yr (save ~17%)
  • Solo devs, bootstrapped teams
  • <$2M ARR
  • 1 production deployment
  • Email support
  • 12 months updates
Enterprise
Custom
Starts ~$799/mo
  • SLA (99.9% uptime)
  • On-prem / air-gapped
  • SOC2 / HIPAA ready
  • Dedicated engineer
  • Custom NLI fine-tunes

Perpetual license: $1,299 one-time · First 50 licensees: 50% off first year

Stop hallucinations before your users see them

Install in 30 seconds. Score your first response in 60.

pip install director-ai[nli,vector]