Own the redaction layer.
Wordcab Redact is private PII, PHI, and PCI detection that ships inside your product. Built with Knowledgator on the open GLiNER model family. Embeddable, fine-tunable, and deployable in the same boundary your team already operates.
Built in partnership with Knowledgator, whose open-source extraction models have crossed 5M+ downloads on Hugging Face. The GLiNER-PII checkpoints we co-authored are open and Apache 2.0; benchmark them on your data before you talk to us. Wordcab Redact is the production build: vertical fine-tunes, hardware-tuned variants, a named engineer on your account, an SLA, and indemnified commercial licensing.
From a strong baseline to your production data.
The open GLiNER-PII reference posts 98% F1 character-level on Knowledgator's published EHR benchmark, ahead of every commercial DLP measured against it. That's a strong starting baseline, and the right place to begin an evaluation.
Every production deployment then meets text that no published benchmark covers in full: domain-specific entity types, tokenization quirks in spoken-form numbers, lowercase blood types in clinical notes, MRNs that share token shape with billing codes, and the entity classes unique to your vertical. Wordcab Redact carries the baseline into your environment, fine-tuned on representative data, packaged for your hardware, and supported.
A strong starting baseline
Open Apache 2.0 GLiNER-PII posts 98% character-level F1 on EHR PII, ahead of Azure Language (50.2%), Presidio (22.3%), and the leading generic LLMs on the same benchmark. Wordcab Redact carries that baseline into vertical fine-tunes for healthcare, finance, legal, and contact-center deployments.
~20× faster than LLM redaction
The pipeline finishes before an LLM finishes its first token. Inline in a voice agent's turn budget, or millions of documents in an overnight batch. Same model, same control plane.
Inside the perimeter
Runs on the same Helm chart as Voice and Think. VPC, on-prem, airgap. No call-home in the critical path, no telemetry to Wordcab on redacted content.
Runs on the hardware you have
~20k tokens/sec on a single L4 with the open reference. Wordcab Redact adds quantized CPU and distilled GPU variants tuned for the boxes you actually operate; edge deployments are a configuration, not a port.
This is the redaction layer.
Entity detection
Find PII, PHI, PCI, credentials, and custom entities in any text.
The GLiNER detection model is zero-shot. Name an entity in the request and the model finds it. Common types ship with vetted defaults. Tenant-specific identifiers like INTERNAL_TICKET_ID or PATIENT_ROOM_NUMBER work without a retrain.
Replacement modes
Detect-only, placeholder replacement, pseudonymization, or character mask.
Pick the mode that fits the workflow. Pseudonymize keeps referential integrity across utterances so downstream analytics stay useful. Mask preserves length and casing for systems that depend on it. Detect returns spans only, so your application controls substitution.
Vertical fine-tunes
Healthcare, financial services, legal, contact center, or your own variant.
The open Apache 2.0 reference is a strong starting baseline. Wordcab Redact carries it forward into the verticals we ship for. A healthcare or finance fine-tune is trained on representative data from that domain, plus the tokenization edge cases every production deployment eventually meets. Lowercase blood types, spoken-form account numbers, MRNs that share token shape with billing codes, and the quiet failures auditors find first.
Same control plane
Helm chart, observability, RBAC, and audit log shared with the rest of the Wordcab stack.
Redact installs as part of the existing Wordcab Helm chart. Prometheus, OpenTelemetry, audit log, and RBAC are the same surfaces your team already operates Voice and Think against. One stack, one upgrade path, one set of dashboards.
Compared against the systems teams actually evaluate.
Character-level F1 on Electronic Health Records, N=376. Source: Knowledgator's open benchmark on the GLiNER-PII reference checkpoint. Vertical fine-tunes are trained on representative data from each domain and measured on yours during Pilot, so the deployment number is the one you ship on.
| System | Type | EHR PII F1 | Where it fits |
|---|---|---|---|
| GLiNER-PII (Knowledgator + Wordcab) | Specialized NER | 98.0% | Open, Apache 2.0. Benchmark on your data, run it yourself. |
| Generic LLM (GPT-class) | Generative | 84.5% | Expensive at scale, slow inline, leaves the perimeter. |
| Open LLM (Llama-class) | Generative | 77.8% | Better than nothing for PHI; not a redaction engine on its own. |
| Azure Language | Hosted classifier | 50.2% | Outside your boundary. Pricing scales with volume. |
| Microsoft Presidio | Pattern + NER | 22.3% | Useful as a baseline; not production-grade on clinical text. |
Numbers above are from Knowledgator's published character-level micro-average benchmark on EHR PII. They establish a strong baseline against generic comparators. Every Pilot includes a redaction eval against your representative text so the numbers you ship on are the ones measured on your data, not on a public test set.
Entity coverage you can point at in audit.
The detection model is zero-shot. Anything in the request is valid. The list below is the vetted set Wordcab benchmarks and supports per vertical. Custom types are a request parameter, not a retraining cycle.
General PII
PERSON· names, aliasesEMAIL,PHONE,URLADDRESS,IP_ADDRESSDATE_OF_BIRTH,AGE,GENDERUSERNAME,PASSWORDORGANIZATION,JOB_TITLE
Healthcare · PHI
MEDICAL_RECORD,HEALTH_PLAN_IDBLOOD_TYPE(incl. lowercase)MEDICATION,DIAGNOSISPROCEDURE,PROVIDER_NAMEFACILITY_NAME,DEVICE_IDBIOMETRIC_ID,LICENSE_NUMBER
Finance · PCI
CARD_NUMBER,CARD_EXPIRYCARD_CVV,ROUTING_NUMBERIBAN,SWIFT,BICACCOUNT_NUMBER,CUSTOMER_IDSSN,TAX_IDCRYPTO_WALLET
Legal, contact-center, and government variants add their own type sets (case identifiers, agent IDs, classification markings). Custom entities are a free-text label in the request body, useful for tenant-specific identifiers like internal ticket IDs or patient room numbers.
Built for the teams making the embed-vs-buy call.
The buyers we ship to are technical product leads at data-security, AI-security, and archive companies, plus the engineering teams in regulated verticals embedding redaction into their own platform.
Detection accuracy is your product.
License or embed the model your DSPM or DLP competes on. Same engine, your packaging.
Frequently asked questions
What's the difference between the open GLiNER-PII model and Wordcab Redact?
Can it handle the edge cases generic DLP misses?
How does it integrate with the rest of the Wordcab stack?
/v1/redact endpoint, same auth, same Webhook Portal for async jobs. Inline mode runs inside the voice agent turn budget; batch mode handles archive backfills overnight on the same cluster.Can we run it fully on-prem or airgapped?
What hardware do we need?
How do we get started?
Ship redaction your audit team can sign off on.
If your platform sends voice, transcripts, or LLM outputs through healthcare, finance, or any regulated workflow, Wordcab Redact is the layer that takes a strong baseline and tunes it to the data your auditors will actually see.
Talk to an Engineer
We usually respond within one business day.