Step 4: Review Detected PII with Confidence Scores - FOI Redaction Walkthrough
Examine AI-detected PII and understand confidence scores for each detection
Great! You've deployed the demo
Now let's walk through what you just deployed and see it in action.
Start WalkthroughChoose your next step
Generate Evidence Pack
Create your business case documentation with what you've learned.
Generate Evidence PackWalkthrough progress
Step 4 of 5 • 3 minutes
Review Detected PII with Confidence Scores
Examine the PII the AI found and understand confidence scores to know when human review is needed.
Screenshot updating - please check back soon
Screenshot updating - please check back soon
Expected outcome
- All PII in the document has been detected
- Each detection shows a confidence score
- You understand high/medium/low confidence thresholds
- You can identify which detections need human review
What you should see
After processing completes, you'll see a results section with detected PII. The exact results depend on which sample document you uploaded:
Understanding confidence scores
Each PII detection includes a confidence score representing how certain the AI is that it found real personal information. This helps you decide which detections to trust and which need human review.
- Meaning
- Very high confidence - PII detected with near certainty. AI is 95%+ sure this is personal information.
- Recommended action
- Auto-redact - Safe for production redaction without human review
- Examples
- Email addresses (99%), phone numbers with standard formats (98%), full names in clear context (97%)
- Meaning
- Likely PII but some uncertainty. AI is 80-94% confident - usually correct but not certain.
- Recommended action
- Flag for review - Quick human check before redaction (5-10 seconds per item)
- Examples
- Partial addresses (no postcode), unusual name formats, ambiguous phone numbers
- Meaning
- Possible false positive. AI is less than 80% confident - may or may not be actual PII.
- Recommended action
- Manual review required - Human judgment needed to confirm or reject
- Examples
- Organization names flagged as person names, street names without numbers, generic email addresses
Production tip: In a real FOI system, you'd set a confidence threshold (e.g., 95%) to automatically redact high-confidence detections and flag lower scores for review. This balances automation with accuracy.
Verify accuracy
Compare the AI detections against the original PDF document:
-
Open the sample PDF you downloaded earlier
Find it in your downloads folder and open it alongside the redaction interface
-
Check each detected PII item
For each detection in the results, locate it in the PDF. Does the text match exactly? Is it actually PII that should be redacted?
-
Look for missed PII (false negatives)
Scan the PDF for any personal information the AI didn't detect. This is rare with our sample documents (97-99% recall rate) but can happen with unusual formats.
-
Check for false positives
Are any detections NOT actually PII? For example, organization names flagged as person names. Low confidence scores often indicate these edge cases.
Key observations
Context-aware detection
AI understands context. "Manchester" in an address is flagged as PII, but "Manchester City Council" is correctly identified as an organization, not personal information.
Multi-page coverage
AI processes the entire document, finding PII regardless of where it appears - headers, footers, tables, or body text across all pages.
Format independence
Phone numbers work with any format: 07712 345678, 0121-456-7890, +44 20 7946 0958. No need to standardize formats before processing.
UK-specific patterns
Correctly identifies UK postcodes (M1 4BT, LS1 5AB), National Insurance numbers (AB 12 34 56 C), and UK address formats including flats and building names.
Calculate your time savings
Based on what you just saw, estimate your council's FOI redaction savings:
Troubleshooting
Some PII not detected (false negatives)
If the AI missed some personal information:
- Check if PII is in an unusual format (all caps, no spaces, abbreviations)
- Verify text is machine-readable (scanned documents may need better OCR)
- Look for handwritten text (Comprehend works best on typed text)
- Check if PII is in tables or complex layouts (may reduce detection accuracy)
- Unusual UK formats may not be recognized (very old postcode styles, etc.)
Our sample documents should detect 97-99% of PII. In production, always perform final human review before publishing FOI responses to catch edge cases.
Non-PII flagged as personal information (false positives)
If the AI flagged text that isn't actually PII:
- Check confidence score - false positives often have lower scores (<90%)
- Common false positives: organization names, building names, generic email addresses (info@council.gov.uk)
- Review the context - is this actually personal or organizational information?
- In production, you'd manually un-mark false positives during review step
- Consider adding common false positives to an allowlist (e.g., council building names)
Setting a higher confidence threshold (95%+) reduces false positives but may increase manual review workload.
Confidence scores seem inaccurate
If confidence scores don't match your expectations:
- Remember: scores reflect AI's certainty, not ground truth accuracy
- Low-quality scans or unusual fonts reduce confidence even if detection is correct
- Uncommon name spellings or address formats may have lower scores despite being accurate
- Confidence is based on statistical patterns from millions of training documents
- In production, track actual accuracy vs confidence scores to calibrate your threshold