AI Security: Overview of Critical Must Haves

Mark Aklian is Chiri’s Chief Information Security Officer and has vast senior security experience including Point72, Bank of America, and many more.

Having curated and battle-tested this checklist for over a year, it’s grown from a simple review aid into a board-ready framework I use with our clients, product teams, vendors, and red-team exercises. It’s been hardened against real incidents, mapped to emerging regs, and trimmed to what actually moves risk. Here’s the zero-BS version πŸ‘‡

AI Security, Privacy & Risk β€” What I Actually Check:

🧠 Architecture transparency
β€’ Where do models run (vendor/self-hosted)?
β€’ Which FMs, fine-tunes, agents?
β€’ Any third-party AI APIs, embeddings, vector DBs?

πŸ” Data flow & retention
β€’ Are prompts/outputs logged? For how long and by whom?
β€’ Used for training/evals? Opt-out controls?

πŸ“š RAG hygiene
β€’ Sources, chunking/metadata, pre-index redaction
β€’ Tenant-scoped indices, per-doc ACLs
β€’ Vector deletion workflows

πŸ” Privacy controls
β€’ Legal basis (GDPR/CCPA), purpose limitation
β€’ DSRs for prompts/vectors/model snapshots
β€’ Handling of sensitive data (ATS/HR/health/finance)

πŸ›‘οΈ Guardrails against abuse
β€’ Prompt-injection defenses, retrieval allow-lists
β€’ Output filtering/citations, jailbreak/misuse protections
β€’ Abstain on low confidence

🧰 Tool/agent safety
β€’ Sandboxing & controlled egress
β€’ Scoped credentials, auditable tokens
β€’ Function-calling with least privilege

πŸ“‹ Governance & change management
β€’ Model cards & documentation
β€’ Prompt/model version control, approvals, rollback
β€’ Audit trails on safety rule changes

πŸ§ͺ Testing & red-teaming
β€’ Pre-prod & continuous evals (accuracy, leakage, injection resilience, bias)
β€’ Independent red-teaming with concrete attack playbooks

πŸ”Ž Explainability & human-in-the-loop
β€’ Source-grounded answers, rationale visibility
β€’ Human review gates for high-impact decisions (hiring, finance, legal)

πŸ”‘ Keys, secrets & spend
β€’ Central LLM gateway, vaulting & rotation
β€’ No secrets in prompts
β€’ Rate limits/quotas, cost anomaly detection

🧭 Regulatory alignment
β€’ Map to NIST AI RMF, ISO/IEC 42001, and (where relevant) EU AI Act
β€’ Sector overlays: EEOC / SEC / FINRA / HIPAA

🚨 AI incident response
β€’ Playbooks: injection, RAG exfil, model/safety regressions
β€’ Forensics: retain prompts/outputs, retrieval snapshots, versions

If your AI implementation (or third party that utilizes AI to enhance their services) can’t answer these with evidence, you’re likely carrying an inordinate amount of risk.

Reach out for guidance. We’re here to help.

Leave a Comment

Your email address will not be published. Required fields are marked *

Scroll to Top