Anthony D. Maio

Staff AI Platform Engineer | LLM Infrastructure & Reliability | Agent Systems & Safety

20 years building production systems across fintech, security, and identity. Now applying that same discipline to LLM infrastructure, agent orchestration, and AI safety—treating oversight as a systems engineering problem, not a policy exercise.

Making Minds is my AI research lab and consultancy—delivering production tooling, open-source models, and peer-reviewed research to clients ranging from early-stage startups to mid-sized industrial organizations.

Agentic AI architectures · Multi-agent coordination · AI coherence & memory · Epistemic stress detection

Flagship

Production tools and models—shipped, installable, used.

pulldown

HTTP-first web content for LLM pipelines

Fetches URLs and converts them to clean, level-controlled Markdown. Four detail levels, batch API, bounded crawl, validator caching, and SSRF guard.

pip install "pulldown[mcp]"
MCP ready 4 detail levels SSRF guard

Cartograph

Map the repo before you burn context

CLI-first repo analysis. Rank files, trace dependency hubs, pull task-scoped context, and hand structured artifacts to Claude Code, OpenClaw, or any agent.

CLI first Claude + OpenClaw

Writing

Long-form analysis, technical walkthroughs, and opinion across Substack, Medium, and Hugging Face.

How to Actually Code With Agents

How to Actually Code With Agents

The velocity trap and the practices that survive it

Your Agent Has Amnesia

Your Agent Has Amnesia

Announcing mnemos: biomimetic memory for agents

Medium → all posts

Hugging Face → profile

The Checkpoint Newsletter

Weekly roundup of developments that matter if you build, deploy, or think critically about AI systems.

  • March 30, 2026 — ARC-AGI-3, Google TurboQuant, OpenAI kills Sora
  • March 21, 2026 — OpenAI acquires Promptfoo, Knuth uses Claude, Mistral Small 4
  • March 20, 2026 — Mistral Small 4, GPT-5.4 Mini/Nano, and the week in releases
Subscribe →

Glossary

Safety & Oversight

HDCS
— Heterogeneous Divergence-Convergence Swarm. Ensemble of diverse AI models that cross-check each other.
CMED
— Cross-Model Epistemic Divergence. Test suite for revealing AI verification blind spots.

Architectures

MRA
— Manifold Resonance Architecture. Detects epistemic stress before generating answers.
C2
— Continuity Core. Layered memory giving stateless AI persistent context.