I’ve always been drawn to systems—not just how they work when they work, but how and why they fail. My academic work focused on time, knowledge, structure, and digital transformation. My operational work builds those things into living processes: automation frameworks, escalation systems, pattern detection, governance. Somewhere in between sits AI.
I’m not interested in AI as hype or promise. I’m interested in where intelligence lives in a system—what gets designed in, what gets forgotten, and what the consequences are when complexity starts to outpace control. In practice, that means looking at real incidents, identifying structural patterns, and building interventions that are simple enough to scale but intelligent enough to adapt.
This space is where I’ll think aloud about that work: the role of AI in enterprise operations, the structural logic behind resilience, the patterns that emerge in failure data, the ways organisations forget what they once knew, and how all of this connects to cognition, memory, and design.
This isn’t theory. It’s not evangelism either. It’s applied systems thinking—with enough abstraction to be reusable, and enough specificity to be real.