Discussion about this post

User's avatar
Neural Foundry's avatar

Excellent curation on systemic thinking. The parallel between your CrowdStrike analysis and the AWS US-East-1 discussion is striking - both highlight how legacy systems with high uptime requirements become untestable in practice. The US-East-1 observation that 'it's hard to test properly – issues often show up only when it goes down' mirrors the high-trust, high-speed environment you described at CrowdStrike. The Vernon piece on metrics resonates strongly. 'If you've arrived at human error for a production incident, you've probably only gone halfway' is exactly right. The use error vs human error distinction is powerful, though I agree it's not as intuitive in software contexts. I'd argue it applies broadly: when a developer ships a bug, was the testig environment adequate? Were code review practices contextual? Did the deployment system align with the cognitive load? The Cat Hicks research on software metrics is game-changing - 'improving software delivery velocity requires systems-level thinking rather than individual-focused interventions' should be posted in every engineering manager's office. The neurodivergent psychological safety table is fantastic. These adaptations (clear expectations, written communication preferences, flexible time structures) actually benefit everyone, not just neurodiverse individuals. It's a great example of how designing for the margins improves the entire system.

Expand full comment

No posts

Ready for more?