▲ | swiftcoder 3 days ago | |||||||||||||||||||||||||||||||
The scenarios in the article are all about mission-critical disaster recovery - we don't even trust the majority of our human colleagues with those scenarios! AI won't make inroads there without humans in the loop, until AI is 100% trustworthy. | ||||||||||||||||||||||||||||||||
▲ | tptacek 3 days ago | parent | next [-] | |||||||||||||||||||||||||||||||
Right, so: having an agent go drop index segments from a search cluster to resolve a volume utilization problem is a bad idea, rather than just suggesting "these old index segments are using up 70% of the storage on this volume and your emergency search cluster outage would be resolved if you dropped them, here's how you'd do that". But there are plenty of active investigative steps you'd want to take in generating hypotheses for an outage. Weakly's piece strongly suggests AI tools not take these actions, but rather suggest them to operators. This is a waste of time, and time is the currency of incident resolution. | ||||||||||||||||||||||||||||||||
▲ | datadrivenangel 3 days ago | parent | prev | next [-] | |||||||||||||||||||||||||||||||
And the author assumes that these humans are going to be very rigorous, which is good for SRE teams, but even then not consistently. | ||||||||||||||||||||||||||||||||
| ||||||||||||||||||||||||||||||||
▲ | topaz0 3 days ago | parent | prev [-] | |||||||||||||||||||||||||||||||
Or it will, and disaster will ensue. |