Post your IT redundancy tales here
Post your IT redundancy tales here
I'm in our daily standup and it's turned into exchanging fucked up sysadmin redundancy tales.
One place I worked lost a machine room. They'd fired people so fast that nobody remembered where the boxes were any more.
I knew, but they didn't ask me. Oh well!
The cycle of IT binge and purge is eternal. Post your tales here.
I have
twothree stories.Company X: Our testbed server room was supported by redundant rooftop AC units, many yards apart. During a storm, a lightning bolt forked (split) One tip.of the bolt hit AC unit one and the other hit AC unit two, killing both cooling units. To make things worse, the server manufacturer did not add a temperature safety shutdown to the units and instead configured them to fan faster the hotter they got. By the time I got there the cable management was warping and melting due to heat.
Company Y: The main datacenter was on tower 2 and the backup datacenter was on tower 1. Most IT staff was present when the planes hit.
EDIT:
Company Z: I started work at a company where they gave me access to a "test" BigIP (unit 3) to use as my own little playground. Prior to my joining the company was run by devs doubling as IT. I deleted the old spaghetti code rules so that I could start from scratch. So, after verifying that no automation was running on my unit (unit 3), I deleted the old rules. Unfortunately the devs/admins forgot to disengage replication on "unit 2" when they gave me "unit 3". So production "unit 2" deleted its rules and told production "unit 1" to do the same. Poof...production down and units offline. I had to drive four hours to the datacenter and code the entire BigIP from scratch and under duress. I quit that job months after starting. Some shops are run so poorly that they end up fostering a toxic environment.
Well, that one dark "redundancy" story...
I don't understand why they had redundancy so physically close.
Whatever affects one has a high risk of affecting the other.
Different regions is a thing for a reason.
company X sounds like the sort of bad shit I remember from a DC my side of the world, which was so frequently broken in various states that occasionally you couldn't even touch the outside doorknobs (the heat would translate from the inside)
Y: oof.
Z: lol, wow. good on ya for leaving. no point sticking around in that kind of disasterzone. it ain't ever gonna get fixed by your efforts.
Company Z:
Oh God. I worked as an NSE with F5 for seven years and had the opportunity to join dozens of "oh God everything's fucked" calls, and autosync was a factor in the majority of them. I'm not sure why you would want half-finished virtual servers pushed onto your active production devices, but damn if people weren't doing that all the blasted time.