One of our ESXi 4.1 servers lost a drive in it's Raid-1 boot array about a week ago. No biggie, bought new drive, inserted new drive today, array began rebuild. Other drive fails during rebuild. "F&%K!" I screamed.
Good news is the VMs are still running as is ESXi' as far as I can tell. I know though as soon ESXi needs to read/write to a log file or something it's all going to come crashing down.
So the question is, since I don't have vmotion and I don't have another suitable host anyway, how long do I have before ESXi dies? Do I have enough time to get a new server (10 days)? Should I put a really powerful working up quick (overnight)? Or is best to shut the VMs down now and deal with outage?
A few bits of information. The VMs are stored completely on a shared NFS array, not on the hosts' physical storage. The host has dual power supplies each tied to separate battery backups each on their own circuit. I don't think power will bring them down. The host is starved for RAM (thanks Exchange) and was planned to be replaced before year's end. Only thing running on the host is ESXi itself. It was loaded to the boot array. Everything else is stored on the NFS systems.