BIOS update bricked server....How was your monday?

drnick5

Active Member
Reaction score
122
Hey Folks,

I went to a clients site to get their server online after an outage. (power failure for a few hours, so I went in to make sure the server came up smooth as this is a pretty new client). A few weeks earlier I was at the site for a problem with the domain controller VM freezing at a black screen. All I had to do was reboot the VM and it came up fine. After looking at the logs and doing some searching I found out that VM freeze was specific to the type of motherboard they had (intel) and the exact CPU they had. There was a known firmware bug that caused VM's to occasionally freeze with this combination of hardware and specific to 2008 R2 VM's (perfect storm!) Intel released a firmware update which included a microcode update to the CPU to resolve this.

So, since I was already down there for an outage, and we were scheduled to come down the following day to install the firmware update, I decided to do the BIOS update while I was there. I shut down the VM's and the host server and rebooted the host off the Intel Deployment assistant DVD. The dvd connected to the internet, found the BIOS update and downloaded it. after installing it said it needed to reboot, this is where things went south. After reboot I was met with a black screen on the host server (no splash screen or anything else indicating its posted).

I tried to move the jumper to the "recover bios" option, this got some text on the screen showing BIOS information, and a blinking cursor, but nothing else. I tried downloading the recovery BIOS form intel, but couldn't get it to load (I just saw this cursor, it never seemed to attempt to read from the USB stick with the recovery BIOS).

I also tried clearing the CMOS and even tried pulling the battery for 10 minutes to let it reset itself. All were met with the same black screen. After phone calls to my vender, it was determined that the board is bricked and a new one is coming over night.

Anyone have something similar happen before? I've been in this game for about 12 years, and have heard plenty of these horror stories but never had it happen to me personally.

As a side question, how do you guys handle firmware updates to servers? Typically I never apply a firmware update unless its to fix a specific problem, which happened to be the case here. I figured the intel deployment assistant was the easiest way to apply this (and it was pretty easy... except for the bricking the server part). Any suggestions on how to avoid something like this in the future?
 
In general, not just servers, but even client machines. Unless there is some issue that is causing trouble, I do not do bios updates for the reason you had. You had a working system and now you don't. Not saying it's your fault, sometimes you have mondays.
 
In general, not just servers, but even client machines. Unless there is some issue that is causing trouble, I do not do bios updates for the reason you had. You had a working system and now you don't. Not saying it's your fault, sometimes you have mondays.

Unless I misunderstood, the OP's customer's DC was unreliable. I'd call that "trouble" .
 
In general, not just servers, but even client machines. Unless there is some issue that is causing trouble, I do not do bios updates for the reason you had. You had a working system and now you don't. Not saying it's your fault, sometimes you have mondays.

Sigh

jiFfM.jpg
 
Yes and if you read the op you'd see that the BIOS update was addressing his exact problem. The need for the update isn't the issue. His method of applying it may be.
 
it never seemed to attempt to read from the USB stick with the recovery BIOS).

Have you tried a different thumb drive? I've seen BIOSes be picky about what types of thumb drives they will boot from. I've not seen it so much lately but that was a common issue back in the Server 2003/XP days when a 512mb thumb drive was an insane amount of storage.
 
I was just making a statement. I realized the op was having a specific issue. Why do you think I said it was not his fault?
 
The one time I saw a BIOS update go south in probably thousands of them over nearly 20 years, it was on a cloner system with some internet based BIOS flash utility. By "internet based"...I mean it's some tool that runs from within the BIOS or a bootable CD and runs the flash tool from RAM. Never felt good about that method...skips too many steps which can avoid corruption or CRC checks.

Have done many tier-1 servers and always use the vendors downloadable program which is saved to HDD, extracted...and run from there.
 
Just a quick update. We FINALLY got everything up and running for them.

I'm still not entirely sure what went wrong with the flash process originally. As I said, I NEVER perform these updates unless its to resolve a specific problem we are having, which was the case here as the DC virtual machine would lock up randomly. This update was set to resolve that issue.

I ended up having to replace the motherboard, but heres the kicker. After driving 2 hours round trip to pick up a board from my distributor. I replaced it but the server still wouldn't boot. Ended up finding out my distributor gave me a board with the original BIOS, however the Xeon chip we were using is V2 version that required an update to be compatible with the chip. So I had to drive back up to my distributor, have them use a V1 CPU they had instock to boot, perform the flash, and then swap it back with my V2 cpu.

Lets hope the rest of the week goes smoothly...
 
See! This computer stuff is easy! :D

I have had a few BIOS update issues in the past but nothing that I wasn't able to recover from; no hard bricks. Lucky I suppose.
 
Back
Top