Linode Forum
Linode Community Forums
 FAQFAQ    SearchSearch    MembersMembers      Register Register 
 LoginLogin [ Anonymous ] 
Forum locked  This topic is locked, you cannot edit posts or make further replies.
Author Message
 Post subject:
PostPosted: Wed Jan 09, 2008 1:59 pm 
Offline
Newbie

Joined: Wed Jan 09, 2008 11:50 am
Posts: 2
How long does it take to move drives from 1 server to another?


Top
   
 
 Post subject:
PostPosted: Wed Jan 09, 2008 2:42 pm 
Offline
Linode Staff
User avatar

Joined: Tue Apr 15, 2003 10:32 pm
Posts: 246
Location: NJ, USA
We have been on and off the phone with the data center all morning. They are totally consumed with support tickets from the power outage and keep telling us they are sending a tech to our cage - which obviously isn't happening.

We're doing our best to resolve the situation.

-Tom


Top
   
 
 Post subject:
PostPosted: Wed Jan 09, 2008 2:43 pm 
Offline

Joined: Wed Jan 09, 2008 2:38 pm
Posts: 1
the only reason i recently got a linode was so that there wouldn't be any downtime.

does this happen every now and then, or is this the first time?


Top
   
 
 Post subject:
PostPosted: Wed Jan 09, 2008 3:04 pm 
Offline
Junior Member

Joined: Tue Sep 25, 2007 9:00 am
Posts: 26
Looks like atlanta23 got power cycled at about 12:00MST today, my server was not responding so I took a look at the dashboard and it said the linode was currently powered off, then a minute later it was showing running and the server was back.


Top
   
 
 Post subject:
PostPosted: Wed Jan 09, 2008 3:24 pm 
Offline
Linode Staff
User avatar

Joined: Tue Apr 15, 2003 10:32 pm
Posts: 246
Location: NJ, USA
atlanta24 Linodes should be up at this time.

atlanta20, atlanta21, atlanta22, and atlanta23 were also rebooted while the tech was in our cage.


Top
   
 
 Post subject:
PostPosted: Wed Jan 09, 2008 3:34 pm 
Offline
Linode Staff
User avatar

Joined: Tue Apr 15, 2003 10:32 pm
Posts: 246
Location: NJ, USA
Here is the RFO Atlanta has supplied:

Quote:
At approximately 4:45 am EST the NAP suffered a power outage lasting approximately 10 seconds from Georgia Power.

The generators fired and came online 15 seconds after the initial outge and the load was transferred to generators which ran for 30 minutes while monitoring the incoming power quality from GA Power at which time the load was transferred back to utility.

One of the UPS's that serves part of the facility suffered a battery outage on 2 different redundant strings which caused it to drop the load.
We installed a second redundant string approximately 9 months ago to minimize the possibility of this type of situation. The batteries in the 2 strings are setup in parallel meaning each is capable of carrying the full load for up to 5 minutes.

All it takes is 1 battery in a string to fail for the entire string to fail. this is the same in all ups systems and is the reason we installed the second string from advice from the manufacturer.

The original string batteries are 1.5 years old and were installed new. The second string is 9 months old and was installed new.

A single battery in the second string failed after 3 batteries in the first string failed.

We turned the generators back on to avoid an interruption during troubleshooting and maintenance and MGE sent a tech onsite within an hour to troubleshoot at which time we discovered the battery issue. we replaced the batteries within an hour of diagnosis and brought the system back onlnine and out of maintenance bypass.

The load is currently protected and all batteries have been tested again.

Both sets of batteries have been maintained and tested by MGE direct service every 6 months under a pm plan that they recommended for proper maintenance and operation.

This was extremely rare and unforseen to have something like this happen.

We are purchasing our own battery tester and will set up a monthly pm on the batteries that we will conduct ourselves in addition to the 6 month pm that MGE does on the UPS as well as the batteries. We are also researching a real time battery monitoring system that can predict battery failure.

Batteries are the weakest link in the system and we feel like we properly followed recommended engineering and maintenance on these systems. - however that will not assure 100% as we found out today in a very rare incident.




Extemporaneous events that continued to affect service during the outage:
one of the main metro e switches that runs the links of our backbone went offline during the outage and during that powerinduced reboot we lost connectivity to half our backbones. we have our backbones split in half - with half going out the east and half out the west side of the building taking dirverse paths across redundant switches to the final interconnect points.
the switch was unstable when it came back online due to a gbic that died and for some odd reason rebooted itself several times about every 10 minutes. we replaced the gbic with a spare we keep onsite.

This caused half the backbones to go up and down and placed a large cpu load on the different core routers we have due to bgp table loads going on - this is very cpu intensive and when you have a lot of up and down it can appear that the network is completely down (it is if you are on a link that is flapping) but the fact is that the entire network was not down but was impacted. this settled down when the switch was stabilized.

We split our backbones up over several different redundant backbone routers.

once this switch was brought back online and stabilized the network stabilized as well.

an access switch that serves 16 servers also died and we replaced it with a spare once we found the issue. we keep spares on site for every piece of network gear we have.

an apc that was only 6 months old and is a dual fed apc from 2 different power sources (including the newer ups) failed and did not come back - we replaced it with an onsite spare. it was bizarre to say the least and of course it powered one of our 3 main dns clusters so we lost dns capacity for an hour.

Most of the issues currently going on are related to server hardware that did not do well in a power reboot situation or need a fsck. we are actively working on them and will not rest until all is well.

Many customers in the facility do have A and B feeds from our power. we offer this through different ups systems / different power panels and different transformers. Some very early customers that purchased a and b feeds when we only had one ups system at the NAP are on the same ups and as such lost power. those customers will be offered a free move on their b feed to the newer ups to increase their power diversity - they simply need to open a ticket.

What are we doing on power in the future?

We have another UPS from MGE on order as of 4 weeks ago that is due to deliver in mid Feb that will increase the diveristy of the power in the facility. We plan on having 2 battery strings on it as well.

We are in the process of installing another set of 5 cummins generators and another 3000 amp transformer which will further diversify our generator and transformer plant - this will be completed in mid february - construction of this is going on currently we took delivery of the switchgear and generators 2 weeks ago. 4 ups/ will be moved to the new power feeed and g enarators to diversify the power source to the UPS . this will give us 100% redundancy on the A / B feeds at that point.

We installed a redundant b feed to our metro e gear and 2 dual fed apcs at our TELX cabinet after TELX suffered a complete UPS failure at 56 marietta 4 months ago. This turned out to be good because there was another complete failure of the B ups 4 weeks ago - but we were not affected since we had a redundant feed from them. the outage affected all customers on the second floor. we would have more than 50% of our network had we not been on dual fed apcs and dual power feeds at the building which would have been bad.

we are increasing the battery pm schedule to monthly from biannual.

we are researching a battery monitoring system for the strings.

we will be taking a fuel delivery this week to restock our main fuel supply

we are examining in depth on of our 4 core metro switch abnormalities this morning and if we do not find a rfo from the manufacturer will be examining replacing it or upgrading to a different more robust solution - which has been in our long term plan but may get moved up.

we will be doing another power examination of our core swithcing routers ( currently 6 of them all with dual fed power ) and our core metro e switches (currently 4 of them) to make sure that our power feeeds are truly redundant and no legacy circuits are there to affect them.

we will be examining our on site spares inventory to make sure we are still at correct levels since we used some items this morning.


We appologize for the outage caused by the failure of hte primary and backup batteries and will continue to provide the best service at an excellent price.
The MGE tech that has all the major accounts in Atlanta including coke and several others told us that this was a very freak occurance with negligible odds of happening and in his opinion we have done everything right on our maintenance and pm and redundancy of the batteries and he would have done the same thing and that there was really nothing he would have recommended different at that point.


Top
   
 
 Post subject:
PostPosted: Wed Jan 09, 2008 3:54 pm 
Offline
Senior Newbie

Joined: Tue Mar 23, 2004 6:15 pm
Posts: 15
Chris; are the Linodes dual supplied from the A and B supplies?


Top
   
 
 Post subject: linode response times?
PostPosted: Wed Jan 09, 2008 4:22 pm 
Offline
Newbie

Joined: Wed Jan 09, 2008 4:13 pm
Posts: 2
Thanks for posting the RFO from the Datacentre, interesting reading and certainly a bad morning for them.

Given that the outage itself was one of those unlucky failures it seems to me the main question is why did it take so long for the fault to be acknowledged and responded to?

The first widespread problems were noticed around 04:41 (UTC-5) but the first acknowledgement that the fault was being worked on wasn't until over 3 hours later at 07:59 when this thread started. That's well below the standard we've come to expect from Linode.

Obviously the router problems detailed in the RFO and the general panic/understaffing in the DC delayed restoration, which is somewhat understandable, but I haven't heard any explanation yet for why a short "problem acknowledged, looking into it" couldn't have been posted early than 3 hours into the outage.

Were linode staff aware of the outage at 04:41?


Top
   
 
 Post subject: Times
PostPosted: Wed Jan 09, 2008 4:24 pm 
Offline
Newbie

Joined: Wed Jan 09, 2008 4:13 pm
Posts: 2
hah, just noticed my times differ a few minutes from what the DC RFO reports, so i'll defer to them...


Top
   
 
 Post subject:
PostPosted: Thu Jan 10, 2008 4:18 am 
Offline
Newbie

Joined: Thu Jan 10, 2008 4:15 am
Posts: 2
Everybody else down again?

Been several hours for me, or doesn't anyone in the US care that the rest of the world might be online while they are sleeping....


Top
   
 
 Post subject:
PostPosted: Thu Jan 10, 2008 6:09 am 
Offline
Linode Staff
User avatar

Joined: Sat Jun 21, 2003 2:21 pm
Posts: 160
Location: Absecon, NJ
My Linode in Atlanta is working fine. Have you checked you Linode in Lish?


Top
   
 
 Post subject:
PostPosted: Thu Jan 10, 2008 6:19 am 
Offline
Newbie

Joined: Thu Jan 10, 2008 4:15 am
Posts: 2
Logged in with Lish and a reboot order that I made a while ago went through - almost exactly after I loged in, but it is back up now.

Thanks


Top
   
 
 Post subject:
PostPosted: Mon Jan 14, 2008 12:34 pm 
Offline
Linode Staff
User avatar

Joined: Tue Apr 15, 2003 10:32 pm
Posts: 246
Location: NJ, USA
http://www.linode.com/forums/viewtopic. ... 3845#13845


Top
   
 
Display posts from previous:  Sort by  
Forum locked  This topic is locked, you cannot edit posts or make further replies.


Who is online

Users browsing this forum: No registered users and 2 guests


You cannot post new topics in this forum
You cannot reply to topics in this forum
You cannot edit your posts in this forum
You cannot delete your posts in this forum

Search for:
Jump to:  
RSS

Powered by phpBB® Forum Software © phpBB Group