home bbs files messages ]

Forums before death by AOL, social media and spammers... "We can't have nice things"

   alt.airports      Just one step above a dirty bus station      8,692 messages   

[   << oldest   |   < older   |   list   |   newer >   |   newest >>   ]

   Message 8,482 of 8,692   
   Deport Indian IT Scabs to All   
   H1B Outsourced IT probably hurt Delta Ai   
   10 Apr 17 05:31:55   
   
   XPost: misc.immigration.usa, comp.sys.ibm.pc.hardware.storage, rec.travel.air   
   XPost: sac.politics   
   From: deport.indian.it.scabs@hpe.com   
      
   Delta Airlines last night suffered a major power outage at its   
   data center in Atlanta that led to a systemwide shutdown of its   
   computer network, stranding airliners and canceling flights all   
   over the world. You already know that. What you may not know,   
   however, is the likely role in the crisis of IT outsourcing and   
   offshoring.   
      
   Whatever the cause of the Delta Airlines power outage, data   
   center recovery pretty much follows the same routine I used 30   
   years ago when I had a PDP-8 minicomputer living in my basement   
   and heating my house. First you crawl around and find the power   
   shut-off and turn off the power. I know there is no power but   
   the point is that when power returns we don’t want a surge   
   damaging equipment. Then you crawl around some more and turn off   
   power to every individual device. Wait in the dark for power to   
   be restored, either by the utility or a generator. Once power   
   has been restored turn the main power switch back on then crawl   
   back to every device, turning them back on in a specific order   
   that follows your emergency plan. You do have an emergency plan,   
   right? In the case of the PDP-8, toggle in the code to launch   
   the boot PROM loader (yes, I have done this is complete   
   darkness). Reboot all equipment and check for errors. Once   
   everything is working well together then reconnect data to the   
   outside world.   
      
   Notice the role in this process of crawling around in the dark?   
   How do you do that when your network technicians are mainly in   
   India?   
      
   Yes, every data center of a certain size has bodies on-site, but   
   few have enough such bodies to do all the crawling required for   
   a building with thousands of individual devices.   
      
   Modern data centers can switch to UPS power very quickly,   
   usually less than 1/30th of a second. They will run on battery   
   power for a few minutes while the generators come on line.   
   Smarter data centers drop power to the HVAC system until the   
   generators are on line and handling the full load. Smarter IT   
   departments also monitor the quality of the electric power   
   coming into the data center. They can see the effect of bad   
   weather on the power grid. When there are storms approaching the   
   area they proactively switch to generator power, which even if   
   it isn’t needed is a good test. Better to fire up the   
   generators, have them go into phase with the utility then take   
   over the load gracefully rather than all at once. It is doubtful   
   that happened last night at Delta.   
      
   Delta Airlines was an IBM outsourcing customer, it may still be   
   today, I don’t know. The company hasn’t returned my call.   
      
   Loss of power in a data center usually triggers a disaster   
   recovery plan. When that happens you have two basic choices:   
   switch to your backup systems somewhere else or fix the outage   
   and recover your primary systems. The problem with going to   
   backup systems is those backups usually do not have capacity for   
   100 percent of the workload so only the most critical functions   
   are moved. Then once everything is fixed you have to move your   
   workload back to your production systems. That is often high   
   risk, a major pain, and takes a lot of effort. So in a   
   traditional disaster recovery setup, the preference will always   
   be to recover the primary services.   
      
   Anything less than a 100 percent service backup isn’t disaster   
   recovery, it is disaster coping.   
      
   Now if the IT support team is thousands of miles away, offshore,   
   the process for restarting hundreds -- perhaps thousands -- of   
   systems can be slow and painful. If you lose the data link   
   between your support team and the data center due to that same   
   power outage your support team can do nothing until the data   
   link is fixed.   
      
   In the old days a smart IT department would put 50 people in the   
   data center with pre-printed documentation on how to recover the   
   systems. They’d go into a massive divide and conquer effort to   
   restart everything. One person can work on several systems at   
   the same time. While debugging the systems the IT team can spot   
   and diagnose network and SAN (Storage Area Network) problems and   
   work shoulder to shoulder with the network team until everything   
   is fixed. Today due to long distance data connections, offshore   
   IT folks can only work on a few systems at a time. All   
   discussions with other teams is done via conference calls.   
   Language issues make the challenges much harder.   
      
   A further problem with this scenario is that network,   
   application, and storage support can be on completely separate   
   contracts with vendors who may not play well together. Some   
   vendors simply refuse to cooperate because cooperation isn’t a   
   contract term.   
      
   Now I don’t know if any of this applies to Delta Airlines   
   because it is too busy to answer a question from little old me,   
   but I’m sure the answers will appear in coming days. Hopefully   
   other IT departments will learn from Delta’s experience.   
      
   https://betanews.com/2016/08/09/outsourced-it-probably-hurt-   
   delta-airlines-when-its-power-went-out/   
      
   --- SoupGate-Win32 v1.05   
    * Origin: you cannot sedate... all the things you hate (1:229/2)   

[   << oldest   |   < older   |   list   |   newer >   |   newest >>   ]


(c) 1994,  bbs@darkrealms.ca