home bbs files messages ]

Forums before death by AOL, social media and spammers... "We can't have nice things"

   rec.autos.driving      Automobile discussion (general)      162,178 messages   

[   << oldest   |   < older   |   list   |   newer >   |   newest >>   ]

   Message 161,685 of 162,178   
   David Fritz to All   
   Ghost in Musk's machines: Software bugs'   
   12 Oct 17 20:02:32   
   
   XPost: alt.politics.obama, alt.autos.toyota, sac.politics   
   XPost: alt.society.liberalism   
   From: david.fritz@vzw.com   
      
   Last year, a dark historical landmark was reached. Joshua Brown became the   
   first confirmed person to die in a crash where the car was, at least in   
   part, driving itself. On a Florida highway, his Tesla Model S ploughed   
   underneath a white truck trailer that was straddling the road, devastating   
   the top half of the car.   
      
   Brown’s crash is well known. But more mundane bugs are finding their way   
   into increasingly software-dependent, semi-autonomous cars. Software   
   problems accounted for nearly 15 per cent of US car recalls in 2015, up   
   from less than five per cent in 2011, according to the most recent report   
   from financial advisors Stout Risius Ross.   
      
   Last year, to name a few examples, Toyota recalled around 320,000 cars   
   after they found “improper programming” could cause airbags and seatbelt   
   pretensioners to activate unbidden. Ford had to recall 23,000 cars because   
   software problems in their electronic windows meant they had excessive   
   “closing force.”   
      
   Despite the latest wave of excitement about artificial intelligence, the   
   fear among some of those in the industry is that bugs could prove a   
   serious hurdle to mass adoption – not least because of the weird,   
   unexpected nature of the accidents they can cause.   
      
   Philip Koopman, an associate professor at Carnegie Mellon University and   
   an expert on autonomous vehicle safety, told The Reg: “I look at the   
   errors, and almost always say: ‘Wow, that should not have happened.’ And   
   the most likely explanation is that they did not follow a safety   
   standard.”   
      
   The “continuous stream” of defects in the car industry signals “underlying   
   problems: they just don’t want to spend the time and effort to get it   
   right,” he argues.   
      
   Car manufacturers contacted by The Reg were unwilling to talk.   
      
   Significantly, many developing autonomous vehicles are hiring developers   
   from Silicon Valley whose backgrounds are in general purpose software –   
   software that, of course, crashes with reasonable frequency. People are   
   not hiring from among the ranks of the airline safety industry.   
      
   “Knowing how to code is not knowing how to be safe,” Koopman says.   
      
   Allegations of poor code go back years. Koopman was an expert witness for   
   plaintiffs in a 2013 court case in Oklahoma that looked into whether   
   computer problems had caused a Toyota Camry to accelerate uncontrollably   
   and crash, killing a passenger in 2007.   
      
   He and another investigator found Toyota’s electronic throttle control   
   system was “just awful.” An 18-month investigation found numerous problems   
   in the software [PDF], including the potential for stack overflow and no   
   protection against bit flips – where ambient radiation in the outside   
   environment can switch a bit. The report concluded Toyota’s code was   
   “spaghetti.”   
      
   The jury decided the electronics had been at fault and awarded $3m in   
   compensation. Toyota stands by the safety of its throttle system, a   
   company spokesman said, pointing out that an earlier official   
   investigation, partly carried out by NASA, did not find any faults.   
      
   Yoav Hollander, founder of Foretellix, a company looking to develop new   
   ways to find bugs in engineered systems, has for a number of years been   
   attending conferences about verifying the safety of autonomous cars (and   
   other autonomous systems). He was not impressed by progress initially,   
   although thinks the situation is now “improving.”   
      
   One of the issues, Hollander says, is that companies are overly focused on   
   preventing what he calls “expected bugs” – where engineers anticipate a   
   problem. This might include making sure that the car cameras can correctly   
   identify a pedestrian wearing a black coat at night.   
      
   But then there are unexpected bugs – problems that no one has thought of   
   or situations that have been overlooked. A car designed in the US but   
   driven in the UK could set off on the right hand side of the road simply   
   because its default location is the US – all because a developer forgot to   
   include an instruction to check location after a memory reset.   
      
   These kinds of “autonomous vehicle only” bugs – mistakes that no human   
   would ever make – will be big news, Hollander says. “People will say:   
   ‘Hey, I’m at the mercy of the vehicle’.”   
      
   The Joshua Brown crash – driving at full speed into a clearly visible   
   trailer – is arguably one such example as it “would never happen to a   
   human being,” Hollander says.   
      
   After the Florida accident, Tesla reportedly wasn’t immediately sure why   
   its autopilot system hadn’t braked. They probed the possibility that the   
   system deliberately ignored the trailer to avoid braking when approaching   
   objects like overhead bridges. This was an idea supported by an   
   investigation [PDF] by the National Highway Traffic Safety Administration.   
      
   A Reg request for clarification from Tesla went unanswered.   
      
   https://www.theregister.co.uk/2017/10/09/bugs_in_autonomous_vehicles/   
      
   --- SoupGate-Win32 v1.05   
    * Origin: you cannot sedate... all the things you hate (1:229/2)   

[   << oldest   |   < older   |   list   |   newer >   |   newest >>   ]


(c) 1994,  bbs@darkrealms.ca