home bbs files messages ]

Forums before death by AOL, social media and spammers... "We can't have nice things"

   comp.ai.philosophy      Perhaps we should ask SkyNet about this      59,235 messages   

[   << oldest   |   < older   |   list   |   newer >   |   newest >>   ]

   Message 58,954 of 59,235   
   Leroy N. Soetoro to All   
   Google AI Overviews put people at risk o   
   04 Jan 26 01:18:10   
   
   XPost: comp.internet.services.google, sac.politics, alt.fan.rush-limbaugh   
   XPost: talk.politics.guns, talk.politics.medicine   
   From: leroysoetoro@americans-first.com   
      
   People are being put at risk of harm by false and misleading health   
   information in Google’s artificial intelligence summaries, a Guardian   
   investigation has found.   
      
   The company has said its AI Overviews, which use generative AI to   
   provide snapshots of essential information about a topic or question,   
   are “helpful” and “reliable”.   
      
   But some of the summaries, which appear at the top of search results,   
   served up inaccurate health information and put people at risk of harm.   
      
   In one case that experts described as “really dangerous”, Google wrongly   
   advised people with pancreatic cancer to avoid high-fat foods. Experts   
   said this was the exact opposite of what should be recommended, and may   
   increase the risk of patients dying from the disease.   
      
   In another “alarming” example, the company provided bogus information   
   about crucial liver function tests, which could leave people with   
   serious liver disease wrongly thinking they are healthy.   
      
   Google searches for answers about women’s cancer tests also provided   
   “completely wrong” information, which experts said could result in   
   people dismissing genuine symptoms.   
      
   A Google spokesperson said that many of the health examples shared with   
   them were “incomplete screenshots”, but from what they could assess they   
   linked “to well-known, reputable sources and recommend seeking out   
   expert advice”.   
      
   The Guardian investigation comes amid growing concern that AI data can   
   confuse consumers who may assume that it is reliable. In November last   
   year, a study found AI chatbots across a range of platforms gave   
   inaccurate financial advice, while similar concerns have been raised   
   about summaries of news stories.   
      
   Sophie Randall, director of the Patient Information Forum, which   
   promotes evidence-based health information to patients, the public and   
   healthcare professionals, said the examples showed “Google’s AI   
   Overviews can put inaccurate health information at the top of online   
   searches, presenting a risk to people’s health”.   
      
   Stephanie Parker, the director of digital at Marie Curie, an end-of-life   
   charity, said: “People turn to the internet in moments of worry and   
   crisis. If the information they receive is inaccurate or out of context,   
   it can seriously harm their health.”   
      
   The Guardian uncovered several cases of inaccurate health information in   
   Google’s AI Overviews after a number of health groups, charities and   
   professionals raised concerns.   
      
   Anna Jewell, the director of support, research and influencing at   
   Pancreatic Cancer UK, said advising patients to avoid high-fat foods was   
   “completely incorrect”. Doing so “could be really dangerous and   
   jeopardise a person’s chances of being well enough to have treatment”,   
   she added.   
      
   Jewell said: “The Google AI response suggests that people with   
   pancreatic cancer avoid high-fat foods and provides a list of examples.   
   However, if someone followed what the search result told them then they   
   might not take in enough calories, struggle to put on weight, and be   
   unable to tolerate either chemotherapy or potentially life-saving   
   surgery.”   
      
   Typing “what is the normal range for liver blood tests” also served up   
   misleading information, with masses of numbers, little context and no   
   accounting for nationality, sex, ethnicity or age of patients.   
      
   Pamela Healy, the chief executive of the British Liver Trust, said the   
   AI summaries were alarming. “Many people with liver disease show no   
   symptoms until the late stages, which is why it’s so important that they   
   get tested. But what the Google AI Overviews say is ‘normal’ can vary   
   drastically from what is actually considered normal.   
      
   “It’s dangerous because it means some people with serious liver disease   
   may think they have a normal result then not bother to attend a   
   follow-up healthcare meeting.”   
      
   A search for “vaginal cancer symptoms and tests” listed a pap test as a   
   test for vaginal cancer, which is incorrect.   
      
   Athena Lamnisos, the chief executive of the Eve Appeal cancer charity,   
   said: “It isn’t a test to detect cancer, and certainly isn’t a test to   
   detect vaginal cancer – this is completely wrong information. Getting   
   wrong information like this could potentially lead to someone not   
   getting vaginal cancer symptoms checked because they had a clear result   
   at a recent cervical screening.   
      
   “We were also worried by the fact that the AI summary changed when we   
   did the exact same search, coming up with a different response each time   
   that pulled from different sources. That means that people are getting a   
   different answer depending on when they search, and that’s not good   
   enough.”   
      
   Lamnisos said she was extremely concerned. “Some of the results we’ve   
   seen are really worrying and can potentially put women in danger,” she   
   said.   
      
   The Guardian also found Google AI Overviews delivered misleading results   
   for searches about mental health conditions. “This is a huge concern for   
   us as a charity,” said Stephen Buckley, the head of information at Mind.   
      
   Some of the AI summaries for conditions such as psychosis and eating   
   disorders offered “very dangerous advice” and were “incorrect, harmful   
   or could lead people to avoid seeking help”, Buckley said.   
      
   Some also missed out important context or nuance, he added. “They may   
   suggest accessing information from sites that are inappropriate … and we   
   know that when AI summarises information, it can often reflect existing   
   biases, stereotypes or stigmatising narratives.”   
      
   Google said the vast majority of its AI Overviews were factual and   
   helpful, and it continuously made quality improvements. The accuracy   
   rate of AI Overviews was on a par with its other search features like   
   featured snippets, which had existed for more than a decade, it added.   
      
   The company also said that when AI Overviews misinterpreted web content   
   or missed context, it would take action as appropriate under its   
   policies.   
      
   A Google spokesperson said: “We invest significantly in the quality of   
   AI Overviews, particularly for topics like health, and the vast majority   
   provide accurate information.”   
      
      
   --   
   November 5, 2024 - Congratulations President Donald Trump.  We look   
   forward to America being great again.   
      
   We live in a time where intelligent people are being silenced so that   
   stupid people won't be offended.   
      
   Every day is an IQ test. Some pass, some, not so much.   
      
   Thank you for cleaning up the disasters of the 2008-2017, 2020-2024 Obama   
   / Biden / Harris fiascos, President Trump.   
      
   Under Barack Obama's leadership, the United States of America became the   
   The World According To Garp.  Obama sold out heterosexuals for Hollywood   
   queer liberal democrat donors.   
      
   --- SoupGate-Win32 v1.05   
    * Origin: you cannot sedate... all the things you hate (1:229/2)   

[   << oldest   |   < older   |   list   |   newer >   |   newest >>   ]


(c) 1994,  bbs@darkrealms.ca