Forums before death by AOL, social media and spammers... "We can't have nice things"
|    sci.med.psychobiology    |    Dialog and news in psychiatry and psycho    |    4,734 messages    |
[   << oldest   |   < older   |   list   |   newer >   |   newest >>   ]
|    Message 3,718 of 4,734    |
|    =?UTF-8?B?4oqZ77y/4oqZ?= to All    |
|    Study delivers bleak verdict on validity    |
|    28 Aug 15 09:18:56    |
      From: bulldog23x@gmail.com              Psychology              Study delivers bleak verdict on validity of psychology experiment results                     Of 100 studies published in top-ranking journals in 2008, 75% of social       psychology experiments and half of cognitive studies failed the replication       test              Psychology experiments are failing the replication test – for good reason               There are many reasons why an experiment might fail to replicate, but more       than this, the study has highlighted some issues with academic publishing and       modern science.        There are many reasons why an experiment might fail to replicate, but more       than this, the study has highlighted some issues with academic publishing and       modern science.                      Photograph: Pere Sanz / Alamy/Alamy       Ian Sample Science editor       @iansample                     Thursday 27 August 2015 14.00 EDT Last modified on Friday 28 August 2015 08.50       EDT                     A major investigation into scores of claims made in psychology research       journals has delivered a bleak verdict on the state of the science.              An international team of experts repeated 100 experiments published in top       psychology journals and found that they could reproduce only 36% of original       findings.              The study, which saw 270 scientists repeat experiments on five continents, was       launched by psychologists in the US in response to rising concerns over the       reliability of psychology research.              The first imperative: Science that isn’t transparent isn’t science       “There is no doubt that I would have loved for the effects to be more       reproducible,” said Brian Nosek, a professor of psychology who led the study       at the University of Virgina. “I am disappointed, in the sense that I think       we can do better.”              “The key caution that an average reader should take away is any one study is       not going to be the last word,” he added. “Science is a process of       uncertainty reduction, and no one study is almost ever a definitive result on       its own.”              All of the experiments the scientists repeated appeared in top ranking       journals in 2008 and fell into two broad categories, namely cognitive and       social psychology. Cognitive psychology is concerned with basic operations of       the mind, and studies tend to        look at areas such as perception, attention and memory. Social psychology       looks at more social issues, such as self esteem, identity, prejudice and how       people interact.              In the investigation, a whopping 75% of the social psychology experiments were       not replicated, meaning that the originally reported findings vanished when       other scientists repeated the experiments. Half of the cognitive psychology       studies failed the same        test. Details are published in the journal Science.              Even when scientists could replicate original findings, the sizes of the       effects they found were on average half as big as reported first time around.       That could be due to scientists leaving out data that undermined their       hypotheses, and by journals        accepting only the strongest claims for publication.              Despite the grim findings, Nosek said the results presented an opportunity to       understand and fix the problem. “Scepticism is a core part of science and we       need to embrace it. If the evidence is tentative, you should be sceptical of       your evidence. We        should be our own worst critics,” he told the Guardian. One initiative now       underway calls for psychologists to submit their research questions and       proposed methods to probe them for review before they start their experiments.              John Ioannidis, professor of health research and policy at Stanford       University, said the study was impressive and that its results had been       eagerly awaited by the scientific community. “Sadly, the picture it paints -       a 64% failure rate even among        papers published in the best journals in the field - is not very nice about       the current status of psychological science in general, and for fields like       social psychology it is just devastating,” he said.              But he urged people to focus on the positives. The results, he hopes, will       improve research practices in psychology and across the sciences more       generally, where similar problems of reproducibility have been found before.       In 2005, Ioannidis published a        seminal study that explained why most published research findings are false.              Marcus Munafo, a co-author on the study and professor of psychology at Bristol       University, said: “I think it’s a problem across the board, because       wherever people have looked, they have found similar issues.” In 2013, he       published a report with        Ioannidis that found serious statistical weaknesses were common in       neuroscience studies.              Scandals prompt return to peer review and reproducible experiments       Nosek’s study is unlikely to boost morale among psychologists, but the       findings simply reflect how science works. In trying to understand how the       world works, scientists must ask important questions and take risks in finding       ways to try and answer them.        Missteps are inevitable if scientists are not being complacent. As Alan Kraut       at the Association for Psychological Science puts it: “The only finding that       will replicate 100% of the time is likely to be trite, boring and probably       already known: yes,        dead people can never be taught to read.”              There are many reasons why a study might not replicate. Scientists could use a       slightly different method second time around, or perform the experiment under       different conditions. They might fail to find the original effect by chance.       None of these would        negate the original finding. Another possibility is that the original result       was a false positive.              Among the experiments that stood up was one that found people are equally       adept at recognising pride in faces from different cultures. Another backed up       a finding that revealed the brain regions activated when people were given       fair offers in a financial        game. One study that failed replication claimed that encouraging people to       believe there was no such thing as free will made them cheat more.              Munafo said that the problem of poor reproducibility is exacerbated by the way       modern science works. “If I want to get promoted or get a grant, I need to       be writing lots of papers. But writing lots of papers and doing lots of small       experiments isn’t        the way to get one really robust right answer,” he said. “What it takes to       be a successful academic is not necessarily that well aligned with what it       takes to be a good scientist.”              More news Topics       Psychology Peer review and scientific publishing Research              [continued in next message]              --- SoupGate-Win32 v1.05        * Origin: you cannot sedate... all the things you hate (1:229/2)    |
[   << oldest   |   < older   |   list   |   newer >   |   newest >>   ]
(c) 1994, bbs@darkrealms.ca