[HN Gopher] How many medical studies are faked or flawed?
       ___________________________________________________________________
        
       How many medical studies are faked or flawed?
        
       Author : PaulHoule
       Score  : 190 points
       Date   : 2023-09-19 16:31 UTC (6 hours ago)
        
 (HTM) web link (web.archive.org)
 (TXT) w3m dump (web.archive.org)
        
       | ponderings wrote:
       | Isn't this an issue with the reviews rather than the publication
       | attempts? Wouldn't the normal (if there is such a thing) approach
       | between people be to credit a reviewer with the quality of their
       | reviews or lack thereof? Basically, who signed off on it?
        
         | tcmart14 wrote:
         | That is my understanding. Papers can be wrong for different
         | reasons, some nefarious, some honest mistakes. The problem I
         | wouldn't say is with the paper itself. It more of people citing
         | it with a lack of peer review or lack of replication. And also
         | peer review being like a stereotypical "LGTM" approval on a
         | pull request.
        
       | PeterStuer wrote:
       | Tldr, most.
        
       | bjourne wrote:
       | We all know it is super-easy to cheat in science, but not much
       | can be done about it. Short of requiring replication for every
       | result published which isn't feasible if the study was costly to
       | produce. And the problem isn't confined to medicine either. How
       | many studies in hpc of the type "we setup this benchmark and our
       | novel algorithm/implementation won!" aren't also faked or flawed?
        
         | gmd63 wrote:
         | Plenty can be done about it. We can start by increasing the
         | standards for experimental proof past words on a piece of
         | paper.
         | 
         | Stream or record your experiment. If a 12 year old has the
         | ability to stream his gaming life on twitch, scientists should
         | be able to record what they are doing in more detail.
         | 
         | You could start a new journal with higher level of prestige
         | that only publishes experiments that adhere to more modern
         | methods of proof.
        
           | gustavus wrote:
           | This seems overly simplistic, I mean part of the reason so
           | many studies are so difficult is because they can happen over
           | the period of months or years. This isn't like your High
           | School chemistry experiment. These studies can be massive
           | longitudinal studies that take years to conclude and work on.
           | 
           | If the answers were easy I'm sure someone would've
           | implemented it already, it turns out those kind of things are
           | hard.
           | 
           | I'd recommend checking out the following article.
           | https://slatestarcodex.com/2014/04/28/the-control-group-
           | is-o...
        
             | gmd63 wrote:
             | I'm interested more in why it specifically would be hard to
             | just show people what you are doing (or did) instead of
             | telling them a story at the end of months or years.
             | 
             | Critical details and oversights can be lost in translation
             | between reality and LaTeX that could be easily pointed out
             | by third party observers.
        
               | MockObject wrote:
               | Maybe experiments can take thousands of hours.
        
           | sfink wrote:
           | > We can start by increasing the standards for experimental
           | proof past words on a piece of paper.
           | 
           | That is overly cynical. We are already well past that
           | standard.
           | 
           | We require words on a piece of paper that were written by
           | somebody from a recognizable institution, or by an AI
           | operated by somebody from a recognizable institution.
        
         | noslenwerdna wrote:
         | Require experimental analysis to be pre-registered? That is how
         | it's done in particle physics, and it works well.
        
         | michaelrpeskin wrote:
         | The "setup this benchmark" happens in medicine all of the time,
         | and it's super insidious and no one sees it. It's hard to tease
         | out.
         | 
         | For example, say you have a new (patentable) drug that you're
         | trying to get approved and replace an old (generic, cheap)
         | drug. You need to prove that the new drug is at least as safe
         | as the old one and/or more effective than the old one. Which
         | sounds reasonable.
         | 
         | Let's say that the old drug is super safe and that the new one
         | may not be. What you can do is set up your RCT so that you
         | surreptitiously underdose the control arm so the safe drug
         | looks less effective. And then you can say that your new drug
         | is more effective than the old one.
         | 
         | Peer review doesn't notice this because you can easily hide it
         | in the narrative of the methods section. I've seen it a couple
         | of times.
         | 
         | So you can easily have "gold standard RCTs" that get the result
         | you want just by subtle changes in the study design.
        
         | soperj wrote:
         | If the study is costly to produce, then it's even more
         | important to replicate it, otherwise you'd wasting a large
         | amount of money on a study with no real sense of whether it is
         | flawed or not.
        
           | mcmoor wrote:
           | I've read that the real reason why researches generated by
           | war crimes are worthless, is because we can't replicate it.
           | Using a result that its validity can't be checked is just a
           | way to disaster. And this is researches with the highest cost
           | (human lives) and we still can't use it.
           | 
           | Lots of "interesting" psychology experiments in the days of
           | yore turns out to have lots of damaging confounding variables
           | and we can't just redo the experiment because, well, we
           | shouldn't.
        
       | MockObject wrote:
       | When I first grasped how many fake papers are floating around, I
       | gnashed my teeth and called for heinous penalties upon the
       | fraudsters.
       | 
       | After a moment, I thought better, and decided that we should
       | actually offer incentive for fraudulent papers! Liars will always
       | have incentive to lie. When their lies are accepted by those who
       | ought to be more skeptical, it's an indictment of the system,
       | which should be more robust in detecting fraud.
        
       | dukeofdoom wrote:
       | The few times I visited a hospital I met people there with
       | complications from previous drugs. One with liver failure. Thanks
       | to that experience no covid shot for me, plus I actually read the
       | pfzier study the judge forced them to release. I now view drug
       | companies on the level of Cartels in Mexico with politicians in
       | their pockets.
        
         | alwayslikethis wrote:
         | > the pfzier study the judge forced them to release
         | 
         | Can you send a link to it?
        
           | dukeofdoom wrote:
           | Google made it impossible to find, found a link on yandex.
           | https://phmpt.org/wp-
           | content/uploads/2021/11/5.3.6-postmarke...
        
       | febeling wrote:
       | I heard somewhere 11% of cancer research can be replicated. Does
       | anyone have numbers with sources for replicatability?
        
       | [deleted]
        
       | TZubiri wrote:
       | 37
        
       | rendang wrote:
       | I'm not nearly qualified to make this argument, but has anyone
       | ever suggested that we collectively do away with the principle
       | that one must publish original research in order to receive a
       | PhD? Maybe in something like entomology, there are enough
       | undescribed beetle species out there to supply myriads of
       | dissertations, but in other fields it seems like you are just
       | incentivizing trivial, useless, or fraudulent research.
        
         | obviouslynotme wrote:
         | That wouldn't fix anything at all. You still get that PhD to do
         | research and publish. The only fix for this is funded and
         | career-advancing randomized replication. We should lean harder
         | into the scientific method, not withdraw from it. Absolutely
         | nothing else will work.
        
         | 2482345 wrote:
         | As a completely unqualified laymen my initial question would be
         | "If you don't reward novelty why would anyone focused on career
         | building want to be novel?" Not that I doubt academia has
         | people who want to push the cutting edge forward (if anything
         | that seems to be the only reason to go into academia vs.
         | private industry usually) but if I was a fresh-faced PhD
         | aspirant I'd want to take the most reliable route to getting my
         | degree and treading ground someone else has already walked
         | seems like a much safer way to do that than novel research if
         | the reward at the end is the same.
         | 
         | But maybe that's a good thing? I can't actually say a reason I
         | think it'd be that terrible except for the profs doing novel
         | research that would lose their some of their student workforce.
        
         | bee_rider wrote:
         | IMO the problem might be more that in general we've got this
         | perception that there's a sort of "academic ranking," or
         | something like that, that puts:
         | 
         | Bachelors < Masters < PhD
         | 
         | Which, of course, not not really the case. A PhD in <field> is
         | a specialization in creating novel research in <field>. In
         | terms of actually applying <field>, a Masters ought to be as
         | prestigious or whatever as a PhD. That it isn't thought of that
         | way seems to indicate, I dunno, maybe we need a new type of
         | super-masters degree (one that gives you a cool title I guess).
         | 
         | Or, this will get me killed by some academics, but let's just
         | align with with the general public seems to think anyway: make
         | a super-masters degree, give it the Dr title, make it the thing
         | that indicates total mastery of a specific field (which is what
         | the general public's favorite Doctors, MDs, have, anyway) (to
         | the extent to which a degree can even indicate that sort of
         | thing, which is to say, not really, but it is as good as we've
         | got). Then when PhDs can have a new title, Philosopher of
         | <field>, haha.
        
       | 0cf8612b2e1e wrote:
       | Those are two wildly different things. Most things are flawed.
       | Deliberately fabricating results is anathema to advancing
       | science.
        
         | ftxbro wrote:
         | They mean flawed to the point of being useless in a way that
         | indicates incompetence or negligence or fraud. Not flawed like
         | they had a typo or a negative result or didn't use MLA
         | formatting in their citation.
        
         | rossdavidh wrote:
         | True that, but the article does distinguish between "flawed"
         | (44%) and "problems that were so widespread that the trial was
         | impossible to trust" (26%).
         | 
         | Really, though, how/why would we expect otherwise? There's
         | nothing in the system to prevent it, and plenty to incentivize
         | it. There's really no good reason to expect (under the current
         | system) that it would not happen (a lot). Without systemic
         | change, it will not get better.
        
         | Randomizer42 wrote:
         | [flagged]
        
         | acc_297 wrote:
         | Yes I work with these IPD spreadsheets every day and they all
         | have typos often in the time/date column (issues with midnight
         | and new years are very common) this can result in an erroneous
         | reporting of something like average drug exposure or clearance
         | if 1 or 2 subjects have uncaught errors in the reported data
         | 
         | I would expect 26% or more studies have these flaws but faked
         | data is a different thing entirely and 26% fake would be
         | incredibly worrying
        
           | rez9x wrote:
           | How many studies are faked and how many just have a
           | 'selective outcome'? Almost every nutritional study I read is
           | at some point sponsored by a group that would benefit from a
           | positive (or negative) outcome. I imagine several studies are
           | run and only the one where the conditions led to a desired
           | outcome are actually published. The researchers may know the
           | results are inconsistent and there is an error somewhere, but
           | it's not in their best interest to find the error and correct
           | it.
        
             | LudwigNagasena wrote:
             | That's why preregistration is important.
        
         | jovial_cavalier wrote:
         | How do you discern between a study that is flawed and one that
         | is faked? For instance, you could use some methodology that you
         | know to be flawed, and allow that flawed methodology to bias
         | your result in a pre-determined direction. If you're ever
         | caught, you just claim it was an honest error, but it functions
         | exactly the same as if you generated the data wholesale.
        
           | qt31415926 wrote:
           | He groups them together because ultimately the result is that
           | the science can't be trusted. He doesn't go so far to claim
           | that one was intentionally faked vs gross incompetence.
        
           | mcmoor wrote:
           | My interpretation is that "flawed" study usually have honest
           | data but wrong interpretation/analysis or wrong data
           | gathering design so to make the conclusion essentially
           | worthless. "Faked" study just fucking lie and may just
           | provide rigged data to support a flawless conclusion.
        
       | passwordoops wrote:
       | My PhD advisors wife's job at a big pharma company was in a
       | department that attempted to reproduce interesting papers. She
       | claimed they only had a 25% _success_ rate (this was early-mid
       | 2000s)
        
       | pella wrote:
       | 2 months ago: https://news.ycombinator.com/item?id=36770624 ( 333
       | points )
        
       | Solvency wrote:
       | I often chuckle that same people who vehemently defend vaccine
       | safety studies also decry studies that show processed seed oils,
       | glyphosate, aspartame, and hundreds of other molecules or
       | compounds are unsafe over time. That's the beauty of medical
       | studies! Whether they're good or bad, your own biases mean you
       | can simply ignore the ones that don't align with your opinions.
       | 
       | Oh, before I forget! Red meat and saturated fat is terrible for
       | you! Wait, no, it's actually sugar that's evil. Vegetables are
       | great for you! Oh wait, most vegetables contain oxalates and
       | other defense, mechanisms and compounds that are actually bad for
       | you overtime.
        
         | rfrey wrote:
         | It's pretty easy to ridicule (and therefore feel superior to,
         | nice bonus!) an intellectual opponent if you get to invent
         | their views out of thin air.
        
         | johndhi wrote:
         | Hmm -- in usual fashion I haven't read the original post here,
         | but I'm guessing it didn't find that RCT trials that big pharma
         | do are quite as useless as most 'medical studies' generally.
         | 
         | Do you think well-funded RCTs (like those that support vaccine
         | safety) are just as weak as any old observational study?
        
           | johndhi wrote:
           | I've now read it.
           | 
           | But my question to the person saying it's problematic to
           | defend vaccine studies and attack food results is: isn't it
           | possible that you feel the research procedures used in one
           | are superior to those used in another?
           | 
           | For example: vaccine safety study looks at 200,000 people and
           | randomly assigns them to use or not use the vaccine.
           | Coffee/red wine study looks at 30 people and surveys them
           | about how they felt last week after drinking coffee/red wine.
           | Looking at these two, I think it's fair to put more trust in
           | the vaccine study.
        
             | misterdad wrote:
             | [dead]
        
         | paulddraper wrote:
         | > Red meat and saturated fat is terrible for you!
         | 
         | Red meat is a little bit bad for longevity. The majority of the
         | reported effect is correlative.
         | 
         | > it's actually sugar that's evil.
         | 
         | Sugar is bad but mostly because it's easy to overeat, and
         | obesity is all around terrible for health.
         | 
         | > Vegetables are great for you! Oh wait, most vegetables
         | contain oxalates and other defense, mechanisms and compounds
         | that are actually bad for you
         | 
         | Cooking removes most oxalates (tho vitamins too, to be fair).
         | But the overall effect of oxalates is relatively minor, except
         | in extreme cases.
         | 
         | ---
         | 
         | Every food source has advantages and disadvantages.
         | 
         | Not being obese is 75% of the health battle.
        
           | ifyoubuildit wrote:
           | This is all stated as if it is fact. It sounds believable to
           | me, but how do you (and the rest of us) know that the sources
           | you got this from aren't a part of the junk being called out
           | by the article?
        
             | paulddraper wrote:
             | Preponderance of the evidence of multiple studies.
             | 
             | Skepticism is reasonable.
        
         | vampirical wrote:
         | I think you're accidentally telling on yourself here. You're
         | looking at somebody getting a result which is surprising to you
         | but rather than being curious about how they might be on to
         | something you're turning off your brain and assuming they're
         | malfunctioning.
         | 
         | Something being in a category, such as "a study", doesn't tell
         | you much about a thing. If you read multiple studies on vaccine
         | safety critically and reason about them and what experts are
         | saying about them, IMO most functional human being are going to
         | reach the same general conclusion about vaccine safety. If you
         | do the same thing on studies about seed oils or aspartame
         | you're also going to come to the conclusion that they're safe!
         | If you're not reaching these same results it doesn't necessary
         | mean you're the one who is malfunctioning but you should
         | seriously consider it and try again to learn what you might not
         | know.
        
         | ifyoubuildit wrote:
         | I dislike "the same people who" comments, but I agree with the
         | sentiment. A lot of us have very little ability to determine
         | the validity of this study over that, but will confidently
         | voice an opinion anyway.
         | 
         | The only way (imo) to stay on firm ground is to acknowledge
         | that someone published a thing saying xyz, and maybe that you
         | are x% convinced by it. Can't get too far out over your skis
         | going that route.
        
         | the_af wrote:
         | I've never met one of those people. In general people who decry
         | vaccines (as a norm, not talking about the covid can of worms)
         | tend to fall into the "alternative medicine" bucket and
         | distrust all science studies, and those who trust vaccines tend
         | to also trust other scientific studies...
        
         | LordKeren wrote:
         | The central thesis of this comment is "all medical studies are
         | equal". They are not.
        
       | gustavus wrote:
       | I can't wait for it to get to the next level of meta.
       | 
       | "How many medical studies of medical studies are faked or flawed?
       | A meta report on meta studies."
       | 
       | https://xkcd.com/2755/
        
       | zackmorris wrote:
       | I'm not as worried about faked/flawed studies as I am about
       | pharmaceutical companies knowingly selling bad drugs to make
       | billions of dollars with no recourse from the FDA. I'll never
       | look at Big Pharma the same way again after watching Painkiller:
       | 
       | https://www.netflix.com/title/81095069
        
       | xhkkffbf wrote:
       | I'm sorry to say that I've personally witnessed some people in
       | the next lab commit fraud. It was investigated and someone was
       | fired, but I can believe that it is all too common.
        
       | jjslocum3 wrote:
       | > ...it would help if journals routinely asked authors to share
       | their IPD
       | 
       | Why couldn't a bad actor just fake the raw data? Isn't that what
       | Climategate was all about?
        
       | ketanmaheshwari wrote:
       | Pretty sure most studies about the effects of coffee and red wine
       | are flawed.
        
         | function_seven wrote:
         | That cynicism you have about those studies is probably because
         | you don't eat enough eggs. Or too many. I'm not sure...
        
           | PaulHoule wrote:
           | At least I get eggs from a neighbor and not "big egg"; I keep
           | thinking about getting my own chickens but that's not the way
           | I want to feed Mr. Fox who lives in my neighborhood too.
        
             | hellotheretoday wrote:
             | My neighbor had 7 chickens and a rooster and lost all of
             | them to mr. Fox. It was a tragedy, the eggs were so good.
        
             | the_af wrote:
             | Well, your free range chickens aren't truly free range if
             | they don't have to fend off the occasional (and free range)
             | fox!
             | 
             | Nothing is more natural and free range than the occasional
             | murder between animals.
        
               | johndhi wrote:
               | I like where this thread is going.
        
               | cratermoon wrote:
               | My free range chickens are protected by free range fox
               | predators that are communal with them. Sometimes they
               | will eat a chicken or two, but mostly they let the
               | chickens be because they keep the free range parasites in
               | check.
        
               | PaulHoule wrote:
               | I definitely count on stinging insects that live in holes
               | in my house (or that make holes) to keep other stinging
               | insects away.
               | 
               | What I've been told is that opossums are much worse than
               | foxes in the sense that a fox will usually eat a chicken
               | or two to survive but opossums seem to freak out and will
               | kill all the hens in a henhouse in one go.
               | 
               | I've often wished I could talk with my cats but more than
               | ever I wish I could ask them what they knew about the
               | fox. There is this lady
               | 
               | https://www.youtube.com/@debs3289
               | 
               | who meets them in the street, has them come to her door,
               | and feeds them chicken (!) Secretly I imagine that the
               | fox is really a
               | 
               | https://en.wikipedia.org/wiki/Kitsune
               | 
               | and my wife is always reminding me that it has just one
               | tail, not nine.
        
               | the_af wrote:
               | I have to say that I find foxes beautiful. Then again,
               | I'm not a farmer nor do I keep chickens. I'm a city boy.
        
               | PaulHoule wrote:
               | The animals we're mostly concerned about at our farm are
               | horses and cats. I don't think there is anything left in
               | North America that can trouble a horse, unless you count
               | mosquito-transmitted infectious diseases. I hear foxes
               | are not dangerous to cats but I believe we've lost some
               | to coyotes.
               | 
               | Mostly we've had people around, either tenants or
               | neighbors, who keep chickens so we don't have to. I'll
               | say the eggs from a small scale chicken operation taste a
               | lot better than commercial eggs.
               | 
               | I definitely thought about trying to draw in the fox but
               | as much as that British lady makes it look easy on
               | Youtube Shorts the legends are that foxes can cause a lot
               | of trouble.
        
               | the_af wrote:
               | > _I believe we 've lost some to coyotes_
               | 
               | Am I out of place if I say I also like coyotes?
               | (Remember: city boy, so nothing is at stake for me here.
               | I also like wolves!)
               | 
               | And yeah, Japanese folklore has taught me that it's best
               | to avoid kitsune. Though they sometimes turn into magical
               | women who help you?
        
         | the_af wrote:
         | What!? Everyone knows a glass of wine once a day is good for
         | your heart, but also even a single glass a year is way too much
         | and causes irreversible damage ;)
         | 
         | The old adage applies: "everything nice is either illegal or
         | bad for your health". Or both, I would add.
        
           | willmeyers wrote:
           | Everything in moderation, including moderation
        
           | Clubber wrote:
           | Or will produce a baby.
        
           | zzo38computer wrote:
           | I think that something can have both good and bad effects.
        
             | the_af wrote:
             | Yes, of course. But have you read journalists reporting on
             | science findings? It's always an extreme "scientists now
             | claim one glass of wine a day is good for you!" and
             | "scientists discovered that even one glass of wine will
             | ruin your life forever".
             | 
             | Never the middle ground, it's always a shocking new finding
             | "by science" (spoiler: scientists seldom say the things
             | newspapers and pop-science/nutrition & health articles
             | claim they say).
        
           | jylam wrote:
           | I'm nice so your point doesn't stand. Or does it ? I'm not
           | illegal at least.
        
             | the_af wrote:
             | Well, _you_ would say so even if you were illegal!
        
           | sampo wrote:
           | > "everything nice is either illegal or bad for your health"
           | 
           | And is known to cause cancer in the State of California.
        
             | tarxvf wrote:
             | Well that's alright then, I'm in New Hampshire.
        
         | waihtis wrote:
         | why coffee and red wine specifically?
        
           | blfr wrote:
           | Popular subjects, soft disciplines, frequent contradicting
           | results.
        
           | Ekaros wrote:
           | It might be that those are common enough in shared datasets
           | or when datasets are collected. So it is easy to draw
           | interferences with them and various other measured factors.
        
           | PaulHoule wrote:
           | Both of those are substances that _should_ be harmful based
           | on the effects of the major ingredients: e.g. caffeine is
           | addictive and when I am deprived of it my schizotypy flares
           | up and I have paranoid episodes (I yell at my wife "why the
           | hell are you always standing where I want to go next?") I
           | have had two doctors tell me three other reasons why I should
           | quit.
           | 
           | Look in the medical literature and it seems outright spammed
           | by reports on the positive effects of caffeine and negative
           | reports on any of the harmful effects one would expect.
           | 
           | Similarly the main active ingredient of red wine (alcohol) is
           | harmful, red wine in particular causes a lot of discomfort,
           | dispepsia, hangovers and other unpleasant effects if you get
           | a bad vintage but look the literature and it is like it will
           | transport you to a blue zone and you will love forever.
           | 
           | And you find those kind of papers spammed in "real" journals,
           | not MDPI or "Frontiers" journals.
        
             | waihtis wrote:
             | Yes but that's not something you can automatically draw
             | inferences from. Exercise is harmful to you in a short
             | enough time interval but benefits you on the long run.
        
             | denimnerd42 wrote:
             | Furan content in coffee too. Very low amount but still.
             | 
             | Coffee prevents headaches for me so I'll always drink it.
             | And no it's not related to physical dependence although at
             | this point the withdrawal will guarantee a headache.
        
               | goosinmouse wrote:
               | I can relate. I never drank anything with caffeine and
               | would get headaches fairly often. Headaches were never
               | too bad or too often to need medical attention but was
               | just normal part of life. I started drinking coffee on
               | road trips and drives over 3 hours long and noticed that
               | my headache coming on would go away right after. Now i
               | drink coffee twice daily and i'll get a headache once a
               | month at most.
        
               | smazga wrote:
               | Yerba mate is my caffeine of choice and I suggested it to
               | my Dad who gets bad migraines. He claims that when he
               | feels a migraine coming on, he can drink a can and it
               | will result in a mild headache instead of forcing him to
               | lie down in a dark room for hours.
        
               | lurquer wrote:
               | I drank five or six cups of coffee a day for decades. I'd
               | even have a cup before going to bed -- that's how
               | tolerant I had become.
               | 
               | Got a mild flu/Covid/cold couple years ago. Better in a
               | week. But, during the illness and since, the slightest
               | bit of caffeine would make be incredibly wired to the
               | point of panic attacks. Had to quit cold turkey. I've
               | tried a cup now and again, and it's the same thing: 6
               | hours of overwhelming anxiety.
               | 
               | Wierd. It's like I became hypersensitive to caffeine.
               | Oddly, though, nicotine doesn't have that effect, and I
               | always figured the two stimulants were similar.
        
               | waihtis wrote:
               | Try yerba mate, like seriously. It gives you a very
               | smooth caffeine-induced motivation boost without any of
               | the anxiety effects.
               | 
               | I frequently switch between that and coffee (coffee has a
               | much more pronounced effect and sometimes you have to
               | grind)
        
               | MavisBacon wrote:
               | Interesting. I was at one point diagnosed with "tension
               | headaches" and prescribed a medication called Fioricet
               | which is a combination of caffeine, acetaminophen, and a
               | barbituate called butalbital. Incredibly effective. I'm
               | sure the harm profile of coffee alone is lower so if it's
               | totally eliminating headaches as a problem for you it's
               | perhaps a better solution, but thought i'd throw that out
               | there as it could be worth seeing a neurologist if the
               | problem becomes unmanageable for you
        
               | [deleted]
        
             | 99_00 wrote:
             | >caffeine is addictive and when I am deprived of it my
             | schizotypy flares up and I have paranoid episodes
             | 
             | Dose size matters.
        
               | PaulHoule wrote:
               | ... and I find it very hard to stay at a low dose. If I
               | quit entirely for a few weeks I could probably manage one
               | small coffee a day for a while but inevitably I'd have a
               | rough night and then I need a small and then another
               | small or maybe just a large and pretty soon I can be
               | drinking two whole carafe a day.
               | 
               | I have the same issue w/ cannabis. Right now I have a few
               | plants (legal) in the garden and also a bag that is going
               | to a friend and I don't care. If I had a little puff
               | though the next day I would want another little puff and
               | another and in a week or so I would be like the guy in
               | the Bob Marley song
               | 
               | https://genius.com/The-toyes-smoke-two-joints-lyrics
        
         | swalsh wrote:
         | I read a lot of studies about gout, there's a lot of studies
         | like "we found participants who consumed x cups of coffee
         | lowered their uric acid" all of them follow the same pattern.
         | They asked people to consume more liquids, and it lowered their
         | uric acids.
         | 
         | I think peeing more lowers your uric acid.
        
       | neilv wrote:
       | You know how (to use a familiar field) most of software
       | development has become going through the motions, churning
       | massive insecure bulk, doing nonsense rituals and making up myths
       | about process and productivity, hopping jobs frequently, cutting
       | corners and breaking rules, etc., and the purpose is usually not
       | to produce trustworthy solutions, but only to further career and
       | livelihood/wealth?
       | 
       | With everything we've been seeing in recent years on HN, about
       | science reproducibility and fraud, and the complaints about
       | commonplace fudging and fraud that you might hear privately from
       | talking with PhDs/students in various fields... I wonder whether
       | science has developed a similar alignment problem.
       | 
       | How many people in science careers are doing trustworthy science?
       | And when they aren't, why not?
        
         | diogenes4 wrote:
         | > You know how (to use a familiar field) most of software
         | development has become going through the motions, churning
         | massive insecure bulk, doing nonsense rituals and making up
         | myths about process and productivity, hopping jobs frequently,
         | cutting corners and breaking rules, etc., and the purpose is
         | usually not to produce trustworthy solutions, but only to
         | further career and livelihood/wealth?
         | 
         | Well furthering wealth is the reason why tech companies exist.
         | The incentives in academia are completely different. You might
         | be right, but i see no reason to expect similar behavior across
         | such drastically different situations.
        
         | [deleted]
        
       | [deleted]
        
       | somenameforme wrote:
       | "For more than 150 trials, Carlisle got access to anonymized
       | individual participant data (IPD). By studying the IPD
       | spreadsheets, he judged that 44% of these trials contained at
       | least some flawed data: impossible statistics, incorrect
       | calculations or duplicated numbers or figures, for instance. And
       | in 26% of the papers had problems that were so widespread that
       | the trial was impossible to trust, he judged -- either because
       | the authors were incompetent, or because they had faked the
       | data."
       | 
       | So, 70% fake/flawed. The finding falls in line with other large
       | scale replication studies in medicine, which have had replication
       | success rates ranging from 11 to 44%. [1] It's quite difficult to
       | imagine why studies where a positive outcome is a gateway to
       | billions of dollars in profits, while a negative outcome would
       | result in substantial losses, might end up being somehow less
       | than accurate.
       | 
       | [1] -
       | https://en.wikipedia.org/wiki/Replication_crisis#In_medicine
        
         | SubiculumCode wrote:
         | I'd like to point out that the study [1] reported 16 trials
         | from the USA with just one "zombie/fatally flawed" trial. e.g.
         | 6%
         | 
         | Most of the studies that were problematic came from China and
         | Egypt.
         | 
         | In other words, nothing new here.
         | 
         | [1] https://associationofanaesthetists-
         | publications.onlinelibrar...
        
         | onlyrealcuzzo wrote:
         | As someone who works in Data Science at FAANG - if you look
         | hard enough - there is something questionably wrong in every
         | step of the data funnel.
         | 
         | And that's when I believe people do have a somewhat best effort
         | to maximize profits. There are plenty of people that only care
         | about career progression and think they can get away with lying
         | and cheating their way to the top. They wouldn't believe that
         | if it didn't work sometimes.
         | 
         | These medical studies are also run mainly to maximize profits,
         | also by some career climbers. They are not run virtuously for
         | the betterment of society.
         | 
         | So I would be astounded if they are as reliable as people might
         | like to believe.
         | 
         | Maybe I'm just being grossly skeptical. Actually, I'd feel
         | better if someone could convince me I'm completely unfounded
         | here.
        
           | verisimi wrote:
           | I don't think that's skeptical.
           | 
           | I do think there is an even worse issue - which is funding.
           | The money incentive means you can fund studies that support
           | whatever you want.
        
           | hotnfresh wrote:
           | I've literally never seen data-driven business decisions that
           | weren't using fatally flawed datasets or methods so bad that
           | you'd be a fool to believe you were getting anything but
           | gibberish out of them, except in trivial cases.
           | 
           | You quickly learn not to be the guy pointing out the problem
           | that means we'll need several people and months or years to
           | gather and analyze data that _would_ allow them to (maybe)
           | support or disprove their conclusion, though. Nobody wants to
           | hear it... because they don't actually care, they just want
           | to present themselves as doing data-driven decision making,
           | for reasons of ego or for (personal, or company) marketing.
           | It's all gut feelings and big personalities pushing companies
           | this way and that, once you cut through the pretend-science
           | shit.
           | 
           | "Yeah, that graph looks great ( _soul dies a little_ ) let's
           | do it"
        
           | ancorevard wrote:
           | "And that's when I believe people do have a somewhat best
           | effort to maximize profits."
           | 
           | Nope. I actually think that if you do scientific research as
           | a company (profit) it may make you less bad/less likely to do
           | fraud compared to academia (non-profit).
           | 
           | Reason is that there are more ways to punish you, employees,
           | board, investors, etc in a profit seeking vehicle, and as a
           | profit seeking vehicle being caught must be part of the
           | profit seeking calculation - in the end, the world of
           | reality/physics will weigh your contribution.
           | 
           | I believe there is evidence that there is more fraudulent
           | scientific research happening in non-profit
           | vehicles/academia. Take for example an area where there are
           | fewer profit seeking companies participating - social
           | sciences. It's dominated by academia. Now look at the
           | replication rate of social sciences.
        
             | screye wrote:
             | I have a simpler reason for the same belief.
             | 
             | You can fake everything except a well designed A/B test. At
             | FAANG scale, a statistically significant A/B test
             | requirement will stop the worst fraud before it hits the
             | user.
        
               | onlyrealcuzzo wrote:
               | And also - you can somewhat take care of bugs by evenly
               | distributing them to your test & control group.
        
             | spicymapotofu wrote:
             | All three points in last paragraph seem wholy unrelated to
             | each other and your larger point. I agreed with first half.
        
           | hackncheese wrote:
           | Objectivity and honesty can be hard to find if all someone
           | cares about is their reputation as a competent researcher or
           | climbing the ladder. What do you think a potential solution
           | would be for this? I feel like even in my own experience,
           | trying something out and it not working feels like failure,
           | when in fact to proclaim it a success or "fix" it is truly
           | what harms both the endeavor for truth and the people reliant
           | on the outcomes of these surveys
        
             | autoexec wrote:
             | > Objectivity and honesty can be hard to find if all
             | someone cares about is their reputation as a competent
             | researcher or climbing the ladder.
             | 
             | It seems like destroying the reputation and career of
             | people who fake science would be a great start. If you're
             | willing to fake data and lie to get results, there will
             | always be an industry who'd love to hire you no matter how
             | tarnished your reputation is. We need a better means to
             | hold researchers accountable and we need to stop putting
             | any amount of faith in any research that hasn't been
             | independently verified through replication.
             | 
             | Today the lobby for orange juice manufactures can pay a
             | scientist to fake research which shows that drinking orange
             | juice makes you more attractive, and then pay publications
             | to broadcast that headline to the world to increase sales.
             | We should have some means to hold publications responsible
             | for this as well.
        
               | cutemonster wrote:
               | > destroying the reputation and career of people who fake
               | science would be a great start
               | 
               | When so many reports are faulty and fraudulent, that
               | might instead be the great start of destroying the
               | careers of those who would have revealed the fraudulent
               | research?
               | 
               | I wonder what'd happen if researchers got compensated and
               | funding based on other things, unrelated to papers
               | published. But what would that be
        
           | shaburn wrote:
           | Really curious how a profitable industry like medical
           | research ranks against less profitible.
        
           | JacobThreeThree wrote:
           | >Actually, I'd feel better if someone could convince me I'm
           | completely unfounded here.
           | 
           | Unfortunately, your skepticism is not unfounded. Those in the
           | industry conclude the same. Take, for instance, the editor in
           | chief of The Lancet:
           | 
           | >The case against science is straightforward: much of the
           | scientific literature, perhaps half, may simply be untrue.
           | Afflicted by studies with small sample sizes, tiny effects,
           | invalid exploratory analyses, and flagrant conflicts of
           | interest, together with an obsession for pursuing fashionable
           | trends of dubious importance, science has taken a turn
           | towards darkness.
           | 
           | https://www.thelancet.com/journals/lancet/article/PIIS0140-6.
           | ..
        
         | cypress66 wrote:
         | You don't need "big pharma" for this. Researchers at
         | universities also do these kinds of things because it helps
         | them advance their careers.
        
           | smu3l wrote:
           | > Researchers at universities also do these kinds of things
           | because it helps them advance their careers.
           | 
           | This is a huge problem and in my opinion is mostly due to bad
           | incentive structures and bad statistical/methodological
           | education. I'm sure there are plenty of cases where there is
           | intentional or at least known malpractice, but I would argue
           | that most bad research is done in good faith.
           | 
           | When I was working on a PhD in biostatistics with a focus on
           | causal inference among other things, I frequently helped out
           | friends in other departments with data analysis. More often
           | than not, people were working with sample sizes that are too
           | small to provide enough power to answer their questions, or
           | questions that simply could not be answered by their study
           | design. (e.g. answering causal questions from observational
           | data*).
           | 
           | In once instance, a friend in an environmental science
           | program had data from an experiment she conducted where she
           | failed to find evidence to support her primary hypothesis.
           | It's nearly impossible to publish null results, and she
           | didn't have funding to collect more data and had to get a
           | paper out of it.
           | 
           | She wound up doing textbook p-hacking; testing a ton of post-
           | hoc hypotheses on subsets of data. I tried to reel things
           | back but I couldn't convince her to not continue because
           | "that's how they do things" in her field. In reality she
           | didn't really have a choice if she wanted to make progress
           | towards her degree. She was a very smart person, and
           | p-hacking is conceptually not hard to understand, but she was
           | incentivized to not understand it or to not look at her
           | research in that way.
           | 
           | * Research in causal inference is mostly about rigorously
           | defining the (untestable) causal assumptions you must make
           | and developing methods to answer causal questions from
           | observational data. Even if an argument can be made that you
           | can make those assumptions in a particular case, there is
           | another layer of modeling assumptions you'll end up making
           | depending on the method you're using. In my experience it's
           | pretty rare that you can really have much confidence that
           | your conclusions about a causal question if you can't run a
           | real experiment.
        
             | bigfudge wrote:
             | It's so interesting to hear you say that. I became
             | disillusioned with causal methods for observational data
             | for similar reasons. You can't often model your way to
             | interesting inferences without an experiment.
        
           | jermaustin1 wrote:
           | >helps them advance their careers
           | 
           | ... into the multi-billion dollar companies GP is talking
           | about.
        
             | Tenoke wrote:
             | A lot who stay purely in Academia do it as well, presumably
             | for the prestige.
        
               | Ilverin wrote:
               | Well you need peer reviewed papers in highish impact
               | journals to get tenure. For some people, the only way
               | they're getting that is by cheating.
        
               | BurningFrog wrote:
               | Doing research has a lottery element. You explore
               | something that might reveal an important discovery. And
               | sometimes it just doesn't.
               | 
               | That doesn't mean you're a bad scientist, just an unlucky
               | one. But it does mean you can't get tenure.
               | 
               | So it's easy to understand why people fake results to
               | secure a career.
        
               | autoexec wrote:
               | > That doesn't mean you're a bad scientist, just an
               | unlucky one. But it does mean you can't get tenure.
               | 
               | That sounds like a really easy problem to solve. Just
               | treat valid science as important regardless of the
               | results. The results shouldn't matter unless they've been
               | replicated and verified anyway.
        
               | vitalurk wrote:
               | Too easy. Define "valid" though.
        
               | autoexec wrote:
               | Valid as in meaningfully peer reviewed to avoid
               | flawed/badly designed studies as well as total garbage
               | (for example https://nerdist.com/article/fake-star-wars-
               | midi-chlorian-pap...) but the gold standard should be
               | replication.
               | 
               | We should reward quality work, not simply the number of
               | research papers (since it's easy to churn out trash) or
               | what the results are (because until they are verified
               | they could be faked).
        
               | autoexec wrote:
               | Peer review is a joke. Too often it's a rubber stamp
               | because there's no accountability for journals that fail
               | to do the job. Unless peer review means something, the
               | standard should change so that published papers only
               | count if they're independently replicated and verified.
        
               | lmm wrote:
               | It's worse than that: if enough people cheat, even the
               | best people can't make the grade without cheating.
        
               | rqtwteye wrote:
               | I would think a successful study is way better for your
               | academic career than a failed study.
        
               | ethanbond wrote:
               | Which is why you fake your data into looking successful,
               | and you don't even go through the effort of publishing
               | the failed studies (leading to its own _additional_
               | problems in understanding what's true and what we know)
        
           | LMYahooTFY wrote:
           | What's your point? We have "big pharma" for this.
           | 
           | Do you think CocaCola and the Sacklers had their own unique
           | ideas shared by no one else? That we've filtered all
           | scrupulous people out of industry?
           | 
           | Scruples are an abstraction at that scale.
        
             | smcin wrote:
             | Is the Coca-Cola mention about their campaign to influence
             | CDC research and policy [0], and pro-obesity astroturf
             | [1][2]?
             | 
             | [0]: Politico "Coca-Cola tried to influence CDC on research
             | and policy, new report states"
             | [https://www.politico.com/story/2019/01/29/coke-obesity-
             | sugar...]
             | 
             | [1]: "Evaluating Coca-Cola's attempts to influence public
             | health 'in their own words': analysis of Coca-Cola emails
             | with public health academics leading the Global Energy
             | Balance Network"
             | https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10200649/
             | 
             | [2]: Forbes: "Emails Reveal How Coca-Cola Shaped The Anti-
             | Obesity Global Energy Balance Network" https://www.forbes.c
             | om/sites/nancyhuehnergarth/2015/11/24/em...
        
         | paulusthe wrote:
         | It's more complex than just that. Sure, there's the people
         | trying to make a dollar who are willing to do bad science in
         | order to get the result they want. But there's also the general
         | publication bias against replication studies - who wants to
         | read them, and who wants to do them (they're not usually seen
         | as prestigious academically: most academics want to test their
         | ideas, not those of others.
         | 
         | And then there's cultural differences in which people sometimes
         | see a negative result as a "failure", don't publish it as a
         | result, and instead skew the data and lie their asses off in
         | order to gain prestige in their career. As long as nobody
         | double checks you, you're good.
        
           | autoexec wrote:
           | > ut there's also the general publication bias against
           | replication studies - who wants to read them, and who wants
           | to do them (they're not usually seen as prestigious
           | academically: most academics want to test their ideas, not
           | those of others.
           | 
           | Academia seems like the idea place for this. Why not require
           | a certain number of replicated studies in order to get a
           | degree? Universities could then be constantly churning out
           | replication studies.
           | 
           | More importantly, why do we bother taking anything that
           | hasn't been replicated seriously? Anyone who publishes a
           | paper that hasn't been verified shouldn't get any kind of
           | meaningful recognition or "credit" for their discovery until
           | it's been independently confirmed.
           | 
           | Since anyone can publish trash, having your work validated
           | should be the only means of gaining prestige in your career.
        
         | NotHowStatsWork wrote:
         | "So, 70% fake/flawed."
         | 
         | I think you miss read that section? Only were 44% fake or
         | flawed acording to this study.
         | 
         | The 26% that were very flawed is a subset of the 44% that were
         | flawed in general. So those precentages should not be added
         | together.
        
           | cycomanic wrote:
           | Oh the irony!
        
         | codingdave wrote:
         | > Carlisle got access to anonymized individual participant data
         | (IPD)
         | 
         | I'm not in the industry so my question might have an obvious
         | answer to those of you who are: How would one go about getting
         | IPD if you wanted to run your own analysis of trial data or
         | other data-driven research?
        
           | misterdad wrote:
           | You'll need to reach out to the study authors with a request.
           | If they are interested (you're going to publish something
           | noteworthy with a citation for them (low chance), you want to
           | bring them in on some funded research (better chance), etc)
           | then they'll push it to their Institutional Review Board (a
           | group of usually faculty and sometimes administrative staff
           | at a University/Hospital/Org) who will review the request,
           | the conditions of the initial data collection, legal
           | restrictions, and then decide if they'll proceed with setting
           | up some sort of IRB agreement / data use agreement. Unless
           | you're a tenured professor somewhere or a respected
           | researcher with some outside group then you probably won't
           | get past any of those steps. Even allegedly anonymized data
           | comes with the risk of exposure (and real penalties) not to
           | mention the administrative overhead (expense, time,
           | attention) that you'll need to be able to cover the cost of
           | through some funded research. That research, btw, will also
           | need to be through some IRB structure. You can tap a private
           | firm that acts as an IRB but that's another process entirely
           | and most certainly requires fat stacks of cash. Legal privacy
           | concerns, ethical concerns, careerism (nobody wants you to
           | find the 'carry the two' you forgot so you can crash their
           | career prospects), bloated expenses (somebody has to pay for
           | all of that paperwork, all those IRB salaries, etc) and etc,
           | etc, etc all keep reproducibility of individual data frozen.
           | Even within the same institution. Within the same team! You
           | have to tread lightly with reproduction.
        
         | [deleted]
        
         | waterheater wrote:
         | Two relevant perspectives to share:
         | 
         | "It is simply no longer possible to believe much of the
         | clinical research that is published, or to rely on the judgment
         | of trusted physicians or authoritative medical guidelines. I
         | take no pleasure in this conclusion, which I reached slowly and
         | reluctantly over my two decades as an editor of the New England
         | Journal of Medicine." -Marcia Angell
         | 
         | "If the image of medicine I have conveyed is one wherein
         | medicine lurches along, riven by internal professional power
         | struggles, impelled this way and that by arbitrary economic and
         | sociopolitical forces, and sustained by bodies of myth and
         | rhetoric that are elaborated in response to major threats to
         | its survival, then that is the image supported by this study."
         | -Evelleen Richards, "Vitamin C and Cancer: Medicine or
         | Politics?"
        
         | Zalastax wrote:
         | Is there not full or at least partial overlap between the 44 %
         | and the 26 %? Which would mean not 70 % but some smaller
         | number?
        
         | morelisp wrote:
         | > The finding falls in line with other large scale replication
         | studies in medicine, which have had replication success rates
         | ranging from 11 to 44%.
         | 
         | These numbers seem almost wholly unrelated. A perfectly good
         | study may be extremely difficult to replicate (or even the
         | original purpose of replication - the experiment _as described
         | in the paper_ may simply not be sufficient); and an attempt at
         | replication (or refutation), successful or not, is under the
         | same pressure to be faked or flawed as the original paper.
        
         | adasdasdas wrote:
         | I also see similar findings in tech where most experiment
         | results are "fudged". In some cases, people run the "same"
         | experiment 5+ times until one is stat sig.
        
         | qt31415926 wrote:
         | I don't think the statement reads that the 44% and the 26%
         | should be additive. Especially given the zombie graphic where
         | it looks like they overlap the 26 on top of the 44, where the
         | orange bar is the 26% and the remaining yellow bar is the 44%
        
         | costigan wrote:
         | I think you misread the article slightly. The 26% of zombie
         | papers are inside the 44% of papers that contained at least
         | some flawed data. Look at the figure just below this quote
         | where the blue bar, indicating papers he thought were ok,
         | covers more than 50%.
        
           | crazygringo wrote:
           | Which, ironically, just shows how easy it is to make mistakes
           | not just in complicated statistical methods, but in basic
           | interpretation of what numbers mean in the first place.
        
             | SubiculumCode wrote:
             | not to mention that everyone here forgets to mention that
             | the majority of submitted papers with suspected fraud came
             | from countries with known issues with fraudulent science
             | publishing. e.g. China. The USA, for example, looked much
             | better (although not perfect) at 6% zombie.
        
       | some_random wrote:
       | While we're on the subject, how many studies period are faked or
       | flawed to the point of being useless? It seems to me that the
       | scientific community's reaction to the replication crisis has
       | been to ignore it.
        
         | systemvoltage wrote:
         | > has been to ignore it
         | 
         | This is putting it very mildly.
        
         | treis wrote:
         | My internal filter is:
         | 
         | Sociology/Psych/Economics are almost all junk. Their
         | conclusions may or may not be correct.
         | 
         | Medical studies are mostly junk. There's way too much financial
         | incentive to show marginal improvement. Theraflu and anti-
         | depressents come to mind. Both show a small effect in studies
         | and launched billion dollar businesses.
         | 
         | Hard science stuff tends to be pretty good. Mostly just
         | outright fraud and they usually end up getting caught.
        
           | fnikacevic wrote:
           | Chemistry still has a ton of issues with people exaggerating
           | numbers that can't be contested like yield and purity of
           | reactions even among the top journals. Straight up fraud is
           | rarer yes.
        
         | RugnirViking wrote:
         | > the scientific community's reaction to the replication crisis
         | has been to ignore it
         | 
         | What can they do? It's an incredibly hard problem to solve.
         | It's like asking why the buisness community has done nothing to
         | adress the housing crisis.
         | 
         | Large scale culture changes, or the entire structure of the way
         | science is conducted and funded, would be the only solutions
        
           | colechristensen wrote:
           | If a university was actually in the business of seeking
           | truth, there is much they could do to solve this.
           | 
           | * working groups "red teams" whose whole job it is to find
           | weaknesses in papers published at the university
           | 
           | * post mortems after finding papers with serious flaws
           | exposing the problem and coming up with constructive
           | corrective actions
           | 
           | * funding / forcing researchers to devote a certain amount of
           | time to replicating significant results
           | 
           | * working groups of experts in statistics and study design
           | available to consult
           | 
           | * systems to register studies, methodologies, data sets, etc.
           | with the aim of eliminating error and preventing post-hoc
           | fishing expeditions for results
           | 
           | The whole-ass purpose of a university is seeking knowledge.
           | They are fully capable of doing a better job of it but they
           | don't because what they actually focus on are things like
           | fundraising, professional sports teams, constructing
           | attractive buildings, and advancing ranking.
           | 
           | Most universities would be better off just firing and not
           | replacing 90% of their administration.
        
             | jltsiren wrote:
             | Universities can't do that, because they lack the
             | expertise. Most of the time, the only people in a
             | university capable of judging a study are the people who
             | did it.
             | 
             | That's because universities are in the business of
             | teaching. Apart from a few rare exceptions, universities
             | don't have the money to hire redundant people. Instead of
             | hiring many experts in the same topic, they prefer hiring a
             | wider range of expertise, in order to provide better
             | learning opportunities for the students.
        
           | some_random wrote:
           | Maybe brush up on undergraduate statistics and experimental
           | design so that they don't execute and publish obviously
           | flawed studies? Perhaps they could apply such basic knowledge
           | to the peer review process which is supposed to catch these
           | things. They could stop defending their peers who publish
           | bogus research, stop teaching about obviously flawed nonsense
           | like the Stanford Prison "Experiment" and hold "scientists"
           | like Philip Zimbardo in contempt. It's like asking how
           | individual cops can make things better, of course they can't
           | fix systematic problems but that doesn't mean there's nothing
           | they can do.
        
           | LudwigNagasena wrote:
           | It doesn't have to be large scale. It can start with a single
           | journal, with a single university, with a single scientist.
           | 
           | Make peer review public. Weight replication studies more.
           | Make conducted peer reviews and replications an important
           | metric for tenure and grants. Publish data and code alongside
           | papers.
        
           | JacobThreeThree wrote:
           | >What can they do?
           | 
           | There's literally thousands of different things they could
           | do.
           | 
           | But why do anything when the real business of academia is in
           | the tax-free hedge funds they operate and the government-
           | subsidized international student gravy train? There's no
           | short-term incentive to change anything.
        
         | adamsb6 wrote:
         | The replication crisis is even a thing because of very strong
         | career incentives.
         | 
         | You might think that publishing about the replication crisis
         | itself would be great for your career, but perhaps not. Maybe
         | the incentives to be able to bullshit your way to a
         | professorship are so great that no one wants to rock the boat.
        
           | rqtwteye wrote:
           | "Maybe the incentives to be able to bullshit your way to a
           | professorship are so great that no one wants to rock the
           | boat."
           | 
           | Our whole economy is fueled by people bullshitting each
           | other.
        
             | wizzwizz4 wrote:
             | Our economy is fuelled by people doing actual work. Large-
             | ish, visible parts are _shaped_ by bullshitters, but that
             | 's a different thing.
        
         | thereisnospork wrote:
         | > It seems to me that the scientific community's reaction to
         | the replication crisis has been to ignore it.
         | 
         | They've always known so there hasn't actually been any new
         | information from which to spur action. In the academic circles
         | I've run in there has always been a strong mistrust of reported
         | results and procedures based on past difficulties with internal
         | efforts replicating results. Basically a right of passage for a
         | grad student to be tasked with replicating work from an
         | impossible paper.
        
           | HarryHirsch wrote:
           | That's the correct answer, but what is baffling is that this
           | is news to so many people frequenting this website. Most
           | everyone posting here has been though a university, but
           | hardly anyone has been involved in the pursuit of research.
        
         | onthecanposting wrote:
         | PIs may not be able to raise money for replication studies.
         | Some of this is a consequence of guidelines for federal funds
         | to prevent waste on duplication of effort.
        
           | some_random wrote:
           | This goes far beyond there not being enough replication
           | studies, the problem is that when replication studies are
           | done the results don't reproduce because the results were
           | bogus. https://en.wikipedia.org/wiki/Replication_crisis
        
             | whatshisface wrote:
             | If the replication studies were done at a reasonable rate
             | there would be no incentive to produce bogus results
             | because you'd be caught before you could go through an
             | entire career as a "successful scientist."
        
           | bee_rider wrote:
           | This points to the real problem and also where the
           | responsibility and interest ought to be to fix it.
           | 
           | There's no replication crisis for academics because they have
           | a meatspace social network of academics; they go to
           | conferences together and know each other. You can just ignore
           | a paper if you know the author is an idiot.
           | 
           | If medical studies are faked, is it a problem? Presumable
           | regulatory agencies are using these studies or something,
           | right? Looks like the FDA and NSF need to fund some more
           | replication studies.
        
         | tppiotrowski wrote:
         | > While we're on the subject, how many studies period are faked
         | or flawed to the point of being useless?
         | 
         | Or how many studies are useless, period? It's like publishing a
         | memoir to Amazon. You can now say "author" on your resume, or
         | when you're introduced or at cocktail parties but nobody finds
         | any value in what you have to say. You can also use ChatGPT
         | because people might not notice.
        
           | some_random wrote:
           | There is always value in expanding the breadth and depth of
           | human knowledge, even if it doesn't seem useful to you, right
           | now. That of course assumes that knowledge is true, which is
           | the crux of the problem now.
        
             | yellowcake0 wrote:
             | Unfortunately, the preponderance of very low value research
             | in the literature puts a significant burden on the
             | scientists who have to sift through a lot of garbage to
             | find what they're looking for. Even if the work is
             | ostensibly correct (much of it is not), it really doesn't
             | do anyone much good, except for the authors of course. But
             | now every undergraduate and every parent's little vanity
             | project at Andover wants a first author contrib., so here
             | we are.
        
       ___________________________________________________________________
       (page generated 2023-09-19 23:00 UTC)