[HN Gopher] Why isn't there a replication crisis in math?
       ___________________________________________________________________
        
       Why isn't there a replication crisis in math?
        
       Author : jseliger
       Score  : 113 points
       Date   : 2022-02-02 18:26 UTC (4 hours ago)
        
 (HTM) web link (jaydaigle.net)
 (TXT) w3m dump (jaydaigle.net)
        
       | svat wrote:
       | Despite the title (a title with a question in it invites people
       | to comment without reading the post, even more than the usual
       | already high level), this is a really good post IMO, with
       | valuable insights into not just mathematics but also the
       | replication crisis elsewhere. (And it _does_ discuss Mochizuki 's
       | claimed proof of the _abc_ conjecture, and links to MathOverflow
       | on formalization with proof assistants, to a recent paper
       | discussing Vladimir Voevodsky 's views, etc.) This from the first
       | part is sound:
       | 
       | > _The replication crisis is partly the discovery that many major
       | social science results do not replicate. But it's also the
       | discovery that we hadn't been trying to replicate them, and we
       | really should have been. In the social sciences we fooled
       | ourselves into thinking our foundation was stronger than it was,
       | by never testing it. But in math we couldn't avoid testing it._
       | 
       | But the post doesn't stop there! The second part of the post
       | (effect sizes etc), with the examples of power posing and the
       | impossibly hungry judges, is even more illuminating. Thanks!
        
       | tabtab wrote:
       | because practical applications of such works are usually several
       | decades away. A bad medical study can kill people today. There's
       | not a lot of incentive to check nor complain about consequences
       | far into the future.
        
       | mjw1007 wrote:
       | I think it's a good article.
       | 
       | But it's worth considering the possibility that mathematics could
       | have fallen, and could still fall, into a state where false
       | results are frequently published, and isn't 'protected' by
       | anything special in the nature of the field or its practitioners.
       | 
       | Just as you might find yourself asking "why did city A fall into
       | the grip of organised crime, and not city B?". You might look for
       | answers in the methods of police recruitment or a strong history
       | of respect for the rule of law or anything like that, but it
       | might turn out that the answer is really just "city A got
       | unlucky".
        
       | zitterbewegung wrote:
       | Math doesn't have a replication crisis at all it has a
       | comprehension crisis.
       | 
       | Since proofs are programs one can basically say that mathematical
       | theorems are incredibly detailed software that is completely open
       | source and invites people to identify programs that don't work
       | and or fix issues.
       | 
       | A famous one is Fermats last theorem which needed a fix but was
       | largely right.
       | 
       | Others have said that it takes 6 months to a year to get
       | published. The other thing with math is the fact that you can get
       | completely scooped and your work is worthless.
       | 
       | Edit: I am using "proofs are programs" very loosely and yes
       | Theorems are much more than programs as other commenters have
       | pointed out.
        
         | jaydaigle wrote:
         | Interesting! My experience is that scooping is less of an issue
         | in math than in any of the science fields I have friends in.
         | Papers are lower-stakes, there's less money involved, and if
         | two of you are working on the same project you can just co-
         | author.
         | 
         | (And if you have an independent paper, that can _also_ get
         | published; your paper is distinct even if the result isn't. I
         | think the PT HOMFLY polynomial was independently proven in like
         | four different papers published within two years (and it's
         | named so that all eight authors get credit).
         | 
         | But also, publication lags shouldn't lead to more scooping,
         | because you can put it up on the arXiv at the beginning of the
         | publication process, not the end. In my experience the paper is
         | treated as "real" once it hits the arXiv; the acceptance is
         | mostly a formality that lets us put it on our promotion packet.
         | 
         | But also, publication times don't lead to scooping generally
         | because you
        
         | bawolff wrote:
         | > Since proofs are programs one can basically say that
         | mathematical theorems are incredibly detailed software that is
         | completely open source and invites people to identify programs
         | that don't work and or fix issues.
         | 
         | I would say its more like pseudocode. There can be quite a
         | large gap between a normal proof, and a machine checkable
         | proof, which is the computer program version.
        
         | pthread_t wrote:
         | > The other thing with math is the fact that you can get
         | completely scooped and your work is worthless
         | 
         | Why math specifically? One would think this applies in
         | virtually all fields.
        
           | zitterbewegung wrote:
           | Increase in development time (publication can take 6 months
           | to a year).
        
         | contravariant wrote:
         | A more notorious example of the comprehension crisis would be
         | Mochizuki's claimed proof of the abc conjecture. So far fairly
         | few people are willing to claim they both understand and agree
         | with the several hundred pages of 'proof'.
        
           | zitterbewegung wrote:
           | I was tempted to use that but Fermat's last theorem is known
           | to the general public for much longer and has a resolution.
        
           | exdsq wrote:
           | https://en.wikipedia.org/wiki/Shinichi_Mochizuki
           | 
           | It's a fun rabbit hole to go down :)
        
         | mb7733 wrote:
         | > Since proofs are programs one can basically say that
         | mathematical theorems are incredibly detailed software that is
         | completely open source and invites people to identify programs
         | that don't work and or fix issues.
         | 
         | I don't think it's that straightforward; proofs in papers are a
         | mix between explanation in natural language and mechanical
         | steps. Not every step of deduction can feasibly be written out.
         | That's part of why computer aided proofs are not that popular
         | in math.
        
         | Quekid5 wrote:
         | An interesting perspective on programs-as-proof is I-forget-
         | his-name-but... the mathematician who made really bold claims,
         | if _only_ you 'd study under his tutelage for a number of years
         | to understand this whole new terminology he invented.
         | 
         | With programs-as-proof it really wouldn't matter. It's either
         | "computer says yes" or "compu'er says noooo".
         | 
         | EDIT: Whoop, sibling post mentioned, it's Mochizuki.
        
           | morelisp wrote:
           | This doesn't seem a very generous description of Mochizuki's
           | work. You don't "need" to study under him for a number of
           | years and there's no evidence he's being obscurantist. The
           | proof is long and has a lot of novel techniques he's
           | invented, and he works primarily in Japanese. You can
           | reasonably side with e.g. Scholze's interpretation without
           | thinking Mochizuki is disingenuous or some kind of scammer.
           | 
           | He's considerably less esoteric than e.g. Grothendieck was
           | even during his more "public" years.
        
             | Quekid5 wrote:
             | I have nothing invested in whether or not any given
             | mathematician is right or wrong. I just picked a random
             | example of a controversial proof -- the point was more that
             | proof-as-computation _could_ settle and any all disputes.
             | 
             | It might not lend more understanding to people not invested
             | in "field X" (or even people who _are_ invested in field
             | X!), but it would be _proof_.
             | 
             | Proof in the current world of math is quite intangible.
        
         | k__ wrote:
         | Aren't programs generally more complex than math proofs?
         | 
         | Like, the more accidental edge cases people produce, the less
         | they understand the program.
        
         | nitwit005 wrote:
         | The math department at my university tried to read through the
         | proof of Fermat's last theorem, as a for-fun activity. They
         | eventually gave up because they realized it would take too much
         | time.
        
         | hgibbs wrote:
         | Having a paper published can take a very very long time, a year
         | is quite short and 6 months is basically the minimum wait. My
         | last paper has taken 2 years to be accepted from the first time
         | I submitted it, despite it being largely the same as the
         | initial submission, accepted as part of my thesis over a year
         | ago, and having several citations. It is very frustrating, and
         | it also means that easier (and less original) work is easier to
         | publish.
        
       | the_watcher wrote:
       | The social science and medical replication crisis seems like it
       | would be far more impactful than a mathematics crisis, right?
       | Politicians, policy-makers, doctors, etc. all make decisions
       | based on potentially flawed or outright incorrect studies in a
       | way that I don't think is true for the equivalents in math,
       | simply because there aren't decisions and policies up for debate
       | related to much of them (if I am wrong about this, please correct
       | me).
        
         | smartscience wrote:
         | If a flawed mathematical paper were used as the basis for what
         | then became a flawed cryptography algorithm, I can see that
         | having impact if the bad guys noticed the flaw first. But yes,
         | I expect examples like that would be comparatively rare.
        
           | not2b wrote:
           | In cryptography the math is almost always the strongest part,
           | and it is the side-channel attacks and implementation
           | mistakes that let the bad guys in. When it is the math, the
           | flaw is often that the algorithm has all the desirable
           | properties proved in a number of papers, but has some
           | exploitable structure that analysts can turn into an attack.
        
       | raphlinus wrote:
       | This is a thoughtful and thought-provoking blog post. I think
       | it's worth asking similar questions of computer science. I think
       | you'll find some math-like patterns -- there's basically no
       | chance Quicksort or any other fundamental algorithm paper will
       | fail to replicate -- and some patterns which will fail to
       | replicate, like in software engineering.
       | 
       | Some of the early results on pseudorandom generators and hash
       | functions aren't holding up well, but I think that's just
       | progress. We understand the problem a whole lot better than we
       | did back then.
       | 
       | Perhaps more interesting is the literature on memory models. The
       | original publications of the Java and C11 memory models had lots
       | of flaws, which took many years to fix (and that process might
       | not be totally done). I worry that there are a bunch of published
       | results that are similarly flawed but just haven't gotten as much
       | scrutiny.
        
         | Analemma_ wrote:
         | The parts of CS that are the most math-like (which include
         | fundamental algorithms) don't have a replication crisis, but
         | the ones that are the most social-science like probably do, or
         | would. I would bet large sums of money that a lot of the
         | literature on stuff like "does OOP lead to better code", "does
         | IDE syntax highlighting lead to fewer bugs" etc. would fail to
         | replicate if anyone bothered trying.
         | 
         | The thing is, the general sense I get is that people in CS
         | already have so little confidence in these results that it's
         | not even considered worth the time to try and refute them.
         | Which doesn't exactly speak well of the field!
        
           | Beldin wrote:
           | Some measurements are interesting and valuable without being
           | replicable. For example, the number of online devices or the
           | number of websites using wordpress. Take the same measurement
           | at a later point in time and the results are different. Yet I
           | wouldn't call those fields maths-like.
        
           | Karrot_Kream wrote:
           | Research into this stuff is very young and so I think it's
           | fair to be skeptical of the results. I'm hoping we'll
           | eventually come up with more rigorous, reproducible results.
        
           | sterlind wrote:
           | I worry about ML papers in particular. models are closely
           | guarded, often impractical to train independently due to
           | ownership of the training/test set, or computing power or
           | details left out of the paper. there's no way to
           | mathematically prove any of it works, either. it's like
           | social science done on organisms we've designed.
        
         | ahelwer wrote:
         | In competitive programming you could basically assume the
         | pseudocode in a paper is not literally correct and requires
         | some tweaking to work, despite a "proof" of its correctness.
         | Particularly with string algorithms.
        
           | sterlind wrote:
           | long time no see!
           | 
           | there's a couple levels there:
           | 
           | rote translating pseudocode into your target language isn't
           | likely to pan out well.
           | 
           | so instead you run the pseudocode in your mind, develop an
           | intuition on how it works, and that's the "replication" bit
           | this post talks about with reviewing math papers.
           | 
           | but both the pseudocode and your code will likely have edge
           | cases you didn't handle. this isn't a problem for math -
           | that's the category of common trivial/easily fixable proof
           | errors that don't really affect the paper. but they're a
           | problem for machines that run them literally.
           | 
           | maybe a good compromise strategy for formal verification is
           | to declare the insight of the algorithm - recurrence relation
           | or whatever - as an axiom, and then use the prover to whack
           | the tricky edge cases.
        
         | AussieWog93 wrote:
         | From my experience in ML, I'd suspect that the "crisis" isn't
         | that the research is false so much as it's useless (algorithm x
         | with parameter set w happens to work well on one particular
         | dataset, conclusion: I have revolutionised the field).
        
           | yodsanklai wrote:
           | This isn't unique to ML. A lot of research is about adding an
           | epsilon to an existing paper, which probably doesn't
           | interesting anyone except a small community working in their
           | very own niche topic.
           | 
           | But does it mean there's a crisis? maybe that's just a way to
           | foster an environment that will let great ideas emerge.
        
             | fshbbdssbbgdd wrote:
             | I'd rather be in the world where we have too many papers
             | tweaking the details of power posing and exactly measuring
             | how much each contributes to the effect. At least we'd know
             | the effect is real.
        
         | yodsanklai wrote:
         | Yes, I'm convinced tons of published results are flawed! I
         | heard top researchers tell their students "don't spend too much
         | time on the proofs, nobody reads them"). And much CS scientific
         | papers don't get a lot of attention. But it's not necessarily
         | bad, other researchers builds on top of this work and results
         | consolidate over time.
        
           | xenonite wrote:
           | Isn't this a misunderstanding? I suspect they rather meant to
           | avoid spending too much time on language in these parts.
        
             | yodsanklai wrote:
             | No, it's not. In that specific case, the supervisor thought
             | the value of the paper didn't lie in the proofs, plus it
             | was a rank B conference. He rather has his student working
             | on a different paper than spending 1 week on the proof.
        
         | skybrian wrote:
         | There was that time when it was discovered that nearly all
         | published binary searches and mergesorts were wrong. [1]
         | 
         | And yet, the concepts of binary search and merge sort are fine.
         | 
         | I think that's quite similar to the situation in math papers?
         | Because math isn't executable, a math paper being
         | "significantly" wrong would be like discovering that a program
         | uses a fatally flawed algorithm and is trying to do the
         | impossible. It can't be fixed.
         | 
         | Programs that can't be fixed seem rare?
         | 
         | [1] https://ai.googleblog.com/2006/06/extra-extra-read-all-
         | about...
        
           | thethirdone wrote:
           | Rather than "wrong", I would describe those implementations
           | as "not fully general". They work perfectly when `n < some
           | large number` as opposed to `n < some large number * 2`. The
           | latter is the best you can do with the function signature,
           | but that is somewhat arbitrary. You could easily choose a 64
           | bit index and exceed all practical needs.
        
       | scoofy wrote:
       | The paper here seems to make absolutely zero distinction between
       | deductive and inductive reasoning, which should be the entire
       | point.
       | 
       | Math is an arbitrary framework, built from arbitrary axioms. It
       | is deductive. Thus, all proofs are simply deduction. The
       | knowledge here is _positive_ knowledge, we can show things are
       | true, false, or undecidable. There may be errors, but those are
       | errors in execution.
       | 
       | Psychology is _not built on a framework_. It is inductive, we are
       | trying to find axioms that map to the data we collect. Thus, all
       | papers are trying to add /build it's arbitrary framework. The
       | only knowledge here is _negative_ knowledge, falsification, we
       | know only what is a failed hypothesis. There _will_ be errors,
       | both in execution, and there will also be statistical errors in
       | experimental results.
       | 
       | The entire point of the replication crisis is that we don't
       | publish or pay attention to results that are boring, so the
       | framework we build is built on skewed data. We don't reject
       | previously popular papers that are now unfalsifiable (the idea
       | that the _now unfalsifiable_ Ingram experiment is still taught in
       | every university psychology dept should be outrageous). The
       | boring results need to be weight statistically against the
       | interesting results, but aren 't, etc. Nobody out there is
       | arguing whether or not the axiom of choice is a _true_ axiom? It
       | sort of doesn 't matter, it can't matter, because it's arbitrary
       | by definition.
       | 
       | You can't have a replication crisis inside of a deductive
       | framework _without changing the framework_. This doesn 't happen
       | too often, but we did see this during the shift from neutonian to
       | einsteinian physics. The study of philosophy of science is fairly
       | obscure, but is the center of this discussion.
        
         | jcranberry wrote:
         | I think you're wrong on some of these points, and the author
         | was trying to make these points but was not explicit.
         | 
         | I remember reading an article in the intelligencer by a
         | mathematician who essentially said there are certain
         | conjectures where if they were to be proved false rather than
         | be thrown into a sea of uncertainty, mathematicians i would
         | quickly move to investigate a readjustment of basic axioms
         | rather than accept that those conjectures are incorrect.
         | 
         | Then there are fields of mathematics around selecting different
         | axioms. Investigating the ramifications of whether you take the
         | undecidable "continuum hypothesis" as true or false. And then
         | theres model theory and such. Presumably they study models of
         | interest and not arbitrary ones.
         | 
         | You're mostly correct that the methodology is mostly deductive
         | but the point is that what we choose to use isn't arbitrary
         | because there are things in math which are more important than
         | axioms as they are things believed to be "real".
        
         | stavros wrote:
         | Exactly, I can't understand what the author was thinking. If a
         | paper shows that 2+2=4, what would the replication problem be?
         | You write the paper out again and 2+2 turns out to equal 5?
         | 
         | "The results can't be replicated" is different from "the logic
         | here is wrong". So different that this article starts from an
         | entirely invalid premise.
        
       | [deleted]
        
       | gus_massa wrote:
       | > _When I've served as a peer reviewer I've read the papers
       | closely and checked all the steps of the proofs, and that means
       | that I have replicated the results._
       | 
       | The side effect is that math papers have an insane long time to
       | publication. Perhaps 6 months, or 1 year or more if you are
       | unlucky.
       | 
       | In physics, the publication time is like 3 month. Something like
       | 1 month for the first review and then two months for making small
       | changes suggested by the referee and discussing with the editor.
       | 
       | As a side^2 effect, some citation index of the journals count
       | only the citations during the first year. But the papers that
       | have the citation are sleeping over the reviewer desk during that
       | year, so the number is lower than the real number.
        
       | lacker wrote:
       | The root of the replication crisis in social sciences is not just
       | that many papers fail to replicate, but that there is no way to
       | clearly resolve a result that fails to replicate. A paper claims
       | that pre-K increases test scores a decade later, another paper
       | claims it doesn't, and there's no clear resolution. The
       | disagreement just festers, with both sides citing research that
       | supports their opinion. The argument often "spills out" into the
       | public sphere.
       | 
       | In mathematics and computer science, there are many errors in
       | published papers. However, once you point out an error, it's
       | usually pretty straightforward to resolve whether it's really an
       | error or not. Often there is a small error which can be fixed in
       | a small way. Exceptions like the abc conjecture are rare.
        
       | amelius wrote:
       | Because math is not science?
       | 
       | https://philosophy.stackexchange.com/questions/14818/is-math...
        
       | curt15 wrote:
       | It did:
       | 
       | https://en.wikipedia.org/wiki/Italian_school_of_algebraic_ge...
        
       | otabdeveloper4 wrote:
       | Because math is not a science.
        
         | iqanq wrote:
         | More like because maths are an actual science.
        
       | dilawar wrote:
       | A bit cynical take is that it is harder to mislead in theory?
       | Your mistakes will get caught and you risk losing your reputation
       | so it less likely that someone will publish a proof without
       | double triple checking. Theory subject has faster verificarion
       | loops.
       | 
       | Verification and review is harder in experimental works and it
       | might take few decades before someone finds an error, or even
       | collect resources to verify the claim. So maybe it's harder to
       | fight 'publish or perish' deamons in experimental sciences?
        
       | keithalewis wrote:
       | "A mathematician's fame is proportional to the number of
       | incorrect proofs they have published."
        
       | deltaonefour wrote:
       | This gets at the heart of the difference between science and
       | math. What is actually science? and what is actually math? You
       | may think you know, but most people don't.
       | 
       | Did you know? That in Science, and therefore reality as we know
       | it... NOTHING can be proven to be true. Proof is the domain of
       | logic and mathematics, NOT science.
       | 
       | There is a huge difference between science and mathematics.
       | 
       | Math is an imaginary game involving logic. You first recursively
       | assume logic is true then you assume some additional things are
       | true called "axioms." You assume the entire universe only
       | encapsulates Logic and the axioms as part of your theory and that
       | is everything that exists in your theoretical universe. You
       | derive or prove statements that you know must consequently be
       | true in your made up universe based off of logic and the
       | "axioms." These statements are called "theorems." That is math...
       | 
       | For example the "pythagorean theorem" is a statement made about
       | geometry assuming that the axioms of geometry are true and logic
       | is true.
       | 
       | So of course if it's a logical game, there's no real replication
       | crisis. All theorems in mathematics are proven true based off of
       | logic. You can't replicate conflicting results if a result is
       | PROVEN to be true via pure logic.
       | 
       | Science is different. Nothing can ever be proven to be true.
       | There are basically two axioms in science. We assume logic is
       | true, just like in math. And we assume that the mathematical
       | theory of probability is a true model for random events happening
       | in the real world. Then we assume everything else about the
       | universe is completely unknown. This is key.
       | 
       | Because the domain of the universe is unknown... at any point in
       | time we can observe something that contradicts our initial
       | hypothesis. ANY point in time. That means even the theory of
       | gravity is forever open to disproof. This is the reason why
       | NOTHING can be proven. To quote Einstein:
       | 
       | "No amount of experimentation can ever prove me right; a single
       | experiment can prove me wrong."
       | 
       | So sure mathematicians can make mistakes... and I know the author
       | is talking about higher level details.... but at its most
       | fundamental level, assuming all other attributes are ideal... it
       | is fundamentally impossible for math to have a replication crisis
       | while science, on the other hand, is fundamentally forever open
       | to these crisis so long as the universe is unknown.
       | 
       | The most interesting thing to me in all of this however is that
       | within science, probability is a random axiom. We have no idea
       | why probability works... it just does. It's the driver behind
       | another fundamental concept: Entropy and the arrow of time. For
       | some strange reason Logic in our universe exists side by side
       | with probability as a fundamental property.
        
         | mcphage wrote:
         | > All theorems in mathematics are proven true based off of
         | logic. You can't replicate conflicting results if a result is
         | PROVEN to be true via pure logic.
         | 
         | You can publish a paper in mathematics that claims to prove
         | something, but is mistaken. A paper claiming that a theorem is
         | proven, is not the same thing as the theorem _being_ proven.
         | However, that 's not often the case--why that is, is an
         | interesting & meaningful question.
        
       | Barrin92 wrote:
       | why not turn the premise from the article around. Instead of
       | suggesting that math might also have a replication crisis, why
       | not question whether the whole replication crisis thing was
       | overblown and effectively more of an ideological attack on a few
       | disciplines.
       | 
       | Errors in science, disagreement and lack of reproducibility I
       | think are common and prevalent but it doesn't necessarily imply
       | that a discipline as a whole doesn't make progress. The obsession
       | with statistical accuracy and 'science as bookkeeping' mentality
       | seems fairly new to begin with, and science did just fine before
       | we even had the means to verify every single thing ever
       | published.
       | 
       | It kind if ignores the dynamic nature of science. Most of what is
       | published probably has close to zero impact regardless of whether
       | its right or wrong, but paradigm changing research generally
       | asserts itself. Science is evolutionary in that sense, it's full
       | of mistakes but stumbles towards correct solutions at uneven
       | tempo. In a sense you can just look at it like VC investment.
       | Nine times out of ten individual things don't work, but the
       | sector overall works, the market economy is full of grifters and
       | failed businesses, but it doesn't matter that much.
       | 
       | So, maybe half of math is bullshit but so is everything else but
       | in math people just say "whatever" until they find something
       | good, whereas in psychology people use it as an opportunity to
       | hack away at it.
        
         | AussieWog93 wrote:
         | The difference between the two examples you have it that
         | venture capital and the market economy openly embrace their
         | flaws, whereas science and academia refuses to acknowledge (or
         | manage) the "humanness" of the system and projects a hyper-
         | enlightened ideal both internally and to the outside world.
        
         | burnished wrote:
         | This isn't turning the problem around in the same sense. In
         | order to effectively turn a problem around you need to use its
         | complement, and then it should be a binary proposition. Your
         | example where you suppose an idealogical attack fails here
         | because it is not the only other explanation, you haven't
         | turned the problem around in the same way that turning "what is
         | the probability this action had an effect" around can become
         | "what is the probability that this action had no effect".
        
         | quanticle wrote:
         | _It kind if ignores the dynamic nature of science. Most of what
         | is published probably has close to zero impact regardless of
         | whether its right or wrong, but paradigm changing research
         | generally asserts itself._
         | 
         | The problem is that psychology, unlike physics, doesn't really
         | have a paradigm. There's the old saying, "Extraordinary claims
         | require extraordinary evidence." But for that heuristic to be
         | effective, you need a standard for what counts as
         | extraordinary. Extraordinary compared to what? In phsyics,
         | there are two well-established paradigms (relativity and
         | quantum mechanics), which establish what counts as ordinary,
         | and what counts as extraordinary. So, for example, if you're
         | making a claim that the distribution of dark matter in the
         | cosmos more clumpy than predicted by existing models, of that
         | the energy level of a particular field is 12 MeV rather than
         | 10, those are ordinary claims, which can be accomodated by
         | tweaks to the existing paradigm. But if you're saying that the
         | speed of light has varied over the history of the universe, or
         | that all subatomic particles are actually tiny vibrating
         | string-like structures, well, that's going to require a lot
         | more evidence.
         | 
         | In psychology, it's much more difficult to have that kind of
         | intuition. Take the concept of priming, for example. Is
         | claiming that people walk more slowly when they're encouraged
         | to think of things that make them feel old extraordinary? It
         | makes a certain sort of intuitive sense, but, on the other
         | hand, there's absolutely no causal mechanism suggested. So when
         | a number of priming studies fail spectacularly under
         | replication [1], I don't know what to think. I don't have a
         | good sense for how much of psychology is overturned by the
         | replication failure, in the same sense that I'd have for
         | physics if it turned out that e.g. the speed of light is a
         | variable rather than a constant.
         | 
         | [1]: https://mindhacks.com/2017/02/16/how-replicable-are-the-
         | soci...
        
           | pdonis wrote:
           | _> In phsyics, there are two well-established paradigms
           | (relativity and quantum mechanics), which establish what
           | counts as ordinary, and what counts as extraordinary._
           | 
           | Actually, I would say that these well-established paradigms
           | establish what counts as _extraordinary_. In other words,
           | relativity and QM are _examples_ of extraordinary claims that
           | we believe _because_ we have extraordinary evidence for them.
           | Both of these theories say all kinds of extraordinary things,
           | and most people who first encounter the theories start out
           | thinking they can 't possibly be true. We believe them not
           | because they are just ordinary, but because we have taken the
           | time and effort to accumulate extraordinary evidence for
           | them.
           | 
           | In that light, the replication crisis in other areas of
           | science is easily explained: they allow extraordinary claims
           | to be published _without_ the extraordinary evidence that
           | those claims would require. So of course many of those claims
           | turn out to be wrong.
        
       | mjfl wrote:
       | because math doesn't do experiments.
        
         | medstrom wrote:
         | Exactly. A math study is not a "study" in the sense of "hey i
         | saw a funny pattern in some data maybe it's a sign of my pet
         | theory" - it's literally already proven when published. There's
         | nothing more to do.
        
         | Jtsummers wrote:
         | Or as noted in the article:
         | 
         | > But one of the distinctive things about math is that our
         | papers aren't just records of experiments we did elsewhere. In
         | experimental sciences, the experiment is the "real work" and
         | the paper is just a description of it. But in math, the paper,
         | itself, is the "real work".
         | 
         | And
         | 
         | > And that means that you can replicate a math paper by reading
         | it.
        
           | horsawlarway wrote:
           | I think that means that the word "experiment" isn't the right
           | term for what most mathematicians do.
           | 
           | I'd say most times it's "modeling", not "experimenting"
        
         | [deleted]
        
         | nimih wrote:
         | https://www.experimentalmath.info/ would beg to differ, I
         | think.
        
         | horsawlarway wrote:
         | This is my take.
         | 
         | Science attempts to describe reality. Math attempts to create
         | rules/axioms.
         | 
         | They're not the same pursuit, although they can often be useful
         | together.
        
         | NovemberWhiskey wrote:
         | This seems the obvious answer. The replication crisis isn't
         | about published material being _wrong_ , it's about the
         | inability to reproduce the results of experiments or studies in
         | a repeatable fashion.
         | 
         | It's not like you make a hypothesis in math and then need to go
         | away and interview a sample of 1,000 circles and report back
         | that, controlling for ellipses that may be misreporting as
         | circles, the ratio of the circumference to the diameter is 3.2
         | +/- 0.1 (p<0.05).
        
         | bunje wrote:
         | "Mathematics is the part of physics where experiments are
         | cheap." - V.I. Arnold
        
       | schuyler2d wrote:
       | You could also argue that Math _had_ its replication crisis in
       | the 17th-19th centuries. E.g. infinite series  "proofs" that were
       | eventually shown to be flawed methodologies.
       | 
       | This and other crises led to grounding modern mathematics with
       | set theory, Zermelo-Fraenkel axioms, etc and understanding what's
       | possible (e.g. Godel's theorem).
       | 
       | Psychology and other social sciences are barely a century old.
        
         | fshbbdssbbgdd wrote:
         | A more recent example is the Italian school of algebraic
         | geometry, where it was discovered in the mid 20th century that
         | many claimed proofs were faulty (and some "proven" results were
         | incorrect).
        
         | zmgsabst wrote:
         | Mathematics _is_ having a replication crisis and people pay so
         | little attention they don't know.
         | 
         | That replication crisis has led to efforts in formal
         | verification such as HoTT, Lean, etc.
         | 
         | https://homotopytypetheory.org/
         | 
         | https://xenaproject.wordpress.com/2021/06/05/half-a-year-of-...
        
           | Victerius wrote:
           | How is math experiencing a replication crisis?
           | 
           | > This site serves to collect and disseminate research,
           | resources, and tools for the investigation of homotopy type
           | theory, and hosts a blog for those involved in its study.
           | 
           | > Exactly half a year ago I wrote the Liquid Tensor
           | Experiment blog post, challenging the formalization of a
           | difficult foundational theorem from my Analytic Geometry
           | lecture notes on joint work with Dustin Clausen.
           | 
           | ???
        
           | dwohnitmok wrote:
           | No by and large mathematics is not having a replication
           | crisis. As the blog post states:
           | 
           | > Question: Was the proof in [Analytic] found to be correct?
           | 
           | > Answer: Yes, up to some usual slight imprecisions.
           | 
           | This has been the case for almost all math formalization
           | efforts. Even when (very rarely) proofs were revealed to be
           | incorrect, the result was salvageable.
        
             | thrown_22 wrote:
             | That's because most math proofs are treading on well
             | understood grounds and only extending them slightly. E.g.
             | it would be like a psychologist asking how an results of a
             | well proven result would be different if all participants
             | wore red shoes.
             | 
             | When you enter truly new grounds mathematicians don't even
             | agree if the distinctions being made have a meaning, let
             | alone if they are true.
        
           | guerrilla wrote:
           | Here's [1] the lecture where Vladimir Voevodsky talked about
           | the problem and his experience with it but like the blog
           | says, he didn't and they don't consider it a crisis. Even
           | HoTT (and other TT) people present it as how things could be
           | much better, not about how things are terrible.
           | 
           | 1. https://youtu.be/E9RiR9AcXeE
        
       | umvi wrote:
       | My completely subjective opinion is that at the highest levels of
       | math, there are only a handful of people that are even capable of
       | peer reviewing, and their time is in high demand.
       | 
       | Wiles's proof of Fermat's Last Theorem is like 120 pages long and
       | he first delivered it disguised as a class to a bunch of grad
       | students who barely understood any of it and hence gave no
       | feedback. Because this is Fermat's Last Theorem which is famous,
       | eventually people in the math community that understood Wiles's
       | work reviewed it and found an error. Had it been a 120 page proof
       | of some not famous problem like random chessboard thought
       | experiments, it probably could go years without anyone seriously
       | looking at it.
        
       | verisimi wrote:
       | "We get away with it becuase we can be right for the wrong
       | reasons--we mostly only try to prove things that are basically
       | true."
       | 
       | Apart from the replication crisis, the other crisis that is not
       | really talked about is funding. Academic funding basically comes
       | from one of 3 (connected) sources - government, corporations and
       | the military. Somehow or other - these 3 sources have pretty much
       | the same or non-conflicting aims. These aims relate to power and
       | control. This is actually the largest crisis, IMO.
       | 
       | Given that information, we can re-assess the quoted statement the
       | author makes.
       | 
       | Perhaps its not that they are proving things that are "basically
       | true". Its that right or wrong do not matter. What does matter is
       | that the answers provided meet the agenda of those funding the
       | study. The answer is not _that_ important as long as it is
       | supportive of whatever agenda is in play. I believe this is the
       | case for the replication crisis in science also.
       | 
       | A replication "crisis" is only a crisis if you are attempting to
       | achieve truth and greater understanding. But truth and
       | understanding are only ostensible reasons, not the actual ones.
       | What these studies are actually doing is creating a parallel
       | construction - the aim is actually for studies to appear
       | 'truthey', without actually being so. What studies should actual
       | do is increase the funder's power, wealth extraction abilities,
       | etc.
       | 
       | If you doubt this and think that truth matters, consider this.
       | Surely we should have cracked the best diet for people by now?
       | But there is no common understanding of what is good or bad to
       | eat - if anything there is more confusion. The reason of course,
       | is that there's no money in recommending whole foods or whatever.
       | However, there is money in drugs to make people 'better'. And
       | money in making diet so confusing that people eat themselves into
       | trouble.
       | 
       | Anyway, if you are in the business of governing or monetising the
       | masses, truth and understanding is the last thing you want. Far
       | better to have a story that gives you control, or extracts money.
       | Such is life under fascist governance (where fascist =
       | corporation + governance working together).
        
       | Iwan-Zotow wrote:
       | Huh?!?
       | 
       | Please read about abc conjecture and whole saga wrt proof,
       | Shinichi Mochizuki, Peter Scholze, etc, etc, etc
        
         | sgillen wrote:
         | Did you read the article? This saga is explicitly mentioned and
         | does not detract from the authors point IMO.
        
       | ocschwar wrote:
       | There is one corner of math that does have a replication crisis.
       | Just as we compare programming languages by how "ergonomic" they
       | are to learn and use, mathematicians do come up with novel
       | notation systems to try to improve the ergonomic state of their
       | field, and since "ergonomics" is another way to say "esthetics",
       | and is proved or disproved by user testing, that is where
       | replication gets hard.
       | 
       | The inventor of category theory's wiring diagrams, for example,
       | has claimed that he could get middle schoolers to understand
       | them. I suspect that success has not been replicated.
        
       | Simplicitas wrote:
       | Steven Pinker takes a decent crack at "Statistical Significance"
       | in his new book Rationality
       | (https://en.wikipedia.org/wiki/Rationality_(book)), which
       | underpins a lot of this and is mentioned in this piece. And I'm
       | still grappling with this part of the book. lol
        
       ___________________________________________________________________
       (page generated 2022-02-02 23:00 UTC)