[HN Gopher] Replace peer review with "peer replication" (2021) ___________________________________________________________________ Replace peer review with "peer replication" (2021) Author : dongping Score : 340 points Date : 2023-08-06 12:33 UTC (10 hours ago) (HTM) web link (blog.everydayscientist.com) (TXT) w3m dump (blog.everydayscientist.com) | geysersam wrote: | Both review and replication has their place. The mistake is | treating researchers and the scientific community as a machine: | "pull here, fill these forms, comment this research, have a gold | star" | | Let people review what they want, where they want, how they want. | Let people replicate when they find interesting and motivating to | work on. | SonOfLilit wrote: | My first thought was "this would never work, there is so much | science being published and not enough resources to replicate it | all". | | Then I remembered that my main issue with modern academia is that | everyone is incentivized to publish a huge amount of research | that nobody cares about, and how I wish we would put much more | work into each of much fewer research directions. | tines wrote: | "Replace peer code review with 'peer code testing.'" | | Probably not gonna catch on. | dongping wrote: | "peer code testing" is already the job of the CI server. As it | is nothing new, it probably is not going to catch on. | fastneutron wrote: | As much as I agree with the sentiment, we have to admit it isn't | always practical. There's only one LIGO, LHC or JWST, for | example. Similarly, not every lab has the resources or know-how | to host multi-TB datasets for the general public to pick through, | even if they wanted to. I sure didn't when I was a grad student. | | That said, it infuriates me to no end when I read a Phys. Rev. | paper that consists of a computational study of a particular | physical system, and the only replicability information provided | is the governing equation and a vague description of the | numerical technique. No discretized example, no algorithm, and | sure as hell no code repository. I'm sure other fields have this | too. The only motivation I see for this behavior is the desire | for a monopoly on the research topic on the part of authors, or | embarrassment by poor code quality (real or perceived). | fabian2k wrote: | I don't see how this could ever work, and non-scientists seem to | often dramatically underestimate the amount of work it would be | to replicate every published paper. | | This of course depends a lot on the specific field, but it can | easily be months of effort to replicate a paper. You save some | time compared to the original as you don't have to repeat the | dead ends and you might receive some samples and can skip parts | of the preparation that way. But properly replicating a paper | will still be a lot of effort, especially when there are any | issues and it doesn't work on the first try. Then you have to | troubleshoot your experiments and make sure that no mistakes were | made. That can add a lot of time to the process. | | This is also all work that doesn't benefit the scientists | replicating the paper. It only costs them money and time. | | If someone cares enough about the work to build on it, they will | replicate it anyway. And in that case they have a good incentive | to spend the effort. If that works this will indirectly support | the original paper even if the following papers don't | specifically replicate the original results. Though this part is | much more problematic if the following experiments fail, then | this will likely remain entirely unpublished. But the solution | here unfortunately isn't as simple as just publishing negative | results, it take far more work to create a solid negative result | than just trying the experiments and abandoning them if they're | not promising. | ebiester wrote: | It's simple but not easy: You create another path to tenure | which is based on replication, or on equal terms as a part of a | tenure package. (For example, x fewer papers but x number of | replications, and you are expected to have x replications in | your specialty.) You also create a grant funding section for | replication which is then passed on to these independent | systems. (You would have to have some sort of randomization | handled as well.) Replication has to be considered at the same | value as original research. | | And maybe smaller faculties at R2s pivot to replication hubs. | And maybe this is easier for some sections of biology, | chemistry and psychology than it is for particle physics. We | could start where cost of replication is relatively low and | work out the details. | | It's completely doable in some cases. (It may never be doable | in some areas either.) | SkyMarshal wrote: | _> x fewer papers but x number of replications, and you are | expected to have x replications in your specialty._ | | Could it be simplified it even further to say x number of | papers, but they only count if they're replicated by others | in the field? | nine_k wrote: | No, the idea is that the same researcher should produce _k_ | papers and _n_ replications, instead of just _k + n_ | published papers. | | I'd argue that since replication is somehow faster than | original research, the requirement would count a | replication somewhat lower than an original paper (say, at | 0.75). | ebiester wrote: | That is my idea... If we opened it up, there's probably | more interesting iterations, such as requiring pre- | registration for all papers, having papers with pre- | registration count as some portion of a full paper even | if they fail so long as the pre-registration passed | scrutiny, having non-replicated papers count as some | portion of a fully replicated paper, and having | replication as a separate category such that there is a | minimum k, a minimum n, and a minimum k+n. | | The non-easy part of this is once we start making changes | to the criteria for tenure, this opens up people trying | to stuff all the solutions for all of the problems that | everyone knows already. (See Above.) Would some one try | to stuff code-available for CS conference papers, for | example? What does it mean for a poster session? At what | point are papers released for pre-print? What does it | mean for the tenure clock or the Ph.D clock? Does it mean | that pre-tenure can't depend on studies that take time to | replicate? What do we do with longitudinal studies? | | I think you're looking at a 50 year transition where you | would have to start simple and iterate. | harimau777 wrote: | Is tenure really as mechanical as "publish this many | papers and you get it"? My impression was that it took | into account things like impact factor and was much more | subjective. If that were the case, then wouldn't you run | into problems with whoever decides tenure paying lip | service to counting replication or failed pre-registered | papers but in practice being biased in favor of original | research? | rapjr9 wrote: | Another approach I've seen actually used in Computer Science | and Physics is to make replication a part of teaching to | undergrads and masters candidates. The students learn how to | do the science, and they get a paper out of replicating the | work (which may or may not support the original results), and | the field benefits from the replication. | Eddy_Viscosity2 wrote: | It's not easy because it isn't simple. How do get all of the | universities to change their incentives to back this? | ebiester wrote: | We agree - the "simple not easy" turn of phrase is speaking | to that point. It is easy once implemented, but it isn't | easy to transition. (I am academia-adjacent by marriage but | closer to the humanities, so I understand the amount of | work it would take to perform the transition.) | MichaelZuo wrote: | This isn't just not easy, it would probably be extremely | political to change the structure of the NSF, National | Labs, all universities and colleges, etc., so | dramatically. | tnecniv wrote: | Your proposal has a whole slew of issues. | | First, people that want to be professors normally do so | because they want to steer their research agenda, not repeat | what other people are doing without contribution. Second, who | works in their lab? Most of the people doing the leg work in | a lab are PhD students, and, to graduate, they need to do | something novel to write up in their dissertation. Thus, they | can't just replicate three experiments and get a doctorate. | Third, you underestimate how specialized lab groups are -- | both in terms of the incredibly expensive equipment it is | equipped with and the expertise within the lab. Even folks in | the same subfield (or even in the same research group!) often | don't have much in common when it comes to interests, | experience, and practical skills. | | For every lab doing new work, you'd basically need a clone of | that lab to replicate their work. | majormajor wrote: | > First, people that want to be professors normally do so | because they want to steer their research agenda, not | repeat what other people are doing without contribution. | | If we're talking about weird incentives and academia you | hit on one of the worst ones right here, I think, since | nothing there is very closely connected to helping students | learn. | | I know that's a dead horse, but it's VERY easy to find | reasons that we shouldn't be too closely attached to the | status quo. | | > For every lab doing new work, you'd basically need a | clone of that lab to replicate their work. | | Hell, that's how startup funding works, or market economies | in general. Top-down, non-redundant systems are way more | fragile than distributed ecosystems. If you don't have the | competition and the complete disconnection, you so much | more easily fall into political games of "how do we get | this published even if it ain't great" vs "how do we find | shit that will survive the competition" | harimau777 wrote: | I think that there's also a lot of | psychological/cultural/political issues that work also need | to be worked out: | | If someone wins the Nobel Prize, do the people who replicated | their work also win it? When the history books are written do | the replicators get equal billing to the people who made the | discovery? | | When selecting candidates for prestigious positions, are they | really going to consider a replicator equal to an original | researcher? | kergonath wrote: | > I don't see how this could ever work, and non-scientists seem | to often dramatically underestimate the amount of work it would | be to replicate every published paper. | | They also tend to over-estimate the effect of peer review | (often equating peer review with validity). | | > If someone cares enough about the work to build on it, they | will replicate it anyway. And in that case they have a good | incentive to spend the effort. If that works this will | indirectly support the original paper even if the following | papers don't specifically replicate the original results. | Though this part is much more problematic if the following | experiments fail, then this will likely remain entirely | unpublished. | | It can also remain unpublished if other things did not work | out, even if the results could be replicated. A half-fictional | example: a team is working on a revolutionary new material to | solve complicated engineering problems. They found a material | that was synthesised by someone in the 1980s, published once | and never reproduced, which they think could have the specific | property they are after. So they synthesise it, and it turns | out that the material exists, with the expected structure but | not with the property they hoped. They aren't going to write it | up and publish it; they're just going to scrap it and move on | to the next candidate. Different teams might be doing the same | thing at the same time, and nobody coming after them will have | a clue. | techdragon wrote: | This waste of effort by way of duplicating unpublished | negative results is a big factor in why replicated results | deserve to be rated more highly than results that have not | been replicated regardless of the prestige of the researchers | or the institutions involved... if no one can prove your work | work was correct... how much can anyone trust your work... | | I have gone down the rabbit hole of engineering research | before and 90% of the time I've managed to find an anecdote | or subsequent research footnotes or actual subsequent | research publications, that substantially invalidated the | lofty claims of the engineers in the 70s or 80s (which is | amazing still despite this, a genuine treasure trove of | research unused and sometimes useful aerospace engineering | research and development) and unfortunately outside the few | proper publications, a lot of the invalidations are not | properly reverse cited research material and I could have | spent a week cross referencing before I spot the link and | realise the unnamed work they are saying they are proving | wrong is actually some footnotes containing the only | published data (before their new paper) on some old work that | has a bad scan copy on the NASA NTRS server under some | obscure title and no related keywords to the topic the | research is notionally about... | | Academic research can genuinely suck sometimes... | particularly when you want to actually apply it. | vibrio wrote: | "They also tend to over-estimate the effect of peer review | (often equating peer review with validity)." | | In my experience, scientists ate comfortably cynical about | peer review- even those that serve as reviewers and editors- | except maybe junior scientists that haven't gotten burned | yet. | renonce wrote: | I don't know how scientists handle peer review but aren't | they fighting with peer review to get their papers | published and apply for PhD and tenure and grants etc with | these publications? | kergonath wrote: | Yes, because we know how the metaphorical sausage is made: | with unpaid reviewers who have many other, more interesting | things to do and often an axe to grind. That is, if they | don't delegate the review to one of their post-docs. | aftoprokrustes wrote: | Post doc? In what kind of utopian field did you work? In | my former institute virtually all papers were written by | PhD candidates, and reviewed by PhD candidates. With the | expected effect on quality (due to lack of experience and | impostor-syndrome-induced "how can I propose to reject? | They are likely better than me"). But the Prof-to- | postdoc-to-PhD-ratio was particularly bad (1-2-15). | kelipso wrote: | I was reviewing papers starting second semester of grad | school with my advisor just signing off on it, so not | even PhD candidates, and it was the same for my lab mates | too. | | Initially we spent probably a few hours on a paper for | peer review because we were relatively unfamiliar with | the field but eventually I spent maybe a couple of hours | doing the review. Wouldn't say peer review is a joke but | it's definitely overrated by the public. | jakear wrote: | It's the general public that equates "peer reviewed" with | "definitely correct, does not need to be questioned". | dongping wrote: | While it is a lot of work, I tend to think that one can then | always publish preprints if they can't wait for the | replication. I don't understand why a published paper should | count as an achievement (against tenure or funding) at all | before the work is replicated. The current model just creates | perverse incentives to encourage lying, P-hacking, and cherry- | picking. This would at least work for fields like machine | learning. | | This is, of course, a naive proposal without too much thought | into it. But I was wondering what I would have missed here. | i_no_can_eat wrote: | and in this proposal, who will be tasked with replicating the | work? | dongping wrote: | In some fields, replication is already the prerequisite to | benchmark the SoTA. So the incentives boil down to | publishing them along with negative results. Or as some | have suggested, make it mandatory for PHD candidates to | replicate. | | Though, it seems that it is possible to game the system, by | creating positive/negative replication intentionally, to | collude with/harm the author. | omgwtfbyobbq wrote: | What about a system where peer replication is required if the | number of citations exceeds some threshold? | p1esk wrote: | Who will be replicating it? Why would I want to set aside my | own research to replicate some claim someone made? How would | this help my career? | Knee_Pain wrote: | Academia's values are not objective. Why is it that | replicating or refuting a study is not seen on par as being | a co-author of said study? There is nothing set in stone | preventing this, only the current academic culture. | p1esk wrote: | Because I want to do original research, and be known for | doing original research. Only if I fail at that, I might | settle for being a guy who reproduces others' work (which | basically means the transition from a researcher to an | engineer). | omgwtfbyobbq wrote: | Whether or not you would be doing original research | depends on whether the cited work can be replicated. | | If the cited work is unable to be replicated, and you try | to replicate but get different results, then you would be | doing original research, and then you can base further | work on your initial original study that came to a | different result. | | On the flip side, if you are able to replicate it, then | you are doing extra work initially, but after replicating | the work you've cited, the work you've done is more | likely to be reproducible by someone else. | | The amount of citations needed to require replication | could itself be a function of how easy it is to replicate | work across an entire field. | | A field where there's a high rate of success in | replicating work could have a higher threshold for | requiring replication compared to a field where it's | difficult to replicate work. | omgwtfbyobbq wrote: | I dunno. Offhand, I guess whoever is citing the work would | need to replicate it, but only if it's cited sufficiently | (overall number of citations, considered foundational, | etc...) | | This could help your career by increasing the probability | that the work you're citing is more likely accurate, and as | a result, your work is also likely more accurate. | RoyalHenOil wrote: | A typical paper may cite dozens or hundreds of other | papers. This does not sound feasible. It honestly seems | like it would worsen the existing problem and force | scientists to focus even more on their own original | research in isolation from others, to avoid the expense | of running myriad replication experiments that they | likely don't have the funding and personnel to do. | boxed wrote: | > I don't see how this could ever work, and non-scientists seem | to often dramatically underestimate the amount of work it would | be to replicate every published paper. | | I don't see how the current system works really either. Fraud | is rampant, and replication crisis is the most common state of | most fields. | | Basically the current system is failing at finding out what is | true. Which is the entire point. That's pretty damn bad. | tptacek wrote: | Fraud seems rampant because you hear about cases of fraud, | but not about the tens of thousands of research labs plugging | away day after day. | mike_hearn wrote: | Unfortunately there's a lot of evidence that fraud really | is very prevalent and we don't hear about it anywhere near | enough. It depends a lot on the field though. | | One piece of evidence comes from software like GRIM and | SPRITE. GRIM was run over psychology papers and found | around 50% had impossible means in them (that could not be | arrived at by any combination of allowed inputs) [1]. The | authors generally did not cooperate to help uncover the | sources of the problems. | | Yet another comes from estimates by editors of well known | journals. For example Richard Horton at the Lancet is no | stranger to fraud, having published and promoted the | Surgisphere paper. He estimates that maybe 50% of medical | papers are making untrue claims, which is interesting in | that this intuition matches the number obtained in a | different field by a more rigorous method. The former | editor of the New England Journal of Medicine stated that | it was "no longer possible to believe much of the medical | research that is published". | | 50%+ is a number that crops up frequently in medicine. The | famous Ioannidis paper, "Why most published research | findings are false" (2005) has been cited over 12,000 | times. | | Marc Andreessen has said in an interview that he talked to | the head of a very large government grant agency, and asked | him whether it could really be true that half of all | biomedical research claims were fake? The guy laughed and | said no it's not true, it's more like 90%. [2] | | Elizabeth Bik uncovers a lot of fraud. Her work is behind | the recent resignation of the head of Stanford University | for example. Years ago she said, _" Science has a huge | problem: 100s (1000s?) of science papers with obvious | photoshops that have been reported, but that are all swept | under the proverbial rug, with no action or only an author- | friendly correction ... There are dozens of examples where | journals rather accept a clean (better photoshopped?) | figure redo than asking the authors for a thorough | explanation."_ In reality there seem to be far more than | mere thousands, as there are companies that specialize in | professionally producing fake scientific papers, and whole | markets where they are bought and sold. | | So you have people who are running the scientific system | saying, on the record, that they think science is overrun | with fake results. And there is some quantitive data to | support this. And it seems to happen quite often now that | presidents of entire universities are being caught having | engaged in or having signed off on rule breaking behavior, | like image manipulation or plagiarism, implying that this | behavior is at least rewarded or possibly just very common. | | There are also whole fields in which the underlying | premises are known to be false so arguably that's also | pretty deceptive (e.g. "bot studies"). If you include those | then it's quite likely indeed that most published research | is simply untrue. | | [1] https://peerj.com/preprints/2064v1/ | | [2] https://www.richardhanania.com/p/flying-x-wings-into- | the-dea... | lliamander wrote: | I agree that most labs are probably not out to defraud | people. But without replication I don't think it's | reasonable to have much confidence in what is published. | magimas wrote: | replication happens over time. For example, when I did my | PhD I wanted to grow TaS2 monolayers on a graphene layer | on an Iridium crystal. So I took published growth | recipees of related materials, adapted them to our setup | and then finetuned the recipee for TaS2. This way I | basically "peer replicated" the growth of the original | paper. I then took those samples to a measurement device | and modified the sample in-situ by evaporating Li atoms | on top (which was the actual paper but I needed a sample | to modify first). I published the paper with the growth | recipee and the modification procedure and other | colleagues then took those instructions to grow their own | samples for their own studies (I think it was MoS2 on | Graphene on Cobalt that they grew). | | This way papers are peer replicated in an emerging manner | because the knowledge is passed from one group to another | and they use parts of that knowledge to then apply it to | their own research. You have to see this from a more | holistic picture. Individual papers don't mean too much, | it's their overlap that generates scientific consesus. | | In contrast, requiring some random reviewer to instead | replicate my full paper would be an impossible task. | He/she would not have the required equipment (because | there's only 2 lab setups in the whole world with the | necessary equipment), he/she would probably not have the | required knowledge (because mine and his research only | partially overlap - e.g. we're researching the same | materials but I use angle-resolved photoemission | experiments and he's doing electronic transport) and | he/she would need to spend weeks first adapting the | growth recipee to the point where his sample quality is | the same as mine. | tptacek wrote: | That's not what publication is about. Publication is a | conversation with other researchers; it is part of the | process of reaching the truth, not its endpoint. | cpach wrote: | People in general (at least on da Internetz) seem to | focus way to much on single studies, and way too little | on meta-studies. | | AFAICT meta-studies is the level where we as a society | really can try to say something intelligent about how | stuff works. If an important question is not included in | a meta-study, we (i.e. universities and research labs) | probably need to do more research on that topic before we | really can say that much about it. | lliamander wrote: | Sure, and scientists need a place to have such | conversations. | | But publication is not a closed system. The "published, | peer-reviewed paper" is frequently an artifact used to | decide practical policy matters in many institutions both | public and private. To the extent that Science (as an | institution in its own right) wants to influence policy, | that influence needs to be grounded in reproducible | results. | | Also, I would not be surprised if stronger emphasis on | reproducibility improved the quality of conversation | among scientists. | vladms wrote: | Maybe replication should (and probably does) happen when | the published thing is relevant to some entity and also | interesting. | | I never seen papers as "truth", but more as | "possibilities". After many other "proofs" (products, | papers, demos, etc.) you can assign some concepts/ideas | the label "truth" but one/two papers from the same group | is definitely not enough. | tnecniv wrote: | Yeah passing peer review doesn't mean that the article is | perfect and to be taken as truth now (and remember, to | err is human; any coder on here has had some long | standing bug that went mostly unnoticed in their code | base). It means it passed the journal's standards for | novelty, interest, and rigor based on the described | methods as a retained by the editor / area chair and peer | reviewers that are selected for being knowledgeable on | the topic. | | Implicit in this process is that the authors are acting | in good faith. To treat the authors as hostile is both | demoralizing for the reviewers (who wants to be that | cynical about their field) and would require extensive | verification of each statement well beyond what is | required to return the review in a timely manner. | | Unless your paper has mathematical theory (and mistakes | do slip through), a publication should not be taken as | proof of something on its own, but a data point. Over | time and with enough data points, a field builds evidence | to turn a hypothesis into a scientific theory. | majormajor wrote: | I think the current system is just measuring entirely the wrong | thing. Yes, fewer papers would be published. But today's goal | is "publish papers" not "learn and disseminate truly useful and | novel things", and while this doesn't solve it entirely, it | pushes incentives further away from "publish whatever pure crap | you can get away with." You get what you measure -> sometimes | you need to change what/how you measure. | | > If someone cares enough about the work to build on it, they | will replicate it anyway. | | That's duplicative at the "oh maybe this will be useful to me" | stage, with N different people trying to replicate. And with | replication not a first-class part of the system, the effort of | replication (e_R) is high. For appealing things, N is probably | > 2. So N X e_R total effort. | | If you move the burden at the "replicate to publish" stage, you | can fix the number of replicas needed so N=2 (or whatever) | _and_ you incentive the orginal researchers to make e_R lower | (which will improve the quality of their research _even before | the submit-for-publication stage_ ). | | I've been in the system, I spent a year or two chasing the tail | of rewrites, submissions, etc, for something that was | detectable as low-effect-size in the first place but I was told | would still be publishable. I found out as part of that that it | would only sometimes yield a good p-value! And everything in | the system incentivized me to hide that for as long as | possible, instead of incentivizing me to look for something | else or make it easy for others to replicate and judge for | themselves. | | Hell, do something like "give undergrads the opportunity to | earn Master's on top of their BSes, say, by replicating (or | blowing holes in) other people's submissions." I would've eaten | up an opportunity like that to go _really really deep* in some | specialized area in exchange for a masters degree in a less- | structured way than "just take a bunch more courses."_ | DoctorOetker wrote: | > [...] non-scientists seem to often dramatically underestimate | the amount of work it would be to replicate every published | paper | | Either "peer reviewed" articles describe progress of promising | results, or they don't. If they don't the research is | effectively ignored (at least until someone finds it | promising). So let's consider specifically output that | described promising results. | | After "peer review" any apparently promising results prompt | other groups to build on them by utilizing it as a step or | building block. | | It can take many failed attempts by independent groups before | anyone dares publish the absence of the proclaimed | observations, since they may try it over multiple times | thinking they must have botched it somewhere. | | On paper it sounds more expensive to require independent | replication, but only because the costs of replication attempts | are hidden until its typically rather late. | | Is it really more expensive if the replication attempts are in | some sense mandatory? | | Or is it perhaps more expensive to pretend science has found a | one-shot "peer reviewed" method, resulting in uncoordinated | independent reproduction attempts that may go unannounced | before, or even after failed replications? | | The pseudo-final word, end of line? | | What about the "in some sense mandatory" replication? Perhaps | roll provable dice for each article, and in-domain sortition to | randomly assign replicators. So every scientist would be | spending a certain fraction of their time replicating the | research of others. The types of acceptable excuses to derelict | these duties should be scrutinized and controlled. But some | excuses should be very valid, for example _conscientious | objection_. If you are tasked to reproduce some of Dr. Mengele | 's works, you can cop out on condition that you thoroughly | motivate your ethical concerns and objections. This could also | bring a lot of healthy criticism to a lot of practices, which | is otherwise just ignored an glossed over for fear of future | career opportunities. | jofer wrote: | Also, don't forget that a lot of replication would | fundamentally involve going and collecting additional samples / | observations / etc in the field area, which is often expensive, | time consuming, and logistically difficult. | | It's not just "can we replicate the analysis on sample X", but | also "can we collect a sample similar to X and do we observe | similar things in the vicinity" in many cases. That alone may | require multiple seasons of rather expensive fieldwork. | | Then you have tens to hundreds of thousands of dollars in | instrument time to pay to run various analysis which are needed | in parallel with the field observations. | | It's rarely the simple data analysis that's flawed and far more | frequently subtle issues with everything else. | | In most cases, rather than try to replicate, it's best to test | something slightly different to build confidence in a given | hypothesis about what's going on overall. That merits a | separate paper and also serves a similar purpose. | | E.g. don't test "can we observe the same thing at the same | place?", and instead test "can we observe something | similar/analogous at a different place / under different | conditions?". That's the basis of a lot of replication work in | geosciences. It's not considered replication, as it's a | completely independent body of work, but it serves a similar | purpose (and unlike replication studies, it's actually | publishable). | b59831 wrote: | [dead] | kshahkshah wrote: | When I looked into this, more than 15 years ago, I thought the | difficult portion wasn't sharing the recipe, but the | ingredients, if you will - granted I was in a molecular biology | lab. Effectively the Material Transfer Agreements between | Universities all trying to protect their IP made working with | each other unbelievably inefficient. | | You'd have no idea if you were going down a well trodden path | which would yield no success because you have no idea it was | well trod. No one publishes negative results, etc. | RugnirViking wrote: | lets be brutally honest with ourselves. | | 99% of all papers mean nothing. They add nothing to the | collective knowledge of humanity. In my field of robotics there | are SOOO many papers that are basically taking three or four | established algorithms/machine learning models, and applying | them to off-the-shelf hardware. The kind of thing any person | educated in the field could almost guess the results exactly. | Hundreds of such iterations for any reasonably popular problems | space (prosthetics, drones for wildfires, museum guide robot) | etc every month. Far more than could possibly be useful to | anyone. | | There should probably be some sort of separate process for | things that actually claim to make important discoveries. I | don't know what or how that should work. In all honesty maybe | there should just be less papers, however that could be | achieved. | indymike wrote: | > 99% of all papers mean nothing. They add nothing to the | collective knowledge of humanity. | | A lot of papers are done as a part of the process of getting | a degree or keeping or getting job. The value is mostly the | candidate showing they have the acumen to produce a paper of | such quality that meets the publisher and peer review | requirements. In some cases, it is to show a future employer | some level of accomplishment or renown. The knowledge for | humanity is mostly the authors ability to get published. | RugnirViking wrote: | well yes. But these should go somewhere else than the | papers that may actually contain significant results. The | problem we have here is that there is an enormous quantity | of such useless papers mixed in with the ones actually | trying to do science. | | I understand that part of the reason for that is that | people need to appear as though they are part of the | "actually trying" crowd to get the desired job effects. But | it is nonetheless a problem, and a large one very worth at | least trying to solve. | staunton wrote: | 99% of science is a waste of time, not just the papers. We | just don't know which 1% will turn out not to be. The point | is that this is making progress. As such, these 99% | definitely _are_ adding to the collective knowledge. Maybe | they add very little and maybe it 's not worth the effort but | it's not nothing. I think one of the effects of AI progress | will be allowing to extract much more of the little value | such publications have (the 99% of papers might not be worth | reading but are good enough for feeding the AI). | [deleted] | throwaway4aday wrote: | What's the value in publishing something that is never | replicated? If no one ever reproduces the experiment and gets | the same results then you don't know if any interpretations | based on that experiment are valid. It would also mean that | whatever practical applications could have come from the | experiment are never realized. It makes the entire pursuit seem | completely useless. | geysersam wrote: | It still has value if we assume the experiment was done by | competent honest people who are unlikely to try to fool us on | purpose and unlikely do have made errors. | | It would be even better if it was replicated of course. | | Depending on what certainty you need you might have to wait | for the result of one or several replications, but that is | application dependent. | wizofaus wrote: | > What's the value in publishing something that is never | replicated? | | Because it presents an experimental result to other | scientists that they may consider worth trying to replicate? | dongping wrote: | Then those unconfirmed results are better put on arxiv, | instead of being used to evaluate the performance of | scientists. Tenure and grant committees should only | consider replicated work. | geysersam wrote: | I don't agree. A published article should not be taken | for Gods Truth no matter if it's replicated or peer | reviewed. | | Lots of "replicated" "peer-reviewed" research have been | found to be wrong. That's fine, it's part of the process | of discovery. | | A paper should be taken for what it is: a piece of | scientific work, a part of a puzzle. | justinpombrio wrote: | > If someone cares enough about the work to build on it, they | will replicate it anyway. | | Well, the trouble is that hasn't been the case in practice. A | lot of the replication crisis was attempting for the first time | to replicate a _foundational_ paper that dozens of other papers | took as true and built on top of, and then seeing said | foundational paper fail to replicate. The incentives point | toward doing new research instead of replication, and that | needs to change. | p1esk wrote: | It is the case in my field (ML): if I care enough about a | published result I try to replicate it. | tnecniv wrote: | This is something very sensible in ML since, you likely | want to use that algorithm for something else (or to extend | / modify it), so you need to get it working in your | pipeline and verify it works by comparing with the | published result. | | In something like psychology that is likely harder, since | the experiment you want to do might be related to but | differ significantly from the prior work. I am no | psychologist, but I'd like to think that they don't take | one study as ground truth for that reason but try to | understand causal mechanisms with multiple studies as data | points. If the hypothesis is correct, it will likely | present in multiple ways. | brightball wrote: | > I don't see how this could ever work, and non-scientists seem | to often dramatically underestimate the amount of work it would | be to replicate every published paper. | | The alternative is a bunch of stuff being published which | people belief as "science" that doesn't hold up under scrutiny, | which undermines the reliability of science itself. The current | approach simply gives people reason to be skeptical. | ImPostingOnHN wrote: | I'm not convinced this proposed alternative is better than | the status quo. It's simply not feasible, no matter how many | benefits one might imagine. | | the concern about skepticism is not irrelevant, but many of | these skeptics also are skeptical of the earth being round, | or older than a few thousand years, or not created by an | omnipotent skylord, and I'm not sure it's actually a | significant concern given the current number and expertise of | those who are skeptical | | so, we can hear their arguments for their skepticism, but | that doesn't mean the arguments are valid to warrant the | skepticism exhibited. And in the end, that's what matters: | skepticism warranted by valid arguments, not just any Cletus | McCletus's skepticism of heliocentrism, as if his opinion is | equal to that of an astrophysicist (it isn't). And you know | what? It isn't necessary to convince a ditch digger that the | earth goes around the sun, if they feel like arguing about | it. | backtoyoujim wrote: | Yes it would indeed mean slowing down and having more | scientists. | | It would mean disruption is no longer a useful tool for human | development. | brnaftr361 wrote: | It may not be. I would be willing to argue that there was a | tipping point and we've long exceeded its boundary - progress | and disruption now is just making finding an equilibrium in | the future increasingly difficult. | | So entering into a paradigm where we test the known space - | especially presently - would 1) help reduce cruft; 2) abate | undersirable forward progress; 3) train the next | generation(s) of scientists to be more diligent and better | custodians of the domain. | ebiester wrote: | I don't necessarily think it would mean more scientists, but | it would mean more expense. You have a moderate number of low | impact papers that people are doing for tenure today - papers | for the purpose of cranking out papers. We are talking about | redirecting efforts but increasing quality of what you have. | jononomo wrote: | If it is not replicated it shouldn't be published, other than | as a provisional draft. I don't care if it hurts your feelings. | sqrt_1 wrote: | FYI there is a at least one science journal that only publishes | reproduced research: | | Organic Syntheses "A unique feature of the review process is | that all of the data and experiments reported in an article | must be successfully repeated in the laboratory of a member of | the editorial board as a check for reproducibility prior to | publication" | | https://en.wikipedia.org/wiki/Organic_Syntheses | throwawaymaths wrote: | > I don't see how this could ever work, | | http://www.orgsyn.org/ | | > All procedures and characterization data in OrgSyn are peer- | reviewed and checked for reproducibility in the laboratory of a | member of the Board of Editors | | Never is a strong word. | indymike wrote: | > This is also all work that doesn't benefit the scientists | replicating the paper. It only costs them money and time. | | Maybe this is what needs to change. If we only reward discovery | and success, then the incentive is to only produce discovery | and success. | johnnyworker wrote: | > If someone cares enough about the work to build on it, they | will replicate it anyway. | | Does it really deserve to be called _work_ if it doesn 't | include the a full, working set of instructions that if | followed to a T allow it to be replicated? To me that's more | like pollution, making it someone else's problem. I certainly | don't see how "we did this, just trust us" can even be | considered science, and that's not because I don't understand | the scientific method, that's because I don't make a living | with it, and have no incentive to not rock the boat. | davidktr wrote: | You just described the majority of scientific papers. A | "working set of instructions" is not really feasible in most | cases. You can't include every piece of hard- and software | required to replicate your own setup. | lliamander wrote: | Sounds like a problem worth solving. | johnnyworker wrote: | Then don't call it science, since it doesn't contribute | anything to the body of human knowledge. | | I think it's fascinating that we can at the same time hold | things like "one is none" to be true, or that you should | write tests first, but with science we already got so used | to a lack of discipline that we just declare it fine. | | It's not hard to not climb a tower you can't get down from. | It's the default, actually. You start with something small | where you can describe everything that goes into | replicating it. Then you replicate it yourself, based on | your own instructions. Before that, you don't bother anyone | else with it. Once that is done, and others can replicate | as well, it "actually exists". | | And if that means the majority of stuff has to be thrown | out, I'd suggest doing that sooner rather than later, | instead of just accumulating scientific debt. | davidktr wrote: | Imagine two scientists, Bob and Alice. Bob has spent the | last 5 years examining a theory thoroughly. Now he can | explain down to the last detail why the theory does not | hold water, and why generations of researchers have been | wrong about the issue. Unfortunately, he cannot offer an | alternative, and nobody else can follow his long winded | arguments anyway. | | Meanwhile, Alice has spent the last 5 years making the | best possible use of the flawed theory, and published a | lot of original research. Sure, many of her publications | are rubbish, but a few contain interesting results. | Contrary to Bob, Alice can show actual results and has | publications. | | Who do you believe will remain in academia? And, | according to public perception, will seem more like an | actual scientist? | tnecniv wrote: | Then Bob has failed. | | Academic science isn't just the doing science part but | the articulation and presentation of your work to the | broader community. If Bob knows this space so well, he | should be able to clearly communicate the issue and, | ideally, present an easily understandable counter example | to the existing theory. | | Technical folks undervalue presentation when writing | articles and presenting at conferences. The burden of | proof is on the presenter, and, unless there's some | incredible demonstration at the end, most researchers | won't have the time or attention to slog through your | mess of a paper to decipher it. There's only so much time | in the day and too many papers to read. | | In my experience, the best researchers are also the best | presenters. I've been to great talks out of my domain | that I left feeling like I understood the importance of | their work despite not understanding the details. I've | also seen many talks in my field that I thought were | awful because the presentation was convoluted or they | didn't motivate the importance of their problem / why | their work addressed it | johnnyworker wrote: | I disagree that Bob doesn't produce actual results, or | that something that is mostly rubbish, but partly | "interesting" is an actual result. We know the current | incentives are all sorts of broken, across the board. | Goodhart's law and all that. To me the question isn't who | remains in academia given the current broken model, but | who would remain in academia in one that isn't as broken. | | To put a point on it, if public distrust of science | becomes big enough, it all can go away before you can say | "cultural revolution" or "fascist strongman". Then | there'd be no more academia, and its shell would be | inhabited by party members, so to speak. I'd gladly | sacrifice the ability of Alice and others like her to | live off producing "mostly rubbish" to at least have a | _chance_ to save science itself. | cycomanic wrote: | This is a very simplistic view. Why do believe QC | departments exist? Even in an industrial setting, | companies make the same thing at the same place on the | same equipment after sometimes years of process | optimisation of well understood technology. This is | essentially a best case scenario and still results fail | to reproduce. How are scientists who work at the cutting | edge of technology with much smaller budgets supposed to | give instructions that can be easily reproduced on first | go? Moreover how are they supposed to easily reproduce | other results? | | That is not to say that scientist should not document the | process to their best ability so it can be reproduced in | principle. I'm just arguing that it is impossible to | easily reproduce other people's results. Again when | chemical/manufacturing companies open another location | they often spend months to years to make the process work | in the new factory. | johnnyworker wrote: | > companies make the same thing at the same place on the | same equipment after sometimes years of process | optimisation of well understood technology. This is | essentially a best case scenario and still results fail | to reproduce. | | We're not talking about 1 of 10 reproduction attempts | failing, we're talking about 100%. And no, companies | don't time and time again try to reproduce something that | has never been reproduced and fail, to then try again, | endlessly. That's just not a thing. | | > it is impossible to easily reproduce other people's | results | | We're also not talking about "easily" reproducing | something, but _at all_. And in principle doesn 't cut | it, it needs to be reproduced in practice. | johngladtj wrote: | You should. | MrJohz wrote: | I work with code, which is about as reproducible as it is | possible to get - the artifacts I produce are literally just | instructions on how to reproduce the work I've done again, | and again, and again. And still people come to me with some | bug that they've experienced on their machine, that I cannot | reproduce on my machine, despite the two environments being | as identical as I can possibly make them. | | I agree that reproduction in scientific work is important, | but it is also apparently impossible in the best possible | circumstances. When dealing with physical materials, inexact | measurements, margins of error, etc, I think we have to | accept that there is no set of instructions that, if followed | to a T, will ever ensure perfect replication. | johnnyworker wrote: | > And still people come to me with some bug that they've | experienced on their machine, that I cannot reproduce on my | machine | | But this is the other way around. Have you ever written a | program that doesn't run _anywhere_ except a single machine | of yours? Would you release it and advertise it and | encourage other people to use it as dependency in their | software? | | If it only runs on one machine of yours, you don't even | know if your code is doing something, or something else in | the machine/OS. Or in terms of science, whether the | research says something about the world, or just about the | research setup. | MrJohz wrote: | I think you misunderstand the point of scientific | publication here (at least in theory, perhaps less so in | practice). The purpose of a paper is typically to say "I | have achieved these results in this environment (as far | as I can tell)", and encourages reproduction. But the | original result is useful in its own right - it tells us | that there may be something worth exploring. Yes, it may | just be a measurement error (I remember the magic faster | than light neutrinos), but if it is exciting enough, and | lots of eyes end up looking, then flaws are typically | found fairly quickly. | | And yes, there are often overly excited press releases | that accompany it - the "advertise it and encourage | others to us it as a dependency" part of it analogy - but | this is typically just noise in the context of scientific | research. If that is your main problem with scientific | publishing, you may want to be more critical of science | journalism instead. | | Fwiw, yes of course I've written code that only runs on | my machine. I imagine everyone has, typically | accidentally. You do it, you realise your mistake, you | learn something from it. Which is exactly what we expect | from scientific papers that can't be reproduced. | johnnyworker wrote: | > But the original result is useful in its own right - it | tells us that there may be something worth exploring. | | I disagree. It shows that when someone writes something | in a text editor and publishes it, others can read the | words they wrote. That's all it shows, by itself. Just | like someone writing something on the web only tells us | that a textarea accepts just about any input. | | And even if it did show more than that, when someone | "explores" it, is the result is more of that, something | that might be true, might not be, but "is worth | exploring"? Then at what point does falsifiability enter | into it? Why not right away? To me it's just another | variation of making it someone else's problem, kicking | the can down the road. | | > if it is exciting enough, and lots of eyes end up | looking, then flaws are typically found fairly quickly. | | If that was true, there wouldn't even be a replication | issue, much less a replication crisis. It's like saying | open source means a lot of people look at the code, if | it's important enough. Time and time again that's proven | wrong, e.g. https://www.zdnet.com/article/open-source- | software-security-... | | > yes of course I've written code that only runs on my | machine. I imagine everyone has | | I wouldn't even know how to go about doing that. Can you | post something that only runs on one of your machines, | and you don't know why? Note I didn't say your machine, I | said _one_ machine of yours. Would you publish something | that runs on one machine of yours but not a single other | one, other than to ask "can anyone tell me why this only | runs on this machine"? I doubt it. | varjag wrote: | > Note I didn't say your machine, I said one machine of | yours. | | This thread discusses _peer_ replication, this is not | even an analogy. | johnnyworker wrote: | If you can't _even_ replicate it yourself, what makes you | think peers could? We are talking about something not | being replicated, not even by the original author. The | most extreme version would be something that you could | only get to run once on the same machine, and never on | any other machine. | MrJohz wrote: | I think you may be seeing the purpose of these papers | differently to me, which may be the cause of this | confusion. | | The way you're describing a scientific publication is as | if it were the end result of the scientific act. To use | the software analogy, you're describing publication like | a software release: all tests have been performed, all CI | workflows have passed, QA have checked everything, and | the result is about to be shipped to customers. | | But talking to researchers, they see publishing more like | making a new branch in a repository. There is no | expectation that the code in that branch already be | perfect (hence why it might only run on one machine, or | not even run at all, because sometimes even something | that doesn't work is still worth committing and exploring | later). | | And just like in software, where you might eventually | merge those branches and create a release out of it, in | the scientific world you have metastudies or other forms | of analysis and literature reviews that attempt to glean | a consensus out of what has been published so far. And | typically in the scientific world, this is what happens. | However, in journalism, this isn't usually what happens, | and one person's experimental, "I've only tested this on | my machine" research is often treated as equivalent to | another person's "release branch" paper evaluating the | state of a field and identifying which findings are | likely to represent real, universal truths. | | Which isn't to say that journalists are the only ones at | fault here - universities that evaluate researchers | primarily on getting papers into journals, and prestige | systems that make it hard to go against conventional | wisdom in the field both cause similar problems by | conflating different levels of research or adding | competing incentives to researchers' work. But I don't | think that invalidates the basic idea of published | research: to present a found result (or non-really), | provide as much information as possible about how to | replicate the result again, and then let other people use | that information to inform their work. It just requires | us to be mindful of how we let that research inform us. | johnnyworker wrote: | > But talking to researchers, they see publishing more | like making a new branch in a repository. | | Well some do, others don't. Like the one who wrote the | article this is a discussion of. | | https://en.wikipedia.org/wiki/Replication_crisis | | > Replication is one of the central issues in any | empirical science. To confirm results or hypotheses by a | repetition procedure is at the basis of any scientific | conception. A replication experiment to demonstrate that | the same findings can be obtained in any other place by | any other researcher is conceived as an | operationalization of objectivity. It is the proof that | the experiment reflects knowledge that can be separated | from the specific circumstances (such as time, place, or | persons) under which it was gained. | | Or, in short, "one is none". One _might_ turn into more | than one, it might not. Until it does, it 's not real. | | more snippets from the above WP article: | | > This experiment was part of a series of three studies | that had been widely cited throughout the years, was | regularly taught in university courses | | > what the community found particularly upsetting was | that many of the flawed procedures and statistical tools | used in Bem's studies were part of common research | practice in psychology. | | > alarmingly low replication rates (11-20%) of landmark | findings in preclinical oncological research | | > A 2019 study in Scientific Data estimated with 95% | confidence that of 1,989 articles on water resources and | management published in 2017, study results might be | reproduced for only 0.6% to 6.8%, even if each of these | articles were to provide sufficient information that | allowed for replication | | I'm not saying it couldn't be fine to just publish things | because they "could be interesting". But the overall | situation seems like quite the dumpster fire to me. As | does software, FWIW. | techas wrote: | Well, you could put incentives to make replication attractive. | Give credit for replication. Give money to the researchers | doing the replication/review. Today we pay an average of | 2000EUR per article, reviewers get 0EUR and the editorial keeps | all for putting a pdf online. I would say there is margin there | to invest in improving the review process. | mandmandam wrote: | It's wild to me that although we _know_ that it was Ghislaine | Maxwell 's daddy who started this incredibly corrupt system, | people hardly mention this fact. | | The US system, and others, even attack people who dare to try | and make science more open. RIP Aaron Swartz, and long live | Alexandra Elbakyan. | sebzim4500 wrote: | >I don't see how this could ever work, and non-scientists seem | to often dramatically underestimate the amount of work it would | be to replicate every published paper. | | I think it would be fine to half the productivity of these | fields, if it means that you can reasonably expect papers to be | accurate. | dmarchand90 wrote: | I believe that, contrary to popular belief, the | implementation of this system would lead to a substantial | increase in productivity in the long run. Here's why: | | Currently, a significant proportion of research results in | various fields cannot be reproduced. This essentially means | that a lot of work turns out to be flawed, leading to wasted | efforts (you can refer to the 'reproducibility crisis' for | more context). Moreover, future research often builds upon | this erroneous information, wasting even more resources. As a | result, academic journals get cluttered with substandard | work, making them increasingly difficult to monitor and | comprehend. Additionally, the overall quality of written | communication deteriorates as emphasis shifts from the | accurate transfer and reproduction of knowledge to the | inflated portrayal of novelty. | | Now consider a scenario where 50% of all research is | dedicated to reproduction. Although this may seem to | decelerate progress in the short term, it ensures a more | consistent and reliable advancement in the long term. The | quality of writing would likely improve to facilitate | replication. Furthermore, research methodology would be | disseminated more quickly, enhancing overall research | effectiveness. | matthewdgreen wrote: | In the current system scientists allocate reproduction | efforts to results that they intend to build on. So if | you've claimed a breakthrough technique for levitating | widgets -- and I think this widget technique can be used to | build spacecraft (or if I think your technique is wrong) -- | then I will allocate precious time and resources to | reproducing your work. By contrast if I don't think your | work is significant and worth following up on, then I | allocate my efforts somewhere else. The advantage is that | more apparently-significant results ("might cure cancer") | tend to get a bigger slice of very limited resources, while | dead-end or useless results ("might slightly reduce | flatulence in cats") don't. This distributed | entrepreneurial approach isn't perfect, but it works better | than central planning. By contrast you could adopt a | Soviet-like approach where cat farts and cancer both share | replication resources, but this seems like it would be bad | for everyone (except the cats.) | advisedwang wrote: | It would be more than just half productivity. Not only do you | have to do the work twice, but you add the delay of someone | else replicating before something can be published and built | upon by others. If you are developer, imagine how much your | productivity would drop going from a 3 minute build to a 1 | day build. | orangepurple wrote: | Terrible analogy. It might take months to come up with an | idea but another should be able to follow your method and | implement it much more quickly than it took you to come up | with the concept and implement it. | magimas wrote: | horrible take. Taking the LK99 situation as an example: | simply copying and adapting a well described growth | recipee to your own setup and lab conditions may take | weeks. And how would you address situations where | measurement setups only exist once on the earth? How | would you do peer replication of LHC measurements? Wait | for 50 years till the next super-collider is built and | someone else can finally verify the results? On a smaller | scale: If you need measurements at a synchrotron | radiation source to replicate a measurement, is someone | supposed to give up his precious measurement time to | replicate a paper he isn't interested in? And is the | original author of a paper that's in the queue for peer | replication supposed to wait for a year or two till the | reviewer gets a beamtime on an appropriate measurement | station? Even smaller: I did my PhD in a lab with a | specific setup that only a single other group in the | world had an equivalent to. You simply would not be able | to replicate these results. | | Peer replication is completely unfeasible in experimental | fields of science. The current process of peer review is | alright, people just need to learn that single papers | standing by themselves don't mean too much. The "peer | replication" happens over time anyway when others use the | same tools, samples, techniques on related problems and | find results in agreement with earlier papers. | evandrofisico wrote: | Usually coming up with a idea is the _easy_ part. For | example, in my PhD project, i started with an idea from | my advisor that he had in the early 2000. | | Implementing the code for the simulation and analysis of | the data? four months, at most. Running the simulation? | almost three years until I had data with good enough | resolution for publishing. | tnecniv wrote: | It's also very easy to come up with bad ideas -- I did | plenty of that and I still do, albeit less than I used | to. Finding an idea that is novel, interesting, and | tractable given your time, skills, resources, and | knowledge of the literature is hard, and maybe the most | important skill you develop as a researcher. | | For a reductive example, the idea to solve P vs NP is a | great one, but I'm not going to do that any time soon! | cycomanic wrote: | I think you don't understand how much work is involved in | just building the techniques and expertise to pull some | experiments off (let's not even talk about the | equipment). | | Even if someone meticulously documents their process, it | could still take months to replicate the results. | | I'm familiar with lithography/nanofabrication and I know | that it is typically the case that a process developed in | one clean-room can not be directly applied to a different | clean room and instead one has to develop a new process | based on what the other results. | | Even in the same lab it can often happen that if you come | back to a process after a longer time, that things don't | work out anymore and quite a bit of troubleshooting | ensues (maybe a supplier for some chemical changed and | even though it should be the same formula it behaves | slightly different). | RoyalHenOil wrote: | Months. Haha. | | I previously worked in agricultural research (in the | private sector), and we spent YEARS trying to replicate | some published research from overseas. And that was | research that had previously been successfully | replicated, and we even flew in the original scientists | and borrowed a number of their PhD students for several | months, year after year, to help us try to make it work. | | We never did get it to fully replicate in our country. We | ended up having to make some pretty extreme changes to | the research to get similar (albeit less reliable) | results here. | | We never did figure out why it worked in one part of the | world but not another, since we controlled for every | other factor we could think of (including literally | importing the original team's lab supplies at great | expense, just in case there was some trace contaminant on | locally sourced materials). | harimau777 wrote: | The issue that I see is: even if halving productivity is | acceptable to the field as a whole; how do you incentivize a | given scientist to put in the effort? | | This seems particularly problematic because it is already | notoriously hard to get tenure and academia is already | notoriously unrewarding to researchers who don't have tenure. | hoosieree wrote: | Half is wildly optimistic. | ImPostingOnHN wrote: | half would only be possible if, for every single paper | published by a given team, there exists a second team just as | talented as the original team, skilled in that specific | package of techniques, just waiting to replicate that paper | coding123 wrote: | Maybe doing an experiment twice, even with a cost that is | double, makes more sense so that we don't all throw away our | coffee when coffee is bad, or throw away our gluten when gluten | is bad, etc... (those are trivial examples) basically the cost | to perform the science in many cases is so minuscule in scale | to how it could affect society. | pvaldes wrote: | One. Doing experiments is yet enough difficult and painful. | | Two. This drain of resources can't be done for free. Somebody | will need to pay twice for half of the research [1], and | faster. Peers will need to be hired and paid, maybe by the | writer's grants. Researchers cant justify to give their own | funds to other teams without a profound change in regulation | and even in that case would be harming their own projects. | | [1] as the valuable experts are now stuck validating things | instead doing their own job | | Would open also a door for foul play. Blocking competitors | teams in molasses just trowing them secondary silly problems | that they know that are a dead end, while the other team work | in the real deal, and take the advantage to win the patent. | mattkrause wrote: | Longer, even! | | Some experiments that study biological development or trained | animals can take a year or more of fairly intense effort to | _start_ generating data. | Maxion wrote: | A year? some data sets take decades to build up before | significant papers can be published on their data. | Replication of the dataset is just not feasible. | | This whole thread just shows how little the average HNer | knows about the academic sciences. | tnecniv wrote: | I know people that had to take a 6+ month trip to Antarctica | for part of their work and others that had to share time on a | piece of experimental equipment with a whole department -- | they got a few weeks per year to run their experiment and had | to milk that for all it's worth. Even if they had funding, | that machine required large amounts of space and staff to | keep it running and they aren't off the shelf products -- | only a few exist at large research centers. | seventytwo wrote: | There would need to be an incentive structure where the first | replications get (nearly) the same credit as the original | publisher. | j45 wrote: | Can every thing be replicated in every field | User23 wrote: | That's the defining characteristic of engineering. If you can't | reliably replicate everything in an engineering discipline then | it's not an engineering discipline. | Hiromy wrote: | Hola te amo | jimmar wrote: | How do you replicate a literature review? Theoretical physics? A | neuro case? Research that relies upon natural experiments? There | are many types of research. Not all of them lend themselves to | replication, but they can still contribute to our body of | knowledge. Peer review is helpful in each of these instances. | | Science is a process. Peer review isn't perfect. Replication is | important. But it doesn't seem like the author understands what | it would take to simply replace peer review with replication. | janalsncm wrote: | I don't think the existence of papers that are difficult to | replicate undermines the value of replicating those that are | easier. | freeopinion wrote: | My mind automatically swapped out the words "peer" for "code". It | took my brain to interesting places. When I came back to the | actual topic, I had accidentally built a great way to contrast | some of the discussion offered in this thread. | dongping wrote: | In the sense of replicating the results, we do have CI servers | and even fuzzers running for our "code replication". | freeopinion wrote: | I don't want to derail the science discussion too much, but | what if you actually had to reproduce the code by hand? Would | that process produce anything of value? Would your habit of | writing i+=1 instead of i++ matter? Or iteration instead of | recursion? | | Would code replication result in fewer use after free, or off | by one than code review? Or would it mostly be a waste of | resources including time? | abnry wrote: | If scientists are going to complain that's its too hard or too | expensive to replicate their studies, then that just shows their | work is BS. | fodkodrasz wrote: | I guess if software developers will complain that it's too hard | or too expensive to thoroughly test their code to ensure | exactly zero bugs at release[1], then that just shows their | work is BS. | | [1]: if you have delivered telco code to Softbank you may have | heard this sentence | abnry wrote: | Replication is not the same thing as zero bugs in software. | alsodumb wrote: | Nah, it doesn't. It just shows that it's time consuming and | expensive to replicate their studies. | abnry wrote: | If that's the case, then don't claim confidence in the work | or make policy decisions based off of it. If there is no | epistemological humility, then yes, it is still BS. | Levitz wrote: | If any study costs X, the study and the replication costs | somewhere in the ballpark of 2*X. This is not trivial. | abnry wrote: | But this is science we are talking about. A one-off lucky | novel result should not be good enough. Why should our | standards and our funding be so low? | Maxion wrote: | Something in switzerland called the Large Hadron Collider comes | to mind. | | I guess we should not talk about the Higgs before someone else | builds a second one and replicates the papers. | abnry wrote: | Physics is generally better since they have good statistical | models and can get six sigma (or whatever) results. | | And replication can be done by the same party (although an | independent party is better), and that may mean many trials. | | And do we even set policy based on existence or non-existent | of higgs bosons? | | I am particularly unhappy with soft sciences in terms of | replication. | azan_ wrote: | What if it REALLY is too expensive? You do realize that there | are studies which literally cost millions of dollars? Getting | funding for original studies is hard enough, good luck securing | additional funds for replication. | snitty wrote: | >If scientists are going to complain that's its too hard or too | expensive to replicate their studies, then that just shows | their work is BS. | | 1 mg of anti-rabbit antibody (a common thing to use in a lot of | biology experiments) is $225 [1]. Outside of things like | standard buffers and growth medium for prokaryotes, this is | going to be the cheapest thing you use in an experiment. | | 1/10th of that amount for anti-flagellin antibody is $372. [2] | | A kit to prep a cell for RNA sequencing is $6-10 per use. | That's JUST isolation of the RNA. Not including reverse | transcribing it to cDNA for sequencing, or the sequencing | itself. [3] | | Let's not even reach things like materials science where you | may be working on an epitaxial growth paper, and there are only | a handful of labs where they could even feasibly repeat the | experiment. | | Or say something with a BSL-3 lab where there are literally | only 15 labs in the US that could feasibly do the work, | assuming they aren't working on their own stuff. [4] | | [1] - https://www.thermofisher.com/antibody/product/Goat-anti- | Rabb... [2] https://www.invivogen.com/anti-flagellin [3] | https://www.thermofisher.com/order/catalog/product/12183018A | [4] https://www.niaid.nih.gov/research/tufts-regional- | biocontain... | NalNezumi wrote: | Imo, A more realistic thing to do is "replicability review" | and/or requirement to submit "methodology map" to each paper. | | The former would be a back and forth between a reviewer that | inquire and ask questions (based on the paper) with the goal to | _reproduce the result_ , but don't have to actually reproduce it. | This is usually good to find out missing details in the paper | that the writer just took for granted everyone in the field knows | (I've met Bio PHD that have wasted Months of their life tracking | up experimental details not mentioned in a paper) | | The latter would be the result of the former. Instead of having | pages long "appendix" section in the main paper, you produce | another document with meticulous details of the | experiment/methodology with every stone turned together with an | peer reviewer. Stamp it with the peer reviewes name so they can't | get away with hand wavy review. | | I've read too many papers where important information to | reproduce the result is omitted. (for ML/RL) If the code is | included I've countless of times found implementation details | that is not mentioned in the paper. In matter of fact, there's | even results suggesting that those details are the make or break | of certain algorithms. [1] I've also seen breaking details only | mentioned in code comments... | | Another atrocious thing I've witnessed is a paper claiming they | evaluated their method on a benchmark and if you check the | benchmark, the task they evaluated on doesn't exit! They forked | the benchmark and made their own task without being clear about | it! [2] | | Shit like this make me lose faith in certain science directions. | And I've seen a couple of junior researcher giving it all up | because they concluded it's all just house of cards. | | [1] https://arxiv.org/abs/2005.12729 | | [2] https://arxiv.org/abs/2202.02465 | | Edit: also if you think that's too tedious/costly, reminder that | publishers rake in record profits so the resources are already | there https://youtu.be/ukAkG6c_N4M | kergonath wrote: | > I've met Bio PHD that have wasted Months of their life | tracking up experimental details not mentioned in a paper | | Same. Now, when I review manuscripts, I pay much more attention | to whether there is enough information to replicate the | experiment or simulation. We can put out a paper with wrong | interpretations and that's fine because other people will | realise that when doing their own work. We cannot let papers | get published if their results cannot be replicated. | | > The latter would be the result of the former. Instead of | having pages long "appendix" section in the main paper, you | produce another document with meticulous details of the | experiment/methodology with every stone turned together with an | peer reviewer. Stamp it with the peer reviewes name so they | can't get away with hand wavy review | | Things that take too much space to go in the experimental | section should go to a electronic supplementary information | document. But then it would be nice if the ESI were appended to | the article when we download a PDF because tracking them is a | pain in the backside. Some fields are better than others about | this, for example in materials characterisation studies it's | very common to have ESI with a whole bunch of data and details. | | Large dataset should go to a repository or a dataset journal, | that way the method is still peer reviewed and the dataset has | a doi and is much easier to re-use. It's also a nice way of | doubling a student's papers count by the end of their PhD. | | > Another atrocious thing I've witnessed is a paper claiming | they evaluated their method on a benchmark and if you check the | benchmark, the task they evaluated on doesn't exit! They forked | the benchmark and made their own task without being clear about | it! [2] | | That's just evil! | Maxion wrote: | > Large dataset should go to a repository or a dataset | journal, that way the method is still peer reviewed and the | dataset has a doi and is much easier to re-use. | | This may be possible in some sciences, but not in | epidemiology or biomed. Often the study is based on tissue | samples owned by some entity, with permission granted only to | some certain entity. | | Datasets in epidemiology are often full of PII, and cannot be | shared publicly for many reasons. | infogulch wrote: | I like the idea of splitting "peer review" into two, and then | having a citation threshold standard where a field agrees that a | paper should be replicated after a certain number of citations. | And journals should have a dedicated section for attempted | replications. | | 1. Rebrand peer review as a "readability review" which is what | reviewers tend to focus on today. | | 2. A "replicability statement", a separately published document | where reviewers push authors to go into detail about the | methodology and strategy used to perform the experiments, | including specifics that someone outside of their specialty may | not know. Credit NalNezumi ITT | analog31 wrote: | Every experimental paper I've ever read has contained an | "Experimental" section, where they provide the details on how | they did it. Those sections tend to be general enough, albeit | concise. | | In some fields, aside from specialized knowledge, good | experimental work requires what we call "hands." For instance, | handling air sensitive compounds, or anything in a condensed or | crystalline state. In my thesis experiment, some of the | equipment was hand made, by me. | | Sometimes specialized facilities are needed. My doctoral thesis | project used roughly 1/2 million dollars of gear, and some of | the equipment that I used was obsolete and unavailable by the | time I finished. | ahmadmijot wrote: | > My doctoral thesis project used roughly 1/2 million dollars | of gear, | | Wow I envy you. My doctoral thesis project spent like... | USD2.5k directly for gears (half of it just to buy lego | bricks to build our own instrument exactly because we can't | afford to buy commercial one lol) | xioxox wrote: | I used a 3 billion dollar space telescope. I don't think | NASA are going to launch another to replicate some of my | results. | janalsncm wrote: | "Concise" isn't good enough. If other scientists are trying | to read through the tea leaves at what you're trying to say | you did, that defeats the entire point of a paper. The | purpose of science is to create knowledge _that other people | can use_ and if people can't replicate your work that's not | science. | analog31 wrote: | I think the point is you don't have to give a complete BOM | that includes where you got the power cables. Each | scientist has to decide what amount of information needs to | be conveyed. Of course this can be abused, or done | sloppily, like anything else. | | A place where you can spread out more is in dissertations. | Mine contained an entire chapter on the experiment, another | on the analysis, and appendices full of source code, | schematics, etc. I happily sent out copies, at my expense. | My setup was replicated roughly 3 times. | User23 wrote: | One thing that everyone needs to remember about "peer review" is | that it isn't part of the scientific method, but rather that it | was imposed on the scientific enterprise by government funding | authorities. It's basically JIRA for scientists. | ahmadmijot wrote: | Quite related: nowadays there is this movement within scientific | researches ie Open Science where the (raw) data from ones | research is open source. And even methods for in-house | fabrication and development together with its source code is open | source (open hardware and open software) | waynecochran wrote: | I spent a lot of my graduate years in CS implementing the details | of papers only to learn that, time and time again, the paper | failed to mention all the short comings and fail cases of the | techniques. There are great exceptions to this. | | Due to the pressure of "publish or die" there is very little | honesty in research. Fortunately there are some who are | transparent with their work. But for the most part, science is | drowning in a sea of research that lacks transparency and | replication short falls. | janalsncm wrote: | I had a very similar experience in my masters. Really made me | think, what exactly are the peers "reviewing" if they don't | even know whether the technique works in the first place. | waynecochran wrote: | I have reviewed many papers and there is never the time to | recreate the work and test. That is why I love the "papers w | code" site. I think every published CS paper should require a | git repo with all their code and experimental data. | cptskippy wrote: | You'll quickly discover when you enter the workforce that the | reasons we have CI/CD, Docker, and virtualization are because | of a similar problem. The dread "it works on my machine" | response. | | CI/CD forces people to codify exactly how to build and deploy | something in order for it to get into a production environment. | Docker and VMs are ways around this by giving people a "my | machine" that can be copied and shared easily. | titzer wrote: | In the PL field, conferences have started to allow authors to | submit packaged artifacts (typically, source code, input data, | training data, etc) that are evaluated separately, typically | post-review. The artifacts are evaluated by a separate committee, | usually graduate students. As usual, everything is volunteer. | Even with explicit instructions, it is hard enough to even get | the same _code_ to run in a different environment and give the | same results. Would "replication" of a software technique | require another team to reimplement something from scratch? That | seems unworkable. | | I can't even _imagine_ how hard it would be to write instructions | for another lab to successfully replicate an experiment at the | forefront of physics or chemistry, or biology. Not just the | specialized equipment, but we 're talking about the frontiers of | Science with people doing cutting-edge research. | | I get the impression that suggestions like these are written by | non-scientists who do not have experience with the peer review | process of _any_ discipline. Things just don 't work like that. | Maxion wrote: | > I get the impression that suggestions like these are written | by non-scientists who do not have experience with the peer | review process of any discipline. Things just don't work like | that. | | Not to mention that the cutting edge in many sciences are | perhaps two-three research groups of 5-30 individuals each in | varying research institutions around the world. | mike_hearn wrote: | Is PL theory actually science? Although we call it computer | science, I don't personally think CS is actually a science in | the sense of studying nature to understand it. Computers are | artificial constructs. CS is a lot closer to engineering than | science. Indeed it's kind of nonsensical to talk about | replicating an experiment in programming language theory. | | For the "hard" sciences, replication often isn't so difficult | it seems. LK-99 being an interesting study in this, where | people are apparently successfully replicating an experiment | described in a rushed paper that is widely agreed to lack | sufficient details. It's cutting edge science but replication | still isn't a problem. Most science isn't the LHC. | | The real problems with replication are found in the softer | fields. There it's not just an issue of randomness or | difficulty of doing the experiments. If that's all there was to | it, no problem. In these fields it's common to find papers or | entire fields where none of the work is replicable even in | principle. As in, the people doing it don't think other people | being able to replicate their work is even important at all, | and they may go out of their way to _stop_ people being able to | replicate their work (most frequently by gathering data in non- | replicable ways and then withholding it deliberately, but | sometimes it 's just due to the design of the study). The most | obvious inference when you see this is that maybe they don't | want replication attempts because they know their claims | probably aren't true. | | So even if peer reviewers or journals were just checking really | basic things like, is this claim even replicable in principle, | that would be a good start. You would still be left with a lot | of papers that replicate fine but their conclusions are still | wrong because their methodology is illogical, or papers that | replicate because their findings are obvious. But there's so | much low hanging fruit. | staunton wrote: | Let's get people to publish their data and code first, shall we? | That's sooo much easier than demanding whole studies to be | replicated... and people still don't do it! | ayakang31415 wrote: | One of the Nobel prizes in Physics was the discovery of Higgs | Boson at LHC. It cost billions of dollars just to build the | facility, and required hundreds of physicists working on it to | just conduct the experiment. You can't replicate this. Although I | fully agree that replication must come first when it is | reasonably doable. | TrackerFF wrote: | Seems to have been hugged to death. | | But - a quick counterexample - as far as replication goes: What | if the experiments were run on custom made or exceedingly | expensive equipment? How are the replicators supposed to access | that equipment? Even in fields which are "easy" to replicate - | like machine learning - we are seeing barriers of entry due to | expensive computing power. Or data collection. Or both. | | But then you move over to physics, and suddenly you're also | dealing with these one-off custom setups, doing experiments which | could be close to impossible to replicate (say you want to | conduct experiments on some physical event that only occurs every | xxxx years or whatever) | pajushi wrote: | Why shouldn't we hold science more accountable? | | "Science needs accounting" is a search I had saved for months | which really resonates with the idea of "peer replication." | | In accounting, you always have checks and balances, you never are | counting money alone. In many cases, accountants duplicate their | work to make sure that it is accurate. | | Auditors are the corollary to the peer review process. They're | not there to redo your work, but to verify that your methods and | processes are sound. | paulpauper wrote: | this would not apply to math or something subjective such as | literature. only experimental results need to be replicated. | Nevermark wrote: | Reproducibility would become a much higher priority if electronic | versions of papers are required (by their distributors, archives, | institutions, ...) to have reproduction sections, which the | authors are encouraged to update over time. | | UPDATABLE COVER PAGE: | | Title Authors | | Abstract Blah, blah, ... | | State of reproduction: Not reproduced. | Successful reproductions: ...citations... Reproduction | attempts: ...citations... Countering reproductions: | ...citations... | | UPDATABLE REPRODUCTION SECTION ATTACHED AT END | | Reproduction resources: Data, algorithms, | processes, materials, ... | | Reproduction challenges: Cost, time, one-off | events, ... | | Making this stuff more visible would help reproducers validated | the value of reproduction to their home and funding institutions. | | Having a standard section for this, with an initial state of "Not | reproduced" provides more incentive for original workers to | provide better reproduction info. | | For algorithm and math work the reproduction could be served best | with downloadable executable bundle. | gordian-not wrote: | The incentive should be to clear the way for tenure track | | The junior faculty will clear the rotten apples at the top by | finding flaws in their research and then will win the tenure that | was lost in return | | This will create a nice political atmosphere and improve science | user6723 wrote: | I remember showing someone raw video of a Safire plasma chamber | keeping the ball of plasma lit for several minutes. They said | they would need to see a peer reviewed paper. The presumption | brought about by the enlightenment era that everyone should get a | vote was a mistake. | dongping wrote: | https://web.archive.org/web/20230130143126/https://blog.ever... | moelf wrote: | I wish we can replicate the LHC | Maxion wrote: | No talking about the Higgs before that happens, apparently. | kergonath wrote: | We will, don't worry. | janalsncm wrote: | For a while Reddit had the mantra "pics or it didn't happen". | | At least in CS/ML there needs to be a "code or it didn't happen". | Why? Papers are ambiguous. Even if they have mathematical | formulas, not all components are defined. | | Peer replication in these fields is an easy low hanging fruit | that could set an example for other fields of science. | simlan wrote: | That is too simplistic. You underestimate the depth of | academia. Sure the latest break through Alzheimers study or | related research would benefit from a replication. Which is | done out of commercial interest anyway. | | But your run of the mill niche topic will not have the dollars | behind it to replicate everyones research.just because CS/AI | research is very convenient to replicate does not mean this can | be extended to all research being done. | | That is exactly why peer review exists to weed out the | implausible and low effort/relevance work. It is not fraud | proof because it was not designed to be. | hedora wrote: | The website dies if I try to figure out who the author ("sam") | is, but it sounds like they are used to some awful backwater of | academia. | | They have this idea that a single editor screens papers to decide | if they are uninteresting or fundamentally flawed, then they want | a bunch of professors to do grunt work litigating the correctness | of the experiments. | | In modern (post industrial revolution) branches of science, the | work of determining what is worthy of publication is distributed | amongst a program committee, which is comprised of reviewers. The | editor / conference organizers pick the program committee. There | are typically dozens of program committee members, and authors | and reviewers both disclose conflicts. Also, papers are | anonymized, so the people that see the author list are not | involved in accept/reject decisions. | | This mostly eliminates the problem where work is suppressed for | political reasons, etc. | | It is increasingly common for paper PDFs to be annotated with | badges showing the level of reproducibility of the work, and | papers can win awards for being highly reproducible. The people | that check reproducibility simply execute directions from a | separate reproducibility submission that is produced after the | paper is accepted. | | I argue the above approach is about 100 years ahead of what the | blog post is suggesting. | | Ideally, we would tie federal funding to double blind review and | venues with program committees, and papers selected by editors | would not count toward tenure at universities that receive public | funding. | jltsiren wrote: | The computer science practice you describe is the exception, | not the norm. It causes a lot of trouble when evaluating the | merits of researchers, because most people in the academia are | not familiar with it. In many places, conference papers don't | even count as real publications, putting CS researchers at a | disadvantage. | | From my point of view, the biggest issue is accepting/rejecting | papers based on first impressions. Because there is often only | one round of reviews, you can't ask the authors for | clarifications, and they can't try to fix the issues you have | identified. Conferences tend to follow fashionable topics, and | they are often narrower in scope than what they claim to be, | because it's easier to evaluate papers on topics the program | committee is familiar with. | | The work done by the program committee was not even supposed to | be proper peer review but only the first filter. Old conference | papers often call themselves extended abstracts, and they don't | contain all the details you would expect in the full paper. For | example, a theoretical paper may omit key proofs. Once the | program committee has determined that the results look | interesting and plausible and the authors have presented them | in a conference, the authors are supposed to write the full | paper and submit it to a journal for peer review. Of course, | this doesn't always happen, for a number of reasons. | cycomanic wrote: | While I agree with the general sentiment of the paper and | creating incentives for more replication is definitely a good | idea, I do think the approach is flawed in several ways. | | The main point is that the paper seriously underestimates the | difficulty and time it requires to replicate experiments in many | experimental fields. Who will decide which work needs to be | replicated? Should capable labs somehow become bogged down with | just doing replication work? Even if they don't find the results | not interesting? | | In reality if labs find results interesting enough to replicate | they will try to do so. The current LK-99 hurrah is a perfect | example of that, but it happens on a much smaller scale all the | time. Researchers do replicate and build on other work all the | time, they just use that replication to create new results (and | acknowledge the previous work) instead of publishing a "we | replicated paper". | | Where things usually fail is in publication of "failed | replication" studies, and those are tricky. It is not always | clear if the original research was flawed or the people trying to | reproduce made an error (again just have a look at what's | happening with LK-99 at the moment). Moreover, it can be | politically difficult to try to publish a "fail to reproduce" | result if you are small unknown lab, if the original result came | from a big known group. Most people will believe that you are the | one who made the error (and unfortunately big egos might get in | the way, and the small lab will have a hard time). | | More generally, in my opinion the lack of replication of results | is just one symptom of a bigger problem in science today. We (as | in society) have essentially turned the scientific environment | increasingly competitive, under the guise of "value for tax payer | money". Academic scientists now have to constantly compete for | grant funding, publish to keep the funding going. It's incredibly | competitive to even get in ... At the same time they are supposed | to constantly provide big headlines for university press | releases, communicate their results to the general public and | investigate (and patent) the potential for commercial | exploitation. No wonder we see less cooperation. | eesmith wrote: | > the real test of a paper should be the ability to reproduce its | findings in the real world. ... | | > What if all the experiments in the paper are too complicated to | replicate? Then you can submit to [the Journal of Irreproducible | Results]. | | Observational science is still a branch of science even if it's | difficult or impossible to replicate. | | Consider the first photographs of a live giant squid in its | natural habitat, published in 2005 at | https://royalsocietypublishing.org/doi/10.1098/rspb.2005.315... . | | Who seriously thinks this shouldn't have been published until | someone else had been able to replicate the result? | | Who thinks the results of a drug trial can't be published until | they are replicated? | | How does one replicate "A stellar occultation by (486958) 2014 | MU69: results from the 2017 July 17 portable telescope campaign" | at | https://ui.adsabs.harvard.edu/abs/2017DPS....4950403Z/abstra... | which required the precise alignment of a star, the trans- | Neptunian object 486958 Arrokoth, and a region in Argentina? | | Or replicate the results of the flyby of Pluto, or flying a | helicopter on Mars? | | Here's a paper I learned about from "In The Pipeline"; "Insights | from a laboratory fire" at | https://www.nature.com/articles/s41557-023-01254-6 . | | """Fires are relatively common yet underreported occurrences in | chemical laboratories, but their consequences can be devastating. | Here we describe our first-hand experience of a savage laboratory | fire, highlighting the detrimental effects that it had on the | research group and the lessons learned.""" | | How would peer replication be relevant? | phpisthebest wrote: | I think in some of those cases you have conclusions drawn from | raw data that could be replicated or reviewed. For example many | teams use the same raw data from Large Colliders, or JWT, or | other large science projects to reach competiting conclusions. | | Yes in a perfect world we would also replicate the data | collection but we do not live in a perfect world | | Same is true for Drug Trials, there is always a battle over | getting the raw data from drug trails as the companies claim | that data is trade secret, so independent verification of drug | trails is very expensive but if the FDA required not just the | release of redacted conclusions and supporting redacted data | but 100% of all data gathered it would be alot better IMO | | For example the FDA says it will take decades to release the | raw data from the COVID Vaccine trials.. Why... and that is | after being forced to do so via a law suit. | eesmith wrote: | > For example many teams use the same raw data from Large | Colliders, or JWT, or other large science projects to reach | competiting conclusions. | | Yes, but why must the first team wait until the second is | finished before publishing? | | What if you are the only person in the world with expertise | in the fossil record of an obscure branch of snails? You | spend 10 years developing a paper knowing that the next | person with the right training to replicate the work might | not even be born yet. | | Other paleontologists might not be able to replicate the | work, but still tell if it's publishable - that's what they | do now, yes? | | > but we do not live in a perfect world | | Alternatively, we don't live in a perfect world which is why | we have the current system instead of requiring replication | first. | | Since the same logic works for both cases, I don't think it's | persuasive logic. | | > the FDA says it will take decades | | Well, that's a tangent. The FDA is charged with protecting | and promoting public health, not improving the state of | scholarly literature. | | And the FDA is only one of many public health organizations | which carried out COVID vaccine trials. | msla wrote: | With some of the things, but admittedly not most of the things | you mentioned, there's a dataset (somewhere) and some code run | on that dataset (somewhere) and replication would mean someone | else being able to run that code on that dataset and get the | same results. | | Would this require labs to improve their software environments | and learn some new tools? Would this require labs to give up | whatever used to be secret sauce? That's. The. Point. | counters wrote: | In practice this is happening in many disciplines, for most | research, on a daily basis. What _isn't_ happening is that | the results of these replications are being independently | peer reviewed, because that isn't incentivized. However, when | replication fails for whatever reason, it usually leads to | insights that themselves lead to stronger scientific work and | better publications later on. | eesmith wrote: | > someone else being able to run that code on that dataset | and get the same results. | | I think when people talk about "replicate" they mean | something more than that. | | The dataset could contain coding errors, and the analysis | could contain incorrect formulas and bad modeling. | Reproducing a bad analysis, successfully, provide no | corrective feedback. | | I know for one paper I could replicate the paper's results | using the paper's own analysis, but I couldn't replicate the | paper's results using my analysis. | | > Would this require labs to give up whatever used to be | secret sauce? That's. The. Point. | | That seems to be a very different Point. | | Newton famously published results made from using his secret | sauce - calculus - by recasting them using more traditional | methods. | | In the extreme cas, I could publish the factors for RSA-1024 | without publishing my factorization method. "I prayed to God | for the answer and He gave them to me." You can verify that | result without the secret sauce. | | I mean, people use all sorts of methods to predict a protein | structure, including manual tweaking guided by intuition and | insight gained during a reverie or day-dream (a la Kekule) | which is clearly not reproducible. Yet that final model may | be publishable, because it may provide new insight and | testable predictions. | msla wrote: | My point is that we can, apparently, improve the baseline | expectations in the parts of science where this kind of | reproducibility is possible. That isn't all science, | granted, but it is some science. It isn't a panacea, | granted, but it could guard against some forms of | misconduct or honest error some of the time. The self- | correcting part of science only works when there's | something for it to work on, so open data and runnable code | ought to improve that self-correction mechanism. | eesmith wrote: | Understood. | | But my point is this linked-to essay appears not only to | exclude some areas of good science, but to suggest that | any topics which cannot be replicated before publication | is only worthy of publication in the Journal of | Irreproducible Results. | | I gave examples to highlight why I disagree with author's | opinion. | | Please do not interpret this to mean I do not think | improvement is possible. | kergonath wrote: | > Who seriously thinks this shouldn't have been published until | someone else had been able to replicate the result? | | Nobody, obviously. You cannot reproduce a result that hasn't | been published, so no new phenomenon is replicated the moment | it is first published. The problem is not the publication of | new discoveries, it's the lack of incentives to confirm them | once they've been published. | | In your example, new observations of giant squids are still | massively valuable even if not that novel anymore. So new | observations should be encouraged (as I am sure they are). | | > Or replicate the results of the flyby of Pluto, or flying a | helicopter on Mars? | | Well, we should launch another probe anyway. And I am fairly | confident we'll have many instances of aircrafts in Mars' | atmosphere and more data than we'll know what to do with it. We | can also simulate the hell out of it. We'll point spectrometers | and a whole bunch of instruments towards Pluto. These are not | really good examples of unreproducible observations. | | Besides, in such cases robustness can be improved by different | teams performing their own analyses separately, even if the | data comes from the same experimental setup. It's not all black | or white. Observations are on a spectrum, some of them being | much more reliable than others and replication is one aspect of | it. | | > How would peer replication be relevant? | | How would you know which aspects of the observed phenomena come | from particularities of this specific lab? You need more than | one instance. You need some kind of statistical and factor | analyses. Replication in this instance would not mean setting | actual labs on fire on purpose. | | It's exactly like studying car crashes: nobody is going to kill | people on purpose, but it is still important to study them so | we regularly have new papers on the subject based on events | that happened anyway, each one confirming or disproving | previous observations. | eesmith wrote: | > Nobody, obviously. You cannot reproduce a result that | hasn't been published, .. The problem is not the publication | of new discoveries, it's the lack of incentives to confirm | them once they've been published. | | Your comment concerns post-publication peer-replication, yes? | | If so, it's a different topic. The linked-to essay | specifically proposes: | | ""Instead of sending out a manuscript to anonymous referees | to read and review, preprints should be sent to other labs to | actually replicate the findings. Once the key findings are | replicated, the manuscript would be accepted and published."" | | That's _pre-publication_ peer-replication, and my comment was | only meant to be interpreted in that light. | kergonath wrote: | > That's pre-publication peer-replication, and my comment | was only meant to be interpreted in that light. | | Sorry I might have gone mixed up between threads. | | Yeah, pre-publication replication is nice (I do it when I | can and am suspicious of some simulation results), but is | not practical at scale. Besides, the role of peer review is | not to ensure results are right, that is just not | sustainable for referees. | hinkley wrote: | Is there space in the world for a few publications that only | publish replicated work? Seems like that would be a reasonable | compromise. Yes you were published, but were you published in | Really Real Magazine? Get back to us when you have and we'll | discuss. | hospadar wrote: | I assume that the goal here is to reduce the number of not- | actually-valid results that get published. Not-actually-valid | results happen for lots of reasons (whoops did experiment wrong, | mystery impurity, cherry picked data, not enough subjects, | straight-up lie, full verification expensive and time consuming | but this looks promising) but often there's a common set of | incentives: you must publish to get tenure/keep your job, you | often need to publish in journals with high impact factor [1]. | | High impact journals [6] tend to prefer exciting, novel, and | positive results (we tried new thing and it worked so well!) vs | negative results (we mixed up a bunch of crystals and absolutely | none of them are room-temp superconductors! we're sure of it!). | | The result is that cherry picking data pays, leaning into | confirmation bias pays, publishing replication studies and | rigorous but negative results is not a good use of your academic | inertia. | | I think that creating a new category of rigor (i.e. journals that | only publish independently replicated results) is not a bad idea, | but: who's gonna pay for that? If the incentive is you get your | name on the paper, doesn't that incentivize coming up with a | positive result? How do you incentivize negative replications? | What if there is only one gigantic machine anywhere that can find | those results (LHC, icecube, etc, a very expensive spaceship)? | | There might be easier and cheaper pathways to reducing bad papers | - incentivizing the publishing of negative results and | replication studies separately, paying reviewers for their time, | coming up with new metrics for researchers that prioritize | different kinds of activity (currently "how much you're cited" | and "number of papers*journal impact" things are common, maybe a | "how many results got replicated" score would be cool to roll | into "do you get tenure"? See [3] for more details). PLoS | publish. | | I really like OP's other article about a hypothetical "Journal of | One Try" (JOOT) [2] to enable publishing of not-very-rigorous- | but-maybe-useful-to-somebody results. If you go back and read OLD | OLD editions of Philosophical Transactions (which goes back to | the 1600's!! great time, highly recommend [4], in many ways the | archetype for all academic journals), there are a ton of wacky | submissions that are just little observations, small experiments, | and I think something like that (JOOT let's say) tuned up for the | modern era would, if nothing else, make science more fun. Here's | a great one about reports of "Shining Beef" (literally beef that | is glowing I guess?) enjoy [5] | | [1] https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6668985/ [2] | https://web.archive.org/web/20220924222624/https://blog.ever... | [3] https://www.altmetric.com/ [4] | https://www.jstor.org/journal/philtran1665167 [5] | https://www.jstor.org/stable/101710 [6] | https://en.wikipedia.org/wiki/Impact_factor, see also | https://clarivate.com/ | throwawaymaths wrote: | How about we create a Nobel prize for replication. One impressive | replication or refutation from last decade (that holds up) gets | the prize split up to three ways among the most important | authors. | 37326 wrote: | [flagged] | elashri wrote: | Great, but who is going to fund the peer replication?. The | economics of research now doesn't even provide a compensation for | peer review process time. | nine_k wrote: | Maybe the numerous complaints about the crisis of science are | somehow related to the fact that scientific work is severely | underpaid. | | The pay difference between research and industry in many areas | is not even funny. | matthewdgreen wrote: | The purpose of science publications is to share new results with | other scientists, so others can build on or verify the | correctness of the work. There has always been an element of | "receiving credit" to this, but the communication aspect is what | actually matters _from the perspective of maximizing scientific | progress._ | | In the distant past, publication was an informal process that | mostly involved mailing around letters, or for a major result, | self-publishing a book. Eventually publishers began to devise | formal journals for this purpose, and some of those journals | began to receive more submissions than it was feasible to publish | or verify just by reputation. Some of the more popular journals | hit upon the idea of applying basic editorial standards to reject | badly-written papers and obvious spam. Since the journal editors | weren't experts in all fields of science, they asked for | volunteers to help with this process. That's what peer review is. | | Eventually bureaucrats (inside and largely outside of the | scientific community) demanded a technique for measuring the | productivity of a scientist, so they could allocate budgets or | promotions. They hit on the idea of using publications in a few | prestigious journals as a metric, which turned a useful process | (sharing results with other scientists) into [from an outsider | perspective] a process of receiving "academic points", where the | publication of a result appears to be the end-goal and not just | an intermediate point in the validation of a result. | | Still other outsiders, who misunderstand the entire process, are | upset that intermediate results are sometimes incorrect. This | confuses them, and they're angry that the process sometimes | assigns "points" to people who they perceive as undeserving. So | instead of simply accepting that _sharing results widely to | maximize the chance of verification_ is the whole point of the | publication process, or coming up with a better set of promotion | metrics, they want to gum up the essential sharing process to | make it much less efficient and reduce the fan-out degree and | rate of publication. This whole mess seems like it could be | handled a lot more intelligently. | nine_k wrote: | For sharing results widely, there's arxiv. The problem is that | the fanout is now overwhelming. | | The public perception of a publication in a prestigious journal | as the established truth does not help, too. | isaacremuant wrote: | > The public perception of a publication in a prestigious | journal as the established truth does not help, too. | | it's not so much the public perception but what | govs/media/tech and other institutions have pushed down so | that the public doesn't question whatever resulting policy | they're trying to put forth. | | "Trust the science" means "Thou shalt not question us, simply | obey". | | Anyone with eyes who has worked in institutions knows that | bureocracy, careerism and corruption are intrinsic to them. | casualscience wrote: | Most of this is very legit, but this | | > Still other outsiders, who misunderstand the entire process, | are upset that intermediate results are sometimes incorrect. | This confuses them, and they're angry that the process | sometimes assigns "points" to people who they perceive as | undeserving. So instead of simply accepting that sharing | results widely to maximize the chance of verification is the | whole point of the publication process, or coming up with a | better set of promotion metrics, they want to gum up the | essential sharing process to make it much less efficient and | reduce the fan-out degree and rate of publication. | | Does not represent my experience in the academy at all. There | is a ton of gamesmanship in publishing. That is ultimately the | yardstick academics are measured against, whether we like it or | not. No one misunderstands that IMO, the issue is that it's a | poor incentive. I think creating a new class of publication, | one that requires replication, could be workable in some fields | (e.g. optics/photonics), but probably is totally impossible in | others (e.g. experimental particle physics). | | For purely intellectual fields like mathematics, theoretical | physics, philosophy, you probably don't need this at all. Then | there are 'in the middle fields' like machine learning which in | theory would be easy to replicate, but also would be | prohibitively expensive for, e.g. baseline training of LLMs. | Maxion wrote: | And on the extreme end you have the multi-decade longitudinal | studies in epidemiology / biomedicine that would be more-or- | less impossible to replicate. | [deleted] | sebastos wrote: | Very well put. This is the clearest way of looking at it in my | view. | | I'll pile on to say that you also have the variable of how the | non-scientist public gleans information from the academics. | Academia used to be a more insular cadre of people seeking | knowledge for its own sake, so this was less relevant. What's | new here is that our society has fixated on the idea that | matters of state and administration should be significantly | guided by the results and opinions of academia. Our enthusiasm | for science-guided policy is a triple whammy, because 1. | Knowing that the results of your study have the potential to | affect policy creates incentives that may change how the | underlying science is performed 2. Knowing that results of | academia have outside influence may change WHICH science is | performed, and draw in less-than-impartial actors to perform it | 3. The outsized potential impact invites the uninformed public | to peer into the world of academia and draw half-baked | conclusions from results that are still preliminary or | unreplicated. Relatively narrow or specious studies can gain a | lot of undue traction if their conclusions appear, to the | untrained eye, to provide a good bat to hit your opponent with. | Maxion wrote: | A significant problem we face today is the way research, | especially in academia, gets spotlighted in the media. They | often hyper-focus on single studies, which can give a skewed | representation of scientific progress. | | The reality is that science isn't about isolated findings; | it's a cumulative effort. One paper might suggest a | conclusion, but it's the collective weight of multiple | studies that provides a more rounded understanding. Media's | tendency to cherry-pick results often distorts this nuanced | process. | | It's also worth noting the trend of prioritizing certain | studies, like large RCTs or systematic reviews, while | overlooking smaller ones, especially pilot studies. Pilot | studies are foundational--they often act as the preliminary | research needed before larger studies can even be considered | or funded. By sidelining or dismissing these smaller, | exploratory studies, we risk undermining the very foundation | that bigger, more definitive research efforts are built on. | If we consistently ignore or undervalue pilot studies, the | bigger and often more impactful studies may never even see | the light of day. | dmbche wrote: | Your analysis seems to portray all scientists as pure hearted. | May I remind you of the latest Stanford scandal where the | president of Stanford was found to have manipulated data? | | Today, publications do not serve the same purpose as they did | before the internet. It is trivial today to write a convincing | paper without research and getting that | published(www.theatlantic.com/ideas/archive/2018/10/new-sokal-h | oax/572212/&sa=U&ved=2ahUKEwjnp5mRtsiAAxVwF1kFHesBDC8QFnoECAkQA | g&usg=AOvVaw0t_Bo31BrT5D9zHBdmNAqi). | matthewdgreen wrote: | No subset of humanity is "pure hearted." Fraud and malice | will exist in everything people do. Fortunately these | fraudulent incidents seem relatively rare, when one compares | the number of reported incidents to the number of | publications and scientists. But this doesn't change | anything. The benefit of scientific publication is _to make | it easier to detect and verify incorrect results_ , which is | exactly what happened in this case. | | I understand that it's frustrating it didn't happen | instantly. And I also understand that it's deeply frustrating | that some undeserving person accumulated status points with | non-scientists based on fraud, and that let them take a high- | status position outside of their field. (I think maybe you | should assign some blame to the Stanford Trustees for this, | but that's up to you.) None of this means we'd be better off | making publication more difficult: it means the metrics are | bad. | | PS When a TFA raises something like "the replication crisis" | and then entangles it with accusations of deliberate fraud | (high profile but exceedingly rare) it's like trying to have | a serious conversation about automobile accidents, but | spending half the conversation on a handful of rare incidents | of intentional vehicular homicide. You're not going to get | useful solutions out of this conversation, because it's | (perhaps deliberately) misunderstanding the impact and causes | of the problem. | mike_hearn wrote: | Fraud isn't exceedingly rare :( It only seems that way | because academia doesn't pay anyone to find it, reacts to | volunteer reports by ignoring it, and the media generally | isn't interested. | | Fraud is so frequent and easy to find that there are | volunteers who in their spare time manage to routinely | uncover not just individual instances of fraud but entire | companies whose sole purpose is to generate and sell fake | papers on an industrial scale. | | https://www.nature.com/articles/d41586-023-01780-w | | Fraud is so easy and common that there are a steady stream | of journals which publish entire editions consisting of | nothing but AI generated articles! | | https://www.nature.com/articles/d41586-021-03035-y | | Despite being written as a joke over a decade ago, you can | page through an endless stream of papers that were | generated by SciGen - a Perl script - and yet they are | getting published: | | https://pubpeer.com/search?q=scigen | | The problem is so prevalent that some people created the | Problematic Paper Screener, a tool that automatically | locates articles that contain text indicative of auto- | generation. | | https://dbrech.irit.fr/pls/apex/f?p=9999:1:::::: | | This is all pre-ChatGPT, and is just the researchers who | can't be bothered writing a paper at all. The more serious | problem is all the human written fraudulent papers with bad | data and bad methodologies that are never detected, or only | detected by randos with blogs or Twitter accounts that you | never hear around. | dmbche wrote: | Thanks you - just discovered Scigen, these links are | incredible | dmbche wrote: | For your analogy on car accidents - a notable difference | between both is that in the case of car accidents, we are | able to get numbers on when, how and why they happen and | then make conclusions from that. | | In this case, we are not even aware of most events of | fraud/"bad papers"/manipulation - the "crisis" is that we | are losing faith in the science we are doing - results that | were cornerstones of entire fields are found to be | nonreproducible, making all the work built on top of it | pointless.(psychology, cancer, economics, etc - I'm being | very broad) | | At this point, we don't know how deep the rot goes. We are | at the point of recognizing that it's real, and looking for | solutions. For car accidents, we're past that - we're just | arguing about what are the best solutions. For the | replication crisis, we're trying to find a way forward. | | Like that scene in The Thing, where they test the blood? | We're at the point where we don't know who to trust. | | Ps: what's a tfa? | [deleted] | 6510 wrote: | Seems like a great way for "inferior" journals to gain | reputation. Counting citations seems a pretty silly formula/hack. | How often you say something doesn't affect how true it is. | SubiculumCode wrote: | Scientist publishes paper based on ABCD data. | | Replicator: Do you know how much data I'll need to collect? | 11,000 particpants followed across multiple timepoints of MRI | scanning. Show me the money. | petesergeant wrote: | Definitely something that needs large charitable investment, | but charities like that do exist, eg Wellcome Trust | SubiculumCode wrote: | Like 290+ million, just to get started. | jhart99 wrote: | Replication in many fields comes with substantial costs. We are | unlikely to see this strategy employed on many/most papers. I | agree with other commenters that materials and methodology should | be provided in sufficient detail so that others could replicate | if desired. | leedrake5 wrote: | Peer Review is the right solution to the wrong problem: | https://open.substack.com/pub/experimentalhistory/p/science-... | | On replication, it is a worthwhile goal but the career incentives | need to be there. I think replicating studies should be a part of | the curriculum in most programs - a step toward getting a PhD in | lieu of one of the papers. | vinnyvichy wrote: | Fear of the frontier.. that's why instead of people getting | excited to look for new rtsp superconductor candidates, we get | a lot of talk downplaying the only known one. Strong link vs | weak link reminds me of how some cultures frown on stimulants | while other cultures frown on relaxants. | nomilk wrote: | https://web.archive.org/web/20230130143126/https://blog.ever... | the_arun wrote: | Thank you. Currently the original article is throttled. | | Seems like article is not about software code. | fodkodrasz wrote: | How would you peer-replicate observation of a rare, or unique | event, for example in astronomy? | lordnacho wrote: | Either get your own telescope and gather your own data, or if | only one telescope captured a fleeting event, take that data | and see if the analysis turns out the same. | GuB-42 wrote: | Peer review is not the end. When replication is particularly | complex or expensive, peer review may just a way to see if the | study is worth replicating. | hgsgm wrote: | The problem is equating publication with truth. | | Publication is a _starting point_ , not a _conclusion_ | | Publication is submitting your code. It still needs to be tested, | rolled out, evaluated, and time-tested. | miga wrote: | Peer review does not serve to assure replication, but assure | readability and comprehensibility of the paper. | | Given that some experiments cost billions to conduct, it is | impossible to implement "Peer Replication" for all papers. | | What could be done is to add metadata about papers that were | replicated. | kergonath wrote: | Barriers to publication should be lower for replication | studies, I think that's the main problem. | | If someone wants to spend some time replicating something | that's only been described in a paper or two, that is valuable | work for the community and should be encouraged. If the person | is a PhD student using that as an opportunity to hone their | skills, it's even better. It's not glamorous, it's not | something entirely new, but it is _useful_ and _important_. And | this work needs to go to normal journals, otherwise there's | just be journals dedicated to replication and their impact | factor will be terrible and nobody will care. | s1artibartfast wrote: | They're basically no barriers to publication. There are a | number of normal journals that publish everything submitted | if it appears to be honest research. | kergonath wrote: | Not nice journals, though. At least not in my experience | but that's probably very field-dependent. It's not uncommon | to get a summary rejection letter for lack of novelty and | that is one aspect they stress when they ask us to review | articles. | s1artibartfast wrote: | But novelty IS what makes those journals nice and | prestigious in the first place. It is the basis of their | reputation. | | It's basically a catch 22. We want replication in | prestigious journals, but any Journal with replications | becomes less novel and prestigious. | | It all comes down to what people value about journals. If | people valued replication more than novelty, replication | journals would be the prestigious ones. | | It all comes back to the fact that doing novel science is | considered more prestigious than replication. | Institutions can play all kinds of games to try to make | it harder for readers to tell novelty apart from | replication, but people will just find new ways to signal | and determine the difference. | | Let's say we pass a law that prestigious journals must | published 50% replications. The Prestige from publishing | in that journal will just shift to publishing in that | journal with something like first demonstration in the | title or publishing in that journal Plus having a high | citation or impact value. | | It is really difficult to come up with the system or | institution level solution when novelty is still what | individuals value. | | As long as companies and universities value innovation, | figure out ways to determine which scientists are | innovative, and value them more | strangattractor wrote: | Maybe add people as special authors/contributors to the | original work. | | There always seems to be a contingent of people that think that | anything less than %100 solution is inadequate so nothing is | done. Peer review has proven itself inadequate and people hang | on to it tooth and nail. Some disciplines should require | replication on everything - I won't name Psychology or Social | Sciences in general but the failure to replicate rate for some | is unacceptable. | ebiester wrote: | Let's not make perfect be the enemy of good. We may never be | able to replicate every field, but we could start many fields | today. It means changing our values to make replication as a | valid path to tenure and promotion and a required element of | Ph.D studies. | julienreszka wrote: | >Experiments that cost billions to conduct | | If you can't replicate them it's like they didn't happen | anyways | thfuran wrote: | So no experiments have happened because I don't have a lab, | and CERN is just an elaborate ruse? | kergonath wrote: | It's a bit more subtle than that. Not all papers are equal | and I'd trust an article from a large team where error and | uncertainty analysis has been done properly (think the Higgs | boson paper) over a handful of dodgy experiments that are | barely documented properly. | | But yeah, in the grand scheme of things if it hasn't been | replicated, then it hasn't been proven, but some works are | credible on their own. | tnecniv wrote: | Ah yes, if I can't run the LHC at home, none of the work | there happened | mathisfun123 wrote: | >Peer review does not serve to assure replication, but assure | readability and comprehensibility of the paper. | | I have had a paper rejected twice in a row over the last year. | Both times the comments include something like "paper was very | well-wriiten; well-written enough that an undergrad could read | it". | | Peer review ensures the gates are kept. | NalNezumi wrote: | Isn't readability and comprehensibility the job of the | editor/journal to check. (after all they're actually paid) | maybe not for conference, but peer review is more for checking | if the methodology, scope, claim, direction, conclusion and | relevances is sound&trustable. | | At least that's my understanding | hedora wrote: | In CS, the editor / journal don't do those things. Instead, | the reviewers do. (Sometimes reviewers "shepherd" papers to | help fix readability after acceptance). | | Also, most work goes to conferences; journals typically | publish longer versions of published works. | kergonath wrote: | The editor is often not the right person to decide based on | technical details. Most often, articles they receive anre | outside their field of expertise and they don't really have a | way of deciding if a section is comprehensible or not. It's | very difficult for an outsider to know what bit of jargon is | redundant and what bit is actually important to make sense of | the results. So this bit of readability check falls to the | referees. | | In theory editors (or rather copyeditors, the editors | themselves have to handle too many papers to do this sort of | thing) should help with things like style, grammar, and | spelling. In practice, quality varies but it is often subpar. | kkylin wrote: | Highly dependent on journal / field. In mine (mathematics) | most associate editors work for free, same as reviwers. The | reviewer do all the things you say, and in addition try to | ensure readability & novelty. Most journals do have | professional copy editing, but that's separate from the | content review. | | I don't know how refereed conference proceedings work (we | don't really use these). The only journals I know of that | have professional editors (i.e., editors who are not active | researchers themselves) are Nature and affiiliated journals, | but someone more knowledgeble should correct me here. ___________________________________________________________________ (page generated 2023-08-06 23:00 UTC)