Thank you Matt for bringing up this
topic and for everybody who articulated
their opinions. This is a very
interesting debate, which I find
particularly enlightening, given that I
am one of those people that has chosen
to park their manuscript in arXiv in the
foreseeable future.
If I can try to summarize the gist of
the opinions, then it seems that every
researcher tends to come up with a set
of heuristics to try and determine
whether a particular publication is
worth their time and effort without
actually reading it. These include the
reputation of the publication platform,
but are additionally influenced greatly
by the authors' perceived reputation,
their affiliation(s), the level of
presentation (aesthetics, language,
structure, bibliography, etc.), the type
and extent of the claims made, their
novelty, their topicality, and for older
papers, the number of citations they
received. We would like to think that
all these give a pretty good idea of
whether a paper is worthy even before
reading the abstract. Just like any
other endeavor, a judgment error here
can be a false negative - ignoring a
good paper, which could have advanced
other results and ideas, and could have
saved repeating work, or realizing that
what you have been working on had been
already done by somebody else. The
judgment error here can also be a false
positive - giving undeserved attention
to an unworthy paper, which may result
in waste of time, money, and escalate to
wrongly citing it and basing further
false claims upon it - a potential
embarrassment. To some, there is a
pedagogical point to make here, since
the risk in a false positive is so high
that it is also critical to warn others
against it.
I'd like to offer another perspective
about the role and usefulness of arXiv,
as I have personally experienced it,
which goes beyond its preprint
repository function. As it relates to a
specific work, it may not be easily
generalizable, although I think it
highlights the shades of grey involved
in the process of doing different forms
of science.
First, publishing on arXiv has liberated
me from adhering to the standard article
format and allowed for keeping a more
organic structure that made better sense
for the writing and topic - neither a
book nor an article, a hybrid between
theory and experiment, something that
does not clearly belong in any specific
journal.
Second, it has allowed for some
relaxation of the usual cautiousness of
completely refraining from any
speculation. While this may be an
obvious red flag for some readers, I
think it's fair play as long as the act
of speculation is clearly stated and the
ensuing logical flow is kept in check.
Third, it has made the question of who
can review the material moot. Every
reader is a reviewer in their own right
and must be able to trust their own
judgment. Producing a document that may
not be reviewable in the traditional
sense because of its length
and interdisciplinarity has very limited
options for publication. One such option
is to publish it as a book or a thesis,
if suitable reviewers can be found. Many
walls can be hit here. Another option is
to break it down to multiple papers and
send them to different journals, which
would take many years and hoops to jump
(a good example is de Boer's "Auditory
physics" trilogy from 1980, 1984 and
1991, although I don't know the back
story of this series). The benefit in
going through this usual process may be
the increase in trust in the relevance
and correctness of the material that the
readers should have, while they can also
enjoy a better presentation (fewer
errors, better focus, etc.). The cost of
adhering to traditional format would be
many years of delay and loss of
precision in the message, as I envision
and would like to communicate. It may
also be the loss of precedence if
someone else has come up with similar
ideas at the same time - not at all an
uncommon thing in the history of science
(e.g., Darwin and Wallace).
The alternative was to use arXiv for
publication (it could have been another
repository). Critically, it provides an
agreed upon stamp of authorship with a
publication date. At the very least, it
has non-zero reputation in several
scientific fields, there is very
rudimentary control by its staff of what
goes into it, initial author affiliation
(or reference by affiliated people) is
required, and it allows for version
updates. More importantly, it relies on
trust in the judgment of the few that
would be willing to invest time in
reading the manuscript, so they can
decide for themselves whether it is a
worthwhile piece, or one that should
have never seen the light of day and be
forgotten. I believe it is a more adult
way to treat the readers, who should be
capable to assess the quality of the
work after decades of education, without
being prescribed a nominal map of where
"bad science" necessarily lies that must
be avoided at all costs.
Whichever strategy of reading and
publishing is embraced, there is going
to be no one-rule-fits-all here, and
every scholar has to be comfortable with
their own choices, obviously. All have
clear merits and none is completely
infallible.
Adam.
On Thu, May 25, 2023, at 9:34 PM, Ole C
Bialas wrote:
Thank you
Dan, Alain and everyone else for this
important debate. I
think its
essential that we, as a field, have a
constructive debate
about
publishing models because it feels
like the current model of
for-profit
publishing is unsustainable and will
hopefully be replaced by
something
better.
I agree
with most of Dan's arguments in
defense of preprints although I
think
that the boost in speed and citations
is the weakest just because
I think
there is usually no inherent
time-pressure to most of our
research
- after all, it's not like we are
developing vaccines for a
global
pandemic or something.
More
importantly, preprints provide open
access for readers and authors
and
removes gatekeepers. The latter may
allow the publishing of research
that goes
against widely accepted standards in
style, design,
methodology
and so on but this kind of heterodoxy
is something I
personally
welcome. Of course, I value the
critique of experts but in
the
current system I don't really get this
critique. Instead, I just get
the
information that someone, who is
probably and expert on the matter
and may
or may not have spent a lot of time on
this particular paper,
saw it
fit for publication.
I am not
convinced by Alain's argument that the
current peer-review
process
is a safeguard against bad science. As
Dan suggested, there is a
good
amount of research showing the
ineffectiveness of the current
review
system. There may even be the danger
that certain publications
are taken
at face value, instead of being
assessed critically, just
because
they appeared in a reputable journal.
Thus, peer-review may
provide a
false sense of security, much like the
use of helmets in
American
Football caused an increase in
traumatic brain injury because
it lead
players to charge head first into each
other.
The only
time I noticed a truly bad effect of
preprints was during the
pandemic,
when media outlets picked up on flawed
corona related research
( masks
don't work etc.) and then reported it
as facts without
understanding
or explaining what a preprint is.
I think
that it would be useful to have a
review process that is open,
transparent
and detached from publishing, like
movie reviews written on
pages
such as imdb. In this way, scientist
could not only access and
cite the
research itself but also critical
reviews of that research.
This
would also allow young scientists such
as myself to get more
insight
into the secretive world of academic
publishing. Of course
coming up
with a good architecture that sets the
right incentives for
such a
system is no trivial task but I don't
see clinging to the status
quo of
publishing as a viable option on the
long run.
Again,
thank you all for adding to this
debate!
All the
best,
Ole
Am
25.05.2023 11:51 schrieb Goodman,
Daniel F M:
>
Alain,
>
> You
write about preprints as if they're
some new thing with potentially
>
dangerous unknowable consequences, but
they've been around and used
>
extensively (particularly in maths and
physics) for over 30 years at
> this
point (arXiv was founded in 1991).
Most major funders and journals
>
recognise preprints, probably the
majority of funders now have open
>
access requirements that can be
fulfilled with preprints, and a few
are
> even
mandating their use. It's actually not
much younger than the
>
widespread use of peer review, which
didn't become a de facto standard
>
until the 1960s-1970s (Nature didn't
use it until 1973 for example).
>
> When
you say you're not convinced by
arguments about speed or number of
>
citations, I guess you mean about the
net benefits not about the facts?
>
Because the data is really start:
papers in biology which originally
>
appeared as preprints get 36% more
citations
>
immediate and long lasting
>
> To
make the argument clearer, let's break
it down into the different
>
roles that preprints can have.
>
> The
first role is what preprints can do in
the period following the
>
publication of a paper in a journal.
In this case, posting a preprint
> of
> a
paper fulfills open access
requirements and makes it possible for
the
>
whole world to read your paper,
including the general public, and
>
people
> at
less wealthy universities and
countries that cannot afford the
>
journal subscription. I cannot see any
coherent argument against this.
> It's
a disgrace that the public pays for
science but is not able to
>
access the results of the work they
paid for, and it is only a
>
hindrance
> to
scientific progress to gate access to
knowledge.
>
> The
second role is what preprints can do
in the time between the
>
journal
>
accepting the paper and making it
available. This is purely about speed
> of
publication but I can't see any reason
why you wouldn't want this
>
speed? I just went to the most recent
issue of JASA and looked at the
>
first three papers as a rough sample,
and this delay was 3 weeks, 3.5
>
weeks and 6.5 weeks. It's not years,
but might make the difference in
>
someone's job or grant application.
>
> The
third role is where I guess you mostly
disagree Alain, the time
>
period between publishing the preprint
and journal acceptance. But I
>
don't really see any conflict here. If
you don't want to read preprints
> and
prefer to wait then just don't read
them. But they will have value
> for
other readers (like me) who accept the
limitations, and they have
>
great value for the authors (36% more
citations for example). For
>
reference, for my sample of JASA
papers above, the times from first
>
submission to journal publication were
22 weeks, 27 weeks, and 38
>
weeks.
>
> I
would dispute the strength of the
quality control you mention though.
> A
study of peer review at the BMJ with
deliberate major and minor
>
errors
>
found that on average peer reviewers
picked up on 2.6 to 3 of 9 major
>
errors deliberately introduced
> sort
of quality control, but not enough to
mean that you can
>
uncritically read peer reviewed
papers.
>
> And
on the other hand, there is also a
downside to only reading peer
>
reviewed work in that you are subject
to editorial and reviewer biases.
> A
PNAS study found that a paper
submitted with a Nobel prize winner as
>
author was recommended for acceptance
by 20% of reviewers, but the very
> same
paper with an unknown student as
author was only recommended for
>
acceptance 2% of the time
>
> More
controversially perhaps, I think there
is a potential fourth role
> for
preprints that are never submitted to
a journal. This is very
>
common
> in
maths, physics and computer science
and works well there. I think it
>
would work even better when combined
with a post-publication peer
>
review
>
platform that made reviews open,
prominently displayed with an
>
at-a-glance summary, and easily
accessible. But that's an argument for
>
another day!
>
> Dan
>
>
------ Original Message ------
> Date
25/05/2023 09:01:43
>
Subject Re: arXiv web of trust
>
>>
Dan, all,
>>
>>
I'm not convinced by arguments about
speed of 'publication', number of
>>
citations, or algorithmic suggestions.
Think 'fake news' and the
>>
impact of recommendation algorithms on
the quality of information,
>>
minds, and the state of the world.
>>
>>
The review process can be seen as
quality control. A product maker
>>
that eliminates that phase can deliver
them faster, introduce jazzier
>>
products, make more money, and
dominate the market. Peer-review -
like
>>
product quality control - doesn't
eliminate all flaws, but it may make
>>
them less likely and easier to spot
and eliminate.
>>
>>
I suspect there is a generational
dimension to this debate. The three
>>
of us that argued most strongly in
defence of the review process have
>>
(or have had) a well-established
career. How could we not defend the
>>
practices that got us there? Someone
struggling to gain recognition,
>>
and a job, may be tempted by
mechanisms that bypass those
practices.
>>
Fair enough, but beware. It might be a
bit like tearing down the walls
>>
and ripping up the floor to feed the
boiler.
>>
>>
The debate may become moot with the
introduction of AI-based tools to
>>
assist writing and reviewing. Why not
use similar tools to read the
>>
papers too, and understand them, and
produce new science (of possibly
>>
better quality)? This sounds great,
except that I don't see much room
>>
for a human scientist in that loop.
So much for your careers.
>>
>>
I find the generational issue
unnerving, personally. For the first
>>
time in my life, I'm old and the
others are new. It takes some
>>
getting used to.
>>
>>
Alain
>>
>>
>>
>>
>>>
On 24 May 2023, at 15:42, Goodman,
Daniel F M
>>>
>>>
I have no hesitation in calling a
preprint a "publication". There's
>>>
no magic in peer review that makes it
not count as published before
>>>
this process. Even the word preprint
is archaic now given how many
>>>
journals are online only.
>>>
>>>
Personally, I now primarily read
preprints because most of the work
>>>
in the areas I'm interested in appears
a year or two earlier as
>>>
preprints than in a journal. It's much
more exciting and progress can
>>>
be much faster when there isn't a
multi year between doing work and
>>>
seeing how others make use of it. I
just had an email from someone
>>>
asking if they could cite a tweet of
mine that had inspired them to
>>>
do some work and this sort of thing is
great! Why should we accept
>>>
years of delay between each increment
of progress?
>>>
>>>
Of course, reading preprints means you
have to cautious. But, I
>>>
always treat papers I read critically
whether they've been through
>>>
peer review or not, and I would
encourage everyone to do the same.
>>>
Peer review is of very uneven quality,
based on quantitative studies
>>>
and based on my own experience as a
reviewer reading the other
>>>
reviews. Terrible papers with glaring
errors get through peer review.
>>>
So I don't think we can uncritically
accept the results of peer
>>>
reviewed papers, and in practice most
scientists don't. We criticise
>>>
peer reviewed papers all the time.
It's this process of review or
>>>
feedback after publication that is the
real scientific process, and
>>>
it would be much easier if the reviews
were made available so we
>>>
could more easily judge for ourselves.
The sooner we move to a system
>>>
of open and transparent post
publication peer review like the
systems
>>>
Etienne is talking about, the better.
>>>
>>>
I do agree with Alain's point that
there are too many papers to read
>>>
them all, but for me that's not an
argument for the traditional
>>>
approach to peer review but for
experimenting with different
>>>
approaches to recommending papers.
Again personally, I find I have a
>>>
higher hit rate with algorithmic
suggestions from Semantic Scholar
>>>
and from things I see posted on social
media than I do from going
>>>
through journal table of contents
(which I still do out of habit).
>>>
>>>
And as a last point to encourage
preprints, the evidence shows that
>>>
papers that are first available as a
preprint get cited more overall.
>>>
And if that doesn't convince you I
don't know what will. ��
>>>
>>>
Dan
>>>
>>>
---
>>>
This email was written on my phone,
please excuse my brevity.
>>>
>>>
Sent: Wednesday, 24 May 2023 10:38
>>>
Subject: Re: [AUDITORY] arXiv web of
trust
>>>
>>>
Thanks for opening this nice debate,
Max!
>>>
>>>
I side with Brian for the need of
serious peer-review, but I am less
>>>
sure how this can be achieved
nowadays. Publishers are increasingly
>>>
pressuring reviewers to work fast
because their business model relies
>>>
on volume, and there seems to be
little cost to publishing poor
>>>
quality papers. With the ever
precarisation of research, it takes a
>>>
very strong faith in the ethos of
scientific integrity to remain a
>>>
thorough reviewer.
>>>
>>>
If we accept that, as a consequence of
this pressure, there are more
>>>
flawed papers that pass the review
process, it would mean that we, as
>>>
consumers of the literature, should be
more cautious when citing
>>>
articles. We should more critically
examine what we cite, and sort of
>>>
perform our own review. But of course,
that's also very time
>>>
consuming... and it is also very
inefficient at the scale of the
>>>
community: me *not* citing an article
because I found that it is
>>>
potentially flawed will not prevent
others from citing it, and the
>>>
effort I will have put in reviewing it
will be largely wasted.
>>>
>>>
So I do believe that there is a strong
benefit in having more open
>>>
discussions about papers, and in some
cases, the fact that they are
>>>
published or not in the traditional
sense, may be partially
>>>
irrelevant. We definitely don't want
to turn the scientific community
>>>
into social media, where a few
arbitrary influencers get to decide
>>>
what's worthy and what isn't. But
there are now places where
>>>
scientific arguments can be shared,
and reflections can be had,
>>>
constructively.
>>>
>>>
That's what we tried to do for the
last edition of the International
>>>
Symposium on Hearing, but hosting the
papers as "pre-print" (for lack
>>>
of a better term) freely available on
Zenodo
>>>
publically available on PubPeer (and
more can be added; here's an
>>>
example:
>>>
Contributors are still able to publish
their articles in the
>>>
traditional sense, and hopefully the
published version will be
>>>
connected to the ISH version in some
form so that users can view the
>>>
history and comments. In others words,
there is much benefit for the
>>>
two systems to co-exist (we can get
rid of private publishers,
>>>
though, and switch to decentralized
institutional ones).
>>>
>>>
Remains the problem raised by Alain:
as readers, how do we deal with
>>>
the volume? While publishers have been
selling us "reputation" in the
>>>
form of journals in very much
overrated ways (such as impact
factors,
>>>
and what not), it is true that
journals do have a curating role that
>>>
should not be underestimated. This
being said, editors do not
>>>
actively seek authors to steer
publications towards a specific topic
>>>
(besides Frontiers' take it all
harassment approach). It is still the
>>>
authors that decide to submit to a
specific journal or another. As a
>>>
result, following the JASA TOC gives
us access to a semi-random
>>>
sample of what's going on in the
field. It does offer,
>>>
stochastically, some degree of
protection against confirmation bias
>>>
in literature search (whereby you only
look for papers that confirm
>>>
your idea). I wonder if automatic
suggestions of "related papers"
>>>
could achieve something similar in
other venues?
>>>
>>>
Cheers,
>>>
-Etienne
>>>
>>>
>>>
--
>>>
Etienne Gaudrain, PhD
>>>
>>>
Lyon Neuroscience Research Centre /
Auditory Cognition and
>>>
Psychoacoustics (CAP)
>>>
CNRS UMR5292, Inserm U1028, Université
Lyon 1
>>>
Centre Hospitalier Le Vinatier -
Bâtiment 462 - Neurocampus
>>>
95 boulevard Pinel, 69675 Bron Cedex,
France
>>>
>>>
>>>
>>>
On Wed, 24 May 2023 at 10:56, Alain de
Cheveigne
>>>
Hi Jonathan, all,
>>>
>>>
Here's a different perspective.
>>>
>>>
First of all, the issue of peer review
should be distinguished from
>>>
that of publishers shaving the wool
off our backs (more below).
>>>
>>>
Peer review offers functions that we
miss out on in the preprint
>>>
model. Weeding out junk is one,
improving papers (and the ideas in
>>>
them) is another. A third is reducing
the bulk of things to read.
>>>
>>>
The last might seem counterintuitive:
surely, more is better? The
>>>
thing is, we have limited time and
cognitive bandwidth. Lack of time
>>>
is the major obstacle to keeping
abreast, and lack of time of the
>>>
potential audience is what prevents
our ideas having an impact. You
>>>
painstakingly work to solve a major
problem in the field, write it up
>>>
carefully, and no one notices because
attention is carried away by
>>>
the tweet cycle.
>>>
>>>
The review/journal model helps in
several ways. First, by
>>>
prioritizing things to read (as an
alternative to the random - or
>>>
otherwise biased - selection induced
by lack of time). Second, by
>>>
improving the readability of the
papers: more readable means less
>>>
time per paper means more attention
for other papers - including
>>>
possibly yours. Third, by organizing -
however imperfectly - the
>>>
field.
>>>
>>>
For example, you can (or could) keep
abreast of a topic in acoustics
>>>
by scanning JASA and a few other
journals. With the preprint/twitter
>>>
model the 'field' risks being
shattered into micro-fields, bubbles,
>>>
or cliques.
>>>
>>>
My experience of the review process is
- as everyone's - mixed. I
>>>
remember intense frustration at the
reviewer's dumbness, and despair
>>>
at ever getting published. I also
remember what I learned in the
>>>
process. Almost invariably, my papers
were improved by orders of
>>>
magnitude (not just incrementally).
>>>
>>>
I also spend a lot of time reviewing.
I find it a painful process,
>>>
as it involves reading (I'm a bit
dyslexic), and trying to understand
>>>
what is written and - to be helpful to
the author - what the author
>>>
had in mind and how he/she could
better formulate it to get the
>>>
message across, and avoid wasting the
time of - hopefully - countless
>>>
readers. It does involve weeding out
some junk too.
>>>
>>>
Science is not just about making new
discoveries or coming up with
>>>
radically new ideas. These are few and
far between. Rather, it's a
>>>
slow process of building on other
people's ideas, digesting, tearing
>>>
down, clearing the rubble, and
building some more. The review
process
>>>
makes the edifice more likely to
stand. Journals play an important
>>>
role in this accumulation, even if
most content is antiquated and
>>>
boring. It's a miracle that some
journals have done this over
>>>
decades, even centuries.
>>>
>>>
Which brings back to the issue of
money, impact factors, and
>>>
careers. Lots to say about that,
mostly depressing, but mainly
>>>
orthogonal from the peer-review issue.
>>>
>>>
Alain
>>>
>>>
>>>
>>>
>>>
>>>
wrote:
>>>
>
>>>
> Matt,
>>>
>
>>>
> In this context I would avoid the
term “publishing”, since that
>>>
has such a different meaning for so
many people, but I personally do
>>>
take advantage of posting preprints on
a public server (like arXiv)
>>>
almost every chance I get.
>>>
>
>>>
> Preprints (preprint = a fully
written paper that is not (yet)
>>>
published) have been useful for many
decades, originally in physics,
>>>
as a way of getting one's research
results out in a timely manner.
>>>
Other key benefits are that it
establishes primacy of the research
>>>
findings, that it is citable in other
researchers' papers, and that
>>>
it can be promoted by social media
such as this listserve (more below
>>>
on this). But the biggest benefit is
typically getting the paper out
>>>
into the world for others to learn
from, without having to wait based
>>>
on the whims of publishers and
individual reviewers. If most of your
>>>
published papers get accepted
eventually, and the most important
>>>
findings don’t get cut in the review
process, then preprints are
>>>
something you should definitely
consider. Reviewers often make
>>>
published papers better, but maybe not
so much better that it’s worth
>>>
waiting many months for others to see
your results.
>>>
>
>>>
> arXiv is the oldest website for
posting preprints, and if its
>>>
Audio and Speech section is active,
that might be a good place to
>>>
post your preprints. But there may be
other options for you. As an
>>>
auditory neuroscientist I typically
use bioRxiv (e.g., "Changes in
>>>
Cortical Directional Connectivity
during Difficult Listening in
>>>
Younger and Older Adults”
>>>
also use PsyArXiv if the topic is more
perceptual than neural (e.g.,
>>>
“Attention Mobilization as a Modulator
of Listening Effort: Evidence
>>>
about promoting your research on
social media?]
>>>
>
>>>
> I’m sure others have opinions
too.
>>>
>
>>>
> Jonathan
>>>
>
>>>
>
>>>
wrote:
>>>
>>
>>>
>> Is anyone publishing on arXiv
at the moment ? It seems that to
>>>
publish there they rely on a web of
trust.
>>>
>>
>>>
>> There is an Audio and Speech
section of arXiv which would suit
>>>
our community.
>>>
>>
>>>
>> thanks
>>>
>>
>>>
>> Matt
>>>
>
>>>
> --
>>>
> Jonathan Z. Simon (he/him)
>>>
> University of Maryland
>>>
> Dept. of Electrical &
Computer Engineering / Dept. of
Biology /
>>>
Institute for Systems Research
>>>
> 8223 Paint Branch Dr.
>>>
> College Park, MD 20742 USA
>>>
> Office: 1-301-405-3645, Lab:
1-301-405-9604, Fax: 1-301-314-9281
>>>
>
>>>
>
>>