[Corpora-List] 1st cfp: 4REAL 2018 - LREC Workshop on Replicability and Reproducibility

António Branco antonio.branco at di.fc.ul.pt
Mon Nov 20 13:19:53 CET 2017

[apologies for duplicates; please help do disseminate]

================= FIRST CALL FOR PAPERS================

4REAL Workshop 2018

Workshop on Replicability and Reproducibility of Research Results in Science and Technology of Language http://4real2018.di.fc.ul.pt

12 May 2018 Miyazaki, Japan

Collocated with LREC 2018 11th Language Resources and Evaluation Conference http://lrec2018.lrec-conf.org

Important dates:

15 January 2018: Deadline for submissions to 4REAL workshop 9 February 2018: Notification of authors of submissions to 4REAL workshop 1 March 2018: Deadline for camera-ready 9-11 May 2018: LREC main conference (Wednesday to Friday) 12 May 2018: 4REAL workshop (Saturday)

Call for papers:

Reproduction and replication of research results are at the heart of the validation of scientific knowledge and of the scientific endeavor. Reproduction of results entails arriving at the same overall conclusions that is, to appropriately validate a set of results, scientists should strive to reproduce the same answer to a given research question by different means, e.g. by reimplementing an algorithm or evaluating it on a new dataset, etc. Replication has a more limited aim, typically involving running the exact same solution or approach under the same conditions in order to arrive at the same output result.

Despite their key importance, reproduction and replication have not been sufficiently encouraged given the prevailing procedures and priorities for the reviewing, selection and publication of research results.

The immediate motivation for the increased interest on reproducibility and replicability is to be found in a number of factors, including the realization that for some published results, their replication is not being obtained (e.g. Prinz et al., 2011; Belgley and Ellis, 2012); that there may be problems with the commonly accepted reviewing procedures, where deliberately falsified submissions, with fabricated errors and fake authors, get accepted even in respectable journals (e.g. Bohannon, 2013); that the expectation of researchers vis a vis misconduct, as revealed in inquiries to scientists on questionable practices, scores higher than one might expect or would be ready to accept (e.g. Fanelli, 2009); among several others.

This workshop seeks to foster the discussion and the advancement on a topic that has been given insufficient attention in the research area of language processing tools and resources and that has been an important topic emerging in other scientific areas, continuing the objectives of the first edition of the 4REAL workshop, at LREC 2016. We are thus inviting submissions of articles that present cases, either with positive or negative results, of actual replication or reproduction exercises of previous published results in our area.

Specific topics within the scope of the call include, but are not limited to:

– Reproduction and replication of previously published systems, resources, results, etc.

– Analysis of reproducibility challenges in system-oriented and user-oriented evaluation.

– Reproducibility challenges on private or proprietary data.

– Reproducibility challenges on ephemeral data, like streaming data, tweets, etc.

– Reproducibility challenges on online experiments.

– Reproducibility in evaluation campaigns.

– Evaluation infrastructures and Evaluation as a Service (EaaS).

– Experiments on data management, data curation, and data quality.

– Reproducible experimental workflows: tools and experiences.

– Data citation: citing experimental data, dynamic data sets, samples, etc.

We are interested also in articles discussing the challenges, the risk factors, the appropriate procedures, etc. specific to our area or that should be adopted, or adapted from other neighboring areas, including methodologies for monitoring, maintaining or improving citation of language resources and tools and to assess the importance of data citation for research integrity. This includes also of course the new risks raised by the replication articles themselves and their own integrity, in view of the preservation of the reputation of colleagues and works whose results are reported has having been replicated, etc.

“Replicability Focus” in Language Resources and Evaluation Journal:

Best papers will be invited to be submitted to the Language Resources and Evaluation Journal under its new “Replicability Focus”.

Organization committee:

António Branco (University of Lisbon) Nicoletta Calzolari (ILC) Khalid Choukri (ELRA)

Program committee:

Aljoscha Burchardt (DFKI) António Branco (University of Lisbon) Gertjan van Noord (University of Groningen) Joseph Mariani (CNRS/LIMSI) Khalid Choukri (ELRA) Maria Gavrilidou (ILSP) Marko Grobelnik (Jozef Stefan Institute) Marko Tadic (University of Zagreb) Nancy Ide (Vassar College) Nicoletta Calzolari (ILC) Patrick Paroubek (CNRS/LIMSI)) Piek Vossen (VU University Amsterdam) Senja Pollak (Jozef Stefan Institute) Simon Krek (Jozef Stefan Institute) Stelios Piperidis (ILSP) Thierry Declerck (DFKI) Yohei Murakami (Language Grid Japan)

“Identify, Describe and Share your LRs!” initiative:

Describing your Language Resources (LRs) in the LRE Map is now a normal practice in the submission procedure of LREC (introduced in 2010 and adopted by other conferences). To continue the efforts initiated at LREC 2014 about “Sharing LRs” (data, tools, web-services, etc.), when submitting a paper, authors will have the possibility to upload LRs in a special LREC repository. This effort of sharing LRs, linked to the LRE Map for their description, may become a new “regular” feature for conferences in our field, thus contributing to creating a common repository where everyone can deposit and share data.

As scientific work requires accurate citations of referenced work so as to allow the community to understand the whole context and also replicate the experiments conducted by other researchers, LREC 2018 endorses the need to uniquely Identify LRs through the use of the International Standard Language Resource Number (ISLRN, www.islrn.org), a Persistent Unique Identifier to be assigned to each Language Resource. The assignment of ISLRNs to LRs cited in LREC papers will be offered at submission time.


Begley and Ellis, 2012, Drug development: Raise standards for preclinical cancer research, Nature Bohannon, John, 2013, Who’s Afraid of Peer Review?, Science Fanelli,2009 How Many Scientists Fabricate and Falsify Research? Prinz, et al., 2011, Believe it or not: how much can we rely on published data on potential drug targets?, Nature Reviews Drug Discovery 10, 712.

More information about the Corpora mailing list