1st Workshop on Scholarly Document Processing and Shared Tasks (SDP 2020) @ EMNLP 2020

SDP 2020


Data Mining & Analysis Databases & Information Systems



You are invited to participate in the 1st Workshop on Scholarly Document Processing (SDP 2020) to be held in conjunction with the 2020 Conference in Empirical Methods in Natural Language Processing (EMNLP 2020) on November 11 or 12 in Punta Cana, Dominican Republic.
The workshop will consist of a research track and a shared task track. The shared task track includes the 6th edition of the CL-SciSumm shared task (https://github.com/WING-NUS/scisumm-corpus/) and two new summarization tasks -- CL-LaySumm and LongSumm -- geared towards easier access to scientific methods and results.
The tentative submission deadline is July 15, 2020.
SDP is a continuation of the BIRNDL (https://philippmayr.github.io/BIRNDL-WS/) and WOSP (https://wosp.core.ac.uk/) workshop series.
Call for papers:
Workshop Date and Venue: November 11/12, Punta Cana, Dominican Republic
Website: https://ornlcda.github.io/SDProc/
** Introduction **
In addition to the long-standing challenge faced by scholars of keeping up with the growing literature in their own and related fields, they must now compete with malign pseudo-science and dis-information in informing public policy and behavior. This has stimulated workshops and research focused on enhancing search, retrieval, summarization, and analysis of scholarly documents. However, the general research community on scholarly document processing remains fragmented, and efforts towards natural language understanding of scholarly text that is central to vastly improve all the said downstream applications are not widespread.
To address these gaps, we, the organizers of BIRNDL and WOSP workshops, propose the first Workshop on Scholarly Document Processing. We seek to reach to the broader NLP and AI/ML community to pool the distributed efforts to improve scholarly document understanding and enable intelligent access to the published research. The goal of SDP is two-fold: to increase collaboration between communities interested in leveraging knowledge stored in scientific literature and data and to establish SDP as the single-focused primary venue for the field.
We seek to appeal to the mainstream NLP and ML community working on SDP tasks – which are NLP tasks – to publish at SDP as we seek to establish SDP as the integrated premier venue. We have established a steering committee to help us turn SDP into a conference in the forthcoming years.
** Topics of Interest **
We invite submissions from all communities interested in natural language processing, information retrieval, and data mining problems in scientific documents; and in processing scientific documents for easier access to various audiences. The topics of interest include, but are not limited to:
* Information extraction, text mining and parsing scholarly literature
* Reproducibility and peer review
* Lay Summarization (i.e., summaries created for non-experts) of individual and collections of scholarly documents
* Discourse modeling and argument mining
* Summarization and question-answering for scholarly documents
* Semantic and network-based indexing, search and navigation in structured text
* Graph analysis/mining including citation and co-authorship networks
* Analysing and mining of citation contexts for document understanding and retrieval
* New scholarly language resources and evaluation
* Connecting and interlinking publications, data, tweets, blogs or their parts
* Disambiguation, metadata extraction, enrichment, and data quality assurance for scholarly documents
* Bibliometrics, scientometrics, and altmetrics approaches and applications
* Other aspects of scientific workflows including open access/science, and research assessment
* Infrastructures for accessing scientific publications and/or research data
** The 6th Computational Linguistics Scientific Document Summarization Shared Task (CL-SciSumm 2020) ** (Organisers: Muthu Kumar Chandrasekaran)
CL-SciSumm is the first medium-scale shared task on scientific document summarization, with over 500 annotated documents. Last year's CL-SciSumm shared task introduced large scale training datasets, both annotated from ScisummNet and auto-annotated. For the task, Systems were provided with a Reference Paper (RP) and 10 or more Citing Papers (CPs) that all contain citations to the RP, which they used to summarise RP. This was evaluated against abstract and human written summaries on ROUGE.
The task is defined as follows:
* Given: A topic consisting of a Reference Paper (RP) and Citing Papers (CPs) that all contain citations to the RP. In each CP, the text spans (i.e., citances) have been identified that pertain to a particular citation to the RP.
* Task 1A: For each citance, identify the spans of text (cited text spans) in the RP that most accurately reflect the citance. These are of the granularity of a sentence fragment, a full sentence, or several consecutive sentences (no more than 5).
* Task 1B: For each cited text span, identify what facet of the paper it belongs to, from a predefined set of facets.
* Task 2 (optional bonus task): Finally, generate a structured summary of the RP from the cited text spans of the RP. The length of the summary should not exceed 250 words.
This year, CL-SciSumm '20 will have two new tracks: LaySumm and LongSumm.
** CL-LaySumm 2020: The 1st Computational Linguistics Lay Summary Challenge Shared Task ** (Organisers: Anita De Waard, Ed Hovy)
To ensure and increase the relevance of science for all of society and not just a small group of niche practitioners, researchers have been increasingly tasked by funders and publishers to outline the scope of their research for the general public by writing a summary for a lay audience, or lay summary. The LaySumm summarization task considers automating this responsibility, by enabling systems to automatically generate lay summaries. A lay summary explains, succinctly and without using technical jargon, what the overall scope, goal and potential impact of a scientific paper is.
The corpus for this task will comprise full-text papers with lay summaries, in a variety of domains, and from a number of journals. Elsevier will make available a collection of Lay Summaries from a multidisciplinary collection of journals, as well as the abstracts and full text of these journals.
The task is defined as follows:
* Given: A full-text paper, its Abstract, and a Lay Summary of a given paper
* Task: For each paper, generate a Lay Summary of the specified length
Evaluation
The Lay Summary Task will be scored by using several ROUGE metrics to compare the system output and the gold standard Lay Summary. As a follow-up to the intrinsic evaluation, we will crowdsource a number of automatically generated lay summaries to a panel of judges and a lay audience. Details of the crowdsourcing evaluation will be announced with the sharing of the final test corpus on July 1st.
All nominated entries will be invited to publish a paper in Open Access (Author-Payment Charges will be waived) in a selected Elsevier publication. Authors will be asked to provide an automatically generated lay summary of their paper, together with their contribution.
** LongSumm 2020: Shared Task on Generating Long Summaries for Scientific Documents ** (Organisers: Michal Shmueli-Scheuer, Guy Feigenblat)
Most of the work on scientific document summarization focuses on generating relatively short summaries (250 words or less). While such a length constraint can be sufficient for summarizing news articles, it is far from sufficient for summarizing scientific work. In fact, such a short summary resembles more to an abstract than to a summary that aims to cover all the salient information conveyed in a given text. Writing such summaries requires expertise and a deep understanding in a scientific domain, as can be found in some researchers’ blogs.
The LongSumm task opted to leverage blogs created by researchers in the NLP and Machine learning communities and use these summaries as reference summaries to compare the submissions against.
The corpus for this task includes a training set that consists of 1705 extractive summaries and around 700 abstractive summaries of NLP and Machine Learning scientific papers. These are drawn from papers based on video talks from associated conferences (Lev et al. 2019 TalkSumm) and from blogs created by NLP and ML researchers. In addition, we create a test set of abstractive summaries. Each submission is judged against one reference summary (gold summary) on ROUGE and should not exceed 600 words.
** Submission Information **
Authors are invited to submit full and short papers with unpublished, original work. Submissions will be subject to a double-blind peer review process. Accepted papers will be presented by the authors at the workshop either as a talk or a poster. All accepted papers will be published in the workshop proceedings.
The submissions should be in PDF format and anonymized for review. All submissions must be written in English and follow the EMNLP 2020 formatting requirements.
Long paper submissions: up to 8 pages of content, plus unlimited references.
Short paper submissions: up to 4 pages of content, plus unlimited references.
Final versions of accepted papers will be allowed 1 additional page of content so that reviewer comments can be taken into account.
Shared Task registration: Participants of all shared tasks need to register here before March 31st, 2020
** Important Dates **
Research track:
Submission deadline – July 15, 2020
Notification of Acceptance – August 17, 2020
Camera-ready submission due – August 31, 2020
Workshop – November 11 or 12, 2020
Shared task track:
Training set release – Feb 15, 2020
Deadline for registration and short systems description – March 31, 2020
Test set release (Blind) – July 1, 2020
System runs due – August 1, 2020
Preliminary system reports due – August 16, 2020
Camera-ready submission due – August 31, 2020
Workshop – November 11 or 12, 2020
The dates are at this stage indicative only and can change.
** Keynote Speakers **
Kuansang Wang, Managing Director, Microsoft Research Outreach Academic Services
The second keynote speaker will be announced shortly
** Journal Extension **
In the past, the accepted authors were invited to submit an extended version of their work to a special issue of a selected journal. The organizers are currently in the process of identifying appropriate journals to host a similar special issue this year. Relevant updates including topics and requirements for this special issue will be shared on the workshop website in due time.
** Organizing Committee **
Muthu Kumar Chandrasekaran, Amazon, Seattle, USA
Anita de Waard, Elsevier, USA
Guy Feigenblat, IBM Research AI, Haifa Research Lab, Israel
Dayne Freitag, SRI International, San Diego, USA
Tirthankar Ghosal, Indian Institute of Technology Patna, India
Drahomira Herrmannova, Oak Ridge National Laboratory, USA
Eduard Hovy, Research Professor, LTI, Carnegie Melon University, USA
Petr Knoth, Open University, UK
David Konopnicki, IBM Research AI, Haifa Research Lab, Israel
Philipp Mayr, GESIS – Leibniz Institute for the Social Sciences, Germany
Robert M. Patton, Oak Ridge National Laboratory, USA
Michal Shmueli-Scheuer, IBM Research AI, Haifa Research Lab, Israel
Dominika Tkaczyk, Crossref, UK
** Steering Committee **
Edward Fox, Professor, Department of Computer Science and Director, Digital Library Research Laboratory, Virginia Tech
C. Lee Giles, David Reese Professor, College of Information Sciences and Technology, Pennsylvania State University
Min-Yen Kan, Associate Professor, School of Computing, National University of Singapore
Dragomir Radev, A. Bartlett Giamatti Professor of Computer Science, Yale University
Jie Tang, Professor and Associate Chair of the Department of Computer Science and Technology, Tsinghua University
Alex Wade, Group Technical Program Manager, Chan Zuckerberg Initiative
Kuansang Wang, Managing Director, Microsoft Research Outreach Academic Services
Bonnie Webber, Professor, School of Informatics, University of Edinburgh
** Programme Committee **
Akiko Aizawa, National Institute of Informatics, Japan
Colin Batchelor, Cambridge, UK
Joeran Beel, Trinity College Dublin, Ireland
Katarina Boland, GESIS, Germany
Guillaume Cabanac, University of Toulouse, France
Cornelia Caragea, University of illinois at Chicago, US
Zeljko Carevic, GESIS, Germany
Tanmoy Chakraborty, IIIT Delhi, India
Richard Eckart de Castilho, TU Darmstadt, Germany
Helena Deus, Elsevier Labs
Daniel Duma, University of Edinburgh, UK
Ed A. Fox, Virginia Tech, USA
Norbert Fuhr, University of Duisburg, Germany
C. Lee Giles, Penn State University, USA
Bela Gipp, University of Wuppertal, Germany
Goran Glavas, University of Mannheim, Germany
Hannaneh Hajishirzi, University of Washington, USA
Monica Ihli, University of Tennessee Knoxville, USA
Ameni Kacem, iCOVER, France
Roman Kern, Graz University of Technology, Austria
Atsushi Keyaki, Denso IT Laboratory Inc, Tokyo, Japan
Martin Klein, Los Alamos National Laboratory, USA
Ilia Kuznetsov, Techn. Univ. Darmstadt, Germany
Birger Larsen, Aalborg University, Denmark
Anne Lauscher, University of Mannheim, Germany
Paolo Manghi, National Research Council of Italy, Italy
Bruno Martins, University of Lisbon, Portugal
Norman Meuschke, University of Wuppertal, Germany
Diego Molla-Aliod, Macquarie University, Australia
Preslav Nakov, Qatar Computing Research Inst., Qatar
Federico Nanni, University of Mannheim, Germany
Jumana Nassour, Ben-Gurion University, Israel
Paco Nathan, Derwen Inc., US
Manabu Okumura, Tokyo Institute of Technology, Japan
Francesco Osborne, Open University, UK
Arzucan Ozgur, Bogazici University, Turkey
Sujit Pal, Elsevier Labs
Rajesh Piryani, South Asian University, India
Silvio Peroni, University of Bologna, Italy
Sujit Pal, Elsevier Labs
Animesh Prasad, Amazon, Cambridge, UK
Horacio Saggion, Universitat Pompeu Fabra, Spain
Angelo Antonio Salatino, The Open University, UK
Philipp Schaer, TH Cologne, Germany
Vivek Kumar Singh, Banaras Hindu University, India
Kazunari Sugiyama, Kyoto University, Japan
Saeed Ul Hassan, IT University, Pakistan
Lucy Vanderwende, University of Washington, USA
Stephen Wan, CSIRO, Australia
Bonnie Webber, University of Edinburgh, UK
Ivana Williams, Chan Zuckerberg Initiative, USA
Dietmar Wolfram, University of Wisconsin-Milwaukee, USA
Jian Wu, Old Dominion University, USA
More details available on the workshop website:
https://ornlcda.github.io/SDProc/