View on GitHub

SimpleText@CLEF-2022

SimpleText is a new lab organised as a part of CLEF-2022 conference, initiated by CLEF initiative.

SimpleText@CLEF-2022 Home


Home | Call for papers | Important dates | Tasks | Tools | Program | Publications | Organisers | Contact | CLEF-2023


SimpleText: Automatic Simplification of Scientific Texts

Lab topic and goals

SimpleText tackles technical challenges and evaluation challenges by providing appropriate data and benchmarks for text simplification.
We propose the following tasks:

To face these challenges, SimpleText aims to answer the following research questions:
RQ1 — What textual expression carrying information should be simplified (document and passage to be included in the simplified summary)?
RQ2 — What kind of background information should be provided (what terms should be contextualised giving a definition and/or application)? What information is the most relevant or helpful?
RQ3 — How to improve the readability of a given short text (e.g. by reducing vocabulary and syntactic complexity) with acceptable rate of information distortion?

Significance for the field & Relevance to CLEF

Being science literate is an important ability for people. It is one of the keys for critical thinking, objective decision-making and judgment of the validity and significance of findings and arguments, which allows discerning facts from fiction. Thus, having a basic scientific knowledge may also help maintain one’s health, both physiological and mental. The COVID-19 pandemic provides a good example of such a matter. Understanding the issue itself, being aware of and applying social distancing rules and sanitary policies, choosing to use or avoid particular treatment or prevention procedures can become crucial. In the context of a pandemic, the qualified and timely information should reach everyone and be accessible. That is what motivates projects such as EasyCovid.

However, scientific texts are often hard to understand as they require solid background knowledge and use tricky terminology. Although there were some recent efforts on text simplification, removing such understanding barriers between scientific texts and general public in an automatic manner is still an open challenge. SimpleText Lab brings together researchers and practitioners working on the generation of simplified summaries of scientific texts. It is a new evaluation lab that follows up the SimpleText-2021 Workshop. All perspectives on automatic science popularisation are welcome, including but not limited to: Natural Language Processing (NLP), Information Retrieval (IR), Linguistics, Scientific Journalism, etc.

Scenarios

The goal is to create a simplified summary of multiple scientific documents based on a given query which provides the user with an instant simplified summary on a specific topic he/she is interested in or to generate a daily digest, for example for ArXiv.

Evaluation setup, metrics, and tasks

Dataset: We use the Citation Network Dataset: DBLP+Citation, ACM Citation network. An elastic search index is provided to participants accessible through a GUI API. This Index is adequate to:
• Apply basic passage retrieval methods based on vector or language IR models
• Generate Latent Dirichlet Allocation models,
• Train Graph Neural Networks for citation recommendation as carried out in Stellar Graph for example,
• Apply deep bidirectionnal transformers for query expansion…

The data is two-fold: Medecine and Computer Science, two domains among the most popular in forums like ELI5.

English queries: For this edition queries are a selection of recent press titles from The Guardian enriched with keywords manually extracted from the content of the articles. It has been checked that each keyword allows to extract at least 5 relevant abstracts. The use of these keywords is optional.
Input format for all tasks:
• Topics are in the MD format
• Full text articles from The Guardian (link, folder with full texts in the MD format)
• ElasticSearch index on the data server
• DBLP full dump in the JSON.GZ format
• DBLP abstracts extracted for each topic in the following MD format (doc_id, year, abstract)

How to Cite

If you extend or use this work, please cite the paper where it was introduced:

Liana Ermakova, Eric SanJuan, Jaap Kamps, Stéphane Huet, Irina Ovchinnikova, Diana Nurbakova, 
Sílvia Araújo, Radia Hannachi, Elise Mathurin, and Patrice Bellot. 2022. 
Overview of the CLEF 2022 SimpleText Lab: Automatic Simplification of Scientific Texts. 
In Experimental IR Meets Multilinguality, Multimodality, and Interaction: 13th International 
Conference of the CLEF Association, CLEF 2022, Bologna, Italy, September 5–8, 2022, Proceedings. 
Springer-Verlag, Berlin, Heidelberg, 470–494. https://doi.org/10.1007/978-3-031-13643-6_28

Paper

Dowload .BIB