SimpleText@CLEF-2021 Pilot tasks
Home | Call for papers | Important dates | Pilot tasks | |
Program | Publications | Organisers | Contact |
SimpleText Pilot Task Guidelines
We invite you to submit both automatic and manual runs! Manual intervention should be reported.
Pilot Task 3: Scientific text simplification – Language simplification
The goal of this pilot task is to provide a simplified version of text passages. Participants will be provided with queries and abstracts of scientific papers. The abstracts can be split into sentences as in the exemple https://guacamole.univ-avignon.fr/nextcloud/index.php/s/SQTdS2Yowf9dxNa.
Output format:
Simplified passages in a TSV tabulated file with the following fields:
- run_id: Run ID starting with team_id_
- manual: Whether the run is manual {0,1}
- topic_id: Topic ID
- topic_text: Topic text
- source_passage: Source passage text
- simplified_passage: Text of the simplified passage
run_id manual topic_id topic_text doc_id source_passage simplified_passage
Evaluation
The simplified passages will be evaluated manually with eventual use of aggregating metrics.
OUTPUT example:
run_id | manual | topic_id | topic_text | doc_id | source_passage | simplified_passage |
---|---|---|---|---|---|---|
BTU | 1 | 1 | Digital assistants like Siri and Alexa entrench gender biases, says UN | 3003409254 | Automated decision making based on big data and machine learning (ML) algorithms can result in discriminatory decisions against certain protected groups defined upon personal data like gender, race, sexual orientation etc. Such algorithms designed to discover patterns in big data might not only pick up any encoded societal biases in the training data, but even worse, they might reinforce such biases resulting in more severe discrimination. | Automated decision-making may include sexist and racist biases and even reinforce them because their algorithms are based on the most prominent social representation in the dataset they use. |