Cette page n'est disponible qu'en anglais !
cofr
)Neural network automatic coreference system adapted to French, published in the following paper:
download code view github repo
On this page:
This repository contains an adaptation of the Kantor and Globerson's coreference resolution tool (described in Kantor and Globerson, 2019) to French. The work is described in the following paper:
Wilkens Rodrigo, Oberle Bruno, Landragin Frédéric, Todirascu Amalia (2020). French coreference for spoken and written language, Proceedings of the 12th Edition of the Language Resources and Evaluation Conference (LREC 2020), Marseille, France.
For training and evaluation, we used two French corpora:
We converted the corpora into the jsonlines
format used by the original system. The Democrat corpus, originally composed of 10000-word-long chunks of longer texts has been split down to 2000-word-long documents. We kept only the texts from the 19th century to the 20th century. The Ancor corpus has been cut into thematic sections. The corpus used are available for download below.
We have trained three models:
fr_mentcoref
: the model has been trained to detect mentions (all mentions, including singletons) and to resolve coreference,fr_ment
: the model has been trained to detect mentions (all mentions) only,fr_coref
: the model has been trained to resolve coreference only (oracle mentions are to be fed to the system in order to use this model).We report the results of three experiments:
fr_mentcoref
: this is the original end-to-end setup, in which one model both detects mentions and resolves coreference,fr_coref
: this is used to evaluate the coreference resolution task proper (as opposed to the combined two subtasks of mention detection and coreference resolution). Oracle mentions are given to the system,fr_ment
, specialized for mention detection, and fr_coref
, specialized for coreference resolution.Features must be extracted before training or evaluating the model (this step is called bertification
since it uses BERT (Bidirectional Encoder Representations from Transformers)). We tested two window sizes, the default one (511) and the one proposed by Kantor and Globerson (129). The bigger window yields slightly better results, but the bertification process is longer (about 8x). All our pretrained models use the bigger window size.
default window size (511) | small window size (129) | |||||
---|---|---|---|---|---|---|
fr_mentcoref |
fr_coref | fr_ment + fr_coref | fr_mentcoref | fr_coref | fr_ment + fr_coref | |
dem1921 (mention identification) | 76.36 | 99.96 | 88.95 | 75.88 | 99.96 | 88.58 |
dem1921 (conll f1) | 66.00 | 85.04 | 75.00 | 65.38 | 84.84 | 74.97 |
ancor (mention identification) | 52.75 | 99.99 | 88.06 | 52.37 | 99.99 | 88.03 |
ancor (conll f1) | 50.50 | 88.75 | 75.65 | 50.48 | 88.45 | 75.47 |
Note that these results include singleton mentions, which are available in both French corpora, but not in the standard corpus used for evaluation of English systems (OntoNotes/CoNLL 2012).
These scores were obtained with the official scorer script used for the CoNLL 2012 evaluation compaign.
References:
First install the library requirements. By default, we use tensorflow-gpu
. Install tensorflow
if you want to use your CPU.
Examples below use Democrat. If you prefer to train and evaluate with Ancor, use the setup_{corpus,models}_ancor.sh
scripts and adapt the experiments.conf
configuration file (search for ancor
and democrat
and comment/uncomment parameters according to your corpus).
To reproduce our results, please run the following commands:
bash -x -e setup_all.sh
bash -x -e setup_corpus_dem1921.sh
bash -x -e setup_models_dem1921.sh
bash -x -e extract_bert_features.sh test.french.jsonlines evaluate
python3 evaluate.py fr_ment,fr_coref test.french.jsonlines output.jsonlines
This will download the corpus and our pretrained models. For more information on the corpora we used for training, see here.
Note that the longest part is the bertification and should be done with a GPU. As said before, you can reduce the window size to 129. This would yield slightly lower results but will dramatically decrease the bertification time.
Also note that this is not the official evaluation from the scorer.pl
script used for the CoNLL-2012 evaluation campaign. For that, you need to convert the output.jsonlines
file to the CoNLL format and to use the scorer script.
To convert to CoNLL, please see the corefconversion
project.
To train new models, please run the following commands:
bash -x -e setup_all.sh
bash -x -e setup_corpus_dem1921.sh
bash -x -e setup_training.sh
python3 train.py train_fr_ment
python3 train.py train_fr_coref
The training corpora are available here.
First, you will need to download BERT Multilingual:
wget https://storage.googleapis.com/bert_models/2018_11_23/multi_cased_L-12_H-768_A-12.zip
unzip multi_cased_L-12_H-768_A-12.zip
rm multi_cased_L-12_H-768_A-12.zip
To make predictions with our pretrained models, you will need to convert your corpus in the jsonlines
format (see the corefconversion
project). For a demo, you can use one the document of the Democrat corpus:
bash -x -e setup_all.sh
bash -x -e setup_models_dem1921.sh
bash -x -e setup_corpus_dem1921.sh
head -n 5 test.french.jsonlines | shuf | head -n 1 > myfile.jsonlines
python3 predict.py fr_ment,fr_coref myfile.jsonlines mypredictions.jsonlines
To convert to the text or CoNLL formats, please see the corefconversion
project. You may also want to import the annotated text into a software such as TXM for exploitation. Please see this script to learn how to do this.
Here is an example. This is an except from the Chartreuse de Parme (Stendhal), from Wikisource. This text is NOT part of the Democrat corpus, so it was totally unseen during the development and the training phases:
See detailed instructions for more information.
GPU
environment variable, which the code treats as shorthand for CUDA_VISIBLE_DEVICES
.This work is published under the terms of the Apache 2.0 licence. See the LICENSE
file for more information.
This work has been developed within the Democrat project (DEscription et MOdélisation des Chaînes de Référence: outils pour l'Annotation de corpus (en diachronie et en langues comparées) et le Traitement automatique) and the Alector project (Aide à la lecture pour enfants dyslexiques et faibles lecteurs), from the French National Research Agency (ANR) (ANR-16-CE28-0005 and ANR-15-CE38-0008, respectively).