-
Notifications
You must be signed in to change notification settings - Fork 20
Home
André Pires edited this page Apr 21, 2017
·
36 revisions
This wiki documents the development process for my master's thesis, named Entity and relation extraction from web content.
First, the HAREM dataset was used to perform NER using available tools, namely Stanford NER, NLTK, OpenNLP and spaCy.
- brat: annotation tool and annotated SIGARRA's news
- datasets: Keeps the datasets used
-
scripts:
- extra: scripts not yet used
- evaluation: scripts to compute the evaluation of all tools, using the conlleval script
-
filter-harem: scripts to manipulate HAREM dataset
- harem-to-opennlp: transform HAREM in opennlp input format
- harem-to-standoff: transform HAREM in standoff format, used in spacy
- harem-to-stanford: transform HAREM in conll format, used in stanfordNER
- src: source files for scripts
- run-scripts: commands to run scripts
-
tools:
- nltk: folder to keep nltk related data/scripts
- open-nlp: folder to keep open-nlp related data/scripts
- spacy: folder to keep spacy related data/scripts
- stanford-ner: folder to keep stanford NER related data/scripts
All programs were intended to be ran across HAREM with four different categories:
- Categories: use only categories
- Types: use only types
- Subtypes: use only subtypes
- Filtered: use filtered categories
Taking into account only the categories, the results, ordered by F-measure, were:
- Stanford CoreNLP: 53.53%
- OpenNLP: 53.63%
- SpaCy: 46.81%
- NLTK: 28.33%
Results for categories:
Tool | Precision | Recall | F-measure |
---|---|---|---|
Stanford CoreNLP | 55.67% | 51.05% | 53.22% |
OpenNLP | 55.43% | 51.94% | 53.63% |
SpaCy | 51.21% | 43.10% | 46.81% |
NLTK | 28.16% | 28.64% | 28.33% |
F-measure for all levels:
Tool | Categories | Types | Subtypes | Filtered |
---|---|---|---|---|
Stanford CoreNLP | 53.22% | - | - | 58.82% |
OpenNLP | 53.63% | 48.53% | 50.74% | 57.44% |
SpaCy | 46.81% | 44.04% | 37.86% | 49.22% |
NLTK | 28.33% | 24.88% | 20.08% | 30.32% |