Skip to content

Commit

Permalink
Add BibTeX citation to readme
Browse files Browse the repository at this point in the history
  • Loading branch information
benjaminvdb authored Dec 7, 2020
1 parent ed4e760 commit 812e9d0
Showing 1 changed file with 22 additions and 0 deletions.
22 changes: 22 additions & 0 deletions README.md
Original file line number Diff line number Diff line change
Expand Up @@ -136,6 +136,28 @@ v2: Removed advertisements from reviews and increased dataset size to 118,516.

v1: Initial release

## Citation

Please use the following citation when making use of this dataset in your work.

```
@article{DBLP:journals/corr/abs-1910-00896,
author = {Benjamin van der Burgh and
Suzan Verberne},
title = {The merits of Universal Language Model Fine-tuning for Small Datasets
- a case with Dutch book reviews},
journal = {CoRR},
volume = {abs/1910.00896},
year = {2019},
url = {http://arxiv.org/abs/1910.00896},
archivePrefix = {arXiv},
eprint = {1910.00896},
timestamp = {Fri, 04 Oct 2019 12:28:06 +0200},
biburl = {https://dblp.org/rec/journals/corr/abs-1910-00896.bib},
bibsource = {dblp computer science bibliography, https://dblp.org}
}
```

## Acknowledgements

This dataset was created for testing out the [ULMFiT](https://arxiv.org/abs/1801.06146) (by Jeremy Howard and Sebastian Ruder) deep learning algorithm for text classification. It is implemented in the [FastAI](https://github.com/fastai/fastai) Python library that has taught me a lot. I'd also like to thank [Timo Block](https://github.com/tblock) for making his [10kGNAD](https://github.com/tblock/10kGNAD) dataset publicly available and giving me a starting point for this dataset. The dataset structure based on the [Large Movie Review Dataset](http://ai.stanford.edu/~amaas/data/sentiment/) by Andrew L. Maas et al. Thanks to [Andreas van Cranenburg](https://github.com/andreasvc) for pointing out a problem with the dataset.
Expand Down

0 comments on commit 812e9d0

Please sign in to comment.