A set of resources to perform the evaluation reported in our Automated interlinking of speech radio archives paper.
- kiwi-evaluation.py : script to run the TopN evaluation described in our paper.
- data/editorial-data : ground truth editorial data on a dataset of 132 items from BBC Programmes.
- data/automated-tags : a set of automated tags derived by the framework described in our paper.
- data/automated-transcripts : a set of automated transcripts, generated using CMU Sphinx, a HUB4 acoustic model and a Gigaword-derived language model.
The data was crawled from BBC Programmes on the 16th of May, 2012. Each file is named according to the following pattern: barcode_pid.json, where the barcode is used as an identifier across our different datasets, and the pid is the identifier of that programme on the BBC web site. For example, X0903717_p002h45s.json can be found here. This data holds the editorial tags we are evaluating against.
Each barcode.json holds the automatically derived tags for the programme identified by the barcode. The JSON has the following shape:
[ { "score": score, "link": DBpedia URI }, ... ]
This array is ordered by score descending.
Each sub-directory corresponds to a single programme, which barcode is the name of the directory. Within each sub-directory, there is one JSON file for a 2 minutes chunk of the programme. For example transcript-0.json will hold the automated transcript for the first chunk and transcript-1.json will hold the automated transcript for the second chunk.
The JSON has the following shape:
[ "full transcript", [ [ term, start, end, score 1, score 2 ], ... ]
Start and end are in seconds and score 1 and 2 respectively captures the acoustic model score and the language model score.
Running the evaluation with results from the automated tagging described in our paper.
$ python evaluation.py
- Fork this repository.
- Generate JSON files for your automated tags according to the format described above.
- Replace the content of the data/automated-tags directory with your tags.
- Run the evaluation script.
You can handle multiple results via git branch
.
$ git checkout -b myResults1
$ git commit -a
$ git checkout myResults1 && ./evaluation.py
See 'COPYING' and 'AUTHORS' files. The license in 'COPYING' only applies to the Python code and the automated transcripts and tags. The editorial data is under the same non-commercial license as the BBC Programmes data.
@inproceedings{ldow2012raimondetal,
author = {Yves Raimond and
Chris Lowis and
Roderick Hodgson and
Jonathan Tweed},
title = {Automated semantic tagging of speech audio},
booktitle = {WWW (Companion Volume)},
year = {2012},
pages = {405-408},
ee = {http://doi.acm.org/10.1145/2187980.2188060},
crossref = {DBLP:conf/www/2012c},
bibsource = {DBLP, http://dblp.uni-trier.de}
}