Abstract
Cascaded speech recognition (SR) and information extraction (IE) could support the best practice for clinical handover and release clinicians' time from writing documents to patient interaction and education. However, high requirements for processing correctness evoke methodological challenges and hence, processing correctness needs to be carefully evaluated as meeting the requirements. This overview paper reports on how these issues were addressed in a shared task of the eHealth evaluation lab of the Conference and Labs of the Evaluation Forum (CLEF) in 2016. This IE task built on the 2015 CLEF eHealth Task on SR by using its 201 synthetic handover documents for training and validation (appr. 8; 500 + 7; 700 words) and releasing another 100 documents with over 6; 500 expert-Annotated words for testing. It attracted 25 team registrations and 3 team submissions with 2 methods each. When using the macro-Averaged F1 over the 35 form headings present in the training documents for evaluation on the test documents, all participant methods outperformed all 4 baselines, including the organizers' method (F1 = 0:25), published in 2015 in a top-Tier medical informatics journal and provided to the participants as an option to build on, a random classifier (F1 = 0:02), and majority classifiers for the two most common classes (i.e., NA to filter out text irrelevant to the form and the most common form heading, both with F1 > 0:00). The top-2 methods (F1 = 0:38 and 0:37) had statistically significantly (p > 0:05, Wilcoxon signed-rank test) better performance than the third-best method (F1 = 0:35). In comparison, the top-3 methods and the organizers' method (7th) had F1 of 0.81, 0.80, 0.81, and 0.75 in the NA class, respectively.
Original language | English |
---|---|
Title of host publication | 2016 Working Notes of Conference and Labs of the Evaluation Forum |
Editors | Krisztian Balog, Linda Cappellato, Nicola Ferro, Craig Macdonald |
Place of Publication | Online |
Publisher | CEUR Workshop Proceedings |
Pages | 1-14 |
Number of pages | 14 |
Volume | 1609 |
Publication status | Published - 2016 |
Event | 7th International Conference of the CLEF Association, CLEF 2016 - Evora, Evora, Portugal Duration: 5 Sept 2016 → 8 Sept 2016 http://clef2016.clef-initiative.eu/ (Conference website) |
Publication series
Name | CEUR WS- 1609 - CLEF2016 Working Notes |
---|---|
Publisher | CEUR Workshop |
Volume | 1609 |
ISSN (Print) | 1613-0073 |
Conference
Conference | 7th International Conference of the CLEF Association, CLEF 2016 |
---|---|
Abbreviated title | CLEF 2016 |
Country/Territory | Portugal |
City | Evora |
Period | 5/09/16 → 8/09/16 |
Other | CLEF 2016 is the seventh CLEF conference continuing the popular CLEF campaigns which have run since 2000 contributing to the systematic evaluation of information access systems, primarily through experimentation on shared tasks. Building on the format first introduced in 2010, CLEF 2016 consists of an independent peer-reviewed conference on a broad range of issues in the fields of multilingual and multimodal information access evaluation, and a set of labs and workshops designed to test different aspects of mono and cross-language Information retrieval systems. Together, the conference and the lab series will maintain and expand upon the CLEF tradition of community-based evaluation and discussion on evaluation issues |
Internet address |
|