Open Research will be unavailable from 10.15am - 11am on Saturday 14th March 2026 AEDT due to scheduled maintenance.
 

Task 1 of the CLEF ehealth evaluation lab 2016: Handover information extraction

dc.contributor.authorSuominen, Hannaen
dc.contributor.authorZhou, Liyuanen
dc.contributor.authorGoeuriot, Lorraineen
dc.contributor.authorKelly, Liadhen
dc.date.accessioned2025-06-29T19:33:50Z
dc.date.available2025-06-29T19:33:50Z
dc.date.issued2016en
dc.description.abstractCascaded speech recognition (SR) and information extraction (IE) could support the best practice for clinical handover and release clinicians' time from writing documents to patient interaction and education. However, high requirements for processing correctness evoke methodological challenges and hence, processing correctness needs to be carefully evaluated as meeting the requirements. This overview paper reports on how these issues were addressed in a shared task of the eHealth evaluation lab of the Conference and Labs of the Evaluation Forum (CLEF) in 2016. This IE task built on the 2015 CLEF eHealth Task on SR by using its 201 synthetic handover documents for training and validation (appr. 8; 500 + 7; 700 words) and releasing another 100 documents with over 6; 500 expert-Annotated words for testing. It attracted 25 team registrations and 3 team submissions with 2 methods each. When using the macro-Averaged F1 over the 35 form headings present in the training documents for evaluation on the test documents, all participant methods outperformed all 4 baselines, including the organizers' method (F1 = 0:25), published in 2015 in a top-Tier medical informatics journal and provided to the participants as an option to build on, a random classifier (F1 = 0:02), and majority classifiers for the two most common classes (i.e., NA to filter out text irrelevant to the form and the most common form heading, both with F1 > 0:00). The top-2 methods (F1 = 0:38 and 0:37) had statistically significantly (p > 0:05, Wilcoxon signed-rank test) better performance than the third-best method (F1 = 0:35). In comparison, the top-3 methods and the organizers' method (7th) had F1 of 0.81, 0.80, 0.81, and 0.75 in the NA class, respectively.en
dc.description.statusPeer-revieweden
dc.format.extent14en
dc.identifier.issn1613-0073en
dc.identifier.otherORCID:/0000-0002-4195-1641/work/207330483en
dc.identifier.scopus84984820786en
dc.identifier.urihttp://www.scopus.com/inward/record.url?scp=84984820786&partnerID=8YFLogxKen
dc.identifier.urihttps://hdl.handle.net/1885/733765455
dc.language.isoenen
dc.relation.ispartofseries2016 Working Notes of Conference and Labs of the Evaluation Forum, CLEF 2016en
dc.sourceCEUR Workshop Proceedingsen
dc.subjectComputer Systems Evaluationen
dc.subjectData Collectionen
dc.subjectInformation Extractionen
dc.subjectMedical Informaticsen
dc.subjectNursing Recordsen
dc.subjectPatient Handoffen
dc.subjectPatient Handoveren
dc.subjectRecords as Topicen
dc.subjectSoftware Designen
dc.subjectSpeech Recognitionen
dc.subjectTest-set Generationen
dc.subjectText Classificationen
dc.titleTask 1 of the CLEF ehealth evaluation lab 2016: Handover information extractionen
dc.typeConference paperen
dspace.entity.typePublicationen
local.bibliographicCitation.lastpage14en
local.bibliographicCitation.startpage1en
local.contributor.affiliationSuominen, Hanna; School of Computing, ANU College of Systems and Society, The Australian National Universityen
local.contributor.affiliationZhou, Liyuan; School of Computing, ANU College of Systems and Society, The Australian National Universityen
local.contributor.affiliationGoeuriot, Lorraine; LIGen
local.contributor.affiliationKelly, Liadh; Trinity College Dublinen
local.identifier.ariespublicationu4334215xPUB1690en
local.identifier.citationvolume1609en
local.identifier.pured7373a4e-348a-4eab-b06f-4d1de31d57c9en
local.identifier.urlhttps://www.scopus.com/pages/publications/84984820786en
local.type.statusPublisheden

Downloads