Abstract | ||
---|---|---|
Up to now, for most endoscopical computer aided celiac disease diagnosis approaches, image regions showing discriminative features have to be manually extracted by the physicians, prior to their automatized classification. This is obligatory to get idealistic and reliable data which is free from strong image degradations. On the one hand such a human interaction during endoscopy is subjective, expensive and tedious, but on the other hand state-of-the-art fully automatized selection corresponds to decreased classification accuracies compared to experienced human experts. In this work, a fully automatized approach is introduced which exploits the availability of a significant number of subimages within one original endoscopic image. A weighted decision-level and a weighted feature-level fusion method are introduced and investigated with respect to the achieved classification accuracies. The outcomes are compared with simple decision-level and feature-level fusion methods and the manual and the automatized patch selection. Finally, we show that the proposed feature-level fusion method outperforms all other automatized methods and comes close to manual patch selection. |
Year | DOI | Venue |
---|---|---|
2014 | 10.1007/978-3-319-11752-2_55 | PATTERN RECOGNITION, GCPR 2014 |
Field | DocType | Volume |
Pattern recognition,Computer-aided,Computer science,Human interaction,Speech recognition,Artificial intelligence,Information fusion,Discriminative model | Conference | 8753 |
ISSN | Citations | PageRank |
0302-9743 | 1 | 0.35 |
References | Authors | |
17 | 3 |
Name | Order | Citations | PageRank |
---|---|---|---|
Michael Gadermayr | 1 | 74 | 15.65 |
Andreas Uhl | 2 | 1958 | 223.07 |
Andreas Vécsei | 3 | 167 | 18.36 |