Help Privacy Policy Disclaimer
  Advanced SearchBrowse





Human Computing and Crowdsourcing Methods for Knowledge Acquisition


Kondreddi,  Sarath Kumar
Databases and Information Systems, MPI for Informatics, Max Planck Society;
International Max Planck Research School, MPI for Informatics, Max Planck Society;

Fulltext (restricted access)
There are currently no full texts shared for your IP range.
Fulltext (public)
There are no public fulltexts stored in PuRe
Supplementary Material (public)
There is no public supplementary material available

Kondreddi, S. K. (2014). Human Computing and Crowdsourcing Methods for Knowledge Acquisition. PhD Thesis, Universität des Saarlandes, Saarbrücken. doi:10.22028/D291-26564.

Cite as: https://hdl.handle.net/11858/00-001M-0000-0024-3C3D-F
Ambiguity, complexity, and diversity in natural language textual expressions
are major hindrances to automated knowledge extraction. As a result
state-of-the-art methods for extracting entities and relationships from
unstructured data make incorrect extractions or produce noise. With the advent
of human computing, computationally hard tasks have been addressed through
human inputs. While text-based knowledge acquisition can benefit from this
approach, humans alone cannot bear the burden of extracting knowledge from the
vast textual resources that exist today. Even making payments for crowdsourced
acquisition can quickly become prohibitively expensive.
In this thesis we present principled methods that effectively garner human
computing inputs for improving the extraction of knowledge-base facts from
natural language texts. Our methods complement automatic extraction techniques
with human computing to reap the benefits of both while overcoming each other�s
limitations. We present the architecture and implementation of HIGGINS, a
system that combines an information extraction (IE) engine with a human
computing (HC) engine to produce high quality facts. The IE engine combines
statistics derived from large Web corpora with semantic resources like WordNet
and ConceptNet to construct a large dictionary of entity and relational
phrases. It employs specifically designed statistical language models for
phrase relatedness to come up with questions and relevant candidate answers
that are presented to human workers. Through extensive experiments we establish
the superiority of this approach in extracting relation-centric facts from
text. In our experiments we extract facts about fictitious characters in
narrative text, where the issues of diversity and complexity in expressing
relations are far more pronounced. Finally, we also demonstrate how interesting
human computing games can be designed for knowledge acquisition tasks.