The Best Explanation: Beyond Right and Wrong in Question Answering

Publikation: Bog/antologi/afhandling/rapportPh.d.-afhandlingForskning

Standard

The Best Explanation : Beyond Right and Wrong in Question Answering. / Johannsen, Anders Trærup.

Det Humanistiske Fakultet, Københavns Universitet, 2013. 187 s.

Publikation: Bog/antologi/afhandling/rapportPh.d.-afhandlingForskning

Harvard

Johannsen, AT 2013, The Best Explanation: Beyond Right and Wrong in Question Answering. Det Humanistiske Fakultet, Københavns Universitet.

APA

Johannsen, A. T. (2013). The Best Explanation: Beyond Right and Wrong in Question Answering. Det Humanistiske Fakultet, Københavns Universitet.

Vancouver

Johannsen AT. The Best Explanation: Beyond Right and Wrong in Question Answering. Det Humanistiske Fakultet, Københavns Universitet, 2013. 187 s.

Author

Johannsen, Anders Trærup. / The Best Explanation : Beyond Right and Wrong in Question Answering. Det Humanistiske Fakultet, Københavns Universitet, 2013. 187 s.

Bibtex

@phdthesis{80b2fef8257e44b4a4ffe6dfcc9c1c1a,
title = "The Best Explanation: Beyond Right and Wrong in Question Answering",
abstract = "There are right and wrong answers, but there are also ways of answering questions which are helpful and some which are not, even when they convey the same information. In this thesis we present a data-driven approach to automatically recognizing the good answers. A good answer lays out its information so that it is easy to read and understand. And answer structure matters—imagine assembling a piece of IKEA furniture if the order of instructions is scrambled. In general, text is structured by discourse relations, and discourse markers (DMs, e.g. however, moreover, then) are the most apparent and reliable signs of this structure. In the thesis we use DMs to model aspects of answer structure. Unfortunately, standard discourse processing software make two unrealistic assumptions about DMs, making it hard to apply to community-generated data. They are a) that gold-standard annotations are available for feature generation; and b) that DMs form a closed class for which we have labeled examples of all members. We challenge those assumptions, showing that a) in the absence of gold annotations, state-of-the-art performance can be obtained with much simpler features; and b) sharing features between DMs based on similarity via word embeddings gives an error reduction of at least 20% on unknown DMs, compared to no sharing or sharing by part-of-speech. Structure-building expressions are often more complex than the simple DMs discussed above and could even be discontinuous (e.g not only X but also Y). However, discovering such patterns automatically is a very hard search problem. As an alternative, we generate representations based on regular expressions using data elicited from workers on Mechanical Turk. Using these complex expressions for answer ranking gives an error reduction of 24% compared to a bag-of-words model. We introduce the task of ranking answers across domains, learning from questions and answers collected from community Q&A sites (cQA). In one experiment we show that importance sampling, where training data is sampled according to similarity between questions, leads to significant improvements over an uninformed sampling strategy.",
author = "Johannsen, {Anders Tr{\ae}rup}",
year = "2013",
language = "English",
publisher = "Det Humanistiske Fakultet, K{\o}benhavns Universitet",
address = "Denmark",

}

RIS

TY - BOOK

T1 - The Best Explanation

T2 - Beyond Right and Wrong in Question Answering

AU - Johannsen, Anders Trærup

PY - 2013

Y1 - 2013

N2 - There are right and wrong answers, but there are also ways of answering questions which are helpful and some which are not, even when they convey the same information. In this thesis we present a data-driven approach to automatically recognizing the good answers. A good answer lays out its information so that it is easy to read and understand. And answer structure matters—imagine assembling a piece of IKEA furniture if the order of instructions is scrambled. In general, text is structured by discourse relations, and discourse markers (DMs, e.g. however, moreover, then) are the most apparent and reliable signs of this structure. In the thesis we use DMs to model aspects of answer structure. Unfortunately, standard discourse processing software make two unrealistic assumptions about DMs, making it hard to apply to community-generated data. They are a) that gold-standard annotations are available for feature generation; and b) that DMs form a closed class for which we have labeled examples of all members. We challenge those assumptions, showing that a) in the absence of gold annotations, state-of-the-art performance can be obtained with much simpler features; and b) sharing features between DMs based on similarity via word embeddings gives an error reduction of at least 20% on unknown DMs, compared to no sharing or sharing by part-of-speech. Structure-building expressions are often more complex than the simple DMs discussed above and could even be discontinuous (e.g not only X but also Y). However, discovering such patterns automatically is a very hard search problem. As an alternative, we generate representations based on regular expressions using data elicited from workers on Mechanical Turk. Using these complex expressions for answer ranking gives an error reduction of 24% compared to a bag-of-words model. We introduce the task of ranking answers across domains, learning from questions and answers collected from community Q&A sites (cQA). In one experiment we show that importance sampling, where training data is sampled according to similarity between questions, leads to significant improvements over an uninformed sampling strategy.

AB - There are right and wrong answers, but there are also ways of answering questions which are helpful and some which are not, even when they convey the same information. In this thesis we present a data-driven approach to automatically recognizing the good answers. A good answer lays out its information so that it is easy to read and understand. And answer structure matters—imagine assembling a piece of IKEA furniture if the order of instructions is scrambled. In general, text is structured by discourse relations, and discourse markers (DMs, e.g. however, moreover, then) are the most apparent and reliable signs of this structure. In the thesis we use DMs to model aspects of answer structure. Unfortunately, standard discourse processing software make two unrealistic assumptions about DMs, making it hard to apply to community-generated data. They are a) that gold-standard annotations are available for feature generation; and b) that DMs form a closed class for which we have labeled examples of all members. We challenge those assumptions, showing that a) in the absence of gold annotations, state-of-the-art performance can be obtained with much simpler features; and b) sharing features between DMs based on similarity via word embeddings gives an error reduction of at least 20% on unknown DMs, compared to no sharing or sharing by part-of-speech. Structure-building expressions are often more complex than the simple DMs discussed above and could even be discontinuous (e.g not only X but also Y). However, discovering such patterns automatically is a very hard search problem. As an alternative, we generate representations based on regular expressions using data elicited from workers on Mechanical Turk. Using these complex expressions for answer ranking gives an error reduction of 24% compared to a bag-of-words model. We introduce the task of ranking answers across domains, learning from questions and answers collected from community Q&A sites (cQA). In one experiment we show that importance sampling, where training data is sampled according to similarity between questions, leads to significant improvements over an uninformed sampling strategy.

M3 - Ph.D. thesis

BT - The Best Explanation

PB - Det Humanistiske Fakultet, Københavns Universitet

ER -

ID: 95356609