4.3 Answer extraction
a subject of debate. These are some top results:
— U M good : “Most Classicists would agree that, whether
In this phase, the clustered documents are filtered
there was ever such a composer as "Homer" or not, the
based on the user model and answer sentences are
Homeric poems are the product of an oral tradition [. . . ]
located and formatted for presentation.
Could the Iliad and Odyssey have been oral-formulaic po-
UM-based filtering The documents in the clus-
ems, composed on the spot by the poet using a collection of
ter tree are filtered according to their reading diffi-
memorized traditional verses and phases?”
culty: only those compatible with the UM’s read-
— U M med : “No reliable ancient evidence for Homer –
ing level are retained for further analysis 6 .
[. . . ] General ancient assumption that same poet wrote Il-
iad and Odyssey (and possibly other poems) questioned by
Semantic similarity Within each of the retained
many modern scholars: differences explained biographi-
documents, we seek the sentences which are se-
cally in ancient world (e g wrote Od. in old age); but simi-
mantically most relevant to the query by applying
larities could be due to imitation.”
the metric in (Alfonseca et al., 2001): we rep-
— U M poor : “Homer wrote The Iliad and The Odyssey
resent each document sentence p and the query
(at least, supposedly a blind bard named "Homer" did).”
q as word sets P = {pw 1 , . . . , pw m } and Q =
In the three results, the problem of attribution of
{qw 1 , . . . , qw n }. The distance from p to q is then
the Iliad is made clearly visible: document pas-
dist q (p) = P 1 ≤i≤m min j [d(pw i , qw j )], where
sages provide a context which helps to explain the
d(pw i , qw j ) is the word-level distance between
controversy at different levels of difficulty.
pw i and qw j based on (Jiang and Conrath, 1997).
6 Evaluation
Ranking Given the query q, we thus locate
in each document D the sentence p ∗ such that
Since YourQA does not single out one correct an-
p ∗ = argmin p∈D [dist q (p)]; then, dist q (p ∗ ) be-
swer phrase, TREC evaluation metrics are not suit-
comes the document score. Moreover, each clus-
able for it. A user-centred methodology to assess
5The likelihood is estimated using the formula:
how individual information needs are met is more
L
i,D = P
appropriate. We base our evaluation on (Su, 2003),
w∈DC(w, D) · log(P (w|lm
i)), where w is a
word in the document, C(w, d) is the number of occurrences
which proposes a comprehensive search engine
of w in D and P (w|lm
i) is the probability with which w
evaluation model, defining the following metrics:
occurs in lm
i6However, if their number does not exceed a given thresh-
Bạn đang xem 4. - BÁO CÁO KHOA HỌC ADAPTIVITY IN QUESTION ANSWERING WITH USER MODELLING AND A DIALOGUE INTERFACE PPTX