Lec15 Qa
Lec15 Qa
Ques-on Answering
Wei Xu
(many slides from Greg Durrett)
QA is very broad
‣ Factoid QA: what states border Mississippi?, when was Barack Obama
born?
‣ Lots of this could be handled by QA from a knowledge base, if we had a
big enough knowledge base
‣ “Ques<on answering” as a term is so broad as to be meaningless
‣ Is P=NP?
‣ What is 4+5?
‣ What is the transla=on of [sentence] into French? [McCann et al.,
2018]
2
Classical Ques-on Answering
‣ Form seman-c representa-on from seman-c parsing, execute against
structured knowledge base
Q: “where was Barack Obama born”
????
Mary
‣ Uses fixed-size
representa-ons for the
final predic-on, mul-class
classifica-on
START END
query U
Sij = hi · uj
passage H
Seo et al. (2016)
Bidirec-onal Agen-on Flow
Each passage
word now “knows
about” the query
What was Marie Curie the first female recipient of ? [SEP] One of the most famous people born in Warsaw was Marie …
‣ BERT: transformer-based
approach with pretraining
on 3B tokens
SQuAD 2.0 SOTA: Spring 2019
SQuAD SOTA: Spring 19
‣ SQuAD 2.0: harder dataset
because some ques<ons
are unanswerable
‣ Industry contest
27
SQuAD 2.0 SOTA: Fall
SQuAD SOTA: Today 2019
‣ Performance is very
saturated
28
SQuAD 2.0 SOTA: Today
SQuAD SOTA: Today
‣ Performance is very
saturated
29
TriviaQA
‣ Totally figuring this
out is very challenging
‣ Coref:
the failed campaign
movie of the same name
‣ “Which film…”: can iden<fy movies and then spot keywords that
are related to the ques<on
31
Takeaways
‣ Many flavors of reading comprehension tasks: cloze or actual ques-ons,
single or mul--sentence
‣ Complex agen-on schemes can match queries against input texts and
iden-fy answers