CS474 Intro to NLP. List Questions. Context Task. Definition Questions. Question Answering. List questions

Similar documents
Towards open-domain QA. Question answering. TReC QA framework. TReC QA: evaluation

The answer (circa 2001)

Question answering. CS474 Intro to Natural Language Processing. History LUNAR

CS674 Natural Language Processing

Modeling Sequence Data

Information Retrieval. CS630 Representing and Accessing Digital Information. What is a Retrieval Model? Basic IR Processes

Question Answering Approach Using a WordNet-based Answer Type Taxonomy

Final Project Discussion. Adam Meyers Montclair State University

Question Answering Using XML-Tagged Documents

CS54701: Information Retrieval

CS473: Course Review CS-473. Luo Si Department of Computer Science Purdue University

Shrey Patel B.E. Computer Engineering, Gujarat Technological University, Ahmedabad, Gujarat, India

Manning Chapter: Text Retrieval (Selections) Text Retrieval Tasks. Vorhees & Harman (Bulkpack) Evaluation The Vector Space Model Advanced Techniques

Automatic Summarization

Answer Extraction. NLP Systems and Applications Ling573 May 13, 2014

Making Sense Out of the Web

CS 6320 Natural Language Processing

Information Retrieval: Retrieval Models

A Survey Of Different Text Mining Techniques Varsha C. Pande 1 and Dr. A.S. Khandelwal 2

Natural Language Processing. SoSe Question Answering

INTERNATIONAL JOURNAL OF COMPUTER ENGINEERING & TECHNOLOGY (IJCET) CONTEXT SENSITIVE TEXT SUMMARIZATION USING HIERARCHICAL CLUSTERING ALGORITHM

Search Engines. Information Retrieval in Practice

Natural Language Processing SoSe Question Answering. (based on the slides of Dr. Saeedeh Momtazi)

Query Difficulty Prediction for Contextual Image Retrieval

Chapter 27 Introduction to Information Retrieval and Web Search

Information Retrieval CSCI

Chapter 4. Processing Text

Information Retrieval

Question Answering Systems

Information Retrieval CS Lecture 01. Razvan C. Bunescu School of Electrical Engineering and Computer Science

Ranking in a Domain Specific Search Engine

Semantics Isn t Easy Thoughts on the Way Forward

Introduction to Text Mining. Hongning Wang

What is this Song About?: Identification of Keywords in Bollywood Lyrics

Building Test Collections. Donna Harman National Institute of Standards and Technology

Natural Language Processing SoSe Question Answering. (based on the slides of Dr. Saeedeh Momtazi) )

Taming Text. How to Find, Organize, and Manipulate It MANNING GRANT S. INGERSOLL THOMAS S. MORTON ANDREW L. KARRIS. Shelter Island

Let s get parsing! Each component processes the Doc object, then passes it on. doc.is_parsed attribute checks whether a Doc object has been parsed

To search and summarize on Internet with Human Language Technology

QANUS A GENERIC QUESTION-ANSWERING FRAMEWORK

CHAPTER 5 SEARCH ENGINE USING SEMANTIC CONCEPTS

It s time for a semantic engine!

Topics for Today. The Last (i.e. Final) Class. Weakly Supervised Approaches. Weakly supervised learning algorithms (for NP coreference resolution)

Optimal Query. Assume that the relevant set of documents C r. 1 N C r d j. d j. Where N is the total number of documents.

More on indexing CE-324: Modern Information Retrieval Sharif University of Technology

slide courtesy of D. Yarowsky Splitting Words a.k.a. Word Sense Disambiguation Intro to NLP - J. Eisner 1

Domain-specific Concept-based Information Retrieval System

Exam Marco Kuhlmann. This exam consists of three parts:

Department of Electronic Engineering FINAL YEAR PROJECT REPORT

Boolean Model. Hongning Wang

Information Retrieval. (M&S Ch 15)

Classification and Comparative Analysis of Passage Retrieval Methods

Ngram Search Engine with Patterns Combining Token, POS, Chunk and NE Information

Semantic Extensions to Syntactic Analysis of Queries Ben Handy, Rohini Rajaraman

Dynamic Feature Selection for Dependency Parsing

Multimedia Information Systems

Cross Lingual Question Answering using CINDI_QA for 2007

TIC: A Topic-based Intelligent Crawler

Entity and Knowledge Base-oriented Information Retrieval

CS 572: Information Retrieval. Lecture 1: Course Overview and Introduction 11 January 2016

Text Mining: A Burgeoning technology for knowledge extraction

Dmesure: a readability platform for French as a foreign language

Syntactic Analysis. CS345H: Programming Languages. Lecture 3: Lexical Analysis. Outline. Lexical Analysis. What is a Token? Tokens

A Method for Semi-Automatic Ontology Acquisition from a Corporate Intranet

NTUBROWS System for NTCIR-7. Information Retrieval for Question Answering

Ranked Retrieval. Evaluation in IR. One option is to average the precision scores at discrete. points on the ROC curve But which points?

Knowledge Retrieval. Franz J. Kurfess. Computer Science Department California Polytechnic State University San Luis Obispo, CA, U.S.A.

Knowledge Engineering with Semantic Web Technologies

Finding Related Entities by Retrieving Relations: UIUC at TREC 2009 Entity Track

WEB SEARCH, FILTERING, AND TEXT MINING: TECHNOLOGY FOR A NEW ERA OF INFORMATION ACCESS

QUALIFIER: Question Answering by Lexical Fabric and External Resources

Precise Medication Extraction using Agile Text Mining

Learning to Answer Biomedical Factoid & List Questions: OAQA at BioASQ 3B

Information Retrieval

Sheffield University and the TREC 2004 Genomics Track: Query Expansion Using Synonymous Terms

GTE: DESCRIPTION OF THE TIA SYSTEM USED FOR MUC- 3

ResPubliQA 2010

NLP Final Project Fall 2015, Due Friday, December 18

M erg in g C lassifiers for Im p ro v ed In fo rm a tio n R e triev a l

S A I. Semantic Search via XML Fragments: A High-Precision Approach to IR. Jennifer Chu-Carroll. Carroll IBM T.J. Watson Research Center.

R 2 D 2 at NTCIR-4 Web Retrieval Task

Information Retrieval

Information Retrieval and Web Search

ΕΠΛ660. Ανάκτηση µε το µοντέλο διανυσµατικού χώρου

CMPSCI 646, Information Retrieval (Fall 2003)

To access a search engine go to the search engine s web site (i.e. yahoo.com).

Web Search Engine Question Answering

Parsing tree matching based question answering

Natural Language Processing Tutorial May 26 & 27, 2011

How to.. What is the point of it?

An Evaluation of Information Retrieval Accuracy. with Simulated OCR Output. K. Taghva z, and J. Borsack z. University of Massachusetts, Amherst

TEXT CHAPTER 5. W. Bruce Croft BACKGROUND

SEARCH ENGINE INSIDE OUT

Natural Language Processing

A CASE STUDY: Structure learning for Part-of-Speech Tagging. Danilo Croce WMR 2011/2012

Improving Synoptic Querying for Source Retrieval

modern database systems lecture 4 : information retrieval

Watson & WMR2017. (slides mostly derived from Jim Hendler and Simon Ellis, Rensselaer Polytechnic Institute, or from IBM itself)

Compilers and Interpreters

Natural Language Processing with Deep Learning CS224N/Ling284

Transcription:

CS474 Intro to NLP List questions List Questions Question Answering Last class History Open-domain QA Basic system architecture Today Finish basic system architecture Predictive indexing methods Pattern-matching methods Can t just rely on a single document Performance TREC 2003: F.40 TREC 2004: F.62 Who is Colin Powell? What is mold? Definition Questions Context Task Track a target discourse object through a series of questions Hard to evaluate Who is the audience? Evaluation requires matching concepts in the desired response to concepts in a system response TREC 2003: Audience: questioner is an adult, a native speaker of English, and an average reader of US newspapers Results: F.55 Performance TREC 2004 Factoids:.84 initial;.74 non-initial Lists:.62 F Other:.46 F

Question answering Overview and task definition History Open-domain question answering Basic system architecture [Cardie et al., ANLP 2000] Predictive indexing methods Pattern-matching methods question document collection IR IR Subsystems Basic system architecture documents, text passages Linguistic Filters Filters answer hypotheses 1 st Guess 2 nd Guess 3 rd Guess 4 th Guess 5 th Guess System architecture: document retrieval QA as document retrieval question Document Retrieval document collection IR Subsystems documents, text passages Linguistic Filters Filters answer hypotheses 1 st Guess 2 nd Guess 3 rd Guess 4 th Guess 5 th Guess Question SMART 10-word chunker Guesses Document The chaotic development that is gobbling up the Amazon rain forest could finally be reined in with a new plan developed by officials of Amazon countries. Sixty percent of the Amazon, the world s largest tropical rain forest, lies in Brazil, but the forest also Japan will not fund construction of the final segment of a controversial highway through the Amazon rain forest in Brazil. Sen. Bob Kasten of Wisconsin said Japan s ambassador to the United States, Nobuo Matsunga, assu

Document retrieval only Corpus Baseline evaluation TREC-8 development corpus (38 questions) TREC-8 test corpus (200 questions) Development (38) Test (200) Correct MAR Correct MAR Smart 3 3.33 29 3.07 question System architecture: passage retrieval document collection Doc Doc Retrieval Passage Retrieval IR Subsystems documents, text passages Linguistic Filters Filters answer hypotheses 1 st Guess 2 nd Guess 3 rd Guess 4 th Guess 5 th Guess MAR = Mean Answer Rank Passage retrieval Query-dependent text summarization Which country has the largest part of the Amazon rain forest? [The chaotic development that is gobbling up the Amazon rain forest could finally be reined in with a new plan developed by leading scientists from around the world.] [ That s some of the most encouraging news about the Amazon rain forest in recent years, said Thomas Lovejoy, an Amazon specialist.] [ It contrasts markedly with a year ago, when there was nothing to read about conservation in the Amazon. ] [Sixty percent of the Amazon, the world s largest tropical rain forest, lies in Brazil.] [Salton et al.] Extract passages that best summarize each document w.r.t. the query Query-dependent text summarization Basic algorithm 1. Decide on a summary length (10% of document length). 2. Use standard ad-hoc retrieval algorithm to retrieve top documents. 3. Treat each sentence/paragraph in top N documents as a document itself. Use standard document similarity equations to assign a similarity score to the sentence/paragraph. 4. Return highest-scoring sentences/paragraphs as the summary, subject to the length constraint.

Passage retrieval Query-dependent text summarization Which country has the largest part of the Amazon rain forest? [The chaotic development that is gobbling up the Amazon rain forest could finally be reined in with a new plan developed by leading scientists from around the world.] [ That s some of the most encouraging news about the Amazon rain forest in recent years, said Thomas Lovejoy, an Amazon specialist.] [ It contrasts markedly with a year ago, when there was nothing to read about conservation in the Amazon. ] [Sixty percent of the Amazon, the world s largest tropical rain forest, lies in Brazil.] [Salton et al.] Sort summary extracts across top k documents answer hypotheses ordered list of summary extracts QA as query-dependent text summarization Question Sentence reordering 10-word chunker Guesses SMART Summaries Summaries The chaotic development that is gobbling up the Amazon rain forest could finally be reined in with a new plan developed by officials of Amazon countries. Sixty percent of the Amazon, the world s largest tropical rain forest, lies in Brazil, but the forest also Japan will not fund construction of the final segment of a controversial highway through the Amazon rain forest in Brazil. Sen. Bob Kasten of Wisconsin said Japan s ambassador to the United States, Nobuo Matsunga, assu Evaluation: text summarization Evaluation: text summarization Development (38) Test (200) Correct MAR Correct MAR Smart 3 3.33 29 3.07 Text Summarization 4 2.25 45 2.67 MAR = Mean Answer Rank Summarization method can limit performance Development corpus In only 23 of the 38 developments questions (61%) does the correct answer appear in the summary for one of the top k=7 documents Test corpus In only 135 of the 200 developments questions (67.5%) does the correct answer appear in the summary for one of the top (k=6) documents

Linguistic filters System architecture: linguistic filters 50 byte answer length effectively eliminates how or why questions almost all of the remaining question types are likely to have noun phrases as answers development corpus: 36 of 38 questions have noun phrase answers consider adding at least a simple linguistic filter that considers only noun phrases as answer hypotheses question document collection Doc Doc Retrieval Passage Retrieval documents, text passages Noun Noun phrase phrase filter filter Linguistic filters answer hypotheses 1 st Guess 2 nd Guess 3 rd Guess 4 th Guess 5 th Guess The noun phrase filter QA using the NP filter Which country has the largest part of the Amazon rain forest? ordered list of summary extracts [The huge Amazon rain forest] is regarded as vital to [the global environment]. [Japan] will not fund [the construction] of [the final segment] of [a controversial highway] through [the Amazon rain forest] in [Brazil], according to [a senior Republican senator]. ordered list of NPs answer hypotheses Question 10-word chunker Guesses SMART Sentence reordering Summaries NP finder [Cardie & Pierce ACL98] NPs Summaries The chaotic development that is gobbling up the Amazon rain forest could finally be reined in with a new plan developed by officials of Amazon countries. Sixty percent of the Amazon, the world s largest tropical rain forest, lies in Brazil, but the forest also Japan will not fund construction of the final segment of a controversial highway through the Amazon rain forest in Brazil. Sen. Bob Kasten of Wisconsin said Japan s ambassador to the United States, Nobuo Matsunga, assu

Chunking answer hypotheses: BAD Which country has the largest part of the Amazon rain forest? Evaluation: NP filter Development (38) Test (200) Correct MAR Correct MAR Smart 3 3.33 29 3.07 Text Summarization 4 2.25 45 2.67 Question-Answering System Japan Brazil a new plan Amazon countries A section Northwestern Brazil A plan the Amazon region eight surrounding countries A Brazilian company union leader his modest wooden house The people defense Brazilian wealth the international market TS + NPs 7 2.29 50 2.66 MAR = Mean Answer Rank Using NP finder of Cardie & Pierce (1998) ~94% precision and recall on Wall Street Journal text How much does the (unnatural) NP chunking help? Without it, only 1 and 20 questions answered for each corpus, respectively NP filter is extracting good guesses, but better linguistic processing is needed to promote the best guesses to the top of the ranked guess list Semantic class checking Semantic class checking Question SMART Summaries Sentence reordering NPs NP finder Semantically compatible with question type? Semantic class checker 10-word chunker Guesses The chaotic development that is gobbling up the Amazon rain forest could finally be reined in with a new plan developed by officials of Amazon countries. Sixty percent of the Amazon, the world s largest tropical rain forest, lies in Brazil, but the forest also Japan will not fund construction of the final segment of a controversial highway through the Amazon rain forest in Brazil. Sen. Bob Kasten of Wisconsin said Japan s ambassador to the United States, Nobuo Matsunga, assu Approximate question type using question word Who is the president of the U.S.? person Which country has the largest part of the Amazon rain forest? country Where is the Connecticut River? state? county? country? location? What fabric should one use to make curtains? fabric??? Check that head noun (i.e. the last noun) of answer NP is of the same type a man = person Massachusetts = state, location

Semantic type checking Evaluation: semantic class filter Use lexical resource to determine semantic compatibility WordNet! Proper names handled separately since they are unlikely to appear in WordNet Small set (~20) rules entity, something object, physical object location region district, territory administrative district country, state, nation South American country Brazil Development (38) Test (200) Correct MAR Correct MAR Smart 3 3.33 29 3.07 Text Summarization 4 2.25 45 2.67 TS + NPs 7 2.29 50 2.66 TS + NPs + Semantic Type 21 1.38 86 1.90 MAR = Mean Answer Rank Weak syntactic and semantic information allows large improvements Problems? Sources of error Question answering Smart TS Ling Filters Overview and task definition History 6% 31% 40% 36% Open-domain question answering Basic system architecture Predictive indexing methods 63% 24% Slides based on those of Jamie Callan, CMU Pattern-matching methods Development questions (38) Test questions (200)

Indexing with predictive annotation Predictive annotation Some answers belong to well-defined semantic classes People, places, monetary amounts, telephone numbers, addresses, organizations Predictive annotation: index a document with concepts or features that are expected to be useful in (many) queries E.g. people names, location names, addresses, etc. Add additional operators for use in queries E.g. Where does Ellen Vorhees work? Ellen Vorhees NEAR/10 *organization Predictive annotation How is annotated text stored in the index? Treat <$QA-token, term> as meaning that $QA-token and term occur at the same location in the text Or use phrase indexing approach to index as a single item Issues for predictive annotation What makes a good QA-token? Question that would use the token Can be recognized with high reliability (high precision) Occurs frequently enough to be worth the effort How do you want the system to make use of the QA-tokens? Filtering step? Transform original question into an ad-hoc retrieval question that incorporates QA-tokens and proximity operators? Common approaches to recognizing QA-tokens Tables, lists, dictionaries Heuristics Hidden Markov models

Advantages and disadvantages + Most of the computational cost occurs during indexing Allows use of more sophisticated methods + Annotator has access to complete text of document Important for recognizing some types of features Must know ahead of time which types of features/concepts are likely to be important Increases size of index considerably E.g. by an order of magnitude if many features Used (in varying amounts) by almost all open-domain Q/A systems Question answering Overview and task definition History Open-domain question answering Basic system architecture Predictive indexing methods Pattern-matching methods Slides based on those of Jamie Callan, CMU Simple pattern-based QA Observation: there are many questions but fewer types of questions Each type of question can be associated with Expectations about answer string characteristics Strategies for retrieving documents that might have answers Rulesfor identifying answer strings in documents Example Who is the President of Cornell? Expectation: answer string contains a person name Named entity identification Search query: president Cornell *PersonName Rule: *PersonName, President of Cornell Matches David Skorten, President of Cornell Answer = David Skorten

Question analysis Input: the question Output Search query Answer expectations Extraction strategy Requires Identifying named entities Categorizing the question Matching question parts to templates Method: pattern-matching Analysis patterns still created manually Question analysis example Who is Elvis? Question type: who Named-entity tagging: Who is <personname>elvis</person-name> Analysis pattern: if question type = who and question contains <person-name> then Search query doesn t need to contain a *PersonName operator Desired answer probably is a description Likely answer extraction patterns Elvis, the X» Elvis, the king of rock and roll the X Elvis» the legendary entertainer Elvis Frequency of question types on an Internet search engine 42% what 21% where 20% who 8% when 8% why 2% which 0% how Question analysis Relative difficulty of question types Whatis difficult What time What country Whereis easy Whois easy Whenis easy Whyis hard Whichis hard Howis hard Example: What is Jupiter? 1. What We Will Learn from Galileo 2. The Nature of Things: Jupiter s shockwaves How a comet s bombardment has sparked activity on Earth 3. Jupiter-Bound Spacecraft Visits Earth on 6-Year Journey 4. STAR OF THE MAGI THEORIES ECLIPSED? 5. Marketing & Media: Hearst, Burda to Scrap New Astrology Magazine 6. Greece, Italy Conflict On Cause Of Ship Crash That Kills 2, Injures 54 7. Interplanetary Spacecraft To `Visit` Earth With LaserGraphic 8. A List of Events During NASA s Galileo Mission to Jupiter 9. SHUTTLE ALOFT, SENDS GALILEO ON 6-YEAR VOYAGE TO JUPITER 10. Rebuilt Galileo Probe readied For Long Voyage To Jupiter

Answer extraction Select highly ranked sentences from highly ranked documents Perform named-entity tagging (or extract from index) and perform part of speech tagging The/DT planet/nn <location>jupiter/nnp</location> and/cc its/prp moons/nns are/vbp in/in effect/nn a/dt mini-solar/jj system/nn,/, and/cc <location>jupiter/nnp</location> itself/prp is/vbz often/rb called/vbn a/dt star/nn that/in never/rb caught/vbn fire/nn./. Apply extraction patterns the/dt X Y, Y=Jupiter -> the planet Jupiter -> planet Simple pattern-based Q/A: assessment Extremely effective when Question patterns are predictable Fairly few patterns cover the most likely questions Could be several hundred Not much variation in vocabulary Simple word matching works The corpus is huge (e.g., Web) Odds of finding an answer document that matches the vocabulary and answer extraction rule improves Somewhat labor intensive Patterns are created and tested manually Common problem: matching questions to answers Document word order isn t exactly what was expected Solution: soft matching of answer patterns to document text Approach: use distance-based answer selection when no rule matches E.g. for What is Hunter Rawlings address? Use the address nearest to the words Hunter Rawlings User the address in the same sentence as Hunter Rawlings Common problem: matching questions to answers Answer vocabulary doesn t exactly match question vocabulary Solution: bridge the vocabulary mismatch Approach: use WordNet to identify simple relationships astronaut is a type of person astronaut and cosmonaut are synonyms

Common problem: improving the set of retrieved documents Sometimes the IR system can t find any documents that have answers (even though the right documents are in the corpus) Solution: get a broader set of documents Approach: if answer extractor fails to find an answer, send the question back to the search engine with instructions to widen the search Assumes answer extractors can tell when they fail Approach: use a variety of retrieval strategies to retrieve documents E.g., all words within one sentence, then all words within one paragraph, then within same document, E.g. add synonyms to query or do query expansion Simple, but much higher computational expense Common problem: improving answer extraction patterns Word sequence patterns have limited power Solution: create patterns that use syntactic information Partial syntactic parsing of documents Is this noun the subject or the object of the sentence? Allows more complex patterns Question: Who shot Kennedy? Who implies a person that should be subject of answer sentence/clause Kennedy should be direct object of answer Pattern: <subject> shot Kennedy Matching text: Oswald shot Kennedy Common problem: selecting/ranking the answer Multiple answer candidates Solutions Features used to represent answer candidates Frequency Distance to question words Location in answer passage(s) Selection functions Created manually Learned from training data