Scalable Information Extraction and Integration - PowerPoint PPT Presentation

1 / 76
About This Presentation
Title:

Scalable Information Extraction and Integration

Description:

Scalable Information Extraction and Integration Eugene Agichtein Microsoft Research Emory University Sunita Sarawagi IIT Bombay – PowerPoint PPT presentation

Number of Views:223
Avg rating:3.0/5.0
Slides: 77
Provided by: euge148
Category:

less

Transcript and Presenter's Notes

Title: Scalable Information Extraction and Integration


1
Scalable InformationExtraction and Integration
  • Eugene Agichtein Microsoft Research ? Emory
    University
  • Sunita Sarawagi IIT Bombay

2
The Value of Text Data
  • Unstructured text data is the primary source of
    human-generated information
  • Citeseer, comparison shopping, PIM systems, web
    search, data warehousing
  • Managing and utilizing text information
    extraction and integration
  • Scalability a bottleneck for deployment
  • Relevance to data mining community

3
Example Answering Queries Over Text
For years, Microsoft Corporation CEO Bill Gates
was against open source. But today he appears to
have changed his mind. "We can be open source. We
love the concept of shared source," said Bill
Veghte, a Microsoft VP. "That's a super-important
shift for us in terms of code access. Richard
Stallman, founder of the Free Software
Foundation, countered saying
Select Name From PEOPLE Where Organization
Microsoft
PEOPLE
Name Title Organization Bill Gates
CEO Microsoft Bill Veghte VP
Microsoft Richard Stallman Founder Free
Soft..
Bill Gates Bill Veghte
(from William Cohens IE tutorial, 2003)
4
Managing Unstructured Text Data
  • Information Extraction from text
  • Represent information in text data in a
    structured form
  • Identify instances of entities and relationships
  • Main approaches and architectures
  • Scaling up to large collections of documents
    (e.g., web)
  • Information Integration
  • Combine/resolve/clean information about entities
  • Entity Resolution Deduplication
  • Scaling Up Batch mode/algorithmic issues
  • Connections between Information Extraction and
    Integration
  • Coreference Resolution
  • Deriving values from multiple sources
  • (Web) Question Answering

5
Part I Tutorial Outline
  • Overview of Information Extraction
  • Entity tagging
  • Relation extraction
  • Scaling up Information Extraction
  • Focus on scaling up to large collections (where
    data mining and ML techniques shine)
  • Other dimensions of scalability

6
Information Extraction Components
7
Information Extraction Tasks
  • Extracting entities and relations this tutorial
  • Entities named (e.g., Person) and generic (e.g.,
    disease name)
  • Relations entities related in a predefined way
    (e.g., Location of a Disease outbreak)
  • Common extraction subtasks
  • Preprocessing sentence chunking, syntactic
    parsing, morphological analysis
  • Creating rules or extraction patterns manual,
    machine learning, and hybrid
  • Applying extraction patterns to extract new
    information
  • Postprocessing and complex extraction not
    covered
  • Co-reference resolution
  • Combining Relations into Events and Facts

8
Related Tutorials
  • Previous information extraction tutorials
    consult for more details
  • R. Feldman, Information Extraction Theory and
    Practice, ICML 2006http//www.cs.biu.ac.il/feldm
    an/icml_tutorial.html
  • W. Cohen, A. McCallum, Information Extraction and
    Integration an Overview, KDD 2003
    http//www.cs.cmu.edu/wcohen/ie-survey.ppt
  • A. Doan, R. Ramakrishnan, S. Vaithyanathan,
    Managing Information Extraction, SIGMOD06
  • N. Koudas, D. Srivastava, S. Sarawagi, Record
    Linkage Similarity Measures and Algorithms,
    SIGMOD 2006

9
Entity Tagging
  • Identifying mentions of entities (e.g., person
    names, locations, companies) in text
  • MUC (1997) Person, Location, Organization,
    Date/Time/Currency
  • ACE (2005) more than 100 more specific types
  • Hand-coded vs. Machine Learning approaches
  • Best approach depends on entity type and domain
  • Closed class (e.g., geographical locations,
    disease names, gene protein names) hand coded
    dictionaries
  • Syntactic (e.g., phone numbers, zipcodes)
    regexes
  • Others (e.g., person and company names) mixture
    of context, syntactic features, dictionaries,
    heuristics, etc.
  • Almost solved for common/typical entity types
  • Non-syntactic entities computationally expensive

10
Example Extracting Entities from Text
  • Useful for data warehousing, data cleaning, web
    data integration

Address
4089 Whispering Pines Nobel Drive San Diego CA
92122
1
Ronald Fagin, Combining Fuzzy Information from
Multiple Systems, Proc. of ACM SIGMOD, 2002
Citation
Segment(si) Sequence Label(si)
S1 Ronald Fagin Author
S2 Combining Fuzzy Information from Multiple Systems Title
S3 Proc. of ACM SIGMOD Conference
S4 2002 Year
11
Hand-Coded Methods
  • Easy to construct in many cases
  • e.g., to recognize prices, phone numbers, zip
    codes, conference names, etc.
  • Easier to debug maintain
  • Especially if written in a high-level language
    (as is usually the case) e.g.,
  • Easier to incorporate / reuse domain knowledge
  • Can be quite labor intensive to write

From Avatar
12
Example of Hand-Coded Entity Tagger
Ramakrishnan. G, 2005, Slides from Doan et al.,
SIGMOD 2006
Rule 1 This rule will find person names with a
salutation (e.g. Dr. Laura Haas) and two
capitalized words
lttokengt INITIALlt/tokengt lttokengtDOT
lt/tokengt lttokengtCAPSWORDlt/tokengt lttokengtCAPSWORDlt/
tokengt
Rule 2 This rule will find person names where two
capitalized words are present in a Person
dictionary
lttokengtPERSONDICT, CAPSWORD lt/tokengt lttokengtPERSON
DICT, CAPSWORDlt/tokengt
CAPSWORD Word starting with uppercase, second
letter lowercase E.g., DeWitt will
satisfy it (DEWITT will not)
\pUpper\pLower\pAlpha1,25 DOT
The character .
13
Hand Coded Rule Example Conference Name
These are subordinate patternswordOrdinals"(?
firstsecondthirdfourthfifthsixthseventheig
hthninthtentheleventhtwelfththirteenthfourte
enthfifteenth)"my numberOrdinals"(?\\d?(?1s
t2nd3rd1th2th3th4th5th6th7th8th9th0th)
)"my ordinals"(?wordOrdinalsnumberOrdinals
)"my confTypes"(?ConferenceWorkshopSymposiu
m)"my words"(?A-Z\\w\\s)" A word
starting with a capital letter and ending with 0
or more spacesmy confDescriptors"(?internation
al\\sA-Z\\s)" .e.g "International
Conference ...' or the conference name for
workshops (e.g. "VLDB Workshop ...")my
connectors"(?onof)"my abbreviations"(?\\(
A-Z\\w\\w\\W\\s?(?\\d\\d)?\\))"
Conference abbreviations like "(SIGMOD'06)" The
actual pattern we search for.  A typical
conference name this pattern will find is "3rd
International Conference on Blah Blah Blah
(ICBBB-05)"my fullNamePattern"((?ordinals\\s
wordsconfDescriptors)?confTypes(?\\sconnec
tors\\s.?\\s)?abbreviations?)(?\\n\\r\\.lt
)"
Given a
ltdbworldMessagegt, look for the conference
pattern
lookForPattern(dbworldMessag
e, fullNamePattern)
In a given
ltfilegt, look for occurrences of ltpatterngt
ltpatterngt is a regular expression
sub
lookForPattern     my (file,pattern) _at__
14
Gene Protein Tagger AliBaba
  • Extract gene names from PubMed abstracts
  • Use Classifier (Support Vector Machine - SVM)
  • Corpus of 7500 sentences
  • 140.000 non-gene words
  • 60.000 gene names
  • SVMlight on different feature sets
  • Dictionary compiled from Genbank, HUGO, MGD, YDB
  • Post-processing for compound gene names

15
Some Hand Coded Entity Taggers
  • FRUMP DeJong 82
  • CIRCUS / AutoSlog Riloff 93
  • SRI FASTUS Appelt, 1996
  • MITRE Alembic (available for use)
  • Alias-I LingPipe (available for use)
  • OSMX Embley, 2005
  • DBLife Doan et al, 2006
  • Avatar Jayram et al, 2006

16
Machine Learning Methods
  • Can work well when training data is easy to
    construct and is plentiful
  • Can capture complex patterns that are hard to
    encode with hand-crafted rules
  • e.g., determine whether a review is positive or
    negative
  • extract long complex gene names

From AliBaba
  • The human T cell leukemia lymphotropic virus
    type 1 Tax protein represses MyoD-dependent
    transcription by inhibiting MyoD-binding to the
    KIX domain of p300.
  • Can be labor intensive to construct training data
  • Question how much training data is sufficient?

17
Popular Machine Learning Methods for IE
  • Naive Bayes
  • SRV Freitag-98, Inductive Logic Programming
  • Rapier Califf Mooney-97
  • Hidden Markov Models Leek, 1997
  • Maximum Entropy Markov Models McCallum et al,
    2000
  • Conditional Random Fields Lafferty et al, 2000
  • Implementations available
  • Mallet (Andrew McCallum)
  • crf.sourceforge.net (Sunita Sarawagi)
  • MinorThird minorthird.sourceforge.net (William
    Cohen)

For details Feldman, 2006 and Cohen, 2004
18
Example of State-based ML Method
19
Extracted Entities Resolving Duplicates
Document 1 The Justice Department has officially
ended its inquiry into the assassinations of John
F. Kennedy and Martin Luther King Jr., finding
no persuasive evidence'' to support conspiracy
theories, according to department documents. The
House Assassinations Committee concluded in 1978
that Kennedy was probably'' assassinated as the
result of a conspiracy involving a second gunman,
a finding that broke from the Warren Commission
's belief that Lee Harvey Oswald acted alone in
Dallas on Nov. 22, 1963. Document 2 In 1953,
Massachusetts Sen. John F. Kennedy married
Jacqueline Lee Bouvier in Newport, R.I. In 1960,
Democratic presidential candidate John F. Kennedy
confronted the issue of his Roman Catholic faith
by telling a Protestant group in Houston, I do
not speak for my church on public matters, and
the church does not speak for me.' Document 3
David Kennedy was born in Leicester, England in
1959.  Kennedy co-edited The New Poetry
(Bloodaxe Books 1993), and is the author of New
Relations The Refashioning Of British Poetry
1980-1994 (Seren 1996). 
From Li, Morie, Roth, AI Magazine, 2005
20
Important Problem, Addressed in Part II
  • Appears in numerous real-world contexts
  • Plagues many applications
  • Citeseer, DBLife, AliBaba, Rexa, etc.

21
Outline
  • Overview of Information Extraction
  • Entity tagging
  • Relation extraction
  • Scaling up Information Extraction
  • Focus on scaling up to large collections (where
    data mining and ML techniques shine)
  • Other dimensions of scalability

22
Relation Extraction Disease Outbreaks
  • Extract structured relations from text

May 19 1995, Atlanta -- The Centers for Disease
Control and Prevention, which is in the front
line of the world's response to the deadly Ebola
epidemic in Zaire , is finding itself hard
pressed to cope with the crisis
Disease Outbreaks in The New York Times
Date Disease Name Location
Jan. 1995 Malaria Ethiopia
July 1995 Mad Cow Disease U.K.
Feb. 1995 Pneumonia U.S.
May 1995 Ebola Zaire
Information Extraction System (e.g., NYUs
Proteus)
23
Example Protein Interactions
We show that CBF-A and CBF-C interact with each
other to form a CBF-A-CBF-C complex and that
CBF-B does not interact with CBF-A or CBF-C
individually but that it associates with the
CBF-A-CBF-C complex.
24
Relation Extraction
  • Typically require Entity Tagging as preprocessing
  • Knowledge Engineering
  • Rules defined over lexical items
  • ltcompanygt located in ltlocationgt
  • Rules defined over parsed text
  • ((Obj ltcompanygt) (Verb located) () (Subj
    ltlocationgt))
  • Proteus, GATE,
  • Machine Learning-based
  • Learn rules/patterns from examples
  • Dan Roth 2005, Cardie 2006, Mooney 2005,
  • Partially-supervised bootstrap from seed
    examples
  • Agichtein Gravano 2000, Etzioni et al., 2004,
  • Recently, hybrid models Feldman2004, 2006

25
Example Extraction Rule NYU Proteus
26
Example Extraction PatternsSnowball AG2000
lts 0.7gt ltin 0.7gt ltheadquarters 0.7gt
LOCATION
ORGANIZATION
lt- 0.75gt ltbased 0.75gt

LOCATION
ORGANIZATION
27
Accuracy of Information Extraction
Feldman, ICML 2006 tutorial
  • Errors cascade (error in entity tag ? error in
    relation extraction)
  • This estimate is optimistic
  • Holds for well-established tasks
  • Many specific/novel IE tasks exhibit lower
    accuracy

28
Outline
  • Overview of Information Extraction
  • Entity tagging
  • Relation extraction
  • Scaling up Information Extraction
  • Focus on scaling up to large collections (where
    data mining and ML techniques shine)
  • Other dimensions of scalability

29
Dimensions of Scalability
  • Efficiency/corpus size
  • Years to process a large collections (centuries
    for Web)
  • Heterogeneity/diversity of information sources
  • Requires many rules (expensive to apply)
  • Many sources/conventions (expensive to maintain
    rules)
  • Accessing required documents
  • Hidden Web databases are not crawlable
  • Number of Extraction Tasks (not covered)
  • Many patterns/rules to develop and maintain
  • Open research area

30
Scaling Up Information Extraction
  • Scan-based extraction
  • Classification/filtering to avoid processing
    documents
  • Sharing common tags/annotations
  • General (keyword) index-based techniques
  • QXtract, KnowItAll
  • Specialized indexes
  • BE/KnowItNow, Linguists Search Engine
  • Parallelization/Adaptive Processing
  • IBM WebFountain, Googles Map/Reduce
  • Application Question Answering
  • AskMSR, Arranea, Mulder

31
Scan
Output Tokens

Extraction System
Text Database
  1. Extract output tokens
  1. Process documents
  1. Retrieve docs from database
  • Scan retrieves and processes documents
    sequentially (until reaching target recall)
  • Execution time Retrieved Docs (R P)

Time for processing a document
Time for retrieving a document
32
Efficient Scanning for Information Extraction
  • 80/20 rule use few simple rules to capture
    majority of the cases PRH2004
  • Train a classifier to discard irrelevant
    documents without processing GHY2002
  • Share base annotations (entity tags) across
    multiple tasks

33
Filtered Scan
Output Tokens

Extraction System
Text Database
filtered
  1. Extract output tokens
  1. Process documents
  1. Retrieve docs from database
  • Scan retrieves and processes all documents (until
    reaching target recall)
  • Filtered Scan uses a classifier to identify and
    process only promising documents(e.g., the
    Sports section of NYT is unlikely to describe
    disease outbreaks)
  • Execution time Retrieved Docs ( R F
    P)

Time for processing a document
Time for retrieving a document
Time for filteringa document
34
Exploiting Keyword and Phrase Indexes
  • Generate queries to retrieve only relevant
    documents
  • Data mining problem!
  • Some methods in literature
  • Traversing Query Graphs AIG2003
  • Iteratively refine queries AG2003
  • Iteratively partition document space Etzioni et
    al., WWW 2004
  • Case studies QXtract, KnowItAll

35
Simple Strategy Iterative Set Expansion
Output Tokens

Text Database
Extraction System
Query Generation
  1. Extract tokensfrom docs
  1. Process retrieved documents
  1. Augment seed tokens with new tokens
  1. Query database with seed tokens

(e.g., ltMalaria, Ethiopiagt)
(e.g., Ebola AND Zaire)
  • Execution time Retrieved Docs (R P)
    Queries Q

Time for answering a query
Time for retrieving a document
Time for processing a document
36
Querying Graph
AIG2003
Tokens
Documents
t1
d1
  • The querying graph is a bipartite graph,
    containing tokens and documents
  • Each token (transformed to a keyword query)
    retrieves documents
  • Documents contain tokens

ltSARS, Chinagt
d2
t2
ltEbola, Zairegt
t3
d3
ltMalaria, Ethiopiagt
t4
d4
ltCholera, Sudangt
t5
d5
ltH5N1, Vietnamgt
37
Recall Limit Reachability Graph
Reachability Graph
Tokens
Documents
t1
t1
d1
t2
t3
d2
t2
t3
d3
t4
t5
t4
d4
t1 retrieves document d1 that contains t2
t5
d5
Upper recall limit determined by the size of
the biggest connected component
38
Reachability Graph for DiseaseOutbreaks
39
Connected Components Visualization
DiseaseOutbreaks, New York Times 1995
40
Getting Around Reachability Limit
  • KnowItAll
  • Add keywords to partition documents into
    retrievable disjoint sets
  • Submit queries with parts of extracted instances
  • QXtract
  • General queries with many matching documents
  • Assumes many documents retrievable per query

41
QXtract AG2003
User-Provided Seed Tuples
Seed Sampling
  1. Get document sample with likely negative and
    likely positive examples.
  2. Label sample documents usinginformation
    extraction systemas oracle.
  3. Train classifiers to recognizeuseful
    documents.
  4. Generate queries from classifiermodel/rules.

Information Extraction
Classifier Training
Query Generation
Queries
42
KnowItAll Architecture
Slides Zheng Shao, UIUC
Web Pages
Search Engine Interface
  • System Work Flow

Rule
Rule template
Extractor
NP1 such as NPList2 head(NP1)
plural(name(Class1)) properNoun(head(each(NPList
2))) gt instanceOf(Class1,head(each(NPList2)))
NP1 such as NPList2 head(NP1) countries
properNoun(head(each(NPList2))) gt instanceOf(Coun
try,head(each(NPList2))) Keywords countries
such as
Assessor
Database
43
KnowItAll Architecture (Cont.)
Frequency
Search Engine Interface
  • System Work Flow

Web Pages
Rule
Extractor
Extracted Information
the United Kingdom and Canada India North Korea,
Iran, India and Pakistan Japan Iraq, Italy and
Spain
Country AND the United Kingdom Countries such as
the United Kingdom
Assessor
Knowledge
the United Kingdom Canada India North Korea Iran
Discriminator Phrase
Country AND X Countries such as X
Database
44
Using Generic Indexes Summary
  • Order of magnitude scale-up in corpus size
  • Indexes are approximate (queries not precise)
  • Require many documents to retrieve
  • Can we do better?

45
Index Structures for Information Extraction
  • Bindings Engine CE2005
  • Indexes of entities CGHX2006, IBM Avatar
  • Other systems (not covered)
  • Linguists search engine (P. Resnik et al.)
    indexes syntactic structures
  • FREE Indexing regular expressions J. Cho et al.

46
Bindings Engine (BE) Slides Cafarella 2005
  • Bindings Engine (BE) is search engine where
  • No downloads during query processing
  • Disk seeks constant in corpus size
  • queries phrases
  • BEs approach
  • Variabilized search query language
  • Pre-processes all documents before query-time
  • Integrates variable/type data with inverted
    index, minimizing query seeks

47
BE Query Support
  • cities such as ltNounPhrasegt
  • President Bush ltVerbgt
  • ltNounPhrasegt is the capital of ltNounPhrasegt
  • reach me at ltphone-numbergt
  • Any sequence of concrete terms and typed
    variables
  • NEAR is insufficient
  • Functions (e.g., head(ltNounPhrasegt))

48
BE Operation
  • Like a generic search engine, BE
  • Downloads a corpus of pages
  • Creates an index
  • Uses index to process queries efficiently
  • BE further requires
  • Set of indexed types (e.g., NounPhrase), with a
    recognizer for each
  • String processing functions (e.g., head())
  • A BE system can only process types and functions
    that its index supports

49
Index design
  • Search engines handle scale with inverted index
  • Single disk seek per term
  • Mainly sequential reads
  • Disk analysis
  • Seeks require 5 ms, so only 200/sec
  • Sequential reads transfer 10-40 MB/sec
  • Inverted index minimizes expensive seeks BE
    should do the same
  • Parallel downloads are just parallel, distributed
    seeks still very costly

50
as
billy
cities
friendly
give
mayors
nickels
seattle
such
words
51
Query such as
docs
docid0
docid1
docid2
dociddocs-1

as
billy
cities
friendly
give
mayors
nickels
seattle
such
words
  1. Test for equality
  2. Advance smaller pointer
  3. Abort when a list is exhausted

docs
docid0
docid1
docid2
dociddocs-1

322
Returned docs
52
such as
as
billy
cities
friendly
give
mayors
nickels
seattle
such
words
In phrase queries, match positions as well
53
Neighbor Index
  • At each position in the index, store neighbor
    text that might be useful
  • Lets index ltNounPhrasegt and ltAdj-Termgt

I love cities such as Philadelphia.
AdjT love
54
Neighbor Index
  • At each position in the index, store neighbor
    text that might be useful
  • Lets index ltNounPhrasegt and ltAdj-Termgt

I love cities such as Philadelphia.
AdjT cities NP cities
AdjT I NP I
55
Neighbor Index
Query cities such as ltNounPhrasegt
I love cities such as Philadelphia.
AdjT Philadelphia NP Philadelphia
AdjT such
56
cities such as ltNounPhrasegt
docs
pos0
pos1
dociddocs-1
posdocs-1
docid0
docid1

as
billy
cities
friendly
give
mayors
nickels
philadelphia
such
words
19



posns
pos0
pos1
pospos-1
posns
pos0
neighbor0
pos1
neighbor1
pospos-1
12
neighbor1
str1
neighbors
blk_offset
neighbor0
str0
NPright
Philadelphia
3
ltoffsetgt
AdjTleft
such
In doc 19, starting at posn 8
I love cities such as Philadelphia.
  1. Find phrase query positions, as with phrase
    queries
  2. If term is adjacent to variable, extract typed
    value

57
Asymptotic Efficiency Analysis
  • k concrete terms in query
  • B bindings found for query
  • N documents in corpus
  • T indexed types in corpus

Query Time (in seeks) Index Space
BE O(k) O(N T)
Std Model O(k B) O(N)
  • B and N scale together k often small T often
    exclusive

58
Experiment 2 KnowItAll on BE
Num Extractions Std Imp/ Google BE Speedup
10k 5,976s
50k 29,880s
150k 89,641s
59
Experiment 2 KnowItAll on BE
Num Extractions Std Imp/ Google BE Speedup
10k 5,976s 95s 63x
50k 29,880s 95s 314x
150k 89,641s N/A N/A
60
BE Summary
  • Significant improvement over generic indexes
  • Index size grows linearly with number of types
  • Some ML-based patterns (e.g., HMMs, CRFs,
    character models) not supported
  • Can we use it for general QA, RE tasks?

61
Similar Approach CGHX2006
  • Support relationship keyword queries over
    indexed entities
  • Top-K support for early termination

62
Indexing Thousands of Entity Types
  • Slides from Chakrabarti et al., WWW 2006

63
Workload-Driven Indexing
64
Selecting Types to Index
65
Parallelization/Adaptive Processing
  • Parallelize processing
  • IBM WebFountain GCG2004
  • Googles Map/Reduce
  • Select most efficient access strategy
  • Cost Estimation and Optimization IAJG2006

66
Map/Reduce Framework
67
Map/Reduce Framework
  • General framework
  • Scales to 1000s of machines
  • Implemented in Nutch
  • Maps easily to information extraction
  • Map phase
  • Parse individual documents
  • Tag entities
  • Propose candidate relation tuples
  • Reduce phase
  • Merge multuple mentiones of same relation tuple
  • Resolve co-references, duplicates

68
Cost Optimizer for Text-Centric Tasks
69
Other Dimensions of ScalabilityManaging Complex
Features CNS2006
R. Fagin and J. Helpern, Belief,
awareness, reasoning, In AI 1998
Many large tables
Authors
Ronald Fagin
Steve Cook
S. Sudarshan
S. Chakrabarti
Nick Koudas
R. K. Narayan
E. F. Codd
J. Widom
  1. Batch up to do better than individual top-k?
  2. Find top segmentation without top-k matches for
    all segments?

70
Other Dimensions of ScalabilityExtraction
Pattern Discovery Konig and Brill, KDD 2006
  • Use suffix array to efficiently explore candidate
    patterns

71
Application Web Question Answering
  • AskMSR does not use patterns
  • Simplicity ? scalability (cheap to compute
    n-grams)
  • Challenge do better than n-grams on web QA

72
Summary
  • Brief overview of information extraction from
    text
  • Techniques to scale up information extraction
  • Scan-based techniques (limited impact)
  • Exploiting general indexes (limited accuracy)
  • Building specialized index structures (most
    promising)
  • Scalability is a data mining problem
  • Querying graphs ? link discovery
  • Workload mining for index optimization
  • Must be optimized for specific text mining
    application?

73
Related Challenges
  • Duplicate entities, relation tuples extracted
  • Missing values
  • Extraction errors
  • Information spans multiple documents
  • Combining relation tuples into complex events

74
Break
  • Eugene Agichtein, Microsoft Emory University
  • http//www.mathcs.emory.edu/eugene/
  • eugene_at_mathcs.emory.edu
  • Next Scalable Information Integration
  • Core set of techniques to enable large-scale IE,
    text mining
  • Sunita Sarawagi

75
References
  • AGI2005 E. Agichtein, Scaling Information
    Extraction to Large Document Collections, IEEE
    Data Engineering Bulletin, 2005
  • AG2003 E. Agichtein and L. Gravano.
    Querying text databases for efficient information
    extraction. ICDE 2003
  • AIG 2003 E. Agichtein, P. Ipeirotis, and L.
    Gravano, Modeling Query-Based Access to Text
    Databases, WebDB 2003
  • CDS2005 l J. Cafarella, D. Downey, S.
    Soderland, and Oren Etzioni. KnowItNow Fast,
    scalable information extraction from the web.
    (HLT/EMNLP), 2005.
  • CE2005 M. J. Cafarella and O. Etzioni. A
    search engine for natural language applications.
    (WWW), 2005
  • CNS2006 A. Chandel, P.C. Nagesh, and S.
    Sarawagi. Efficient batch top-k search for
    dictionary-based entity recognition. ICDE 2006
  • CRW2005 S. Chaudhuri, R. Ramakrishnan, and G.
    Weikum. Integrating db and ir technologies What
    is the sound of one hand clapping?, CIDR 2005.
  • CGHX2006 K. Chakrabarti, V. Ganti, Jiawei Han,
    D. Xin, Ranking Objects Based on Relationships,
    SIGMOD 2006
  • CPD 2006 S. Chakrabarti, Kriti Puniyani and
    Sujatha Das, Optimizing Scoring Functions and
    Indexes for Proximity Search in Type-annotated
    Corpora. WWW 2006

76
References II
  • DBB2002 S. Dumais, M. Banko, E. Brill, J. Lin
    and A. Ng (2002). P. Bennett, S. Dumais and E.
    Horvitz (2002). Web question answering Is more
    always better? SIGIR 2002
  • GHY2002 R. Grishman, S. Huttunen, and R.
    Yangarber. Information extraction for enhanced
    access to disease outbreak reports. Journal of
    Biomedical Informatics, 2002.
  • GCG2004 D. Gruhl, L. Chavet, D. Gibson, J.
    Meyer, P. Pattanayak, A. Tomkins, and J. Zien.
    How to build a WebFountain An architecture for
    very large-scale text analytics. IBM Systems
    Journal, 2004.
  • IAJG2006 Ipeirotis, E. Agichtein, P. Jain,
    and L. Gravano, To Search or to Crawl Towards a
    Query Optimizer for Text-Centric Tasks, SIGMOD
    2006
  • KRV2004 R. Krishnamurthy, S. Raghavan, S.
    Vaithyanathan, H. Zhu, Avatar A Database
    Approach to Semantic Search, SIGMOD 2006
  • PRH2004 P. Pantel, D. Ravichandran, and E.
    Hovy. Towards terascale knowledge acquisition. In
    Conference on Computational Linguistics (COLING),
    2004.
  • PE2005 P. Resnik and A. Elkiss. The
    linguists search engine An overview
    (demonstration). In ACL, 2005.
  • PDT2001 P.D. Turney. Mining the web for
    synonyms PMI-IR versus LSA on TOEFL. In European
    Conference on Machine Learning (ECML), 2001.
  • C. König and E. Brill, Reducing the Human
    Overhead in Text Categorization, KDD 2006
Write a Comment
User Comments (0)
About PowerShow.com