Transcript ppt
Concept Hierarchy
Induction
by Philipp Cimiano
Objective
Structure information into categories
Provide a level of generalization to define
relationships between data
Application: Backbone of any ontology
Overview
Different approaches of acquiring
conceptual hierarchies from text corpus.
Various clustering techniques.
Evaluation
Related Work
Conclusion
Machine Readable Dictionaries
Entries: ‘a tiger is a mammal’, or ‘mammals
such as tigers, lions or elephants’.
exploit the regularity of dictionary entries.
the head of the first NP - hypernym.
Example
Exception
Exception
is-a (corolla, part)………..is a NOT VALID
is-a (republican, member) ……….. is a NOT VALID
is-a (corolla, flower)………..is a NOT VALID
is-a (republican, political party)………..is a NOT VALID
Exception
Alshawis solution
Results using MRDs
Dolan et al. - 87% of the hypernym
relations extracted are correct
Calzolari cites a precision of > 90%
Alshawi - precision of 77%
Strengths And Weaknesses
Correct, explicit knowledge
Robust basis for ontology learning
Weakness- domain independent
Lexico-Syntactic patterns
Task: automatically learning hyponym
relations from the corpora.
'Such injuries as bruises, wounds and broken bones'
hyponym (bruise, injury)
hyponym (wound, injury)
hyponym (broken bone, injury)
Hearst patterns
'Such injuries as bruises, wounds and broken bones'
Requirements
Occur frequently in many text genres.
Accurately indicate the relation of interest.
Be recognizable with little or no preencoded knowledge
Strengths And Weaknesses
Identified easily and are accurate
Weakness:
patterns appear rarely
is-a relation do not appear in Hearst style
pattern
Distribution Similarity
'you shall know a word by the company it
keeps’ [Firth, 1957].
semantic similarity of words – similarity of
the contexts.
Using distribution similarity
Strengths And Weaknesses
reasonable concept hierarchy.
Weakness:
Cluster tree lacks clear and formal interpretation
Does not provide any intentional description of
concepts
Similarities may be accidental (sparse data)
Formal Concept Analysis (FCA)
FCA output
Similarity measures
Smoothing
Evaluation
Semantic cotopy (SC).
Taxonomy overlap (TO)
Evaluation Measure
100% Precision Recall
Low Recall
Low Precision
Results
Results
Results
Results
Strengths And Weaknesses
FCA generates formal concepts
Provides intentional description
Weakness:
Size of the lattice can get exponential in the size
spurious clusters
Finding appropriate labels for the cluster
Problems with Unsupervised
Approaches to Clustering
Data sparseness leads to spurious
syntactic similarities
Produced clusters can’t be appropriately
labeled
Guided Clustering
Hypernyms directly used to guide
clustering
WordNet
Hearst
Agglomerative clustering
Similarity Computation
Ten most similar terms of the tourism reference taxonomy
The Hypernym Oracle
Three sources
WordNet
Hearst
patterns matched in a corpus
Hearst patterns matched in the World Wide
Web
Record hypernyms and amount of
evidence found in support of hypernyms.
WordNet
Collect hypernyms found in any
dominating synset containing term, t
Include number of times the hypernym
appears in a dominating synset
Hearst Patterns (Corpus)
Record number of isa-relations found
between two terms
Hearst Patterns (WWW)
Download 100 Google abstracts for each
concept and clue:
Evidence
Total Evidence for Hypernyms:
•time: 4
•vacation: 2
•period: 2
Clustering Algorithm
1.
2.
3.
Input a list of terms
Calculate the similarity between each
pair of terms and sort from highest to
lowest
For each potential pair to be clustered
consult the oracle.
Consulting the Oracle case 1
If term 1 is a hypernym of term 2 or viceversa:
Create
appropriate subconcept relationship.
Consulting the Oracle case 2
Find the common hypernym for both terms
with greatest evidence.
If one term has already been classified:
t’ = h
h is a hypernym of t’
t’ is a hypernym of h
Consulting the Oracle case 3
Neither term has been classified:
Each
term becomes a subconcept of the
common hypernym.
Consulting the Oracle case 4
The terms do not share a common
hypernym:
Set
aside the terms for further processing.
r-matches
For all unprocessed terms, check for rmatches (i.e. ‘credit card’ matches
‘international credit card’)
Further Processing
If either term in a pair is already classified
as t’, the other term is classified under t’ as
well.
Otherwise place both terms under the
hypernym of either term with the most
evidence.
Any unclassified terms are added under
the root concept.
Evaluation
Taxonomic overlap (TO)
ignore
leaf nodes
Sibling overlap (SO)
measures
quality of clusters
Evaluation
Tourism domain:
Lonely
Planet
Mecklenburg
Finance domain:
Reuters-21578
Tourism Results—TO
Finance Results—TO
Tourism Results—SO
Finance Results—SO
Human Evaluation
Future Work
Take word sense into consideration for the
WordNet source.
Summary
Hypernym guided agglomerative clustering
works pretty good.
Better
than the “Golden Standard”
Good human evaluation
Provides labels for clusters
No spurious similarities
Faster than agglomerative clustering
Learning from Heterogeneous
Sources of Evidence
Many ways to learn concept hierarchies
Can we combine different paradigms?
Any
manual attempt to combine strategies
would be ad hoc
Use supervised learning to combine
techniques
Determining relationships with
machine learning
Example: Determine if a pair of words has
an “isa” relationship
Feature 1:
Matching patterns in a corpus
Given two terms t1 and t2 we record how
many times a Hearst-pattern indicating an
isa-relation between t1 and t2 is matched
in the corpus
Normalize by maximum number of Hearst
patterns found for t1
Example
This provided the best F-measure with a
single-feature classifier
Feature 2:
Matching patterns on the web
Use the Google API to count the matches
of a certain expression on the Web
Feature 3:
Downloading webpages
Allows for matching expressions with a more
complex linguistic structure
Assign functions to each of the Hearst patterns
to be matched
Use these “clues” to decide what pages to
download
Download 100 abstracts matching the query
“such as conferences”
Example
Feature 4:
WordNet – All senses
Is there a hypernym relationship between
t1 and t2?
Can be more than one path from the
synsets of t1 to the synsets of t2
Feature 5:
WordNet – First sense
Only consider the first sense of t1
Feature 6:
“Head”- heuristic
If t1 r-matches t2 we derive the relation
isa(t2,t1)
e.g.
t1 = “conference”
t2 = “international conference”
isahead(“international conference”,”conference”)
Feature 7:
Corpus-based subsumption
t1 is a subclass of t2 if all the syntactic
contexts in which t1 appears are also
shared by t2
Feature 8:
Document-based subsumption
t1 is a subclass of term t2 if t2 appears in
all documents in which t1 appears
# of pages where t1 and t2 occur
# of pages where t1 occurs
Example
Naïve Threshold Classifier
Used as a baseline
Classify an example as positive if the
value of a given feature is above some
threshold t
For each feature, the threshold has been
varied from 0 to 1 in steps of 0.01
Baseline Measures
Evaluation
Classifiers
Naïve
Bayes
Decision Tree
Perceptron
Multi-layer perceptron
Evaluation Strategies
Undersampling
Oversampling
Try different threshold values other than 0.5
Introducing a cost matrix
Add additional examples to the minority class
Varying the classification threshold
Remove a number of majority class examples (non-isa
examples)
Different penalties for different types of misclassification
One Class SVMs
Only considers positive examples
Results
Results (cont.)
Discussion
The best results achieved with the one-class SVM
(F = 32.96%)
More
than 10 points above the baseline classifier
average (F = 21.28%) and maximum (F = 21%)
strategies
More than 14 points better than the best singlefeature classifier (F = 18.84%) using the isawww
feature
Second best results obtained with a Multilayer
Perceptron using oversampling or undersampling
Discussion
Gain insight from finding which features were
most used by classifiers
Used this information to modify features and
rerun experiments
Summary
Using different approaches is useful
Machine learning approaches outperform naïve
averaging
Unbalanced character of the dataset poses a
problem
SVMs (which are not affected by the imbalance)
produce the best results
This approach can show which features are the
most reliable as predictors
Related Work
Taxonomy Construction
Lexico-syntactic
patterns
Clustering
Linguistic
approaches
Taxonomy Refinement
Taxonomy Extension
Lexico-syntactic patterns
Hearst
Iswanska et al. – added extra patterns
Poesia et al. – anaphoric resolution
Ammad et al. – applying to specific domains
Etzioni et al. – patterns matched on the www
Cederburg and Widdows – precision improved with
Latent Semantic Analysis
Others working on learning patterns automatically
Clustering
Hindle
Pereira et al.
group nouns semantically
derive verb-subject and verb-object dependencies from a 6
million word sample of Associated Press news stories
top-down soft clustering algorithm with deterministic annealing
words can appear in different clusters (multiple meanings of
words)
Caraballo
bottom-up clustering approach to build a hierarchy of nouns
uses conjunctive and appositive constructions for nouns derived
from the Wall Street Journal Corpus
Clustering (cont.)
The ASIUM System
The Mo'K Workbench
Grefenstette
Gasperin et al.
Reinberger et al.
Lin et al.
CobWeb
Crouch et al.
Haav
Curran et al.
Terascale Knowledge Acquisition
Linguistic Approaches
Linguistic analysis exploited more directly rather than
just for feature extraction
OntoLT - use shallow parser to label parts of speech and
grammatical relations (e.g. HeadNounToClass-ModToSubClass,
which maps a common noun to a concept or class)
OntoLearn - analyze multi-word terms compositionally with
respect to an existing semantic resource (Word-Net)
Morin et al. - tackle the problem of projecting semantic relations
between single terms to multiple terms (e.g. project the isarelation between apple and fruit to an isa-relation between apple
juice and fruit juice)
Linguistic Approaches
Sanchez and Moreno – download first n hits for a
search word and process the neighborhood
linguistically to determine candidate modifiers for the
search term
Sabou - inducing concept hierarchies for the purpose
of modeling web services (applies methods not to full
text, but to Java-documentation of web services)
Taxonomy Refinement
Hearst and Schutze
Widdows
Madche, Pekar and Staab
Alfonseca et aL
Taxonomy Extension
Agirre et al.
Faatz and Steinmetz
Turney
Conclusions
Compared different hierarchical clustering
approaches with respect to:
effectiveness
speed
traceability
Set-theoretic approaches, as FCA, can
outperform similarity-based approaches.
Conclusions
Presented an algorithm for clustering
guided by a hypernym oracle.
More efficient than agglomerative
clustering.
Conclusions
Used machine learning techniques to
effectively combine different approaches
for learning taxonomic relations from text.
A learned model indeed outperforms all
single approaches.
Open Issues
Which similarity or weighting measure should be chosen
Which features should be considered to represent a
certain term
Can features be aggregated to represent a term at a
more abstract level
How should we model polysemy of terms
Can we automatically induce lexico-syntactic patterns
(unsupervised!)
What other approaches are there for combining different
paradigms; and how can we compare these
Questions