A Probabilistic Representation of Systemic Functional Grammar
Download
Report
Transcript A Probabilistic Representation of Systemic Functional Grammar
A Probabilistic Representation of
Systemic Functional Grammar
Robert Munro
Department of Linguistics,
SOAS,
University of London
Outline
Introduction
Functions in the nominal group
Machine learning
Testing framework
Classification vs unmarked function
Gradational realization
Delicacy
Conclusions
2
Introduction
An exploration of the ability of machine
learning to learn and represent functional
categories as fundamentally probabilistic
Gauged in terms of the ability to:
computationally learn functions from labeled
examples and apply to new texts.
represent functions probabilistically: a gradation
of potential realization between categories.
explore finer layers of delicacy.
3
Functions in the nominal group
Functions:
Deictic, Ordinative, Quantitative, Epithet,
Classifier, Thing (Halliday 1994)
Deictic Ordin.
The
first
Quant.
Epith.
Class.
Thing
three
tasty
red
wines
Gradations:
Here, ‘red’ functions also functions as an Epithet.
The uptake of such marked classifiers will not be
uniform.
Overlap does not necessarily limit significance.
4
Machine Learning
Machine learning:
computational inference from specific examples.
A learner named Seneschal was developed
for the task here:
probabilistic
seeks sub-categories (improves both
classification and analysis)
allows categories to overlap
not too dependent on the size of the data set
5
Machine Learning
The task here:
Given categories &
with known values for x
and y, infer a probabilistic model (potentially with
sub-categories) that can classify new examples:
x
?
?
?
?
6
?
?
?
y
Machine Learning
It is important that attributes (x,y,z...) :
represent features that distinguish functions
can be discovered automatically (for large scales)
are meaningful for analysis…?
Compared to manually constructed parsers:
greater scales than are practical
more features/dimensions than are possible
(100’s are common)
7
Testing Framework
The model was learned from 10,000 labeled
words from Reuters sports newswires from
1996
23 features:
part-of-speech and its context
punctuation
group / phrase contexts
collocational tendencies
probability of repetition
8
Testing Framework
Accuracy:
The ability to correctly identify the dominant
function in 4 test corpora (1,000 words each):
1.
2.
3.
4.
9
Reuters Sports Newswires (1996)
Reuters Sports Newswires (2003)
Bio-informatics abstracts
Extract from Virginia Woolf’s ‘The Voyage Out’
Testing Framework
Gradational model of realization:
calculated as the probability of a word realizing
other functions, averaged between all clusters.
Finer layers of delicacy:
Manual analysis of clusters found within a
function.
10
Unmarked function
Unmarked function: function defined by only
part-of-speech (POS) and word order.
eg: adjective = Epithet, non-final noun = Classifier
Previous functional parsers have assumed that
most instances are unmarked:
POS taggers are almost 100% accurate
word order is trivial
…so the problem is solved?
11
Unmarked function
This is a false assumption.
Across the corpora:
< 40% of non-final adj’s realized Epithets
< 50% of Classifiers were nouns
44% of Classifiers were ‘marked’!
12
Unmarked function
This task halved the classification error:
25%
Un m a r k e d B a s e lin e
le a r n in g r a te
e rro r
20%
15%
10%
5%
13
p e rc e n t o f tra in in g s e t
%
90
%
80
%
70
%
60
%
50
%
40
%
30
%
20
%
10
1%
0%
Gradational Realization
Thing
With relationships
Nominal
Although
functions
describedexisting
are
as typically
probabilistic,
between
represented
all
deterministically:
functions
Class.
Deictic Ordin.
The
first
Quant.
Epith.
Class.
Epith.
Thing
three
tasty
red
wines
Quant.
Ordin.
Deictic
14
Delicacy
Demonstrative
Deictic
Possessive
Ordinative
Numerative
Tabular
Quantitative
Discursive
Epithet
Expansive
Classifier
Hyponymic
First Name
Named Entity
Intermediary
Last Name
Thing
Group-Releasing
Nominative
non-Nom.
Nominal
15
Stated
Described
Delicacy
More delicate functions for Classifiers
(Matthiessen 1995) :
Hyponymic: describing a taxonomy or general
‘type-of’ relationship eg: ‘red wine’, ‘gold medal’,
‘neural network architecture'
Expansive: expands the description of the
Head. eg: ‘knee surgery’, ‘optimization problems',
‘sprint champion’,
16
Delicacy
DDe
emico
tin
Po c
ss
Or
d
ntic
meoic
DeD
ss
Po
d
Or
.
dlainr
Obru
Ta
c
Dis
ith
E p ant .
s
Qpuan
Ex
n
po
Hy
e
am t
F.N ithe
Epm e
a
I.N
e
am
L.N
r
m ifie
No lass
C N
nNo
ted
Sta
r
sc g
DeThin
TOar
bduin
la.
Dis r
c
Ep
ith
EQxua
pant
ns.
Hy
po
n
F.N
a
E m
I.Npith e
amet
e
L.N
am
No e
Clm
Noassif
n- N ier
Sta
ted
De
Thscr
ing
17
Delicacy
More delicate descriptions can be found:
more features
more instances / registers
other algorithms / parameters
Methodology can be applied to:
other parts of a grammar
other languages
18
Conclusions
Gradational modeling of functional realization
is desirable
Sophisticated methods are necessary for
computationally modeling functions:
Markedness is common
Machine learning is a useful tool and
participant in linguistic analysis.
19
Thank you
Acknowledgments:
Geoff Williams
Sanjay Chawla
The slides and extended paper will be
published at:
www.robertmunro.com/research/
20