More about HKUST
Natural Language Processing: A Neural Network View
PhD Qualifying Examination
Title: "Natural Language Processing: A Neural Network View"
by
Mr. Yuchen YAN
Abstract:
We survey modern neural network models in Natural Language Processing
(NLP), which has made drastic breakthroughs in the field. Neural network
can model language better than statistical N-gram models, due to the
special structure of recurrent neural network (RNN). The RNN-based
architecture can further be extended to handle machine translation. In
fact, this is the main building block of Google Translation and Microsoft
Translation. From their models, we illustrate that the alignment problem
is the bottleneck. Luckily, such problem may be solved by our next model:
TRAAM.
Neural network can model grammar as good as our hand-crafted grammar
rules. The model can also be extend to model bi-lingual grammar. Since
there are no hand-crafted bi-lingual grammar rules, the representational
power of neural network becomes valuable. In this survey we show that
TRAAM, bi-lingual grammar augmented with neural network, can be considered
as a good candidate for the alignment solution.
The ultimate goal of natural language processing to understand natural
language. Semantic Role Labeling (SRL) is a structural representation of
semantics. In this survey we study the current best neural network parser:
SENNA, and compare it against the commonly used parser, ASSERT (which is a
statistical one). We show the limitation of SENNA, and how it can be
improved by applying RNN. On the other hand, SRL is a hand-crafted
structure, although it is carefully designed, we can still go beyond it. A
recent neural network architecture, generative adversarial network (GAN),
is capable of generating sentences from random noise. In this survey we
analyze its potential in semantics representation.
We show the strengths and reveal the limitations of neural networks by
studying their applications in machine translation, and analyzing their
performances in semantics representation. Interestingly, in both machine
translation and semantics representation, generative networks, TRAAM and
GAN, are showing stronger potential than discriminative networks.
Generative networks learn patterns from unlabeled data; in contrast,
discriminative networks have to be provided correct solutions, like
syntactic tree, or SRL parse. With generative networks, there are still
plenty of open questions waiting to be solved, plenty of opportunities
lying to be seized.
Date: Wednesday, 16 August 2017
Time: 2:00pm - 4:00pm
Venue: Room 2612A
Lifts 31/32
Committee Members: Prof. Dekai Wu (Supervisor)
Prof. Fangzhen Lin (Chairperson)
Prof. Andrew Horner
Prof. Nevin Zhang
**** ALL are Welcome ****