Skip to main content
Browse by:
GROUP

CS-ECE Colloquium: Bringing Structural Information into Neural Network Design

Event Image
Icon calendar
Monday, February 12, 2018
Icon time
12:00 pm - 1:00 pm
Icon speaker
Yangfeng Ji, postdoc researcher at University of Washington
Icon series
CS-ECE Colloquium

Deep learning is one of most popular learning techniques used in natural language processing (NLP). A central question in deep learning for NLP is how to design a neural network that can fully utilize the information from training data and make accurate predictions. A key to solve this problem is to design a better network architecture. In this talk, I will present two examples from my work on how structural information from natural language helps design better neural network models. The first example shows adding coreference structures of entities not only helps different aspects of text modeling, but also improves the performance of language generation; the second example demonstrates structures of organizing sentences into coherent texts can help neural networks build better representations for various text classification tasks. Along the lines of this topic, I will also propose a few ideas for future work and discuss some potential challenges.