Skip to content

AMLC 2019: Dive into Deep Learning for Natural Language Processing

Notifications You must be signed in to change notification settings

robertopossidente/AMLC19-GluonNLP

 
 

Repository files navigation

AMLC 2019: Dive into Deep Learning for Natural Language Processing

Time: Friday, July 12, 2019
Location: Meeting Center 02.100, [2031 7th Ave, Seattle, WA 98121](https://goo.gl/maps/LHWeYRDMYPvNKw3n6)

Presenter: Haibin Lin, Leonard Lausen, Xingjian Shi, Haichen Shen, He He, Sheng Zha

AWS Icon   AmazonAI Icon   Neo Icon   Apache Incubator Icon   MXNet Icon   Gluon Icon   TVM Icon

Abstract

Deep learning has rapidly emerged as the most prevalent approach for training predictive models for large-scale machine learning problems. Advances in the neural networks also push the limits of available hardware, requiring specialized frameworks optimized for GPUs and distributed cloud-based training. Moreover, especially in natural language processing (NLP), models contain a variety of moving parts: character-based encoders, pre-trained word embeddings, long-short term memory (LSTM) cells, transformer layers, and beam search for decoding sequential outputs, among others.

This introductory and hands-on tutorial walks you through the fundamentals of machine learning and deep learning with a focus on NLP. We start off with a crash course on deep learning for NLP with GluonNLP, covering data, automatic differentiation, and various model architectures such as convolutional, recurrent, and attentional neural networks. Then, we dive into how context-free and contextual representations help various NLP domains. Throughout the tutorial, we start off from the basic classification problem, and progress into how it can be structured to solve various NLP problems such as sentiment analysis, question answering, machine translation, and natural language generation. Finally, we demonstrate how we can deploy a state-of-the-art NLP model such as BERT on custom hardware such as EC2 A1 instances with the help of TVM.

Agenda

Time Title
13:15-14:15 Natural Language Processing and Deep Learning Basics
14:15-14:25 Break
14:25-15:15 Word Embeddings and Applications of Basic Models
15:15-15:55 Machine Translation and Sequence Generation
15:55-16:35 Contextual Representations with BERT
16:35-16:45 Break
16:45-17:15 Model Deployment with TVM

FAQ

  • Q: How do I get access to the notebooks from the tutorial?
    • There are two notebook instances used in this tutorial. All sessions except model deployment session use the following setup:
      • For setting it up on SageMaker notebook instances, you can find the instructions here.
      • For setting up with Conda, you can use the following Conda environment files: CPU and GPU. (See this guide on creating Conda environment from environment file.)
    • The notebook instances for model deployment can be set up from the following setting:
      • For setting it up on SageMaker notebook instances, you can find the instructions here.

About

AMLC 2019: Dive into Deep Learning for Natural Language Processing

Resources

Stars

Watchers

Forks

Releases

No releases published

Packages

No packages published

Languages

  • Python 69.0%
  • Jupyter Notebook 30.6%
  • Other 0.4%