Program ( V 0.8 )
Time | 11th July | 12th July | 13th July | 14th July | 15th July | 16th July |
---|---|---|---|---|---|---|
08.00 AM to 09.00 AM | Breakfast | Breakfast | Breakfast | Breakfast | Breakfast | Breakfast |
09.00 AM to 09.30 AM | Registration | |||||
09.30 AM to 10.00 AM | ||||||
10.00 AM to 10.30 AM | ||||||
10.30 AM to 11.00 AM | Tea | Tea | Tea | Tea | Tea | |
11.00 AM to 11.30 AM | Tea | |||||
11.30 AM to 12.30 PM | ||||||
12.30 PM to 02.00 PM | Lunch | Lunch | Lunch | Lunch | Lunch | Lunch |
02.00 PM to 03.00 PM | ||||||
03.00 PM to 03.30 PM | Switching Venue and Settling down | Switching Venue and Settling down | Switching Venue and Settling down | Switching Venue and Settling down | Switching Venue and Settling down | |
03.30 PM to 05.00 PM | ||||||
05.00 PM to 06.30 PM | ||||||
06.30 PM to 07.00 PM | Break | Break | Break | Break | Break | Break |
07.00 PM to 07.30 PM | Poster Session cum Dinner | |||||
07.30 PM to 08.30 PM | Welcome | - | - | |||
08.30 PM to 09.30 PM | Welcome Reception | Dinner Session | Follow-up reading group (Lead by Vidit Jain) | Dinner Session | Dinner Session | Dinner |
09.30 PM to 10.00 PM | - | - | - | - | - |
Day 1 Session 1: Introduction
- Speaker: C. V. Jawahar
- Venue: Himalaya, Lecture Hall - 105
- Scope:
- Introduction to the summer school
- Introduction DL and CV
- Introduction to the topics and keywords;etc
- Keywords:machine learning, deep learning, big data, complex data, challenges, supervised learning, classifiers, support vector machines, kernel trick, neural networks, nonlinear, convex, non-convex, applications
Day 1 Session 2: Overview
- Speaker: Manish Gupta
- Venue: Himalaya, Lecture Hall - 105
- Objective:
- Overview of the topics
- Introduce many keywords
- Enables labs
- Keywords:biological inspiration, neurons, artificial neurons, nonlinearity, activation function, layers, hidden layers, loss function, back propagation, stochastic gradient descent, initialisation, regularization
Day 1 Session 3: Overview of DL libraries
- Speaker: Anoop M. Namboodiri
- Venue: Himalaya, Lecture Hall - 105
- Scope:
Introduction to DL libraries- Torch
- Theano
- Tensorflow
- Caffe
Lab1: Introduction to DL Libraries
- Getting Started
- Basic Image Processing
- Reading and Writing Data
- Understanding basic data structures: Tensors, Blobs, Shared Variables
- Running code on CPU/GPU
- Implementation of multi layer perceptron
- Walk through of basic CNN architecture.
Day 2 Session 1: CNNs
- Speaker: R. Venkatesh Babu
- Venue: Himalaya, Lecture Hall - 105
- Objective:
- Introduction CNN, terminologies, popular architectures
- Features, Finetuning
- Loss functions
- Connect to lab
- Keywords:images, convolutions, pooling, fully connected layers, convolutional layers, invariances, gradients, hierarchical feature learning, deconvolution, weights, visualization, imagenet, lenet, alexnet, vggnet, googlenet, resnet
Day 2 Session 2:
- Speaker:R. Venkatesh Babu
- Venue: Himalaya, Lecture Hall - 105
- Objective:
- Keywords: CNN Features and Fine tuning, Detection, Saliency, Visualization, Stereo, 3D and Optical Flow
Day 2 Session 3: More on CNNs
- Speaker: C. V. Jawahar
- Venue: Himalaya, Lecture Hall - 105
- Objective: How to adapt the CNNs to solve a class of problems?
- Keywords: Alternate loss functions, Triplet Loss, Siamese Networks, Vanishing gradient problem
Lab2: CNN
- Train a network
- Fine-tune a network
- Loading weights in a network
- Extracting features and classification
- Extracting weights and visualize
Day 3 Session 1: RNNs
- Speaker: Chetan Arora
- Venue: Himalaya, Lecture Hall - 105
- Objective:
- Intrdoduction to Reccurrent networks and learning
- LSTM and other popular architectures
- Some applications
- Keywords:sequences, RNN, back-propagation in time, unrolling, exploding and vanishing gradients, LSTM, GRU, BLSTM, applications
Day 3 Session 2: Vision, language hybrid architectures
- Speaker: Vinay P Namboodiri
- Venue: Himalaya, Lecture Hall - 105
- Objective:
- Problem space
- Hybrid (CNN+RNN) solutions
- Captioning
- VQA
- Keywords:multimodal retrieval, cca, hybrid solutions, caption generation, visual question answering, attribute learning, word2vec, dcgan, applications
Day 3 Session 3: Egocentric Action Recognition
- Speaker: Chetan Arora
- Venue: Himalaya, Lecture Hall - 105
- Objective:
Lab3: RNNs
- RNN
- LSTM
Day 4 Session 1: Autoencoders
- Speaker: Anoop M. Namboodiri
- Venue: Himalaya, Lecture Hall - 105
- Scope: unsupervised learning, autoencoders
- Keywords: autoencoders, unsupervised pre-training, denoising autoencoders, variatonal autoencoders, applications
Day 4 Session 2: Face, Pose, Human Activities
- Speaker: Gaurav Sharma
- Venue: Himalaya, Lecture Hall - 105
- Topics
- Part I: Face
- Siamese networks - Deep face, Deepid, VGG Face (Verification)
- Networks for Age, Expressions etc.
- Networks for fiducial points detection
- Part II: Human Actions and Activities
- 3D Conv Nets for Human Actions
- P-CNN for Actions
- Two stream CNN
- RNN variants for Actions
- Part III: Human Pose
- Deep pose
- Multi source deep learning for pose
- DeepCur and DeeperCut (Multi-person pose)
- Part IV: Human Attributes and gestures
- PANDA - Pose aligned networks for human attributes
- Multimodal adaptive gestures (PAMI)
Day 4 Session 3
- Speaker: Gaurav Sharma
- Venue: Himalaya, Lecture Hall - 105
- Objective: Latent Ordinal Model for Video Face Analysis
Lab4: Autoencoders
- Standard Auto-Encoders
- Sparse Auto-Encoders
- Denoising Auto-Encoders
- Stacking of Auto-Encoders
- Convolutional Auto-Encoders
Day 5 Session 1: Optimization for DL
- Speaker: Vineeth N Balasubramanian
- Venue: Himalaya, Lecture Hall - 105
- Objective:
- Associated mathematical and optimization issues
- Keywords: Gradient descent and variants, Backdrop review, Batch, stochastic, mini-batch gradient descent, Loss functions, Cross-entropy, Negative log-likelihood as the most general loss function, Overview of other loss functions (classification, regression, embedding), Challenges, Ill-conditioning, Multiple local optima, saddle points, plateaus, Vanishing/exploding gradients, Slow convergence, Choosing learning rate/parameters, Algorithms to address challenges, Momentum, Nesterov accelerated momentum, Adagrad, Adadelta, RMSProp, Adam, Which one to choose?, Advanced methods, Introduction to second-order methods: Newton method, Conjugate gradient, Natural gradient, Path-SGD, Performance guarantees,
Day 5 Session 2: Practical Issues
- Speaker: Vineeth N Balasubramanian
- Venue: Himalaya, Lecture Hall - 105
- Scope:
- Normalization
- Non-linearities
- Keywords: Simple tricks, Early stopping, Data augmentation, Shuffling and Curriculum learning, Choosing activation functions and target values, Tanh, Softmax, ReLU, Leaky ReLU, Regularization, DropOut, DropConnect, MaxOut, Noise in Data, Label and Gradient, Weight initialization strategies, unsupervised pre-training, Transforming the inputs, Batch normalization
Day 5 Session 3: Symbolic Deep Learning
- Speaker: Shailesh Kumar
- Venue: Himalaya, Lecture Hall - 105
- Scope:
Lab5: Practical Issues
- Batch Normalization
- Dropout / MaxOut
- Update rules
- SGD, Adagrad, Adadelta, Adam, RMSProp
- Weight Regularization
- Weight Initialization
- Data Augmentation
Day 6 Session 1: GPUs
- Speaker: P. J. Narayanan
- Venue: Himalaya, Lecture Hall - 105
- Scope:
- GPU architectures
- GPU programming (CUDA)
- Practical issues in DL and CV
- Kegwords:GPU, parallel processing, throughput, Amadahl's law, graphics pipeline, graphics cards, CUDA, parallelizing neural networks, data parallelism, model parallelism, CUDNN, deep learning libraries, NVIDIA GPUs
Day 6 Session 2: Recent/Advanced topics
- Speaker: Arjun Jain
- Venue: Himalaya, Lecture Hall - 105
- Objective:
- State-of-the-art neural networks for solving vision tasks networks require millions of parameters and billions of arithmetic operations.
- Look into optimizations which can enable running of such networks on low power, memory and compute capability devices such as mobile phones or cars.
- Keywords: Student-Teacher Networks, Deep Compression, SqueezeNet, HashedNets, BinaryConnect, Hardware Acceleration, Fixed-Point Optimization, Compute v/s Bandwidth Limitations, Power Considerations
Day 6 Session 3
Lab6: GPU, CUDA, Extensions
- Custom NN layers and functions
- Introduction to CUDA
- CUDA kernels
- Interfacing CUDA kernels with DNN libraries
Valedictory Session
Talk: A Peek into Recent Advances in Deep Learning at Microsoft
Abstract: Microsoft has been one of the early industry pioneers in the area of Deep Learning (DL) and has engaged with other pioneers from the academic world to create industry-scale successful products/applications in speech recognition as well as in speech translation, object recognition, automatic image captioning, natural language processing, multimodal processing, semantic modeling, web search, contextual entity search, ad selection, and big data analytics. Much of these successes are attributed to the availability of big datasets for training deep models, the powerful general-purpose GPU computing, and the innovations in deep learning architectures and algorithms. In this talk, I will provide an overview of the some of these exciting product areas at Microsoft which currently employ deep learning based technologies. I will also provide a peek into some start-of-the-art research areas, related to DL, currently being pursued at MS where we’re trying to push the envelope of this science.
Bio:
Dr. Manoj Chinnakotla is a Senior Applied Scientist in the Relevance, Natural Language Understanding and Data Sciences group at Microsoft India, Hyderabad. He is also an adjunct faculty at the Language Technology Research Centre (LTRC), IIIT Hyderabad. He did his Ph.D in Information Retrieval (IR) and Natural Language Processing (NLP) from I.I.T. Bombay where he was an Infosys Research Fellow. He has published in several reputed conferences and journals in IR, NLP and ML such as SIGIR, ACL, IJCAI, WWW and ACM TALIP. He also serves as a reviewer and PC member for conferences - IJCAI, ICON, MIKE. His current research interests include Conversational Question Answering, Entity Mining and Machine Learning.
Talk: Deep Learning on Intel Architecture
Speaker: Bharat Kaul
Bio:
Bharat is an engineering manager and heads Intel Labs/Parallel Computing Lab-India.
His research focus is on driving architectural performance leadership for Intel for multi-core/many
architectures along with ensuring leadership in programmer efficiency. Bharat has ~20 years of
industry experience and has been with Intel for 15+ years with experience in both technical and
management leadership positions for product teams in several domains from data networking,
communications, embedded systems and security architectures. Bharat received his B.E. in
Electrical & Electronics from BITS Pilani.