Welcome to the Deep Learning Nanodegree Foundations Program! In this lesson, you‘ll meet your instructors, find out about the field of Deep Learning, and learn how to make the most of the resources Udacity provides.
Every week, you can expect to see this content coming up:
Then, approximately every four weeks you‘ll get a project.
The first week‘s content contains a bit more than an average week, as we‘re covering some introductory material and two topics from Siraj. So you can expect a little less than that going forward. Keep in mind the program is for students of all backgrounds. Some of the material might feel easy for more experienced students, but we‘re covering a lot of different topics and there will be a lot of advanced material.
Here is the list of topics that will be taught throughout the program:
Week 1: Types of machine learning, when to use machine learning, neural network architecture
We’ll start off with a simple introduction to linear regression and machine learning. This will give you the vocabulary you need to understand recent advancements, and make clear where deep learning fits into the broader picture of ML techniques.
We’ll then start exploring neural networks in depth and understand various canonical architectures such as AlexNet, LeNet, and others. We’ll use these neural networks to automatically convert images of numbers into their corresponding digits.
Example network architecture. Source: Andrej Karpathy
Week 2: Cloud computing + sentiment analysis (text classification)
We’ll train deep neural networks in the cloud using GPUs and see how we can use these models on text to do simple sentiment analysis.
Week 3: Math notation + recommendation systems (algebra, calculus, matrix math)
Netflix is just one of many companies that use recommendation systems in their product. We’ll then explore the world of recommendation systems such as those used in Netflix, Amazon, and others. You’ll also be provided a general introduction to the linear algebra that will help you throughout your deep learning coursework.
Netflix is just one of many companies that use recommendation systems in their product.
Week 4: Data preparation (cleaning, regularization, dimensionality reduction)
One of the key parts of applying deep learning in practice is collecting the right type of training data. In this lesson, we’ll explore a variety of techniques to clean and regularize your data so that you can train effective models.
Week 5: Drone image tracking (image classification with CNNs)
Convolutional Neural Networks (CNNs) are currently one of the most exciting advancements in neural networks, given that CNNs can now classify objects in images better than humans can. In this lesson, we’ll learn the intuition behind these networks and use them to track images of drones.
In this lesson, we’ll use Convolutional Neural Networks to perform image classification.
Week 6: Stock prediction (regression with RNNs)
In this lesson, we’ll learn about Recurrent Neural Networks?—?a type of network architecture particularly well suited to time series data. We’ll apply our understanding of these networks on some of the most important time series data we have?—?stock prices!
Recurrent Neural Network architecture. Source: Andrej Karpathy
Week 7: Art generation (transfer learning)
Beyond simply predicting, deep neural networks are now also capable of generating art, music, and images based on samples. In this lesson, we’ll use neural networks to create new art based on artwork we feed in, using a technique known as Style Transfer.
Source: L. Gatys e. al “A Neural Algorithm of Artistic Style” (2015).
Week 8: Music generation (LSTMs applied to Audio)
Neural networks can also be applied to problems in audio, as the famous Wavenet paper by DeepMind has shown. In this lesson, we’ll use a type of Recurrent Neural Network called LSTMs (Long Term Short Term Memory) to generate new pieces of music based on existing samples.
Wavenet. Source: DeepMind
Week 9: Poetry generation (LSTMs applied to NLP)
We’ll similarly extend our domain to include text and language as we use LSTMs to generate novel writing samples based on training data.
Week 10: Language translation (sequence to sequence)
Neural Networks have been a fundamental part of the recent advancements in machine translation. The latest production versions of Google Translate and Baidu Translate both use deep learning architectures to automatically translate text from one language to another. This is done using a process known as Sequence to Sequence Learning, which we will explore in this lesson.
An example Sequence to Sequence Network Architecture. Source: Sutskever et. al
Week 11: Chatbot QA System with voice (sequence to sequence more in-depth)
We’ll further explore Sequence to Sequence learning through building our very own Chatbot QA system that can answer unstructured queries from a user.
Here, you’ll use Deep Learning to build your very own Chatbot. Chatbots Magazine
Week 12: Game bot 2D (Reinforcement Learning via Monte-Carlo tree search)
Some of the most interesting advancements in deep learning have been in the field of Reinforcement Learning, where instead of training on a corpus of existing data, a network learns from live data it receives and adjusts accordingly. We’ll see how to apply Reinforcement Learning to build simple Game-Playing AIs that can win in a wide variety of Atari games.
The classic Atari game of Space Invaders is one of many games you can win with Reinforcement Learning
Week 13: Image compression (Autoencoders)
As recently shown by Google, deep learning can also be used to dramatically improve compression techniques. In this lesson we’ll explore using deep learning to build autoencoders that automatically find sparse representations of data.
Week 14: Data visualization (anomaly detection results in 2D and 3D)
In this lesson, you’ll apply deep learning to detect anomalies in data. This is extremely useful in applications such as fraud prevention with credit cards.
Week 15: Image generation (generative adversarial networks)
As echoed by Yan LeCunn, Generative Adversarial Networks are one of the most fundamental advancements in deep learning. You’ll explore this state of the art concept to generate images that most humans wouldn’t believe are generated by a computer.
Images generated from a Generative Adversarial Network are becoming more and more realistic. Source: OpenAI
Week 16: One-shot learning (Probabilistic Programming) Finally, we’ll look at one-shot learning, where our neural network is able to just learn from one (or a few) example, as opposed to a large amount of data.
Through this curriculum, you will absorb an exciting introduction to some of the most compelling advancements in deep learning! We hope you join us on this journey and we can’t wait to share more of these ideas with you.
【Deep Learning Nanodegree Foundation笔记】第 1 课:INTRODUCTION Welcome
原文:http://www.cnblogs.com/custer/p/6367199.html