Introduction
My note for MOOC lesson Introduction to Deep Learning.
Lecture 2:
2.1 Bag of words
Problem: no sequence. cannot represent the order of words.
This leads to relearn the ruls of language at each point in the sentence.
markov assumption: each state depends only on the last state. $P(x_{t+1} \vert x_t)$
Problem: from far past and future to accurately guess the correct word. So markov not work.
2.2 RNN
- word order
- share parameters across the sequence(W,U weight matrics)
- keep track of long-term dependencies
$x_0$: vector representeation of first word,
$s_0$: cell state at t=0
$s_1$: cell state at t=1
For the $s_1$
For $s_2$:
$s_n$ can contain informmation from all past timesteps
2.4 possible task
language model: what most likely next word. output the possibility of each $s_i$. aka softmax.
classification:sentiment on twitters. output at the end of the learning. Get (negative, nueral, positive) at the last state $s_n$. aka softmax.
machine translation:
encoder -> takes one setence fomr from one laguage you want to translate. output $s_n$ to decoder
decoder -> $s_n$ in encoder is the input for decoder. $s_i$ to each input state of decoder.
2.5 Train: backpropagation
- take the derivative
- shift paremeters in the opposite direction.
2.6 Loss
we make prediction each time, loss at each timestep.
try chain rule:
Consider $s_2$ and $t = 2$:
$s_2$ is dependent on $W$.
2.7 RNN Problem
Problem: because of chain rule, it will vanish gradient.
Now we multiply a lot of small numbers. So the parameters become biased to capture shorter-term dependencies.
要看懂这块建议先看Backward Propagation的推导过程,以及RNN推导和RNN公式。
2.8 RNN Solution
- activation function
- initialization(weights matrix -> $I$ not zeros)
- gated celss(lead to LSTM), complex unit with gates
LSTM的详解可以参看理解LSTM网络。
2.9 Solution for LSTM(translation task)
Give part of the output states of encoder to decoder. Like attention?
Lecture 3: Deep Learning for Computer Vision
Learning is optimization of a function
3.1 CNN layers
一般是3D的数据,前两维是图像,第三维可以理解为RGB这样。
Input:[32x32x3]
CONV:this layer will computer the output of neurons that are connected to local regions in the input. If [32x32x12] this is 12 filters.
RELU: activation function, leave the size unchanged.
POOL: do downsampling operation. result in like [16x16x12]
FC: fully-connected layre computer the class scores. result in like [1x1x10] where the 10 numbers correspond to a class score.
3.2
Depth: number of filters()
Stride: filter step size(也就是每次移动的距离)
padding: zero-pad the input(input的边缘用不用0来进行填充)
Lecture 4: Deep Generative Models
Using learning to generate images/speech/handwriting/language
idea: learn to understand data through generation
4.1 Setup
Discriminative model: give n examples $(x^{(i)}, y^{(i)})$, learn h: $X \to Y$.
Generative model: give n examples $x^{(i)}$, recover $p(x)$.
Maximum-likelihood objective: $\prod_i p_\theta(x) = \sum_i \log p_\theta (x)$
Generation: Sampling from $p_\theta (x)$.
4.2 Autoregressive Models
build a next-step prediction model $p(x_n \vert x_1, \cdots, x_{n-1})$.
If x is discrete, network outputs a probability for each possible value.
If x is continuous, network outputs parameters of simple distribution(Guassian mean and variance).
Generation: sample one step at a time, conditioned on all previous steps.
RNN can also be written as autoregressive language models
e.g. PixelRNN (van der Oord et al. 2016): Autoregressive RNN over pixels in an image
4.3 Problem about Autoregressive Models
autoregressive models are powerful density estimators. but sequential generation can be slow. not reflect the true generating process, not very good for learning representations.
4.4 Autoencoders for representation learning
input to encoder output(bottleneck layer), then latent representation to reconstruction.
So the loss is Reconstruction loss( forces hidden layer to represent information about the input), bottleneck hidden layer forces network learn a compressed latent representation.
4.5 compression as implicit generative modeling
Variational Autoencoders(变分自动机)
Generative extension of autoencoders which allow sampling and estimating probabilities
“Latent variables” with fixed prior distribution $p(z)$
Probabilistic encoder and decoder: $q(z\vert x), p(x \vert z)$.
Trained to maximize a lower bound on log-probability:
$\log p(x) \ge E_{z\sim q(z\vert x)}[\log p(x\vert z) + \log p(z) - \log q(z)]$
4.6 Problems with VAEs
Encoder and decoder’s output distributions are typically limited (diagonal-covariance Gaussian or similar).
Generative Adversarial Networks: make a generative model by having two neural networks compete with each other. 与一般的被用于Supervised Learning任务的深度神经网络不同,GAN同时要训练一个生成网络(Generator)和一个判别网络(Discriminator),前者输入一个noise变量 z ,输出一个伪图片数据$ G(z;θ_g)$ ,后者输入一个图片(real image)/伪图片(fake image)数据 x ,输出一个表示该输入是自然图片或者伪造图片的二分类置信度 $D(x;θ_d)$ ,理想情况下,判别器 D 需要尽可能准确的判断输入数据到底是一个真实的图片还是某种伪造的图片,而生成器G又需要尽最大可能去欺骗D,让D把自己产生的伪造图片全部判断成真实的图片。
具体可以参看【青年学者专栏】解读GAN及其 2016 年度进展。
discriminator: distinguish genuine data
generator: turns random noise into immitation of the data
5. Lecture 5:Multimodal Learning
5.1 What is multimodal Learning
- Multiple modalities are learned jointly
- One modality assists in the learning of another
- Input is one modality, output is another
5.2 Why is multimodal learning hard?
different representations for different data.
Images -> Real-valued, Dense
Text -> Discrete, Sparse
Noisy and missing data
5.2 How to solve
combine seprate models for a single modality at a higher level
pre-train models on single-modality data
Use Embeddings to combine these models.
5.3 Pretrain
- initialize with the weights from another network(better than random)
- low-level features are still be useful.
5.4 Embeddings
A way to represent data.
In deep learning, this is usually a high-dimensional vector.
A neural network can take a piece of data and create a corresponding vector in an embedding space.
A neural network can take embedding vector as an input(word embeddings).
word embeddings: word $\to$ high-dimensional vector.
Can use embeddings to switch between modalities.
5.5 Flickr tagging: model
Guassian model?
Replicated Softmax?
Task: combine a model that can represent images and text. Like you feed a image of a sunset and text sea. It will returns a image of sea sunset.
pretrain unimodal models and combine them at a higher level. pretrain $V_{image}$ and $V_{text}$ speperate.
Two Models: image-specific model and text-specific model.
picture is a vector and word so on.
5.6 Image Captioning
Image $\to$ Vision(Deep CNN) $\to$ Language Generating RNN $\to$ sentence.
5.7 SoundNet
Task: learning sound representation from videos.
- Video $\to$ Raw Waveform $\to$ $g(y)$
- Video $\to$ RGB frames $\to$ Visual Recognition networks $\to$ $f(x; \theta)$
The loss: $D_{KL}(g(y) \Vert f(x;\theta))$
6. Learning to Do
6.1 Learning Task
Supervised Learning: given data, predict labels.
Unsupervised Learning: given data, learn about that data.
Reinforcement Learning: given data, choose action to maximize expected long-term reward.
6.2 Q-learning
$s_0, a_0, r_0, s_1, a_1, r_1, \cdots, s_T, r_T$. is the sequence of states and actions.
Transition function: $P(s_{t+1}, r_t \vert s_t, a_t)$.
Policy Learning: find $\pi(s)$. $a \sim \pi(s)$
Value learning: find $Q(s,a)$. $a = \arg \max_{a’}Q(s,a’)$
maximum expected future rewards starting at state $s_i$. choosing action $a_i$, and then following an optimal policy $\pi^*$.
Define approximate $Q^*$ function.
因为我们是朋友,所以你可以使用我的文字,但请注明出处:http://alwa.info