Skip to content

Commit a351f46

Browse files
committed
update pre
1 parent 663d023 commit a351f46

File tree

2 files changed

+59
-2
lines changed

2 files changed

+59
-2
lines changed

README.md

Lines changed: 52 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -36,18 +36,56 @@ N/A
3636
- Real-Time Continuous Pose Recovery of Human Hands Using Convolutional Networks [[NYU Hand Pose Dataset]][NYU-Hand]
3737

3838
### Week - 8
39-
- Hand Photo Data Collection
39+
- [Hand Video and Images Collection][nyu_hr]
40+
- Additional Articles:
41+
+ SSD: Single Shot MultiBox Detector [[Arxiv]][ssd]
42+
+ R-CNN: Rich Feature Hierarchies for Accurate Object Detection and Semantic Segmentation [[PDF]][rcnn]
43+
+ YOLO:
44+
* You Only Look Once: Unified, Real-Time Object Detection [[Arxiv]][yolopaper]
45+
* [Another YOLO Paper][yolopaper2]
46+
* [Comparison][yolocomparison]
4047

4148
### Week - 9
4249
- Generative Adversarial Nets (GAN) https://arxiv.org/pdf/1406.2661v1.pdf
4350
- Deep Generative Image Models using a Laplacian Pyramid of Adversarial Networks. Emily L. Denton et. al. Nips 2015. https://arxiv.org/abs/1506.05751
51+
- GANs with Feature Extraction Function:
52+
+ ALI
53+
+ BiGAN
54+
+ Adversarial Autoencoder
55+
+ Adversarial Variational Bayes.
4456

4557
### Week - 10 [Variational Auto Encoders]
4658
- Tutorial on Variational Autoencoders(VAEs), Carl Doersch. August, 2016. https://arxiv.org/abs/1606.05908
4759
+ Relevance: sampling from distributions inside deep networks, and can be trained with stochastic gradient descent. VAEs have already shown promise in generating many kinds of complicated data, including handwritten digits, faces, house numbers, CIFAR images, physical models of scenes, segmentation, and predicting the future from static images.
4860
- "Variational Convolutional Networks for Human-Centric Annotations," 13th Asian Conference on Computer Vision, 2016. Tsung-Wei Ke, Che-Wei Lin, Tyng-Luh Liu and Davi Geiger,
4961
+ Relevance: Use of VAEs to annotate automatically images.
5062

63+
### Week - 11 + 12
64+
65+
### Week - 13 [Reinforcement Learning and Markov Decision Process]
66+
67+
### Week - 14 [Recurrent Neural Network]
68+
69+
### Week - 15 [Review Reinforcement Learning and Self-Driving Car]
70+
71+
72+
73+
### Additional Topic - [Object Detection]
74+
- RCNN: Region-based Convolutional Neural Networks
75+
- Fast and Faster R-CNN: https://github.com/rbgirshick/py-faster-rcnn
76+
- YOLO
77+
- SSD
78+
- Soft-NMS: https://github.com/bharatsingh430/soft-nms
79+
80+
### Research Topics:
81+
- Synthetic Gradients [[arxiv]][synthetic_gradients]
82+
- One Shot Learning [[Wiki]][one_shot_learning]
83+
- Improving optimization in GAN
84+
85+
### Additional Resource: [Edit Before Submit:]
86+
机器学习自学者必读的20篇顶级论文导读: https://mp.weixin.qq.com/s/ghMj3OO2yu7IIkQEkvkdVA
87+
常见面试机器学习方法总览: http://www.chinakdd.com/article-oyU85v018dQL0Iu.html
88+
5189

5290
[pre]:https://github.com/lizichen/Machine-Learning-For-Computer-Vision/blob/master/pre-requisite.md
5391
[framework]:https://github.com/lizichen/Machine-Learning-For-Computer-Vision/blob/master/tools-frameworks.md
@@ -69,3 +107,16 @@ N/A
69107
[synergface]: http://rita.osadchy.net/papers/OsadchyLeCunJMLR.pdf
70108
[viola]: http://www.merl.com/publications/docs/TR2004-043.pdf
71109

110+
[nyu_hr]:https://github.com/lizichen/Machine-Learning-For-Computer-Vision/tree/master/NYU_HR
111+
[ssd]:https://arxiv.org/abs/1512.02325
112+
[rcnn]:https://arxiv.org/pdf/1311.2524.pdf
113+
[yolopaper]:https://arxiv.org/abs/1506.02640
114+
[yolopaper2]:https://pjreddie.com/media/files/papers/YOLO9000.pdf
115+
[yolocomparison]:https://pjreddie.com/darknet/yolo/
116+
117+
118+
119+
[one_shot_learning]: https://en.wikipedia.org/wiki/One-shot_learning
120+
[synthetic_gradients]: https://arxiv.org/abs/1703.00522
121+
122+

pre-requisite.md

Lines changed: 7 additions & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -77,7 +77,12 @@
7777
+ If the size of training samples is large, using GD is not effective because in every iteration when updating the parameters, you are running through the complete training set. On the other hand, using SGD will be faster because you only use one training sample and it starts improving itself right away from the first sample.
7878
+ **SGD** often converges much **faster** compared to GD, but the error function is not as well minimized as in the case of GD. Often in most cases, the close approximation that you get in SGD for the parameter values are enough because they reach the optimal values and keep oscillating there.
7979
- ADAM (**Ada**ptive **M**oment Estimation) optimizer: Computes adaptive learning rates for each parameter.
80-
- **ReLU**
80+
- **ReLU** f(x) = max(0, x)
81+
+ Greatly accelerate the convergence of stochastic gradient descent (compared to sigmoid/tanh functions).
82+
* Sigmoid函数的backpropagation很容易出现梯度消失(SGD die out),在Sigmoid接近饱和区,变化太缓慢,导数趋近于0,导致信息丢失。
83+
+ Non-saturating form.
84+
+ ReLU is fragile and can 'die' - to fix this: **Leaky ReLU** & **batch normalization**.
85+
- Maxout
8186
- Softmax
8287
- Logistic Classifier
8388
+ Logistic function: logistic(x) = 1 / (1 + numpy.exp(-x))
@@ -168,6 +173,7 @@
168173

169174
### OTHER TUTORIALS AND REFERENCES:
170175
- Carlos Fernandez-Granda's lecture notes provide a comprehensive review of the prerequisite material in linear algebra, probability, statistics, and optimization.
176+
- **Step-by-Step Backpropagation:** https://mattmazur.com/2015/03/17/a-step-by-step-backpropagation-example/
171177
- Brian Dalessandro's iPython notebooks from DS-GA-1001: Intro to Data Science
172178
- The Matrix Cookbook has lots of facts and identities about matrices and certain probability distributions.
173179
- Stanford CS229: "Review of Probability Theory"

0 commit comments

Comments
 (0)