This is a records of Machine Learning Course of Andrew Ng on Coursera
Linear Regression with One Variable
Model and Cost Function
Linear Algebra Review
Environment Setup Instructions
Multivariate Linear Regression
Computing Parameters Analytically
Submitting Programming Assignments
Classification and Representation Logistic Regression Model Multiclass Classification
Neural Networks
Applications
Cost Function and Backpropagation
Backpropagation in Practice
Application of Neural Networks
Evaluating a Learning Algorithm
Bias vs. Variance
Large Margin Classification
Kernels
SVMs in Practice
Clustering
Dimensionality Reduction
Density Estimation
Building an Anomaly Detection System
Multivariate Gaussian Distribution (Optional)
Predicting Movie Ratings
Collaborative Filtering
Low Rank Matrix Factorization
努力学习中...
- Feature Scaling: 当变量取值范围比较接近时,参数收敛速度比较快,如果变量取值范围很大,就需要进行Feature Scaling(归一化)。一般在(-3,3)范围内都是恶可以接受的。归一化方法可以参考第二周编程练习。
- 除了梯度下降外,Andrew 又讲了几种优化方法——Conjugate,BFGS,L-BFGS.
属性 | 梯度下降 | Conjugate,BFGS,L-BFGS |
---|---|---|
优点 | 理解较为简单 | 不用人为选择学习率;速度快 |
缺点 | 需要人工选择学习率 | 算法复杂 |
- 第七周感觉是目前最有意思的一周,试验了邮件垃圾分类,如果之前没做过会感觉挺有意思。实验中的方法也很方面做一些扩展和优化,自己做个demo还是可以的。
- 第八周是关于无监督学习,讲了K-means聚类和PCA。Andrew建议不要用PCA进行过拟合处理,过拟合的话还是用正则及其他方法。