我是靠谱客的博主 懵懂小熊猫,这篇文章主要介绍Stanford ML - Lecture 6 - Advice for applying machine learning,现在分享给大家,希望可以做个参考。

1. Deciding what to try next

  • Debugging a learning algorithm
    • Suppose you have implemented regularized linear regression to predict housing prices, when you test your hypothesis on a new set of houses, you find that it makes unacceptably large errors in its predictions. What should you try next?
      1. Get more training examples
      2. Try smaller sets of features
      3. Try getting additional features
      4. Try adding polynomial features
      5. Try decreasing 
      6. Try increasing 

2. Evaluating a hypothesis

  • separate data sets into training set (70%) andtest set (30%)
  • Training/Testing procedure for logistic regression
    • learn parameter  from training data
    • compute test set error
    • misclassification error (0/1 misclassification error)

3. Model selection and training/validation/test sets

  • overfitting example
    • the training error is likely to be lower than the actual generalization error
  • model selection
    • select the model that has the lowest test error
  • training set - 60%
  • cross validation set (cv) - 20%
  • test set - 20%
  • training error

  • cross validation error

  • test error
4. Diagnosing bias vs. variance

  • bias (underfit)


  • variance (overfit)

5. Regularization and bias/variance

  • choosing the regularization parameter 


    • ........

6. Learning curves

  • If a learning algorithm is suffering from high bias, getting more training data will not (by itself) help much.
  • If a learning algorithm is suffering from high variance, getting more training data is likely to help.
7. Deciding what to try next (revisited)

  • "small" neural network (fewer parameters, more prone to underfitting)
    • computationally cheapter
  • "large" neural network (more parameters, more prone to overfitting)
    • computationally more expensive
    • use regularization to address overfitting

the definition:

Variance: measures the extent to which the solutions for individual data sets vary around their average, hence this measures the extent to which the function f(x) is sensitive to theparticular choice of data set.

Bias: represents the extent to which the average prediction over all data sets differs from the desired regression function.

variance:估计本身的方差。

bias:估计的期望和样本数据样本希望得到的回归函数之间的差别。

From : http://blog.csdn.net/abcjennifer/article/details/7797502

最后

以上就是懵懂小熊猫最近收集整理的关于Stanford ML - Lecture 6 - Advice for applying machine learning的全部内容,更多相关Stanford内容请搜索靠谱客的其他文章。

本图文内容来源于网友提供,作为学习参考使用,或来自网络收集整理,版权属于原作者所有。
点赞(50)

评论列表共有 0 条评论

立即
投稿
返回
顶部