Reading Report Chapter 1 of PRML

本文深入探讨了概率论在机器学习中的核心作用,包括贝叶斯定理、最大似然估计、决策理论等关键概念,并讨论了这些理论在分类、回归任务中的应用。文章还分析了熵、相对熵、互信息等信息论概念,以及它们在模型选择和高维数据处理中的重要性。

摘要生成于 C知道 ,由 DeepSeek-R1 满血版支持, 前往体验 >

main content

  • probability theory
    • sum rule and product rule
    • Bayes’ theorem
      posterior∝likelihood×priorposterior\propto likelihood\times priorposteriorlikelihood×prior
    • maximum likelihood approah to inferring parameters in a probability distribution using oberved data
    • sum-of-squares error function is a consequence of maximizing likelihood under a Gaussian noise distribution
    • MAP
  • decision theory
    • classification
      • minimizing the misclassification rate
      • minimizing the expected loss
    • regression
      • loss function for regression
    • three methods to solve decision problems and some reasons for why posterior probabilities are required in many cases
  • information theory
    • entropy
      • the relation between entropy and shortest coding length (noiseless coding theorem)
      • interpretation as a measure of disorder
      • entropy over discrete and continuous variables (differential entropy)
    • relative entropy, KL(p∣∣q)≥0KL(p||q)\ge 0KL(pq)0 with equality if and only if p(x)=q(x)p(x) = q(x)p(x)=q(x)
    • mutual information, KL divergence between the joint distribution and the product of the marginals
  • model selection
  • curse of dimensionality

some thinking

The book exhibits much strict mathematical inference and emphasizes on the Bayesian viewpoint, which is very different from Hands-On Machine Learning with Scikit-Learn & TensorFlow, a book I have read previously.
Chapter 1 mainly introduces three basic theories, and bring the Bayesian perspective to each one.
In Bayesian view, probabilities provide a quantification of uncertainty which differs from frequentist viewpoints. Posterior is an adjustment of condition probability using prior probability, which is important in some decision problems such as compensating for class priors where the target class is rare in total data set.
Chapter 1 is only the introduction of the whole book and I believe that with further reading of the book I can have a better comprehension of PRML.

some problems

  • still don’t have a deep understanding of Bayesian viewpoint
  • have much difficulity understanding some mathematical resoning due to lack of some knowledge background such as the calculus of variations
  • too slow reading speed and takes a lot of time to read and understand the meaning

future

  • go on reading to gain overall viewpoints and better understanding of previous chapters
  • enhance some necessary knowledge
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值