【DeepCV】超参数 Hyperparameters

超参数是机器学习算法中预先设定的变量,影响模型的优化过程。神经网络的超参数包括训练和模型两类。训练超参数如学习率、动量和批次大小等,而模型超参数涉及网络结构、隐藏单元数量等。设置超参数通常通过手动、搜索算法或超学习器进行。此外,预处理输入数据、权重初始化和正则化也是优化模型性能的关键因素。

摘要生成于 C知道 ,由 DeepSeek-R1 满血版支持, 前往体验 >

Backto DeepCV

Notes from Neural Network Hyperparameters.

What is Hyperparameters?

Most machine learning algorithms involve “hyperparameters” which are variables set before actually optimizing the model’s parameters. Neural networks can have many hyperparameters, including those which specify the structure of the network itself and those which determine how the network is trained.

Setting the values of hyperparameters can be seen as model selection, i.e. choosing which model to use from the hypothesized set of possible models.

How to set Hyperparameters?

Hyperparameters are often set

  • by hand. experience
  • selected by some search algorithm, such as grid search, random search,
  • optimized by some “hyper-learner”. (hot topic)

Typical Hyperparameter

In particular, we will focus on feed-forward neural nets trained with mini-batch gradient descent.

Trainning Hyperparameters
  • learning rate: determines how quickly the gradient updates follow the gradient direction. If the learning rate is very small, the model will converge too slowly; if the learning rate is too large, the model will diverge.
  • Momentum: A very common technique is to “smooth” the gradient updates using a leaky integrator filter with parameter β \beta β.
  • Loss Function: compares the network’s output for a training example against the intended ground truth output. A common general-purpose loss function is the squared Eclidian distance, given by L = 1 2 ∑ i ( y i − z
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值