Thinking about a paper "A Refinement Approach to Handling Model Misfit in Text Categorization"

    in this paper, i think the most interesting thing is why there is no overfitting? overfitting is inevitable when training on training examples too much. because decision line(surface) fits specified property owned exclusively by these training examples, generality has been lost. in this paper, the authors use a refinement on classifier(Simon Haykin's book <<Neural Networks: A Comprehensive Foundation>> sec 7.5 shows a similar example: boosting by filtering. but difference should be noticed: this paper's method don't vote, it is not a boosting) to improving classifier performance. in the view of decision line, whatever refinement it is, the effect must be adjusted decision line to fitting training examples better. surprising experiments results shows there is no overfitting, authors just show this fact without explanation.
    how could this be? i have noticed that this method generates decision line is different with traditonal line:  you can imagine this "line": root classifier correspond to main line, its children classifers correspond to two smaller lines near the main line and so on... it is similar to fractal!! i guess these decision lines correctly express the STRUCTURE of target(real) decision, because i believe fractal is the nature of nature. but the problem still remained, ok, structure, so what? i guess:
  1 structure is more general than line, however this structure is obtained by training examples but it still has more generality than line. so overfitting is alleviated.
  2 to see gobal from local, this is property of fractal, similarity. training examples decision structure is similar to all examples. however regions contain arbitrary +,-.
  3 this method classifier don't vote, the classification is only determined by leaves nodes, intermidate nodes only do the dispatch job: translate input to a leaves node. this behavior contradict intuition and hard to explaining. leaves nodes should be specified version of intermidate nodes, but it replaces its parents judgements without too much error.
    may be each input is specified input, so correct classification in intermidate nodes can't be viewed as a "perfect classification", it just a "estimate on imperfection".
    further work is how to take advantage of this character of "decision line" if i guess it right. 
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值