开发者:上海品职教育科技有限公司 隐私政策详情

应用版本:4.2.11(IOS)|3.2.5(安卓)APP下载

hemanie · 2021年01月22日

问一道题:NO.PZ201512020300000507 第7小题 [ CFA II ]

* 问题详情,请 查看题干

问题如下:

Assuming a Classification and Regression Tree (CART) model is used to accomplish Step 3, which of the following is most likely to result in model overfitting?

选项:

A.

Using the k-fold cross validation method

B.

Including an overfitting penalty (i.e., regularization term).

C.

Using a fitting curve to select a model with low bias error and high variance error.

解释:

C is correct. A fitting curve shows the trade-off between bias error and variance error for various potential models. A model with low bias error and high variance error is, by definition, overfitted.
A is incorrect, because there are two common methods to reduce over
fitting, one of which is proper data sampling and cross-validation. K-fold cross validation is such a method for estimating out-of-sample error directly by determining the error in validation samples.
B is incorrect, because there are two common methods to reduce over
fitting, one of which is preventing the algorithm from getting too complex during selection and training, which requires estimating an overfitting penalty.

题干是说在Cart模型下,哪种会导致overfitting。但是选项中比如B增加惩罚项,这个从性质上来说,如何能和CART模型一起用呢,感觉和题干没办法融合
1 个答案

星星_品职助教 · 2021年01月22日

同学你好,

CART模型有两种降低overfitting的方式:

1)添加regularization parameter,即在“tree”上添加惩罚项(penalty),使得“tree”不要无限的细化延伸下去。

2)采用“pruning technique”

本题中B选项指得就是第一种方式。例如添加一个参数去控制tree的maximun depth,或者添加一个参数去控制总节点(node)的数量等,这些都是 regularization term或者overfitting penalty.