针对传统的基于信息熵的决策树算法所存在的问题,本研究从粗糙集理论的角度来考虑信息熵,定义了依赖决策熵的概念,并提出一种基于依赖决策熵的决策树算法DTDDE。在DTDDE算法中,采用依赖决策熵的概念来度量每个条件属性的重要性,并选择重要性最大的属性作为当前的分离属性。通过在多个UCI数据集上的实验表明:与现有的决策树算法相比,本研究所提出的算法能够获得更好的分类性能。
To solve the problems of traditional information entropy-based decision tree algorithms,in this paper we consider the issue of information entropy from the view of rough set theory.We define a new concept called dependency decision entropy,and propose a dependency decision entropy-based algorithm(called DTDDE)for constructing decision tree.In algorithm DTDDE,the concept of dependency decision entropy is used to measure the significance of each condition attribute,and the attribute with the maximum significance is selected as the current splitting attribute.Experimental results on several UCI data sets demonstrate that compared with the current decision tree algorithms,our algorithm can obtain better classification performance.