doi.org/10.1023/A:1009869804967

Preview meta tags from the doi.org website.

Linked Hostnames

19

Thumbnail

Search Engine Appearance

Google

https://doi.org/10.1023/A:1009869804967

Partitioning Nominal Attributes in Decision Trees - Data Mining and Knowledge Discovery

To find the optimal branching of a nominal attribute at a node in an L-ary decision tree, one is often forced to search over all possible L-ary partitions



Bing

Partitioning Nominal Attributes in Decision Trees - Data Mining and Knowledge Discovery

https://doi.org/10.1023/A:1009869804967

To find the optimal branching of a nominal attribute at a node in an L-ary decision tree, one is often forced to search over all possible L-ary partitions



DuckDuckGo

https://doi.org/10.1023/A:1009869804967

Partitioning Nominal Attributes in Decision Trees - Data Mining and Knowledge Discovery

To find the optimal branching of a nominal attribute at a node in an L-ary decision tree, one is often forced to search over all possible L-ary partitions

  • General Meta Tags

    83
    • title
      Partitioning Nominal Attributes in Decision Trees | Data Mining and Knowledge Discovery
    • charset
      UTF-8
    • X-UA-Compatible
      IE=edge
    • applicable-device
      pc,mobile
    • viewport
      width=device-width, initial-scale=1
  • Open Graph Meta Tags

    6
    • og:url
      https://link.springer.com/article/10.1023/A:1009869804967
    • og:type
      article
    • og:site_name
      SpringerLink
    • og:title
      Partitioning Nominal Attributes in Decision Trees - Data Mining and Knowledge Discovery
    • og:description
      To find the optimal branching of a nominal attribute at a node in an L-ary decision tree, one is often forced to search over all possible L-ary partitions for the one that yields the minimum impurity measure. For binary trees (L = 2) when there are just two classes a short-cut search is possible that is linear in n, the number of distinct values of the attribute. For the general case in which the number of classes, k, may be greater than two, Burshtein et al. have shown that the optimal partition satisfies a condition that involves the existence of 2 L hyperplanes in the class probability space. We derive a property of the optimal partition for concave impurity measures (including in particular the Gini and entropy impurity measures) in terms of the existence ofL vectors in the dual of the class probability space, which implies the earlier condition.Unfortunately, these insights still do not offer a practical search method when n and k are large, even for binary trees. We therefore present a new heuristic search algorithm to find a good partition. It is based on ordering the attribute's values according to their principal component scores in the class probability space, and is linear in n. We demonstrate the effectiveness of the new method through Monte Carlo simulation experiments and compare its performance against other heuristic methods.
  • Twitter Meta Tags

    6
    • twitter:site
      @SpringerLink
    • twitter:card
      summary_large_image
    • twitter:image:alt
      Content cover image
    • twitter:title
      Partitioning Nominal Attributes in Decision Trees
    • twitter:description
      Data Mining and Knowledge Discovery - To find the optimal branching of a nominal attribute at a node in an L-ary decision tree, one is often forced to search over all possible L-ary partitions for...
  • Item Prop Meta Tags

    3
    • position
      1
    • position
      2
    • position
      3
  • Link Tags

    9
    • apple-touch-icon
      /oscar-static/img/favicons/darwin/apple-touch-icon-6ef0829b9c.png
    • canonical
      https://link.springer.com/article/10.1023/A:1009869804967
    • icon
      /oscar-static/img/favicons/darwin/android-chrome-192x192.png
    • icon
      /oscar-static/img/favicons/darwin/favicon-32x32.png
    • icon
      /oscar-static/img/favicons/darwin/favicon-16x16.png

Links

74