Decision tree: Difference between revisions
CSV import |
CSV import |
||
| Line 36: | Line 36: | ||
[[Category:Statistics]] | [[Category:Statistics]] | ||
{{stub}} | {{stub}} | ||
<gallery> | |||
File:Manual_decision_tree.jpg|Manual decision tree | |||
File:Decision-Tree-Elements.png|Decision tree elements | |||
File:DecisionCalcs.jpg|Decision calculations | |||
File:RiskPrefSensitivity2Threshold.png|Risk preference sensitivity to threshold | |||
File:Beachdecisiontree.png|Beach decision tree | |||
File:Lifeguards.png|Lifeguards | |||
File:Factory2_InfluenceDiagram.png|Factory influence diagram | |||
File:Realnode.jpg|Real node | |||
File:Information_Gain_Tree.jpg|Information gain tree | |||
File:Phi_Function_Tree.jpg|Phi function tree | |||
</gallery> | |||
Latest revision as of 11:23, 18 February 2025
Decision tree is a type of graphical model used in statistics, data mining, and machine learning. It uses a tree-like model of decisions and their possible consequences, including chance event outcomes, resource costs, and utility. It is one way to display an algorithm that only contains conditional control statements.
Overview[edit]
A decision tree is a flowchart-like structure in which each internal node represents a "test" on an attribute (e.g. whether a coin flip comes up heads or tails), each branch represents the outcome of the test, and each leaf node represents a class label (decision taken after computing all attributes). The paths from root to leaf represent classification rules.
Types of Decision Trees[edit]
There are two main types of decision trees: classification tree and regression tree.
- Classification tree analysis is when the predicted outcome is the class (discrete) to which the data belongs.
- Regression tree analysis is when the predicted outcome can be considered a real number (e.g. the price of a house, or a patient's length of stay in a hospital).
Advantages and Disadvantages[edit]
Advantages[edit]
- Decision trees are simple to understand and interpret.
- Requires little data preparation.
- Able to handle both numerical and categorical data.
Disadvantages[edit]
- Decision-tree learners can create over-complex trees that do not generalize the data well.
- Decision trees can be unstable because small variations in the data might result in a completely different tree being generated.


