Generated by GPT-5-mini| ID3 | |
|---|---|
| Name | ID3 |
| Author | Ross Quinlan |
| Year | 1986 |
| Paradigm | Decision tree learning |
| Input | Labeled examples |
| Output | Decision tree |
| Related | C4.5, CART, CHAID, RIPPER |
ID3
ID3 is a decision tree learning algorithm developed for inductive inference from labeled examples. It constructs classification trees by selecting attributes that maximize information gain, producing interpretable models for tasks in domains such as Medical diagnosis, Credit scoring, Speech recognition, and Bioinformatics. The algorithm influenced later systems and standards in machine learning research at institutions such as University of Sydney, AT&T Bell Laboratories, and companies including IBM, Microsoft, and Google.
ID3 was published by Ross Quinlan while at University of Sydney and presented in venues alongside work from researchers at Stanford University, Massachusetts Institute of Technology, Carnegie Mellon University, and University of California, Berkeley. The method uses concepts from Information theory, notably entropy and information gain, to choose splitting attributes; these ideas connect to earlier work by Claude Shannon and later to criteria used in algorithms such as C4.5 and CART. ID3 became widely cited in conferences like International Joint Conference on Artificial Intelligence and journals such as Machine Learning (journal).
ID3 emerged in the mid-1980s amid growth in inductive learning research at institutions including University of Queensland, University of Edinburgh, and University of Cambridge. Quinlan's work built on symbolic machine learning traditions exemplified by systems like PROGOL, FOIL, and research at SRI International. The algorithm influenced subsequent successors including C4.5 and commercial tools from SPSS, SAS Institute, and early Weka (software) distributions from University of Waikato. ID3’s publication paralleled advances in computational resources from vendors such as Sun Microsystems, DEC, and later Intel that enabled wider experimentation.
ID3 takes a set of training examples described by discrete attributes and a target class drawn from a finite set used in tasks at organizations like World Health Organization, Federal Reserve, and NASA. For each node, ID3 computes class entropy using the Shannon formula and selects the attribute with maximum information gain, an approach conceptually linked to measures used in Kolmogorov complexity discussions and analyses at Bell Labs Research. Trees are grown recursively until leaves are pure or no attributes remain; stopping conditions echo practices in studies at MIT Media Lab and Harvard University. ID3 handles categorical attributes directly but requires discretization for continuous variables, a preprocessing performed in toolchains from R Project, MATLAB, and Scikit-learn influenced by methods in papers at NeurIPS and ICML. The output tree can be translated into rule sets or executed in production systems at companies such as Amazon and Facebook.
Quinlan’s successor algorithm C4.5 introduced handling of continuous attributes, pruning, and mechanisms for missing values; these enhancements were compared in benchmark studies at UCI Machine Learning Repository and evaluated against CART by researchers at Columbia University and Princeton University. Other extensions include entropy alternatives such as gain ratio, Gini impurity used in Breiman's work, and approaches in ensemble methods like Random Forests and Gradient Boosting Machines developed at University of Toronto and Microsoft Research. Hybrid systems have combined ID3-like splits with statistical models from Bell Labs and probabilistic frameworks from Stanford and Berkeley leading to methods used in Kaggle competitions and deployments by Netflix.
ID3-style trees have been applied to classification problems at CERN, Centers for Disease Control and Prevention, European Space Agency, and National Institutes of Health. Use cases include diagnostic decision support at Mayo Clinic, risk assessment in banking at JPMorgan Chase, and customer segmentation at Procter & Gamble. Performance comparisons often place ID3 variants favorably for interpretability in benchmarks hosted by UCI Machine Learning Repository and contests organized by Kaggle; however, predictive accuracy is sometimes lower than ensemble methods from Google Research or deep learning models from OpenAI and DeepMind. ID3 trees are valued in regulated fields such as Food and Drug Administration-governed clinical applications for their transparency.
Critiques of ID3 appear in literature from IEEE and ACM venues, noting sensitivity to noisy data, overfitting without pruning, and bias toward attributes with many values as discussed in studies at Cornell University and Yale University. The need for discretization of continuous attributes led to information loss in empirical evaluations by teams at Los Alamos National Laboratory and Lawrence Berkeley National Laboratory. Subsequent methods such as C4.5, CART, and ensemble techniques address many shortcomings but trade off interpretability versus predictive power, a debate present in policymaking fora at European Commission and standards discussions at ISO.
Category:Machine learning algorithms