LLMpediaThe first transparent, open encyclopedia generated by LLMs

Yoav Freund

Generated by GPT-5-mini
Note: This article was automatically generated by a large language model (LLM) from purely parametric knowledge (no retrieval). It may contain inaccuracies or hallucinations. This encyclopedia is part of a research project currently under review.
Article Genealogy
Parent: Reinforcement Learning Hop 4
Expansion Funnel Raw 54 → Dedup 0 → NER 0 → Enqueued 0
1. Extracted54
2. After dedup0 (None)
3. After NER0 ()
4. Enqueued0 ()
Yoav Freund
Yoav Freund
Yoav Freund · CC BY 3.0 · source
NameYoav Freund
Birth date1961
Birth placeJerusalem, Israel
NationalityIsraeli-American
FieldsComputer science, Machine learning, Computational learning theory
WorkplacesUniversity of California, San Diego; Technion – Israel Institute of Technology; AT&T Bell Laboratories
Alma materHebrew University of Jerusalem; University of California, San Diego
Doctoral advisorManfred K. Warmuth
Known forBoosting, AdaBoost, Ensemble methods, Online learning
AwardsGödel Prize, IEEE Fellow, ACM Fellow

Yoav Freund

Yoav Freund is an Israeli-American computer scientist and researcher known for foundational work in machine learning, particularly ensemble methods and boosting. He is a professor at the University of California, San Diego and a co-developer of the AdaBoost algorithm that influenced statistical learning theory, pattern recognition, and applications across industry and academia. Freund's work intersects with theoretical computer science, information theory, and practical algorithm design, impacting research at institutions and companies worldwide.

Early life and education

Born in Jerusalem, Freund studied at the Hebrew University of Jerusalem where he completed undergraduate studies in computer science and related fields. He later pursued doctoral studies at the University of California, San Diego under the supervision of Manfred K. Warmuth, focusing on computational learning theory and algorithmic foundations. During his graduate years Freund collaborated with researchers linked to AT&T Bell Laboratories, engaged with scholars from Massachusetts Institute of Technology, and participated in workshops associated with Neural Information Processing Systems and the International Conference on Machine Learning.

Academic career and positions

Freund held positions at the Technion – Israel Institute of Technology before joining the faculty of the University of California, San Diego. He has been affiliated with research labs and collaborative centers such as AT&T Bell Laboratories, the Center for Neural Computation, and various programs at the National Science Foundation. Freund has served on program committees for conferences including COLT (Conference on Learning Theory), NIPS (NeurIPS), and ICML, and has been invited to give talks at institutions such as Stanford University, Harvard University, Princeton University, and Carnegie Mellon University. His collaborations span researchers from Microsoft Research, Google Research, IBM Research, and academic groups at University of Toronto and University of California, Berkeley.

Research contributions and work on machine learning

Freund is best known as a co-inventor of the AdaBoost algorithm, developed with Robert E. Schapire, which unified ideas from PAC learning, ensemble methods, and margin-based classifiers. AdaBoost's theoretical underpinnings connect to concepts from Vapnik–Chervonenkis theory, support vector machines, and the theory of boosting, and its practical impact influenced tools used at Amazon, Facebook, Apple, and other technology firms. Freund's research explores the convergence properties of boosting, links between boosting and additive modeling, and the design of online learning algorithms such as the Hedge algorithm and variants connected to the framework of multiplicative weights. His work established connections between boosting and information-theoretic constructs like the Kullback–Leibler divergence and techniques related to convex optimization studied at MIT and Princeton University.

Freund contributed to the formal analysis of generalization bounds, sample complexity, and model capacity, tying results to Occam's razor-style bounds, structural risk minimization as promoted by Vladimir Vapnik, and margin theory. He collaborated on algorithms for online prediction, adversarial settings, and bandit problems, interacting with research on algorithms from Leslie Valiant's PAC framework, the Arora–Hazan–Kale multiplicative weights method, and adaptive data analysis. Freund's publications address applications across bioinformatics, computer vision, and natural language processing, influencing systems developed at Genentech, Google DeepMind, and research groups at University College London.

Awards and honors

Freund received the Gödel Prize jointly with collaborators for work on boosting, and has been elected a Fellow of the Association for Computing Machinery and a Fellow of the Institute of Electrical and Electronics Engineers. He has been recognized with awards and invited lectures from organizations including the European Association for Theoretical Computer Science, the International Joint Conference on Artificial Intelligence keynote programs, and honors from his institutions such as named faculty fellowships at University of California, San Diego. Freund's work on AdaBoost and learning theory has been cited in award citations for colleagues who received the Turing Award and other major recognitions.

Selected publications

- Freund, Y.; Schapire, R. E., "A Decision-Theoretic Generalization of On-Line Learning and an Application to Boosting", Proceedings of the European Conference on Computational Learning Theory / COLT and Neural Information Processing Systems (canonical publication establishing AdaBoost concepts). - Freund, Y.; Schapire, R. E., "Experiments with a New Boosting Algorithm", in Machine Learning proceedings and workshops associated with ICML and AAAI. - Freund, Y.; Schapire, R. E.; Singer, Y.; Warmuth, M. K., "Using and Combining Predictors That Specialize", presented at conferences including NIPS and journals covering algorithmic learning theory. - Freund, Y.; Mansour, Y.; Schapire, R. E., works on Adaptive Boosting variants and generalization bounds published in venues such as Journal of Machine Learning Research and proceedings of COLT. - Freund, Y.; Haussler, D., contributions to ensemble and online learning literature appearing in collections from MIT Press and conference volumes from SIAM.

Category:Living people Category:Machine learning researchers Category:University of California, San Diego faculty Category:Israeli computer scientists Category:Fellows of the Association for Computing Machinery