Generated by GPT-5-mini| Chatbots | |
|---|---|
![]() Unknown author · Public domain · source | |
| Name | Chatbots |
| Introduced | 1960s |
| Inventor | Joseph Weizenbaum, Alan Turing |
| Related | Artificial intelligence, Natural language processing, Machine learning |
Chatbots Chatbots are software systems designed to simulate human conversation using Natural language processing, Machine learning, and rule-based engines. Early programs and influential demonstrations by figures such as Alan Turing and Joseph Weizenbaum set foundations that intersect with work at institutions including MIT, Stanford University, and companies like IBM and OpenAI. Contemporary ecosystems involve platforms, research groups, and standards developed by organizations such as Google, Microsoft, Facebook, Amazon, and NVIDIA.
The conceptual roots trace to the Turing test proposed by Alan Turing and to early implementations like ELIZA by Joseph Weizenbaum at MIT and later conversational systems such as PARRY developed by Kenneth Colby at Stanford University. Commercial and research milestones include projects from IBM Watson at IBM Research, dialog systems at AT&T Bell Labs, virtual assistants from Apple (Siri), Microsoft (Cortana), Google (Google Assistant), and virtual agent research at DARPA and European Commission funded labs. Academic conferences and venues such as ACL (conference), EMNLP, NAACL, NeurIPS, ICML, and AAAI have chronicled technical advances, while startups like Rasa (software), SoundHound, CognitiveScale, and large-scale models by OpenAI and DeepMind pushed capabilities. Regulatory and public discourse involved bodies like the European Union and events including hearings in the United States Congress.
Architectures combine components from Natural language processing subfields: tokenization and embeddings (e.g., word2vec, GloVe, BERT), sequence models (e.g., LSTM, GRU), and transformer architectures popularized by Vaswani et al. and implemented in systems like GPT (language model series), BERT by Google Research, and T5. Backends involve frameworks and platforms such as TensorFlow, PyTorch, Hugging Face, ONNX, and inference tooling from NVIDIA and Intel. Dialogue management uses approaches from probabilistic models (e.g., Partially observable Markov decision process) and rule-based engines influenced by research at CMU and SRI International. Data sources include corpora like Penn Treebank, Switchboard Corpus, and web-scale datasets curated by organizations including Common Crawl and academic groups at University of Edinburgh and Stanford NLP Group.
Conversational agents span task-oriented assistants used by Amazon (Alexa), customer-service bots deployed by Salesforce, Zendesk, and SAP, social chat agents in products from Meta Platforms (Facebook Messenger), educational tutors developed at Carnegie Mellon University, therapy-oriented systems influenced by clinical work at Mayo Clinic and Kaiser Permanente, and creative-writing or coding aids from GitHub Copilot (by GitHub and OpenAI). Specialized deployments include multilingual interfaces at UN agencies, virtual receptionists in Hilton and Marriott hospitality chains, and legal or medical triage assistants built by startups like DoNotPay and health-tech firms collaborating with Johns Hopkins University.
Design draws on human–computer interaction research from Don Norman and usability labs at Stanford University and prototype cycles practiced by companies such as IDEO and Frog Design. Development workflows integrate annotation and data pipelines from firms like Scale AI and platforms like Dialogflow by Google Cloud, Microsoft Bot Framework, and open-source stacks including Rasa (software). Evaluation and logging infrastructure borrows from observability tools by Datadog, Splunk, and cloud services from AWS, Azure, and Google Cloud Platform. Teams often include researchers affiliated with University of Cambridge, Oxford University, and corporate labs like DeepMind and Google Brain.
Debates involve bias and fairness concerns raised in work by researchers at MIT Media Lab, Fairness, Accountability, and Transparency (FAccT) community, and cases scrutinized by regulators in the European Commission and Federal Trade Commission. Privacy and data protection intersect with laws such as the General Data Protection Regulation (EU) and discussions in United States Congress and courts. Content moderation and misinformation issues link to actions by Twitter (now X), Meta Platforms, and scrutiny in inquiries like United Kingdom Information Commissioner's Office investigations. Liability debates have engaged legal scholars at Harvard Law School, Yale Law School, and policy groups including Electronic Frontier Foundation and OpenAI advisory panels.
Quantitative evaluation draws on benchmarks and metrics created by research groups at Stanford University, University of Washington, and initiatives like GLUE, SuperGLUE, and dialogue-specific suites such as DSTC challenges and the ConvAI series. Metrics include perplexity, BLEU, ROUGE, METEOR, and task success rates used in papers at ACL (conference), EMNLP, and NeurIPS; human evaluation frameworks are developed by labs at Microsoft Research and Facebook AI Research. Safety and robustness testing employ adversarial methods from OpenAI, DeepMind, and independent testing firms.
Active research continues at intersections with multimodal models by OpenAI, Google Research, and Meta AI, integration with robotics labs at MIT CSAIL and Robotics Institute at Carnegie Mellon University, and alignment and safety work led by groups at Future of Humanity Institute and Center for Human-Compatible AI (UC Berkeley). Progress is expected in lifelong learning research at DeepMind, efficient inference via hardware from NVIDIA and Intel, and policy frameworks advanced by OECD and UNESCO.