The conceptual foundations of Artificial Intelligence (AI) emerged in the 1940s, most notably through Alan Turing’s work on computation. In his influential paper Computing Machinery and Intelligence, Turing (1950) introduced the “Imitation Game” (now the Turing Test) as a pragmatic benchmark for determining whether a machine could mimic human intelligence. Rather than offering a strict definition, Turing reframed intelligence as a functional, computational process. The formal inception of AI as a discipline occurred in 1956 at the Dartmouth Conference, organised by John McCarthy, Marvin Minsky, and others. They proposed that “every aspect of learning or any other feature of intelligence can in principle be so precisely described that a machine can be made to simulate it” (McCarthy et al. 2006). This optimistic vision launched a wave of early research into symbolic AI, which sought to replicate human reasoning through logical rules and formal structures.
During the 1960s and 1970s, symbolic AI dominated the field. Systems such as the General Problem Solver (Newell and Simon 1961) demonstrated rule-based reasoning, while ELIZA (Weizenbaum 1966) simulated therapeutic dialogue through pattern matching. These programs illustrated early successes but were limited to narrow tasks and domains. The 1980s brought expert systems like MYCIN, which could diagnose infections using encoded medical knowledge (Buchanan & Shortliffe 1984). However, reliance on hand-crafted rules proved brittle. The field struggled with ambiguity, uncertainty, and scalability, leading to reduced funding and the first so-called “AI winter.” A significant paradigm shift occurred in the 1990s as AI researchers increasingly adopted statistical and data-driven approaches. Machine learning (ML), defined as the ability of systems to improve through experience, gained traction with the development of algorithms like decision trees and backpropagation for training neural networks (Rumelhart et al. 1986). IBM’s Deep Blue famously defeated chess champion Garry Kasparov in 1997, signalling AI’s potential in constrained domains (Campbell et al. 2002). However, generalisation remained a challenge.
The 2010s marked the advent of deep learning, enabled by advances in computational power, big data, and graphics processing units. A key breakthrough came with AlexNet, a deep convolutional neural network that outperformed all competitors in image recognition tasks (Krizhevsky et al. 2012). This catalysed a wave of progress across vision, speech, and natural language processing. Transformer-based models like BERT (Devlin et al. 2019) and GPT-3 (Brown et al. 2020) demonstrated human-like language capabilities. Reinforcement learning also matured, exemplified by AlphaGo’s 2016 victory over world Go champion Lee Sedol (Silver et al. 2016). These breakthroughs not only redefined AI’s capabilities, but also reignited debates about its ethical, epistemological, and societal implications—especially concerning transparency, resource concentration, and long-term risks (Crawford 2021; Marcus 2018).
In conclusion, the evolution of AI embodies more than just a sequence of technological breakthroughs; it represents a shifting understanding of intelligence itself. From Turing’s theoretical propositions to today’s deep learning systems and generative models, the field has progressively redefined what it means for machines to act, learn, and decide. Yet, each advancement brings new ethical, epistemological, and political questions that cannot be resolved through technical solutions alone. As AI systems increasingly mediate human experience, shape institutions, and influence global power dynamics, it is imperative that their development be guided not only by innovation but also by critical reflection, inclusivity, and democratic accountability.
References:
1. Brown, Tom, Benjamin Mann, Nick Ryder, Melanie Subbiah, Jared D. Kaplan, Prafulla Dhariwal, Arvind Neelakantan, et al. 2020. ‘Language Models Are Few-Shot Learners’. Advances in Neural Information Processing Systems 33: 1877–1901. ^ Back
2. Buchanan, Bruce G., and Edward H. Shortliffe. 1984. Rule-Based Expert Systems: The MYCIN Experiments of the Stanford Heuristic Programming Project. Reading, MA: Addison-Wesley Longman Publishing Co. ^ Back
3. Campbell, Murray, A. Joseph Hoane Jr, and Feng-hsiung Hsu. 2002. ‘Deep Blue’. Artificial Intelligence 134 (1–2): 57–83. ^ Back
4. Crawford, Kate. 2021. The Atlas of AI: Power, Politics, and the Planetary Costs of Artificial Intelligence. New Haven: Yale University Press. https://yalebooks.yale.edu/book/9780300209570/the-atlas-of-ai/ – ^ Back
5. Devlin, Jacob, Ming-Wei Chang, Kenton Lee, and Kristina Toutanova. 2019. ‘BERT: Pre-Training of Deep Bidirectional Transformers for Language Understanding’. In Proceedings of the 2019 Conference of the North American Chapter of the Association for Computational Linguistics: Human Language Technologies, 4171–86. ^ Back
6. Krizhevsky, Alex, Ilya Sutskever, and Geoffrey E. Hinton. 2012. ‘ImageNet Classification with Deep Convolutional Neural Networks’. Advances in Neural Information Processing Systems 25. ^ Back
7. Marcus, Gary. 2018. ‘Deep Learning: A Critical Appraisal’. arXiv preprint. https://arxiv.org/abs/1801.00631 – ^ Back
8. McCarthy, John, Marvin L. Minsky, Nathaniel Rochester, and Claude E. Shannon. 2006. ‘A Proposal for the Dartmouth Summer Research Project on Artificial Intelligence, August 31, 1955’. AI Magazine 27 (4): 12–14. ^ Back
9. Newell, Allen, and Herbert A. Simon. 1961. ‘GPS, A Program That Simulates Human Thought’. In H. Billing (ed.), Lernende Automaten, 109–124. München: R. Oldenbourg. PDF link – ^ Back
10. Rumelhart, David E., Geoffrey E. Hinton, and Ronald J. Williams. 1986. ‘Learning Representations by Back-Propagating Errors’. Nature 323 (6088): 533–536. ^ Back
11. Silver, D., Huang, A., Maddison, C.J., Guez, A., Sifre, L., Van Den Driessche, G., Schrittwieser, J., Antonoglou, I., Panneershelvam, V., Lanctot, M., and Dieleman, S. 2016. ‘Mastering the Game of Go with Deep Neural Networks and Tree Search’. Nature 529 (7587): 484–489. https://doi.org/10.1038/nature16961 – ^ Back
12. Turing, Alan M. 1950. ‘Computing Machinery and Intelligence’. Mind 59 (236): 433–460. ^ Back
13. Weizenbaum, Joseph. 1966. ‘ELIZA—A Computer Program for the Study of Natural Language Communication Between Man and Machine’. Communications of the ACM 9 (1): 36–45. https://doi.org/10.1145/365153.365168 – ^ Back