Editing
History of Artificial Intelligence
(section)
Jump to navigation
Jump to search
Warning:
You are not logged in. Your IP address will be publicly visible if you make any edits. If you
log in
or
create an account
, your edits will be attributed to your username, along with other benefits.
Anti-spam check. Do
not
fill this in!
== <span style="color: #FFFFFF;">Understanding</span> == AI history is best understood as a series of **paradigm shifts** driven by new ideas, new data, and new compute: **1950sβ1960s: Symbolic optimism**. McCarthy, Minsky, and colleagues believed AI was imminent β general intelligence via symbolic logic and search. Programs solved toy problems impressively, leading to overconfident predictions. The reality: logic-based systems couldn't handle real-world ambiguity, noise, or scale. **1970sβ1980s: First winter and expert systems**. After failures to deliver on promises, funding dried up. Expert systems revived interest by encoding specialized knowledge in rule bases β MYCIN (medical diagnosis), XCON (computer configuration). These worked for narrow domains but were brittle, expensive to maintain, and couldn't learn from data. **1990sβ2000s: Statistical machine learning**. The machine learning community, drawing from statistics, proved that data-driven pattern recognition could outperform hand-coded rules for many tasks. SVMs, decision trees, random forests, and boosting algorithms dominated. Feature engineering β manually designing input representations β was the key differentiating skill. **2012βpresent: The deep learning revolution**. AlexNet's breakthrough demonstrated that deep convolutional networks trained end-to-end on GPU clusters could outperform decades of hand-engineered vision systems. This triggered a cascade: ImageNet β deep learning β transformers (2017) β BERT (2018) β GPT-3 (2020) β ChatGPT (2022) β GPT-4 (2023). Each step moved AI from narrow, task-specific systems toward more general capabilities. **What enabled the revolution?**: Three simultaneous improvements: (1) **Compute**: GPU clusters, then TPUs, enabled training at scales previously impossible. (2) **Data**: the internet created unprecedented amounts of labeled and unlabeled data. (3) **Algorithms**: backpropagation + rectified linear units + better initialization made deep networks trainable. </div> <div style="background-color: #8B0000; color: #FFFFFF; padding: 20px; border-radius: 8px; margin-bottom: 15px;">
Summary:
Please note that all contributions to BloomWiki may be edited, altered, or removed by other contributors. If you do not want your writing to be edited mercilessly, then do not submit it here.
You are also promising us that you wrote this yourself, or copied it from a public domain or similar free resource (see
BloomWiki:Copyrights
for details).
Do not submit copyrighted work without permission!
Cancel
Editing help
(opens in new window)
Navigation menu
Personal tools
Not logged in
Talk
Contributions
Create account
Log in
Namespaces
Page
Discussion
English
Views
Read
Edit
View history
More
Search
Navigation
Main page
Recent changes
Random page
Help about MediaWiki
Tools
What links here
Related changes
Special pages
Page information