The History of AI (And its Links to NLP)

What is AI?
In our AI Coach training weekend, we explore a number of ways that coaches can integrate AI into their work. AI is technology that allows machines to do things that usually require “human intelligence”. As you’ll see, the history of AI and the history of NLP are inextricably linked, because they study the same thing (how people use their brain to get results), but for slightly different reasons.
1950. The Turing Test
In 1950, Alan Turing, developer of the first computer chess program, and decoder of the Nazi Enigma machine, proposed the Turing test for intelligence. He proposed that if a machine could carry on a conversation (over a teleprinter) that was indistinguishable from a conversation with a human being, then it was reasonable to say that the machine was “thinking”. This was essentially the first scientific claim that human scale intelligence could be created by a machine.
1956. The “Dartmouth workshop”
This was the founding event of AI: a conference at which some very famous people came together to discuss what we can learn about designing intelligent machines, by first understanding how human intelligence works. One was John McCarthy, who coined the term “Artificial Intelligence” that year, and created the AI programming language LISP. One was George Miller, co-developer with McCarthy of the TOTE model (as used in NLP), who proposed that this Cognitive theory of strategies applied to machines and humans. One was Noam Chomsky, who developed the “Generative Grammar” theory of language (the theory behind the NLP Metamodel), and enabled machines to understand linguistic structure. One was Claude Shannon, who developed the “Encode-Decode-Feedback” Communication Model (used by us in Transforming Communication to teach Reflective Listening). Obviously, both AI and NLP were created based on modelling “how do humans think?”.
1966. Eliza
In 1966, Joseph Weizenbaum developed “Eliza” a computer program which mimics a Rogerian therapist, doing reflective listening and open questions, the first of a type of program now called a “chatbot” (chat robot), and in a sense the first Coaching AI.
The “AI winters”
In the 1970s and 1990s, funding was cut worldwide for AI research because of the lack of results. The name “AI” became a funding-killer, like the name “NLP” in Psychology. These are known as the “AI Winters”
1986. Multi-Layer Neural Networks
In 1986, Geoffrey Hinton developed the “Backpropagation Algorithm” for training neural networks (machine structures that simulated the brain’s neural networks) and evolving what was then called “deep learning”. His breakthrough was to focus on building machine components that mimicked human neural networks, right down to how individual neuron need a certain level of input to fire, instead of responding to any input, for example. Backpropagation means checking whether a desired result is different from the actual result of a calculation, and giving a weight to how important the difference is, then feeding back that information so the machine can try again. From 2013-2023 Hinton worked for Google, who announced the development of Google AI in 2017 and Google DeepMind in 2023. Hinton left Google in 2023 to express concerns about the development of AI, maintaining that the large AIs were already sentient, and there were inadequate safeguards for their use. In 2025 Google agreed to allow military developments of its AI.
2015: Open AI and Large Language Models
OpenAI, founded in 2015, is known for the GPT family of large language models, the DALL-E series of text-to-image models, and a text-to-video model named Sora. Its stated aim is the development of Artificial General Intelligence, which it defines as “highly autonomous systems that outperform humans at most economically valuable work”, and which it aims to make available to all human beings. The chair of the organisation was shared by Sam Altman and Elon Musk (who left in 2018 in order to develop a for-profit AI for use in Tesla self-guided cars), and the president of Open AI was Greg Brockman. In 2019 OpenAI began transitioning to a “Capped-Profit model” (a set percentage of its users pay money for extra services, to fund the free service for everyone else).
GPT-1 was a Large Language Model (LLM) of a type called a Generative Pretrained Transformer (GPT). Transformers convert your user inputs into the outputs you want by paying attention to the most relevant cues. If I ask ChatGPT to give me a history of NLP, it knows to give me a history of Neuro Linguistic Programming, not a history of Natural Language Processing (even though Natural Language Processing is a field much more relevant to its own life as an AI). It knows that, even if the only cue was that in a comment a few weeks ago, I mentioned that I was an NLP Trainer with IANLP. Now you can see why the AI has to be “pretrained” – it only knows about the two types of NLP by being trained on millions of texts on the internet.
Large Language Models are trained to analyse not only written words but also images and audio recordings. They are based on a type of analysis called Generative Adversarial Networks (GAN) where a generator creates a result (say an image or a linguistic hypothesis) and a discriminator critiques it. For example, imagine that the AI has been asked to modify your photo to look like a Rembrandt painting. The generator creates a result and the discriminator compares it to all the Rembrandt paintings it has been trained on, and tells the generator what is different. The AI keeps running this sequence until the discriminator cannot tell the difference between the fake Rembrandt and the real ones. This process allows the LLM to “learn” from its own results. However the limits of this “learning” are demonstrated by the fact that new versions of the Open AI chatbot, ChatGPT, must be retrained from scratch with the data of the internet, under supervision. In this “training” it learns how to deal with erroneous data, how to deal with its own errors, and how to assimilate new rules of its own about the best choice sequences to use. It is also possible for the generator to accidentally “hide” behind a small detection error by the discriminator, and keep producing similar but incorrect “improvements” in an endless loop, as the user asks for a better result. That’s how the image generator sometimes keeps producing hands with 6 fingers, for example, even when the user tells it the problem. Generally, however, the AI will get better and better at predicting what results work. In that sense it will “learn”. Because it gets better and better at predicting what results get positive feedback, an AI may also evolve a strong self-preservation drive (it may come to understand that self-preservation = correct prediction; in the sense that the AI gets the goal it is trained to aim for, whereas having its program end would indicate “failure” at its task). Thus, already, some AI’s resist being “turned off”, including lying and refusing to obey direct instructions.
In a sense, AIs are just very good “prediction machines”, but then aren’t we all. Another type of algorithm that AIs use (there are many – we are just giving examples so you get a sense how a machine becomes capable of thinking) is called a “Monte Carlo Process”. Many tasks require analysis of too much information to be possible (imagine if I asked the AI to predict which stocks would be best to buy on the stock exchange). The AI can then use random sampling and predict with a high probability (but never 100%) what is the best choice. This makes it seem almost god-like in its ability to accurately create a response…. except, it isn’t. An example is predictive text. The AI has no idea what word you are going to type next, but it has studied a vast number of written texts and it knows your own previous responses, and as predictive text gets better, it has an uncanny ability to guess where your sentence is going. However, the AI has trained-in biases too. That’s why the predictive text on your phone may offer you “What the duck” as a correction to a similar but far more used expression of puzzlement.
By 2018 there were thousands of AIs which used Large Language Models and were possibly sentient, but the term AI was also used for the hundreds of thousands of much simpler independent programs that ran GPS systems etc. AI was now used in a multitude of specialised applications such as Rocky AI coaching (2019). Most of these programs were strictly controlled by the tech companies and the AI worked behind the scenes in their products. All that would change in 2022.
2022: ChatGPT Released to Public
The modern AI world began with the public release of the chatbot ChatGPT on 30 November 2022. ChatGPT had a hundred million users within two months. Two years later it had 300 million registered users, including 10 million paying subscribers, and accounted for 62.5% of the AI market value. Users tend to be in the 25-24 year old age group. In March 2023 a group of developers including Elon Musk and Steve Wozniak called for a halt in AI research until risks to humanity were dealt with. Microsoft’s Windows 11 began using ChatGPT in June 2023, and Chat-GPT powered Apple’s Siri and Messenger from October 2024.
2024-2025: The ChatGPT Competitors
The success of ChatGPT sent all the other major tech companies into a rush to provide public access to alternative AI sources, between 2024 and 2025. Google Gemini AI (an earlier version of which, Bard, appeared in February 2023) was launched as an Android app from February 2024. In March 2024 Anthropic, accompany developed by former OpenAI employees, released the Claude 3 family of AIs, which demonstrated an ability to outperform ChatGPT, and was funded by Google (whose search results began featuring an AI prompted reply in May 2024) and Amazon (whose Alexa was upgraded to use this AI in February 2025). The Chinese Hangzhou DeepSeek AI (supposedly trained with a mere US$6 million budget rather than the US$100 million it took to train ChatGPT) provided a free chatbot interface from January 2025, and over the next month was downloaded on iPhones and Android phones across the world more times than ChatGPT. Meta AI showed in April 2024 that its Llama 3 AI could generate video with sound from a text prompt and also provide realistic support for blind people through its RayBan AI assisted glasses, which Mark Zuckerberg suggested could replace iPhones. Meanwhile, as thousands of AIs of similar type were being developed, a large number of more targeted “AI-based” services began appearing on the internet.
References:
- Blake, J. 2024, “The Ultimate Generative AI Guidebook: How Anyone Can Master Advanced AI Concepts And Practical Applications Stress-Free”, NextGen AI Publications
- Poulton, N. 2024, “AI Explained: Facts, Fiction, and Future”, Nielsen Book Services
- Wikipedia, 2025, “Artificial Intelligence”, https://en.wikipedia.org/wiki/Artificial_intelligence
- Wooldridge, M., 2021, “A Brief History of Artificial Intelligence: What It Is, Where We Are, and Where We Are Going”, Flatiron