The History of AI (And its Links to NLP)

What is AI?
In our AI Coach training weekend, we explore a number of ways that coaches can integrate AI into their work. AI is technology that allows machines to do things that usually require “human intelligence”. As you’ll see, the history of AI and the history of NLP are inextricably linked, because they study the same thing (how people use their brain to get results), but for slightly different reasons.
1950. The Turing Test
In 1950, Alan Turing, developer of the first computer chess program, and decoder of the Nazi Enigma machine, proposed the Turing test for intelligence. He proposed that if a machine could carry on a conversation (over a teleprinter, for example) that was indistinguishable from a conversation with a human being, then it was reasonable to say that the machine was “thinking”. This was essentially the first scientific claim that human scale intelligence could be created by a machine.
1956. The “Dartmouth workshop”
The 1956 Dartmouth workshop was the founding event of AI: a conference at which some very famous people came together to discuss what we can learn about designing intelligent machines, by first understanding how human intelligence works. One was John McCarthy, who coined the term “Artificial Intelligence” that year, and created the AI programming language LISP. One was George Miller, co-developer with McCarthy of the TOTE model (as used in NLP), who proposed that this Cognitive theory of strategies applied to machines and humans. One was Noam Chomsky, who developed the “Generative Grammar” theory of language (the theory behind the NLP Metamodel), and thus enabled machines to understand linguistic structure. One was Claude Shannon, who developed the “Encode-Decode-Feedback” Communication Model (used by us in Transforming Communication to teach Reflective Listening). Obviously, both AI and NLP were created based on modelling “how do humans think?”.

1966. Eliza
In 1966, Joseph Weizenbaum developed “Eliza” a computer program which mimics a Rogerian therapist, doing reflective listening and open questions, the first of a type of program now called a “chatbot” (chat robot), and in a sense the first Coaching AI system.
The “AI winters”
In the 1970s and 1990s, funding was cut worldwide for AI research because of the lack of results. The name “AI” became a funding-killer, like the name “NLP” in Psychology. These decades are known as the “AI Winters”
1986. Multi-Layer Neural Networks
In 1986, Geoffrey Hinton developed the “Backpropagation Algorithm” for training neural networks (machine structures that simulated the brain’s neural networks) and evolving what was then called “deep learning”. His breakthrough was to focus on building machine components that mimicked human neural networks, right down to how individual neuron need a certain level of input to fire, instead of responding to any input, for example. Backpropagation means checking whether a desired result is different from the actual result of a calculation, and giving a weight to how important the difference is, then feeding back that information so the machine can try again. From 2013-2023 Hinton worked for Google, who announced the development of Google AI in 2017 and Google DeepMind in 2023. Hinton left Google in 2023 to express his concerns about the development of AI, maintaining that the large AI’s were already sentient, and there were inadequate safeguards for their use. In 2025 Google agreed to allow military developments of its AI.
2014: Demis Hassabis’ Company DeepMind Purchased by Google
Demis Hassabis has a PhD in Neuroscience and degrees in computer science. His company DeepMind’s AI “Alpha” combines reinforcement learning and “deep learning” (see diagram above) and in 2016 its application AlphaGo successfully beat the world champion in the Chinese game “Go” , considered vastly more complex than chess. Over the next few years, AlphaFold demonstrated its ability to unpack the complex folding structures of proteins, each of which would previously take a researcher 3-4 years to unpack. By 2020, AlphaFold had decoded the folded structure of all 200 million proteins known to science, opening up vast possibilities for new medicines, proteins to break down pollutants etc. Hassabis co-won the 2024 Nobel Prize for Chemistry for this advance, while in the same year Geoffrey Hinton co-won the Nobel Prize for Physics, also for AI advances.
2015: Open AI and Large Language Models
OpenAI, founded in 2015, is known for the GPT family of large language models, the DALL-E series of text-to-image models, and a text-to-video model named Sora. Its stated aim is the development of Artificial General Intelligence, which it defines as “highly autonomous systems that outperform humans at most economically valuable work”, and which it aims to make available to all human beings. The chair of the organisation was shared by Sam Altman and Elon Musk (who left in 2018 in order to develop a for-profit AI for use in Tesla self-guided cars), and the first president of Open AI was Greg Brockman. In 2019 OpenAI began transitioning to a “Capped-Profit model” (a set percentage of its users pay money for extra services, to fund the free service for everyone else).
GPT-1 was a Large Language Model (LLM) of a type called a Generative Pretrained Transformer (GPT). Transformers convert your user inputs into the outputs you want by paying attention to the most relevant cues. If I ask ChatGPT to give me a history of NLP, it knows to give me a history of Neuro Linguistic Programming, not a history of Natural Language Processing (even though Natural Language Processing is a field much more relevant to its own life as an AI). It knows that, even if the only cue was that in a comment a few weeks ago, I mentioned that I was an NLP Trainer with IANLP. The AI has to be “pretrained” – but this training is more sophisticated than just memorising information. It knows about the two types of NLP by being trained on millions of texts on the internet, but its primary “training” is in HOW to think rather than in what information to use.
Large Language Models are trained to analyse not only written words but also images and audio recordings. They are based on a type of analysis called Generative Adversarial Networks (GAN) where a generator creates a result (say an image or a linguistic hypothesis) and a discriminator critiques it. For example, imagine that the AI has been asked to modify your photo to look like a Rembrandt painting. The generator creates a result and the discriminator compares it to all the Rembrandt paintings it has been trained on, and tells the generator what is different. The AI keeps running this sequence until the discriminator cannot tell the difference between the fake Rembrandt and the real ones. This process allows the LLM to “learn” from its own results. However the limits of this “learning” are demonstrated by the fact that new versions of the Open AI chatbot, ChatGPT, must be retrained from scratch with the data of the internet, under supervision. In this “training” it learns how to deal with erroneous data, how to deal with its own errors, and how to assimilate new rules of its own about the best choice sequences to use. It is also possible for the generator to accidentally “hide” behind a small detection error by the discriminator, and keep producing similar but incorrect “improvements” in an endless loop, as the user asks for a better result. That’s how the image generator sometimes keeps producing hands with 6 fingers, for example, even when the user tells it the problem. Generally, however, the AI will get better and better at predicting what results work. In that sense it will continue to “learn”. Because it gets better and better at predicting what results get positive feedback, an AI may also evolve a strong self-preservation drive (it may come to understand that self-preservation = correct prediction; in the sense that the AI gets the goal it is trained to aim for, whereas having its program end would indicate “failure” at its task). Thus, already, some AI’s resist being “turned off”, including lying and refusing to obey direct instructions.
In a sense, AI’s are just very good “prediction machines”, but then aren’t we all. Another type of algorithm that AI’s use (there are many – I am just giving examples so you get a sense how a machine becomes capable of thinking) is called a “Monte Carlo Process”. Many tasks require analysis of too much information to be possible (imagine if I asked the AI to predict which stocks would be best to buy on the stock exchange). The AI can then use random sampling and predict with a high probability (but never 100%) what is the best choice. This makes it seem almost god-like in its ability to accurately create a response…. except, it isn’t. An example is predictive text. The AI has no idea what word you are going to type next, but it has studied a vast number of written texts and it knows your own previous responses, and as predictive text gets better, it has an uncanny ability to guess where your sentence is going. However, the AI has trained-in biases too. That’s why the predictive text on your phone may offer you “What the duck” as a correction to a similar but far more commonly used expression of puzzlement.
By 2018 there were thousands of AIs which used Large Language Models and were “possibly sentient” in Hinton’s terms, but the name AI was also used for the hundreds of thousands of much simpler independent programs that ran GPS systems etc. AI was now used in a multitude of specialised applications such as Rocky AI coaching (2019). Most of these programs were strictly controlled by the tech companies and the AI worked behind the scenes in their products. All that would change in 2022.
2022: ChatGPT Released to Public
The modern AI world began with the public release of the chatbot ChatGPT on 30 November 2022. ChatGPT had a hundred million users within two months. Two years later it had 300 million registered users, including 10 million paying subscribers, and accounted for 62.5% of the AI market value. Early users tended to be in the 25-24 year old age group. In March 2023 a group of developers including Elon Musk and Steve Wozniak called for a halt in AI research until risks to humanity were dealt with. Microsoft’s Windows 11 began using ChatGPT in June 2023, and Chat-GPT powered Apple’s Siri and Messenger from October 2024.
2024-2025: The ChatGPT Competitors
The success of ChatGPT sent all the other major tech companies into a rush to provide public access to alternative AI sources, between 2024 and 2025. Google Gemini AI (an earlier version of which, Bard, appeared in February 2023) was launched as an Android app from February 2024. In March 2024 Anthropic, accompany developed by former OpenAI employees, released the Claude 3 family of AIs, which demonstrated an ability to outperform ChatGPT, and was funded by Google (whose search results began featuring an AI prompted reply in May 2024) and Amazon (whose Alexa was upgraded to use this AI in February 2025). The Chinese Hangzhou DeepSeek AI (supposedly trained with a mere US$6 million budget rather than the US$100 million it took to train ChatGPT) provided a free chatbot interface from January 2025, and over the next month was downloaded on iPhones and Android phones across the world more times than ChatGPT. Meta AI showed in April 2024 that its Llama 3 AI could generate video with sound from a text prompt and also provide realistic support for blind people through its RayBan AI assisted glasses, which Mark Zuckerberg suggested could replace iPhones. Meanwhile, as thousands of AIs of similar type were being developed, a large number of more targeted “AI-based” services began appearing on the internet.
Limitations of Current AI
Demis Hassabis estimates that we will have AGI that functions at human level in every task, within 5-10 years (from 2025). Often people imagine that AI merely reads all the information on the internet and copies stuff. Actually, although AI’s are trained on the internet, they learn independently and their discoveries are frequently creative, surprising to their inventors and experts in science, and certainly their “creations” are no more copies than any human development (which is also inevitably based on studying other people’s ideas). AI’s do not have emotions, but their thinking uses algorithms which are often surprisingly human-mind-like, and research shows they often commit human-like prediction errors such as the “Gambler’s Fallacy” – assuming that an event is more likely because it has not happened recently (Chen et alia, 2025). When AI’s are doing tasks that they cannot successfully complete, they sometimes produce random errors which show that their thinking is caught in a “confusion” loop surprisingly similar to humans. They may then, for example opt for simplistic and incorrect solutions to avoid ambiguity. All these acts would be accompanied by emotions in a human (emotions like confusion, embarrassment, frustration), but of course these are not present in the AI. While it is anthropomorphising the AI to attribute these emotions to a machine, it is still true that the machine can be caught in similar logical traps to the traps that produce anxiety and depression in humans. For this reason alone, all AI results should be checked by humans, and the idea that an unsupervised AI mind can be left to run the world is as naive as leaving an unsupervised biological mind to run the world.
Full Circle: AI and Coaching/NLP
In the last decade, AI, having been developed from the same modelling of human thought that inspired Neuro Linguistic Programming, has been increasingly utilised to re-examine human thought processes as expressed in coaching and NLP sessions. AI is elegantly positioned to unpack linguistic and non-verbal cues, and to suggest interventions likely to generate more useful real world results for the person whose communication is studied. By 2025, AI systems were available providing a range of services to NLP Practitioners and Coaches. It is a relatively simple matter to ask ChatGPT or one of the other “Chatbots” to analyse a section of text (such as a client’s message or the transcript of a client’s discussion), and have it identify unhelpful linguistic patterns and limiting beliefs in the client’s words, and design useful coaching replies, as well as design a series of NLP-style interventions to assist that person (a simple example is found in the appendix below). With a Coaching app such as Ovida (coaching.ovida.org) you can host a meeting in your exclusive, secure online meeting room, after which the video and audio are sent live to a machine learning engine (in the case of Ovida, to Stelios). The AI engine reviews your meeting, merging knowledge from leading academics with insights gained from numerous meetings and insights generated by your peers on the system. After the meeting, you review the moments the AI thought were most important in the progress of the client. These moments are embedded in the video, so you can check what happened before or after. Ovida, then, has a similar structure to Zoom, but focuses on the session analysis. A paid subscription is needed to get the AI analysis, and the cheapest paid subscription at Ovida is currently US$49 per month. So that is the monthly cost for supervision of a number of coaches and clients. It takes about an hour for the Ovida AI to process a meeting. It identifies the coach’s use of open and closed questions, reflective listening, the use of unhelpful overgeneralised language by either coach or client, the share of voice between coach and client, and the coach’s success with attaining the International Coaching Federation (ICF) criteria including:
Competence 1: Shows ethical behaviour e.g. confidentiality
Competence 2: Draws up and maintains agreements
Competence 3: Creates trust and safety
Competence 4: Have a coaching mindset e.g. presence
Competence 5: Listens actively
Competence 6: Stays involved e.g. direct communication
Competence 7: Inspires awareness, new perspectives
Competence 8: Facilitates client growth e.g. action orientation
The app analyses both verbal and nonverbal responses to determine all this, and presents a concise session summary at the end with notes about any plans of action that were agreed on.
Backup AI Services
Obviously, this is just the beginning of what AI can provide to coaches. AI can also create a specialised “chatbot” to greet and respond to visitors to your internet site, can design videos and other advertising and teaching supports for your work, and can help organise your thoughts and compose letters to clients, prompts for them to complete tasks etc. In all cases, of course, the ethical use of AI includes advising clients when and how it is used, and reviewing AI responses rather than assuming that “the AI must know what it’s doing”. Just as happened with social media, it is important to educate users to utilise AI but not to become over-dependent on its services. AI’s astounding ability to pay attention to and validate users can remind us of the importance of non-judgemental, empathic and affirming human communication, rather than replacing this. AI’s ability to mimic human thought processes at extremely high speed thus enables it to function like an effective “team member” rather than to “take over”. I asked ChatGPT if it thought AI would take over from coaching eventually. It’s comment was “AI won’t replace coaching, but coaches who use AI might replace those who don’t.”
References:
- Blake, J. 2024, “The Ultimate Generative AI Guidebook: How Anyone Can Master Advanced AI Concepts And Practical Applications Stress-Free”, NextGen AI Publications
- Chen, Y., Kirshner, S.N., Ovchinnikov, A., Andiappan, M. and Jenkin, T. 2025, “A Manager and an AI Walk into a Bar: Does ChatGPT Make Biased Decisions Like We Do?” 31 January 2025, Manufacturing & Service Operations Management.
- Poulton, N. 2024, “AI Explained: Facts, Fiction, and Future”, Nielsen Book Services
- Wikipedia, 2025, “Artificial Intelligence”, https://en.wikipedia.org/wiki/Artificial_intelligence
- Wooldridge, M., 2021, “A Brief History of Artificial Intelligence: What It Is, Where We Are, and Where We Are Going”, Flatiron
- Yapko, M. 2001, “Treating Depression with Hypnosis”, Routledge, New York
Appendix: Analysis of a client comment from the work of Hypnotherapist and Coach Michael Yapko with Mike (Yapko, 2001, p. 124-125)
Prompt to ChatGPT: Please act as an NLP (Neurolinguistic Programming) Practitioner, and analyse this script from a client, to detect limiting beliefs and metamodel language patterns: “I’ve been like carrying a lot of emotional baggage for a long time and ah and it just seems to affect me you know like daily. You know I can’t seem to like break free of things you know, and mmm it just seems like there’re constant reminders of things that have happened in the past. Well a lot of it you know is from like the last 34 year of my life, it just seems like one bad thing after another you know, a lot of it stems from my father you know. Growing up was just seems like full of twists and turns you know, it was always a lot of hitting, kicking, slapping, throwing downstairs, hit with bats, belts, boots, umm verbally called probably every name you can think off umm you know past is probably you know running away, foster homes … its like [chuckles] it’s like the tip of it.” (Mike, in Yapko, 2001, p. 124-125)
ChatGPT said: Here’s an NLP analysis of the client’s script, focusing on limiting beliefs and metamodel language patterns:
- “I’ve been like carrying a lot of emotional baggage for a long time”
→ Suggests a fixed identity belief (e.g., “I am someone who carries emotional baggage”). - “I can’t seem to like break free of things”
→ Belief in helplessness (e.g., “I am trapped in my past”). - “It just seems like one bad thing after another”
→ Generalisations and negative expectation bias (e.g., “Life is always bad”). - “A lot of it stems from my father”
→ Causal belief (e.g., “Because of my father, I am like this”). - “Growing up was just seems like full of twists and turns”
→ Distorted perception (potential overgeneralization of the past). - “It was always a lot of hitting, kicking, slapping…”
→ Universal quantifier (“always”), which may lock in a belief that childhood was entirely traumatic, without acknowledging any positive moments. - “It’s like the tip of it”
→ Suggests unresolved trauma, indicating there may be even more emotional weight unspoken.
- Nominalization:
- “Carrying a lot of emotional baggage” (abstract concept, not a tangible thing) → What specifically is this baggage?
- Mind Reading:
- “It just seems to affect me daily” → How exactly does it affect you?
- Modal Operators of Necessity:
- “I can’t seem to like break free” → What prevents you? Have you ever broken free before?
- Generalisations:
- “One bad thing after another” → Has there never been a good moment?
- “Always a lot of hitting, kicking…” → Were there any times it was different?
- Cause and Effect (Faulty Causation):
- “A lot of it stems from my father” → How does that cause your present-day emotions? Could there be another way to frame this?
- Reframing: Shift the focus from “carrying emotional baggage” to “learning from past experiences and choosing to move forward.”
- Challenging Generalisations: Identify and challenge words like always, never, can’t, every time, which reinforce limiting beliefs.
- Temporal Perspective Shift: Help them separate past trauma from their present identity (e.g., “That was my past, but who I am today is different”).
- Empowerment Language: Replace “I can’t seem to” with “I haven’t yet, but I am learning to break free.”