There are a bunch of very common mistakes and misunderstandings around AI in general, and LLM based AI such as ChatGPT and Bard in particular. When I say ‘very common’ I mean that in terms of statistical likelihood, most members of this community will have at least one of these misunderstandings, and most will have several.
LLM stands for Large Language Model. It is literally a model of how human language works based on a very large sampling. It models the way that language is used. Nothing else. It may seem to demonstrate logic, reasoning, fine judgement, analysis, etc. but that is only because it is predicting the language that would do so. It doesn’t really understand what it writes the way you or I would.
It just predicts the language patterns and words, without understanding the concepts beneath those words.
When we call this AI, remember that it is the Intelligence that is Artificial, as in not real. It seems intelligent. We are still an unknown distance from AGI (Artificial General Intelligence) which is where we artificially make something that is generally intelligent. LLMs alone will not get us to AGI, ever.
When you prompt an LLM-based AI for something that requires logic, or creativity, or understanding, or any kind of analytical skills, it will predict what an answer would look like and generate it. What it won’t do is actually use logic, creativity, or understanding to do so. It simply may correctly predict an answer that did so. This is part of the artificiality of AI. It is a workaround, a hack, to seem intelligent, without actually having to wait unknown years for AGI to become real.
If you ask an LLM based AI to generate ideas for you, it will predict and generate the most likely response. That means it will favour the most published, most used, most thought of ideas. The exact opposite of what you probably intended from such a prompt.
If you ask an LLM-based AI to rate your content on any kind of criteria, whether that be prompting it to analyse and rate the SEO, or the best code, or whatever else, it will again predict the most likely response and generate it. It bases this on its training data, which included Redditt threads, and the Common Crawl. There are few experts and many, many fools in the word, and while hopefully the right answers get used more often than any one wrong answer, the AI cannot differentiate, and just gives you the most predictable response. By design. That is what it is made and built to do.
Current levels of AI are not super-intelligent, nor even actually just very-average human intelligent. There are drunken homeless people in the park that have more actual intelligence than any currently available AI does, (but may not be as widely read, or as specifically trained to generate predictable patterns of words).
AI are systems that appear to be artificially intelligent. They make critical mistakes, often, that by design may appear to be reasonable and fair responses, confidently given, and completely wrong. Every manufacturer and company running AI clearly warn you of this, as they legally must. AI generated responses must be checked and verified with human judgement, or they are just pretty patterns of words.
Just in case you have any doubts at all about the accuracy of what I have said, and you trust ChatGPT, well, then here is ChatGPT telling you the same thing in its own words: LLM AI Limitations