Last Updated: November, 2023
Introduction
Artificial Intelligence (AI) is a branch of computer science that aims to create machines that mimic human intelligence. This didn’t originally mean creating machines that can think for themselves, but rather, creating machines that can be programmed to think or learn in a way that appears intelligent. However, there seems to be very little, or potentially absolutely no difference at all between “actual intelligence” (whatever that means) and “artificial intelligence” - more on this later.
The language I’m using could be seen as anthropomorphising, however, these models are reflections of some ways the human brain works and learns, and I believe the analogies using human-specific language to be well-accurate enough.
Why Does It Work?
The “intelligence” and “reasoning” abilities of any given AI system are not binary - AI is not either intelligent
or not intelligent
- these reasoning abilities exist on a sliding scale, similarly to humans. There is only so much reasoning a particular model can perform, and only so much the model can “remember” (or juggle in its “mind”) at any given time. An AI system’s intelligence
and working memory size
seem to be the constraints on AI ability, however, this is still very high-level, and still hasn’t really answered the question. Funnily enough, we are once again asking "what is intelligence?"
Technically, all these models do is predict the next word (actually “token” instead of “word”, but that’s not too important a distinction at the moment).
Clearly, it seems that by instructing an AI to learn how to effectively model language, it must also learn how to effectively model intelligence itself to be able to excel at modelling language. Stemming from “only” an advanced text prediction algorithm, there is a veritable emergent “ghost-in-the-machine” with a surprisingly organic and authentic display of robust artificial intelligence.
What's On Offer
There's a few conversational LLMs available, with varying usability, function, intended purpose, and ability. Here are a few of the more well-known ones, with some added polemic on my part. This list is in no way comprehensive, and there's a lot more to be said about open-source LLMs, but I digress:
-
ChatGPT-4 (OpenAI)
- Currently, the best option. The UI and general usability of the product is second to none, and majorly helps in squeezing as much productivity out of GPT-4 as much as possible.
- Advanced text generation, image input capabilities, extensive knowledge base up to its last training cut-off in 2023, and improved safety features.
- Applications: Content creation, programming help, educational tools, customer service bots.
- Accessible through OpenAI API, ChatGPT Plus subscription for enhanced features.
-
BLOOM (Hugging Face) - Open Source
- Multilingual capabilities across 46 languages and 13 programming languages, designed for autoregressive text generation with a focus on accessibility and transparency.
- Applications: Text completion, language translation, content generation across multiple languages.
- Free access via Hugging Face, with open-source code and training data available for public use.
-
Gemini (Google)
- Incorporates DeepMind's research for a conversational interaction, supporting multiple languages and offering detailed responses.
- Applications: Web search, content generation, and offering detailed answers to complex queries.
- Accessible by visiting bard.google.com and signing in with a Google account.
-
Bing Chat (Microsoft)
- Utilises OpenAI's GPT-4, offering more up-to-date responses and integration within the Bing search engine for a seamless user experience.
- Applications: Search, chat, content creation, and image generation with more current event information.
- Integrated with Microsoft Edge and available on bing.com for users with a Microsoft account.
-
Claude 3 (Anthropic)
- Emphasises safety and transparency with enhanced recall abilities, capable of processing inputs exceeding 1 million tokens and producing trustworthy responses.
- Applications: Complex task automation, research review, strategy development, and customer-facing experiences.
- Offered in multiple versions like Claude 3 Opus and Sonnet, with specific use cases and cost structures detailed by Anthropic.
The First Pitfall
At some point in the future, I’ll collate more information on open-source LLMs. However, for the time being, I’ll mainly be writing about usage of ChatGPT and GPT, specifically GPT4.
The first pitfall would be not evaluating ChatGPT-Plus, the USD$20/month subscription to ChatGPT. The free version of ChatGPT (ChatGPT-3.5) is a golf buggy – impressive, if before you could only walk or run. The paid version of ChatGPT (ChatGPT-4) is a Ferrari. This is for two reasons:
-
ChatGPT-Free is based only on GPT-3.5, a fine model for understanding text and performing basic intelligent tasks on texts. ChatGPT-Plus, however, is based on GPT-4, a much more recent, and much, much more powerful model. Not only can it understand text and language, but it excels at reasoning, problem solving, intuition, and creativity - it’s much more able to do what you might need it to do. I generally put it this way:
GPT-3.5 could write a single python function at about a 65% success rate – if you work with it to fix the errors, you could probably generally get it to successfully create a python function.
GPT-4 could write a whole python program with about a 95% success rate. If errors are encountered, simply copying and pasting them back into the chat window will generally bump that up to 100%.
-
ChatGPT-Plus allows access to the hugely expanded User Interface. Aside from normal text output (called “inference”), it can also autonomously:
- write and run python code
- browse the internet
- perform proficient computer vision, including recognising and reading text in images
- upload and download files to/from the chat window
- generate visuals and art using Dalle-3
For example, one could simply drag and drop a text file like this one (a dataset from the Australian Bureau of Statistics detailing employment trends across Australian states, by time and by company size) into ChatGPT4, and ask for some interpretation to be done (in this case, calculate the difference between Victoria and the Australian mean) and for some graphs made. It writes some python while I step away to make a coffee, returning coffee in hand to this:

I'll eventually come back to this page to write more, but for now I'll leave it here. Be sure to check out part 2.