LLM Lingo – Terms to know before talking to LLM professionals

Share it with your senior IT friends and colleagues
Reading Time: 3 minutes

With ChatGPT, Large Language Models became the talk of the town. At the same time, there are many terms (LLM Lingo/ LLM Slangs) which become common.

However, this “LLM Lingo” can be overwhelming/ confusing for many non-AI professionals. This is an attempt to define and simplify these LLM terminologies so that we do not feel out of place when we talk to LLM professionals.

Here is the LLM Lingo/ LLM Slang to know before talking to LLM professionals

Large Language Model

Large Language Model aka LLM is an AI model trained on huge text data. For example – ChatGPT.

ChatGPT is the brand name whereas the actual product or technology is LLM.

OpenAI’s LLM is the GPT series, with the latest model being GPT4 and the most renowned model ChatGPT – GPT 3.5.

Learn more about LLMs – https://www.aimletc.com/free-course-introduction-to-large-language-models/

Foundational Models

Foundational models are your base LLMs designed to understand human instructions and generate text based on them.

The foundational models are no longer restricted to text but can also generate images, voices, etc.

Open Source LLMs

There are some proprietary LLMs like GPT4 from OpenAI and there are many open source LLMs as well. Huggingface is a good repository for accessing almost all of these models 

Prompts

Human written instructions for LLMs.

Basically, the instructions that we write for ChatGPT to generate a response is a Prompt.

Prompt Engineering

Techniques to write better prompts.

Why do we need to write better prompts, you asked? 

Because the quality of the input determines the quality of the output.

There are many techniques and frameworks to write better prompts.

Learn Prompt Engineering for free – https://www.aimletc.com/free-course-introduction-to-prompt-engineering/

ChatGPT4

There is no such thing as ChatGPT4. 

ChatGPT, the famous AI tool for OpenAI is based on GPT3 and then OpenAI released its more advanced model GPT4

Also, GPT stands for Generative Pre-Trained Transformers

Tokens

Think of Tokens like a sub-word. 

Normally, 4 tokens are equal to 3 words. ChatGPT was trained on 300 billion tokens

Context Length

It is the maximum number of tokens/ words an LLM model can take to generate a response.

Parameters

OpenAI’s model GPT-3, which is the base model for ChatGPT has 175 billion parameters.

In simple language, parameters are amount of things/ patterns an LLM learns during its training period

RAG (Retrieval Augmented Generation)

We all know that LLMs like ChatGPT have lots of world knowledge and it can generate responses based on that knowledge.

However, what if you want them to answer from your private or proprietary data? This data could be in PDF, YouTube videos, your website, Notion etc. 

This can be done using RAG. 

With RAG, an LLM will chat with your PDF and generate a response. 

Learn more about RAG here – https://www.aimletc.com/free-course-retrieval-augmented-generation/

Vector Database

For an LLM to chat with your data, that data first needs to be converted into word embeddings. 

These embeddings are then stored in a vector database. 

Fine Tuning

With RAG, an LLM refers to your data, but what if you want the model to train on your data?

This can be done through fine-tuning of LLM. 

In simple terms, fine-tuning is the partial training of an LLM after its base training on some specialized data. 

With Fine-tuning, LLMs get more knowledge and can also change their behaviour.

Learn more about Fine-tuning here – https://www.aimletc.com/free-course-finetuning-large-language-models/

Hallucination

In real terms, hallucination means visualizing something that does not exist. 

LLMs also hallucinate and sometimes generate wrong responses. LLM Hallucination is a major problem.

Some experts say, hallucination is an intrinsic nature of LLMs and should not be seen as a problem.

LLM Agents

LLMs became popular but they work in isolation and do not interact with other software applications or other LLMs.

What if we connect LLMs to other software like Google, Wikipedia or other specialized LLMs? They become more useful. They can use their reasoning ability, divide a complex task into smaller tasks and then perform those tasks.

These are known as LLM Agents.

Learn more about LLM Agents – https://www.aimletc.com/what-are-llm-agents-why-are-everyone-talking-about-them/

Featured Image Source

Share it with your senior IT friends and colleagues
Nikhilesh Tayal
Nikhilesh Tayal
Articles: 55