DEV Community

Cover image for ML Type, Algorithm, and Model in common AI applications
Anamika
Anamika

Posted on

ML Type, Algorithm, and Model in common AI applications

Popular AI Applications, and the ML Type, Algorithms, and Models used in/by them.

ChatGPT

anamika-chatgpt

What is ChatGPT

ChatGPT is a very popular GenAI Application, which is now being used by millions of people in some way or the other.
It can be used to write and fix code, ask financial and technical questions, cook, and do a variety of other things.

Model versions
The following table lists the main model versions of ChatGPT, describing the significant changes included with each version:

Version Release date Description Status
GPT-3.5 November 2022 The first ChatGPT version used the GPT-3.5 model. Discontinued
GPT-3.5 Turbo 2023 An improvement over the legacy version of GPT-3.5, GPT-3.5 Turbo in ChatGPT offered better accuracy in responses while using a similar model. Discontinued
GPT-4 March 2023 Introduced with the ChatGPT Plus subscription, the March 2023 version is based on the more advanced GPT-4 model. Active
GPT-4o May 2024 Capable of processing text, image, audio, and video, GPT-4o is faster and more capable than GPT-4, and free within a usage limit that is higher for paid subscriptions.[107] Active
GPT-4o mini July 2024 A smaller and cheaper version of GPT-4o. GPT-4o mini replaced GPT-3.5 in the July 2024 version of ChatGPT.[108] Active
o1-preview September 2024 A pre-release version of OpenAI o1, an updated version that could “think” before responding to requests.[109] Discontinued
o1-mini September 2024 A smaller and faster version of OpenAI o1.[109] Discontinued
o1 December 2024 The full release of OpenAI o1, which had previously been available as a preview.[103] Active
o1 pro mode December 2024 An upgraded version of OpenAI o1 which uses more compute, available to ChatGPT Pro subscribers.[103] Active
o3-mini January 2025 Successor of o1-mini.[110] Active
o3-mini-high January 2025 Variant of o3-mini using more reasoning effort.[110] Active

source: wikipedia

How does ChatGPT gather data?

ChatGPT was trained on a large training dataset that consisted of books, articles, and web pages that are available publicly on the internet.
This is one of the largest training dataset available.
The data collection process was also called Common Crawl, where all the publicly available information was gathered and then fed to ChatGPT.

Training Used:
Unsupervised Learning + Supervised Learning during fine tuning process.

The outcome of the training was a model called Generative pre-trained Transformer model or GPT.
Hence, GPT is the model that powers all the versions of ChatGPT.

ChatGPT is a large language model (LLM), because it has been trained on a large training dataset, which contains billions of instructions.

DALL·E

anamika-dalle

What is DALL·E

DALL·E is an AI model that can generate realistic images, and art from a description in natural language.
It is developed by OpenAI.

How does DALL·E gather data?

The training data for DALL·E consists of a vast collection of text-image pairs sourced from the internet.
These pairs include captions and corresponding images, allowing the model to learn the relationship between the textual description and visual representations.

Training Used:
Unsupervised Learning + Supervised Learning during fine tuning process.

DALL·E is also a large language model (LLM) because it has been fed a large training dataset.
The output model that is used for DALL·E is GPT-3 model, and it is specifically designed for image generation.

GitHub Copilot

anamika-github-copilot

What is GitHub Copilot?

According to GitHub Documentation, "GitHub Copilot is an AI coding assistant that helps you write code faster and with less effort, allowing you to focus more energy on problem solving and collaboration".

GitHub Copilot is an AI tool developed by OpenAI in collaboration with GitHub.
Copilot suggests code as you type, just like having a coding assistant right in your development environment.

How was it trained, and How does it generate code?

GitHub Copilot has been trained on a large dataset on the publicly available code from repositories, coding websites, forums and documentation available on the internet.

Training Used:
Unsupervised Learning + Supervised Learning during fine tuning process.

The outcome of the training was codex model.
GitHub Copilot uses codex model, a descendent of GPT-3, based on transformer architecture.


Thanks for Reading 😊
Please drop a 👍 if you liked the post!

Also, feel free to reach out if you need any other info around this, or any other topic. Will be happy to share.

Top comments (0)