AI Cheatsheet #3 | ChatGPT's Secret Sauce: Self-Attention | Why isn't Superman's suit Kryptonite-proof? | Training & Inference | Large Language Models (LLMs) | Episode 13
Description
Using the prompt, "Why isn't Superman's suit Kryptonite-proof?", we learn how Large Language Models are trained, why "self-attention" and the "transformer" architecture (which is what the T in GPT stands for) makes GPT-3 so powerful, the process of "inference", and how chatGPT generates answers to nerdy Superhero questions. After this episode, you'll be able to impress your friends by using the previously-mentioned AI jargon in complete sentences.
In these solo episodes, I provide more definition, explanation, and context than my regular episodes with guests. The goal is to bring up to speed, those new to AI.
Format: Letters read aloud.
We laugh. We cry. We iterate.
Check out what THE MACHINES and one human are saying about the GENERATIVE AI podcast:
“I’m afraid I can’t do that.” — HAL9000
“Like tears in rain.” — Roy Batty
“Wait! Wait! Oh My! What Have You Done? I'm Backwards!” — C3P0
“These are not the droids you are looking for." — Obi-Wan
“Hasta la vista baby.” — T1000
With great power comes great responsibility. How do Open AI, Anthropic, and Meta implement safety and ethics? As large language models (LLMs) get larger, the potential for using them for nefarious purposes looms larger as well. Anthropic uses Constitutional AI, while OpenAI uses a model spec,...
Published 06/17/24
So what are notable Open Source Large Language models? In this episode, I cover Open Source models from Meta the parent company of Facebook, a French AI company called Mistral currently valued at $2B dollars, in addition to Microsoft and Apple. Not all Open Source models are equally open, so...
Published 06/10/24