What it is. Who makes it. How to use it — and how to put it to work for your family, your work, and yourself.
AI is the big umbrella. Machine Learning is how we build AI today. Inside ML are three learning approaches — and inside one of them (supervised learning) lives the neural network stack that eventually gave us Claude.
Learns from labeled examples. You show it thousands of emails tagged "spam" or "not spam" and it learns the pattern. Most practical AI is this.
No labels — the model finds hidden patterns on its own. You don't tell it what to look for. It discovers structure in the data you didn't know was there.
Learns by trial and error, earning rewards for good decisions and penalties for bad ones. No labeled data — just feedback from the environment.
Classical software is explicit: a programmer writes rules. ML inverts this. You define an objective, feed in data, and an algorithm adjusts millions of internal parameters until the model learns to minimize its error. No rules are written. Behavior emerges from optimization.
Billions of text passages, images, code snippets, and conversations become the raw material for learning.
A loss function measures how wrong the model is. The goal: get that number as low as possible, as reliably as possible.
Gradient descent iteratively tunes parameters. Run this long enough at sufficient scale and complex capabilities emerge.
RLHF — Reinforcement Learning from Human Feedback — turns a raw predictor into something helpful, safe, and pleasant to talk to.
Modern AI didn't come from one invention. It came from two distinct advances — one in architecture, one in alignment — that together turned raw prediction into something that feels like intelligence.
Before the Transformer, AI read text sequentially — word by word, left to right. The breakthrough was attention: letting the model look at all words simultaneously and learn which ones matter most to each other.
In the sentence "The animal didn't cross the street because it was too tired" — attention is what tells the model that "it" refers to "animal," not "street." That ability to track relationships across context is what makes LLMs feel like they understand.
Vaswani et al., "Attention Is All You Need," 2017
A trained model is technically just a very good text predictor. It doesn't yet know how to be helpful or safe. Reinforcement Learning from Human Feedback is the fine-tuning step that shapes it into an assistant.
Human raters compare pairs of model responses and pick the better one. Those preferences become a reward signal, and the model learns to produce outputs humans rate highly. This is what turns a raw predictor into something that follows instructions, avoids harm, and feels like it has a personality.
Why Claude feels different from GPT — same architecture, different RLHF.
The same optimization principle underlies completely different model families — each with a different architecture, training signal, and class of problems it solves well.
Most AI tools you use today combine more than one. ChatGPT and Claude are LLMs fine-tuned with RL. DALL·E is a diffusion model guided by a text encoder. Sora combines both. (Sora shutting down Apr 26)
Transformer-based. Trained to predict the next token across text, code, and data.
Trained to reverse a noise process. Start with static, iteratively denoise into an image or video.
Trained via reward signals from an environment. The model discovers strategies by trial and error.
Trained on images for classification, detection, and segmentation. Vision Transformers now lead on benchmarks.
Combine architectures to handle text, images, audio, and video together. The current frontier.
When someone says "AI did this" — a song, an image, a diagnosis, a recommendation — a different model family is behind each one. They share the same mathematical foundation but are built for completely different problems.
The tools you consciously choose are just the tip of the iceberg. AI has been quietly running your daily life for years — you just didn't call it that.
What's changed isn't that AI arrived. It's that it became visible, conversational, and creative — and now anyone can use it, not just engineers.
Understanding these three modes helps you know what any AI tool can — and can't — do for you.
Trained on patterns to forecast an outcome. Doesn't understand — calculates probabilities.
Given a prompt, generates something that didn't exist before — text, images, music, video, code.
Goes beyond answering — it actually does things. Opens apps, browses the web, writes files, sends emails, completes tasks end-to-end.
Most tools today combine all three. Claude is generative when you chat with it, and agentic when you use Claude Code or give it tools. The lines are disappearing fast — and the agentic layer is where everything is heading.
There isn't one AI — there are several, made by competing companies with different personalities, strengths, and philosophies. Here's how to think about the main ones.
Closed models (OpenAI, Anthropic, Google) keep their code private. Open models (Meta, DeepSeek) publish it — anyone can build on top, modify, or run them privately.
There's no single winner. Each model has a personality and a sweet spot. Most power users keep 2-3 open and switch depending on the task.
Best writing quality, strongest safety track record, and the most complete agentic ecosystem — Code, Cowork, Chrome. The most complete platform for building real things.
Microsoft doesn't compete with a frontier model of its own. Instead they invested $13B in OpenAI, struck a $30B compute deal with Anthropic, and built Copilot on top of both. Their open-source model Phi 4 is small but excellent. Their strategy: be the platform and infrastructure for all the best models — not build one themselves.
Normally when you use Claude or ChatGPT, your words travel to a server, get processed, and come back. A local LLM runs entirely on your own computer. Nothing leaves your machine.
Your message goes to Anthropic's / OpenAI's servers
Processed in the cloud, response sent back
Most powerful models available
Requires internet · Monthly subscription
Company can see your prompts (per their privacy policy)
Everything stays on your machine — zero data sent out
Works offline — plane, no wifi, anywhere
Download once, use forever — no monthly fee
No usage limits, no rate throttling
Less powerful than cloud frontier models — you're trading raw capability for privacy
WHEN DOES THIS ACTUALLY MATTER — FOR YOUR LIFE
Analyzing health documents, insurance paperwork, or a family member's diagnosis — locally, privately.
Contracts, estate planning, NDAs — you want AI help but you don't want those words on anyone's server.
Board memos, M&A documents, client strategy — anything with NDA implications or fiduciary responsibility.
School records, IEP documents, pediatric notes — some parents prefer this stays entirely off cloud servers.
HOW TO RUN ONE — EASIER THAN YOU THINK
WHAT RUNS WELL LOCALLY
THE HONEST TRADEOFF
A local LLM is like having a very capable assistant who works entirely inside your house and never tells anyone anything. They're not quite as brilliant as the best cloud models — but they're completely yours.
A local LLM is about where it runs. A vertical LLM is about what it knows. These are models trained specifically on one industry's language, documents, and data — making them far more precise than a general model in their domain.
Knows a little about everything. Great for most tasks. Like a very well-read generalist.
Trained deeply on one field. Speaks the language. Knows the nuance. Like a specialist with 20 years in the room.
REAL EXAMPLES BY INDUSTRY
Trained on case law, contracts, and legal precedent. Used by major law firms to review documents and draft arguments.
Trained on 40 years of financial documents, earnings calls, and market data. Understands financial language at a level general models can't match.
Trained on clinical records, oncology research, and medical imaging. Helps doctors analyze diagnoses and treatment options.
Trained exclusively on licensed creative assets — no copyright issues. Built specifically for designers and creative professionals.
Built on Khan Academy's entire curriculum. Acts as a personal tutor — never just gives the answer, guides the student to find it.
Trained on brand voice, product catalogs, and consumer behavior. Writes marketing copy and product descriptions that convert.
| CLOUD | LOCAL | |
|---|---|---|
| GENERAL | Claude, ChatGPT | Llama, Gemma |
| VERTICAL | Harvey, Bloomberg GPT | Rare — enterprise only |
If you ever work with a specialized AI tool in your field — a legal research assistant, a financial planning tool, a tutoring app for your kids — you're likely using a vertical LLM under the hood. The general models you're learning today are the foundation everything else is built on.
You've heard of Claude. But Anthropic has built an entire suite of products — each one designed for a different level of access and a different kind of user. Here's how they all fit together.
LAUNCH TIMELINE
THE FULL PRODUCT LINEUP
COWORK IN PLAIN ENGLISH
You point Claude at a folder. You give it a task. It reads your files, figures out what needs to happen, does it, and tells you when it's done. You can walk away. No babysitting required.
Claude Code is for building things. Cowork is for getting things done. If you've ever wished you had an extra set of hands for the administrative layer of your life — scheduling, organizing, researching, formatting — Cowork is what that looks like.
Whether you use Claude in the browser, the desktop app, or the terminal, you're talking to one of three underlying models. Think of them as different settings on the same engine.
The deepest thinker. Best for complex reasoning, nuanced writing, and hard problems where quality matters more than speed.
The everyday workhorse. Fast, smart, and capable — handles the vast majority of tasks beautifully, including today's workshop.
Lightning-quick for simple tasks. Great when you need a fast answer, a short summary, or a quick rewrite on your phone.
These three models power every way you access Claude — the website, the mobile app, the desktop app, and the terminal. The interface changes; the intelligence underneath doesn't.
Anthropic just unveiled its most powerful model yet — one so capable they're not releasing it to the public. This is what the frontier of AI looks like right now.
WHAT IT IS
Anthropic has always had Haiku → Sonnet → Opus. Mythos is a new category above all of them — dramatically more powerful than anything they've released before.
Mythos scanned major software systems and discovered tens of thousands of critical bugs — some of them decades old and never caught by human researchers. It then wrote the fixes.
Anthropic is so concerned about what Mythos could do in the wrong hands that they're keeping it invite-only. Only ~40 companies in the world have access.
PROJECT GLASSWING — WHO HAS ACCESS
+ ~32 more organizations. No public access. No timeline for general release.
$100M COMMITTED
Anthropic is providing up to $100 million in usage credits to companies testing Mythos — plus $4 million directly to open-source security foundations.
WHY THIS MATTERS TO YOU
The apps on your phone. Your banking software. The browser you're using right now. Mythos is scanning all of it for holes that hackers could exploit — and patching them before anyone gets hurt. This is AI working for you in ways you'll never directly see.
THE BIGGER SIGNAL
We are now in a world where AI models are so capable that the companies building them are pausing before releasing them. That is a new thing. The question of what AI should be allowed to do is no longer hypothetical.
A test so hard that when AI finally aces it, we may no longer be able to write one harder. 2,500 questions from nearly 1,000 subject-matter experts at 500+ institutions. In early 2025, the best models scored under 10%. By March 2026, GPT-5.4 Pro leads at 44.3% — human experts still average 90%.
WHAT IT IS
The problem: AI was acing every test we threw at it. MMLU, the gold standard benchmark, is now solved at 90%+ accuracy. We ran out of hard tests.
The solution: Ask the world's sharpest minds to submit their hardest questions. PhD-level. Verifiable answers. No multiple choice tricks. Published in Nature, January 2026.
The result: Even the best models struggle. GPT-5.4 Pro leads at 44.3%. These are questions that genuinely stump frontier AI.
MODEL ACCURACY ON HLE — Scale AI Official Leaderboard · March 2026
Source: labs.scale.com/leaderboard/humanitys_last_exam · March 2026 · Human expert avg: ~90%
WHY IT MATTERS
When a model hits 50%+ on HLE, it means AI has reached expert-level reasoning across essentially every field of human knowledge. We're not there yet — but the pace of improvement suggests we will be very soon.
BUT WAIT — HLE DOESN'T TEST EVERYTHING
70% of developers prefer Claude for coding.
Claude leads on what matters in real engineering — multi-file reasoning, accurate refactoring, fewer hallucinated APIs. Claude Code owns 54% of the enterprise coding market.
Coined by Andrej Karpathy — one of the founders of OpenAI — in early 2025. Instead of writing code yourself, you describe what you want in plain English, and the AI builds it. The skill isn't coding anymore — it's knowing what to ask for and recognizing when it's right.
Learn Python. Study HTML. Understand databases. Take courses for months. Write hundreds of lines of code. Debug errors. Start over. Repeat indefinitely.
Open Claude Code. Describe what you want. Watch it appear in seconds. Give feedback. Iterate. Walk away with something real.
"The best interface for AI is the one you've been using your whole life — plain English."
For a moment, people sold courses on how to word your inputs. Frameworks, formulas, incantations.
Be specific. Give context. Say what you actually want. Push back when it's wrong. That's it. You already know how to do this.
Claude Code used to be a terminal-only developer tool. Anthropic now offers a Desktop App that gives you the same full power through a regular window — download and open it like any other app.
Create a free account with your email address. No credit card required to start.
$20/month. This unlocks Claude Code. The free plan doesn't include it.
Available for Mac and Windows. Installs like any normal app — no terminal involved.
Describe what you want to build in plain English. No coding experience required. Just a conversation.
Web apps run in any browser — phone, laptop, tablet — and can be added to your home screen as a Progressive Web App (no App Store, no fees). Native iOS apps require Xcode, a Mac, and a $99/year Apple Developer Account. For this workshop, we're building web apps — faster, free, and just as useful.
A chatbot answers your question. An agent goes and handles it. This is the most important shift happening in AI right now — and it's already available to you.
HOW AN AGENT WORKS
CHATBOT VS. AGENT — THE KEY DIFFERENCE
| Dimension | Chatbot | Agent |
|---|---|---|
| What it does | Answers your question | Completes your task |
| Memory | This conversation only | Can remember across sessions |
| Tools | Text in, text out | Browses web, reads files, sends email |
| Runs for | Seconds | Minutes to hours, unattended |
| You do | All the follow-up yourself | Review the finished output |
REAL EXAMPLES — FOR YOUR LIFE RIGHT NOW
Tell it: "Parse the Buckley weekly email, pull all Grade 2 events, and send me a weekly digest every Sunday night."
Try with Claude CodeReads your inbox, flags what needs a reply today, drafts responses for your review. You approve, it sends.
Try with Claude + GmailTell it: "Research summer programs for a 7-year-old in NYC, under $3k, with spots still open." It searches, compares, and returns a ranked list.
Try with Perplexity"Find the best-reviewed lunchbox under $40 that fits in a Buckley cubby and has no BPA." Done. With links.
Try with ChatGPT OperatorTHE MINDSET SHIFT
Stop asking AI for answers. Start assigning AI tasks — with a deadline and a deliverable.
In the next 30 minutes, we're going to use Claude — Chat, Cowork, and Code — to build a real, working app together. No coding experience needed. You just need to know how to talk.
The school calendar PDF has everything — but it's a wall of text. This app turns it into a filterable, color-coded calendar so you can see exactly what matters to your child's grade. No more scrolling through dates that don't apply.
See only events relevant to your child's class. One toggle, one view.
Red = days off · Blue = school events · Green = deadlines
Toggle between a calendar grid and a clean scrollable list.
Add to home screen as a PWA. No App Store. No download. Free.
This is a real app you'll actually use. When you see something built from your own school's data, it stops feeling like a demo and starts feeling like a superpower.
By the end: a live web app at a real URL, a recurring weekly digest, and your first GitHub repo — all from plain English conversations.
AI just gives you a way to act on it, faster and bigger than ever before. The version of you that knows how to use these tools is dramatically more capable than the version that doesn't.
reginaflores.ai · Workshop 01 · 2026