📬 Getting this in Promotions? Move it to Primary so you never miss an edition.
IT HootClub — AI Community Newsletter
Hands-on. Career-focused. Future-ready.
Issued 2026-04-11
AI Safety Under Fire
Major lawsuits challenge OpenAI's safety practices while Anthropic's Mythos forces cybersecurity reconsiderations across the industry.
Announcements
Following Up — A Formal Note on AI Authorship
Last week I wrote about what responsible AI-assisted publishing looks like from
where I'm sitting — automation at the wheel for the repeatable and structural,
a human awake and paying attention for the rest. I meant it as a statement of
intent. This is the follow-through.
Yes — this newsletter is written with the assistance of AI. Specifically, one or
more of the same models we discuss and compare in these pages: Gemini, Claude,
and GPT. The news section is scraped and curated by Gemini. The learning items,
quotes, and jokes are pipeline outputs. The Building Intelligence series —
including what you're about to read below — is developed through iterative
research conversations with Claude, edited and shaped by me before a word goes out.
I take full editorial responsibility for everything published here. Nothing ships
without my review. The process is not generate-and-send — it's a back-and-forth
that involves sourced conversations, corrections on both sides, and judgment calls
that are mine to make. The AI is a capable collaborator. The editor is me.
Starting this week, technical content in the Building Intelligence
series will include a sources block so you can verify claims and go deeper
independently. That's the standard I'm holding myself to going forward —
not just transparency about the process, but accountability for the output.
NEW THIS WEEK
Introducing: Building Intelligence 🧠
Starting this week, I'm kicking off a new series called Building Intelligence —
a step-by-step journey through how AI models actually learn, remember, and get built from the ground up.
I'm doing this for the same reasons you probably should be: I'm building products with this
technology, recommending it to clients, and I want to understand exactly what's happening
under the hood — not just how to prompt it. We're going to cover RAG, embeddings, fine-tuning,
training data, hardware requirements, and the full production pipeline. Hands-on. Honest about
what's hard. No skipping the parts that aren't glamorous.
The series runs weekly in its own dedicated section. Each installment builds on the last.
If you've ever wondered how an AI model actually knows something — or why it sometimes
confidently doesn't — this is where we find out together.
Week 1 starts below.
Building Intelligence
BUILDING INTELLIGENCE · WEEK 1
What Does a Model Actually "Know"? — And What Are We Going to Build?
Most AI newsletters teach you how to use AI. This one is going to teach you how to
build it — or at least, how to make it yours.
Over the next several weeks, we're going to take an open-source language model, feed it
data that doesn't exist anywhere in its training, and produce something the big three —
Gemini, Claude, and GPT — simply cannot replicate: a model that genuinely understands the
Milwaukee tech ecosystem. Local colleges and their AI programs. Regional
companies. Events, initiatives, and the people building things here. Ask any of the major
models a specific question about Milwaukee's tech scene right now and you'll get a
generic, often inaccurate answer. That's the gap we're attempting to close.
I want to be upfront about something: I don't know what the outcome of this
will be. I've run a fine-tuning process before — once, on a small dataset,
without fully understanding what I was doing. This series is me learning in public,
documenting every step, every decision, and every result — including the ones that
don't go as planned. If the fine-tuned model turns out to be worse than what you
can already get from ChatGPT, we'll say so. That's the deal.
When it's done, we'll put our model in the ring with the big three and ask them all
the same questions. Side by side. Unedited. You'll see exactly where a small, focused,
locally-trained model wins — and where it doesn't. There are real questions here worth
exploring honestly: Can a smaller model trained on specific data actually outperform a
much larger general model on a narrow domain? And if the big three have already crawled
the same public internet data we're training on — does fine-tuning even give us an
advantage? We're going to find out together.
But here's the larger point — and it matters beyond this newsletter. The outcome of
this project isn't just about Milwaukee or one model. It's a demonstration of something
the industry is still coming to terms with: a well-executed, domain-specific
fine-tuned model can be more effective for a business than a general-purpose model
from one of the big three. Not because it's smarter — it isn't. But because
intelligence applied to the right problem, with the right data, beats raw capability
applied broadly every time.
That's not a guarantee. It's a hypothesis. And proving or disproving it requires
something the big three can't sell you: genuine, targeted work.
Not prompting. Not clicking through settings in a vendor's fine-tuning portal.
Real decisions about data, architecture, and evaluation — made by people who
understand what they're building and why. That's the skill this series is about.
There is real room to stand out in this industry precisely because most organizations
are still treating AI as something you subscribe to rather than something you build.
This isn't a theoretical exercise. The model will run on real hardware — the GPU lab
right here at WCTC — and every step of the process will be documented in this newsletter
as we go. The tools, the decisions, the things that didn't work, and what we learned
from them. If you're a student, a practitioner, or someone advising organizations on
AI adoption, this series is built for you.
Before we get into the project itself, we need to settle one foundational question —
because everything else builds on it: what does an AI model actually know,
and how does it know it?
The honest answer is that it doesn't "know" anything the way you do. A language model
is a massive compression of patterns extracted from text. During training, it reads an
enormous amount of human-written content and learns, statistically, what tends to follow
what. It doesn't store facts in a filing cabinet. It encodes relationships between words,
concepts, and ideas as billions of numerical weights — and when you ask it something,
it reconstructs a plausible response from those weights.
That's why it can be wrong with total confidence. It's not lying — it's pattern-matching
to the most statistically probable answer, even when the correct answer isn't the
probable one.
Here's the analogy that makes this click: imagine someone who read
every book in a library, then the library burned down. They can tell you a lot. They
can reason, summarize, explain, and synthesize. But if you ask them about a document
that was never in the library — or what changed last Tuesday — they'll try to answer
anyway. That confident attempt at an answer they don't actually have is what the field
calls a hallucination. It's a structural limitation, not a bug someone forgot to fix.
This is the core problem we're working on. The big three don't have Milwaukee in their
library — not really. Their training data is so vast that local, regional content gets
diluted to near-zero signal. A sentence about the Milwaukee tech scene that appeared
somewhere on the internet gets averaged across billions of other tokens. It's there,
statistically. It just has almost no influence on what the model actually says.
Fine-tuning is one way to change that. The three techniques that make this possible —
and that we'll be working through over the coming weeks:
Fine-tuning — Taking an existing model and retraining it on new
data so it internalizes specific knowledge. This is the core technique for our project.
It changes what the model structurally knows, not just what it can look up.
Embeddings — The mechanism that converts text into numbers representing
meaning. "Similar" ideas end up numerically close to each other. Understanding this is
essential for understanding why some training data works better than others.
RAG (Retrieval-Augmented Generation) — Giving a model access to
external documents at query time, rather than baking everything into the weights.
We'll cover this as the natural next layer after fine-tuning — and explain when you'd
use one versus the other.
Next week: the model choice. What the numbers actually mean — 7B, 13B, 70B parameters
— and the question nobody seems to ask directly: if bigger models are more capable,
why aren't we just fine-tuning one of those?
Training an AI model is a multi-step process that ranges from gathering data to fine-tuning algorithms. Depending on whether you are building something from scratch or using existing tools, the complexity varies.
Here is a high-level roadmap of how to train an AI model:
---
### 1. Define the Problem Before touching any code, you must decide what you want the AI to do. * **Classification:** Is this email spam or not? Is this a picture of a cat or a dog? * **Regression:** What will the price of a house be in six months? * **Generation:** Can you write a poem or create an image?
### 2. Gather and Prepare Data (The most important step) AI is only as good as the data it learns from. * **Data Collection:** Scrape websites, use public datasets (Kaggle, Google Dataset Search), or use your own company's logs. * **Data Cleaning:** Remove duplicates, fix missing values, and handle outliers. * **Labeling:** If you are doing "Supervised Learning," you need to label your data (e.g., tagging 1,000 photos as "Cat" or "Dog"). * **Data Splitting:** Divide your data into three sets: * **Training set (80%):** Used to teach the model. * **Validation set (10%):** Used to tune the model's settings. * **Test set (10%):** Used at the very end to see how the model performs on data it has never seen.
### 3. Choose an Algorithm/Architecture You don't always have to "invent" an AI. You choose a mathematical structure that fits your problem: * **Linear Regression:** For simple numerical predictions. * **Decision Trees/Random Forests:** For structured data (like spreadsheets). * **Convolutional Neural Networks (CNN):** For images. * **Transformers:** For text and language (this is what powers ChatGPT).
### 4. The Training Process This is where the "learning" happens. You feed the training data into the algorithm. * **Forward Pass:** The model makes a guess. * **Loss Function:** A mathematical formula calculates how wrong the guess was (the "error"). * **Backpropagation & Optimization:** The model adjusts its internal "weights" (settings) to reduce the error for the next guess. * *This cycle repeats thousands or millions of times.*
### 5. Evaluate the Model Once training is done, you test it using your **Test Set**. You look for metrics like: * **Accuracy:** How often was it right? * **Precision/Recall:** Is it missing too many things, or is it giving too many false alarms? * **Overfitting:** A common problem where the model memorizes the training data perfectly but fails to understand new, real-world data.
### 6. Hyperparameter Tuning If the results aren't good, you tweak the "knobs" of the model. These are called **Hyperparameters** (e.g., how fast the model learns, how many layers the neural network has). You re-train and re-evaluate until satisfied.
---
### Two Ways to Get Started
#### Option A: The "From Scratch" Way (Hard) If you want to be a Data Scientist, you learn **Python** and use libraries like: * **Scikit-learn:** For basic machine learning. * **PyTorch** or **TensorFlow:** For deep learning and neural networks.
#### Option B: Transfer Learning (The Modern/Fast Way) You don't need to train a massive model from scratch. You can take a "Pre-trained" model (like GPT-4, Llama 3, or ResNet) and **Fine-tune** it. * **Example:** You take a model that already knows how to recognize objects and you give it 100 photos of *your specific product* to teach it that one new thing.
---
### Tools You Can Use Today 1. **Google Colab:** A free, browser-based tool that lets you write Python and use Google’s powerful GPUs for free. 2. **Hugging Face:** The "GitHub of AI." You can download thousands of pre-trained models here and fine-tune them. 3. **Kaggle:** Great for finding datasets and seeing how others have solved similar problems.
**Where would you like to start? I can explain a specific step in more detail or provide a simple Python code example.**
Anthropic Claude (claude-sonnet-4-20250514)
Training an AI model involves several key steps. Here's a general overview:
## 1. Define Your Problem - Identify what you want the model to do (classify images, predict values, generate text, etc.) - Choose the appropriate type of model (supervised, unsupervised, reinforcement learning)
## 2. Prepare Your Data - **Collect data** relevant to your problem - **Clean and preprocess** the data (handle missing values, normalize, etc.) - **Split data** into training, validation, and test sets (typically 70/15/15 or 80/10/10)
## 3. Choose a Model Architecture - Select appropriate algorithms (neural networks, decision trees, etc.) - Consider pre-trained models for transfer learning - Decide on model complexity based on your data size
## 4. Set Up Training Environment - Choose frameworks (TensorFlow, PyTorch, scikit-learn) - Ensure adequate computing resources (GPU for deep learning) - Set up development environment
## 5. Train the Model - Feed training data to the model - Set hyperparameters (learning rate, batch size, epochs) - Monitor training progress and loss functions - Use validation data to prevent overfitting
## 6. Evaluate and Iterate - Test model performance on unseen data - Analyze metrics relevant to your problem - Adjust hyperparameters or architecture if needed - Retrain as necessary
## Getting Started Resources: - **Beginners**: Start with platforms like Google Colab, Kaggle Learn - **Frameworks**: TensorFlow/Keras, PyTorch, or scikit-learn - **Datasets**: Use public datasets from Kaggle, UCI ML Repository
Would you like me to elaborate on any specific step or type of model?
OpenAI GPT (gpt-4o)
Training an AI model involves several key steps, which can vary depending on the type of model you're developing (e.g., supervised learning, unsupervised learning, reinforcement learning). However, the general process typically involves the following key steps:
1. **Define the Problem**: Clearly outline what you want the AI model to achieve. This could be a classification task, regression, image recognition, language translation, etc.
2. **Collect and Prepare Data**: - **Data Collection**: Gather a dataset relevant to the problem. This data may come from APIs, online repositories, internal databases, or be manually collected. - **Data Cleaning**: Ensure the data is clean and free from errors or inconsistencies. This might involve handling missing data, removing duplicates, or correcting errors. - **Data Preprocessing**: Transform the data into a format suitable for your model. This may include normalization, encoding categorical variables, text processing, image resizing, etc. - **Split the Data**: Divide the data into training, validation, and test sets. A common split is 80/10/10, but this can vary based on the size of the dataset.
3. **Select a Model**: - Choose an appropriate model or algorithm for your problem. This could be a decision tree, neural network, support vector machine, etc. For complex tasks, you might opt for deep learning models like convolutional neural networks (CNNs) or transformers.
4. **Configure the Model**: - Set hyperparameters: These are model parameters set before training begins (e.g., learning rate, number of layers in a neural network, batch size). - Choose an appropriate loss function that reflects the goals of your problem.
5. **Train the Model**: - Feed the training data through the model in iterations (epochs). The model learns by adjusting parameters to minimize the loss function using optimization algorithms like stochastic gradient descent or Adam.
6. **Validate the Model**: - Use the validation set to tune hyperparameters and avoid overfitting. Monitor the model's performance on this set but do not make final decisions based on it.
7. **Evaluate the Model**: - Test the model using the test set (data it has never seen before) to get a sense of how well it performs in the real world.
8. **Optimize and Iterate**: - Analyze results, identify weaknesses, and refine the model. This might involve gathering more data, trying different models, or fine-tuning hyperparameters. - Repeat the training/validation process until satisfactory performance is achieved.
9. **Deploy the Model**: - Once the model meets performance expectations, it can be deployed into a production environment where it interacts with real-world data. - Consider scalability, maintainability, and monitoring for maintenance.
10. **Monitor and Maintain**: Continuously monitor the model's performance to ensure it remains accurate and relevant. This might involve updating the model with new data or adjusting to changes in the input data distribution.
Throughout this process, you'll need a programming environment capable of handling data and training models, such as Python with libraries like TensorFlow, PyTorch, scikit-learn, or Keras. Understanding the domain of your problem is also crucial for making informed decisions about data preprocessing and model selection.
The Learning Loop
DEFINITIONEmbedding
An embedding is a mathematical representation that converts words, sentences, or other data into lists of numbers (vectors) that AI models can understand. Think of it like translating human language into a numerical "coordinate system" where similar concepts are placed closer together - for example, "cat" and "dog" would have similar number patterns because they're both pets.
Source: OpenAI Documentation
TIPPrompt Chaining
Break complex tasks into smaller, sequential prompts where each output feeds into the next prompt. Instead of asking AI to "analyze this data, create a summary, and write recommendations" in one go, use three separate prompts: first analyze, then summarize that analysis, then create recommendations from the summary. This improves accuracy and allows you to refine each step.
Source: Prompt Engineering Institute
TOOLGamma
An AI-powered presentation builder that creates professional slides, documents, and websites from simple text prompts, with smart formatting and design suggestions.
“The expert in anything was once a beginner who refused to give up.”
— Helen Hayes — Helen Hayes was an American actress known as the "First Lady of American Theatre" who won Academy, Emmy, Grammy, and Tony Awards during her career. She overcame significant personal challenges including the loss of her daughter and continued performing well into her 80s. Hayes was celebrated not only for her talent but for her dedication to mentoring young actors and her belief in continuous learning throughout life.
The Nest Jest
Upcoming Events
Event04/13/2026 8:30 AM
Global AI Community Milwaukee AgentConf 2025 — Online (Milwaukee, WI)
**SIGNUP HERE:** [https://globalai.community/agentcamp/milwaukee/](https://globalai.community/agentcamp/milwaukee/) *NOTE: This event is managed outside of Meetup, signup at the link above, meetup signups will not get you in to the conference!* Join us for a free community event focused on AI, where you can learn from experts, network with peers, and explore the latest AI technologies and trends. ### What to Expect: * Deep-Dive Talks from AI pioneers and industry leaders * Technical Workshop [Group: global-ai_milwaukee]
Hands-On : GitHub Copilot SpecKit — Online (Applied Information Sciences, Inc, 11440 Commerce Park Dr # 600, Reston, VA)
**Agenda :** * 4.45 to 5.00 PM ET: Food and Networking * 5.00 to 5.50 PM ET: "Hands-On : GitHub Copilot SpecKit" Hello Everyone! Please join us for our April 14th edition of the AI-ML MeetUp. **Please note this is an in-person meeting and light refreshments/food will be provided. You will need a government-issued ID to enter the facility.** **Title:** Hands-On : GitHub Copilot SpecKit **Description:** Spec-Driven Development **flips the script** on traditional software development. For decad [Group: artificialintelligenceandmachinelearning]
Welcome to MKE Tech Meetup — Penrod – 219 N Milwaukee Street, Milwaukee, WI 53202
A curated social event orienting new tech employees into Milwaukee's tech and innovation ecosystem. Free and open to new tech workers in the region. Street parking available nearby.
Source: MKE Tech Hub Coalition
Event04/22/2026 9:00 AM
PMI-CPMAI Certification Training – 3-Day Bootcamp in Milwaukee, WI — 1433 N Water St, Milwaukee, WI
Master AI in Project Management & project in AI concepts with PMI-CPMAI™. 4-day training, real use cases & exam prep.
Hands-On : Copilot Studio, Microsoft Fabric, Azure AI : Better Together — Online (Online event)
**Hands-On Online Workshop: Copilot Studio, Microsoft Fabric, Azure AI : Better Together** **Date: 09 May 2026, 10 AM to 5 PM Eastern Time** **Level: Beginners/Intermediate** **Registration Link:** https://www.eventbrite.com/e/hands-on-copilot-studio-microsoft-fabric-azure-ai-better-together-tickets-1983680029367?aff=oddtdtcreator **Who Should Attend?** This hands-on workshop is open to developers, senior software engineers, IT pros, architects, IT managers, citizen developers, technology prod [Group: artificialintelligenceandmachinelearning]
Stalking victim sues OpenAI, claims ChatGPT fueled her abuser's delusions and ignored her warnings
A lawsuit alleges that OpenAI ignored three warnings about a dangerous ChatGPT user, including its own mass-casualty flag, while the user stalked and harassed his ex-girlfriend. The case raises serious questions about AI safety monitoring and platform responsibility when users employ ChatGPT for harmful purposes. read more
News2026-04-11 — WIRED
Anthropic's Mythos Will Force a Cybersecurity Reckoning—Just Not the One You Think
Anthropic's new AI model Mythos is being both heralded and feared as a potential hacker's superweapon in cybersecurity circles. Experts view its arrival as a critical wake-up call for developers who have historically treated security as an afterthought in AI development. read more
News2026-04-11 — WIRED
This Startup Wants You to Pay Up to Talk With AI Versions of Human Experts
Onix is launching what it calls a "Substack of bots," featuring digital twins of health and wellness influencers that provide advice 24/7. The platform allows users to interact with AI versions of human experts while potentially promoting their products and services. read more
Tools from Hard AIs
AI Airfare Research
Find cheaper flights using an AI-powered research tool
built by Hard AIs.