Applied · Module 1

Models, parameters and training dynamics

A model is still a function that turns input into output.

48 min 4 outcomes AI Intermediate

Previously

Start with AI Intermediate

Work with evaluation, features, embeddings, and practical model use without the buzzword fog.

This module

Models, parameters and training dynamics

A model is still a function that turns input into output.

Next

Data, features and representation

Raw data is rarely ready for a model.

Progress

Mark this module complete when you can explain it without rereading every paragraph.

Why this matters

If you remember one sentence from this section, keep this one.

What you will be able to do

  • 1 Explain models, parameters and training dynamics in your own words and apply it to a realistic scenario.
  • 2 Training is a loop: predict, measure error, update parameters, repeat with discipline.
  • 3 Check the assumption "Loss matches the decision" and explain what changes if it is false.
  • 4 Check the assumption "Validation is separate" and explain what changes if it is false.

Before you begin

  • Foundations-level vocabulary and concepts
  • Confidence with basic diagrams and section terminology

Common ways people get this wrong

  • Overfitting. The model learns the training set, not the task. It looks strong and fails on new data.
  • Leakage. Information from the future sneaks into inputs. The score rises and trust collapses.

Main idea at a glance

Training loop intuition

Predict, measure error, update, repeat.

Stage 1

Input batch

Each iteration starts with a batch of examples from your training set.

I think batching is essential because it balances gradient stability with computational efficiency.

A model is still a function that turns input into output. At Intermediate level, the useful question is what kind of function it is and what it can store. Modern models are flexible pattern machines. They are not a library of facts. They are a set of learned behaviours shaped by data and training.

The behaviours live in parameters. You can think of them as tiny knobs inside the model. Training turns knobs until the model produces outputs that match the examples often enough.

Use this rule in practice

  1. When you need facts, use retrieval with permissions and traceability

    Ground answers in governed sources instead of expecting the model to memorise truth.

  2. When you need judgement, define a clear decision boundary

    Test explicit failure cases before trusting any automated recommendation.

  3. Define failure early or the system will define it for you

    Operational incidents usually happen where boundaries were implied but never written.

Interactive lab

Glossary Tip

This module includes an interactive practice component. Open the deeper tool or workspace step when you want to test the idea rather than only read it.

More parameters gives the model more capacity. It can represent more subtle patterns. It can also memorise noise if you let it.

Mixture of Experts (MoE)

An architecture where only a fraction of the model's parameters activate for each input. This lets a model keep large total capacity without paying the full compute cost on every token. It is one design option among several, not a guarantee of quality on its own.

Scale matters because the world is messy. If you want a model to handle more languages, more topics, or more edge cases, it usually needs more capacity and more data. Scale is not a free win. It increases cost, increases training time, and increases the number of ways training can go wrong.

Training is a loop where you show examples, let the model predict, measure how wrong it was, and then update the parameters so it is a little less wrong next time. That measure of wrongness is called the loss.

Interactive lab

Glossary Tip

This module includes an interactive practice component. Open the deeper tool or workspace step when you want to test the idea rather than only read it.

Loss is not a moral judgement. It is a learning signal.

Under the hood, the model uses a gradient to decide how to adjust parameters.

Interactive lab

Glossary Tip

This module includes an interactive practice component. Open the deeper tool or workspace step when you want to test the idea rather than only read it.

You do not need the equations yet. The intuition is enough. If the loss is high, the gradient points toward changes that should lower it. If the training setup is stable, repeating this process steadily improves performance.

This is where overfitting shows up.

Interactive lab

Glossary Tip

This module includes an interactive practice component. Open the deeper tool or workspace step when you want to test the idea rather than only read it.

Underfitting is the opposite. The model is too simple or not trained enough, so it cannot learn the pattern even on the training data.

The goal is generalisation.

Interactive lab

Glossary Tip

This module includes an interactive practice component. Open the deeper tool or workspace step when you want to test the idea rather than only read it.

Training is expensive and brittle because small changes can have large effects. Data quality issues, hidden leakage, unstable learning rates, poor shuffling, or a mismatch between training and real inputs can collapse learning. Even when training works, the model can learn shortcuts. It might latch onto the background of an image instead of the object. It might learn the formatting of an email instead of the message. These failures are not rare. They are the default unless you design against them.

What I look for when training “works” but the system still fails

When someone tells me “the model is accurate”, I do not argue. I ask questions. Accurate on what, measured how, and compared to which baseline. Then I ask the question people avoid: what is the model using to be accurate. Shortcut learning is the usual culprit. The model found a pattern that is true in the training data and fragile in reality.

A practical example. Imagine you are classifying “urgent” support tickets. If most urgent tickets were written by the night shift for three months, a model may quietly learn that “night shift writing style” means urgent. It looks brilliant in testing and then falls apart when staffing changes. This is why I care about datasets like I care about code. They contain assumptions, and assumptions are where failures breed.

Training dynamics: good, bad, best practice

Good practice
Keep a baseline model and compare against it. If you cannot beat a simple baseline honestly, you are not ready for complexity.
Bad practice
Treating one metric as proof of general intelligence. Most of the time it is proof you selected an easy target.
Best practice
Write a short model note that includes: the decision, the failure cost, the baseline, the key risks, and the rollback or fallback plan. If it feels like paperwork, remember it is cheaper than the post-incident report.

How to use this level so it actually changes how you work

I am aiming for two audiences at the same time. If you are new, I keep the language calm and concrete. If you are technical, I keep the claims honest and the trade offs explicit.

Good practice
Read one section, do one tool, then write one paragraph of what you observed. If you cannot explain what changed and why, go back. The goal is judgement, not vocabulary collection.
Bad practice
Skimming and declaring it understood because the words look familiar. This is how people ship models that are statistically impressive and operationally useless.
Best practice
Pick one real decision from your world and map every section back to it. What is the input. What is the output. What is the failure cost. What is your fallback. If you can answer those, you are building a system, not a demo.

Mental model

Training loop shape

Training is a loop: predict, measure error, update parameters, repeat with discipline.

  1. 1

    Training data

  2. 2

    Predict

  3. 3

    Loss

  4. 4

    Update

  5. 5

    Parameters

Assumptions to keep in mind

  • Loss matches the decision. If the loss does not reflect what you care about, training improves the wrong behaviour.
  • Validation is separate. If validation is not separate, you cannot tell if you learned the pattern or memorised the dataset.

Failure modes to notice

  • Overfitting. The model learns the training set, not the task. It looks strong and fails on new data.
  • Leakage. Information from the future sneaks into inputs. The score rises and trust collapses.

Key terms

Mixture of Experts (MoE)
An architecture where only a fraction of the model's parameters activate for each input. This lets a model keep large total capacity without paying the full compute cost on every token. It is one design option among several, not a guarantee of quality on its own.

Check yourself

Quick check on models and training dynamics

0 of 9 opened

What is a parameter

A learned number inside the model that shapes how it responds to inputs.

Why does scale matter

More capacity can represent more complex patterns, but it also increases cost and failure modes.

What happens in a training loop

Predict, measure error with a loss, update parameters, then repeat.

What is a loss function used for

It turns wrongness into a signal the model can optimise during training.

What is a gradient in plain terms

A direction for how to change parameters to reduce loss.

What is overfitting

Learning training quirks and failing on new data.

What is generalisation

Performing well on new inputs, not just the training set.

Why is training expensive

It requires many passes over lots of data and many parameter updates.

Scenario. Your validation score improves but production quality drops. Name one realistic cause

Validation does not match production. For example, leakage, a biased split, different input distributions, or a feature that is available in training but not reliably available at inference time.

Artefact and reflection

Artefact

A one-page decision note with assumption, evidence, and chosen action

Reflection

Where in your work would explain models, parameters and training dynamics in your own words and apply it to a realistic scenario. change a decision, and what evidence would make you trust that change?

Optional practice

Apply structured prompt patterns (chain of thought, few-shot, role play) and see how they change model behaviour.

Also in this module

Compare prompt strategies

Put two prompts side by side and compare their structure, clarity and likely effectiveness.

See how models learn

Adjust learning rate, data size and noise to see how a simple model improves or collapses during training.

Build a neural network

Stack layers, choose activations and watch how a network learns to separate classes on a 2D canvas.