Phase Ladder
Study Plan
This is the academy's single map. Follow the phases in order if you are learning AI for the first time. Use them as a repair map if you are preparing for IOAI and entering at your weakest layer. Each phase has an exit gate — you should be able to pass it cold before moving on.
New To AI
Start At Phase 1
Do the phases in order. Do not skip. Each phase produces one workflow habit the next phase assumes.
Preparing For IOAI
Enter At Your Weakest Phase
Read the exit gate for each phase. Enter at the first one you cannot pass cold. Do not restart Phase 1 out of habit.
Already Advanced
Maintenance Mode
When every exit gate is easy, move to Beyond The Academy. The phases are a floor, not a ceiling.
Before Phase 1¶
You should already be able to:
- create and use a Python virtual environment
- run one script from the terminal
- read a short traceback
- tell the difference between a topic page, an example, and a track
If any of that is shaky, start with First Steps and then Getting Started.
First 90 Minutes¶
If you want a single short opening before committing to a full phase:
- read Getting Started
- run one small example from Examples
- inspect one artifact carefully — do not just confirm the script ran
- read the exit gate of Phase 1 below
- if you cannot pass the gate cold, start at Phase 1; otherwise start at the first phase you cannot pass
If you are preparing for IOAI, also open IOAI Competition Surface once to understand how problems are delivered, then come back here.
Phase 1 — Tooling And Inspection¶
Goal. Make arrays, tables, grouped summaries, and plotting feel mechanical enough that the first model does not break your attention.
Do all of this:
- Array Shapes and Axis Operations
- Table Inspection
- Grouped Summaries and Slice Checks
- Feature Matrix Construction
- run
academy/.venv/bin/python academy/examples/numpy-shape-sanity/shape_sanity.py - run
academy/.venv/bin/python academy/examples/course-support-inspection/quick_inspection.py - Track: Python, NumPy, Pandas, Visualization
Exit gate — move on when you can:
- explain row versus column operations without thinking
- inspect missing values and label balance from a cold table in under two minutes
- build a simple feature matrix without losing row alignment
- make one plot that changes what you inspect next
Phase 2 — Honest Validation¶
Goal. Make split discipline, baselines, cross-validation, tuning, calibration, and leakage checks feel non-negotiable. Nothing else in the academy matters until this phase is automatic.
Do all of this:
- Honest Splits and Baselines
- Leakage Patterns
- Cross-Validation
- run
academy/.venv/bin/python academy/examples/validation-baseline-comparison/baseline_comparison.py - run
academy/.venv/bin/python academy/examples/classical-ml-recipes/leakage_patterns_demo.py - Clinic: Public/Private Restraint — do it cold, write the note, then check the reveal
- Clinic: Leakage Or Signal?
- Validation, Leakage, and Model Choice — answer every question before looking at solutions
- Track: scikit-learn Validation and Tuning — full run with artifacts
Exit gate — move on when you can:
- explain why the test set must stay untouched after model selection
- name three leakage patterns without looking them up
- defend a model choice using paired fold deltas, not just the mean
- pass Sheet A: Validation and Leakage Sprint cold
Phase 3 — Model Diagnostics And Classical Choice¶
Goal. Diagnose model behavior and make justified classical model choices. Add geometry, clustering, and representation judgment without losing evaluation discipline.
Do all of this:
- Linear Regression
- Logistic Regression
- K-Nearest Neighbors
- Decision Trees
- K-Means
- PCA
- Learning Curves and Bias-Variance
- Calibration and Thresholds
- Ensemble Methods
- Evaluation Metrics Deep Dive
- Clinic: Overfit Or Underfit?
- Clinic: Ensemble Temptation
- Clinic: Threshold Under Asymmetric Cost
- run
academy/.venv/bin/python academy/examples/classical-ml-recipes/calibration_threshold_demo.py - SVM, Kernels, and Learning Theory
- Track: SVM and Advanced Clustering
Exit gate — move on when you can:
- diagnose overfit vs. underfit from a training curve in under a minute
- choose a threshold when false negative cost is 100× false positive cost
- reject an ensemble that wins by a marginal amount with high operational cost
- justify linear versus nonlinear boundaries on a given dataset
- explain when cross-validation is selection evidence vs. final evidence
Phase 4 — Deep Learning Control¶
Goal. Make training loops, checkpoints, optimizer choices, and overfitting checks feel controlled instead of mysterious. Know when the training loop itself is the bug.
Do all of this:
- PyTorch Training Loops
- Backpropagation
- Activation Functions
- Optimizers and Regularization
- Batch Normalization and Initialization
- Learning Rate Schedulers
- Debugging Deep Learning
- Clinic: Checkpoint Roulette
- run
academy/.venv/bin/python academy/examples/deep-learning-recipes/pytorch_training_loop_demo.py - run
academy/.venv/bin/python academy/examples/deep-learning-recipes/optimizer_regularization_demo.py - Lab:
academy/labs/cnn-from-scratch— receptive fields, translation equivariance, and overfit-one-batch checks - Deep Learning and Checkpoints
- Track: PyTorch Training Recipes — full run with checkpoint selection
Exit gate — move on when you can:
- describe the difference between training and evaluation mode without hesitating
- select the right checkpoint from a 15-epoch log
- explain why training loss is irrelevant for checkpoint selection
- name the first three things to check when a training run diverges
- pass Sheet C: Training and Checkpoint Sprint cold
Phase 5 — Transfer And Representation¶
Goal. Decide when to freeze, probe, fine-tune, or switch representation families instead of retraining everything from scratch. Compare backbones honestly.
Do all of this:
- Transfer and Fine-Tuning
- Convolutional Neural Networks
- Attention and Transformers
- Vision and Text Encoders
- Self-Supervised and Representation Learning
- Metric Learning and Retrieval
- Semi-Supervised Learning
- Clinic: Freeze Or Fine-Tune?
- run
academy/.venv/bin/python academy/examples/deep-learning-recipes/transfer_finetuning_demo.py - Lab:
academy/labs/self-supervised-representation— SimCLR-style pretraining, linear probe at multiple label budgets - Lab:
academy/labs/clip-multi-modal— symmetric InfoNCE, image↔text retrieval asymmetry - Lab:
academy/labs/metric-learning-retrieval— triplet / batch-hard / prototype classifiers, Recall@k on unseen identities - Lab:
academy/labs/semi-supervised-learning— pseudo-labeling vs. label spreading on a 20-labeled + 2000-unlabeled pool - Track: ResNet, BERT, and Fine-Tuning
- Track: Representation Reuse and Embedding Transfer
Exit gate — move on when you can:
- justify frozen features vs. partial fine-tuning vs. full fine-tuning given a data budget
- compare two pretrained backbones on the same downstream task with honest evaluation
- explain when fine-tuning destroys more than it gains
- pass Sheet B: Representation and Geometry Sprint cold
Phase 6 — Decisions Under Constraint¶
Goal. Practice model choice, operating points, and workflow discipline under time, budget, or leaderboard pressure. This is where IOAI readiness is built.
Do all of this:
- Baseline-First Task Solving
- Imbalanced Metrics and Review Budgets
- Selective Prediction and Review Budgets
- Experiments and Ablations
- Reliability Slices
- Clinic: Review Budget Freeze
- run
academy/.venv/bin/python academy/examples/mock-task-recipes/baseline_first_demo.py - run
academy/.venv/bin/python academy/examples/mock-task-recipes/review_budget_demo.py - Track: Vision and Audio Workflows
- Track: Mock Tasks and Timed Workflows — full run under self-imposed time constraint
- Track: Imbalanced Triage and Review Budgets
- Timed Checkpoint Sheets — all three sheets, timed, no notes
Exit gate — move on when you can:
- produce a baseline, one iteration, and a stop decision within a fixed time window
- choose a review budget policy that survives a queue constraint
- defend a stop call when the visible score still has room to improve
- reject a public leaderboard jump that local validation does not support
If you are preparing for IOAI, pair this phase with IOAI Competition Surface for problem-reading, sandbox, and submission practice.
Phase 7 — Specialized Modalities¶
Goal. Extend into specialized modalities and advanced techniques. Enter at your weakest modality — do not read linearly.
Vision And Detection¶
- Vision Augmentation and Shift Robustness
- Object Detection Basics
- Detection and Segmentation
- Density-Based Counting
- Saliency and Explainability
- Lab:
academy/labs/density-map-counting— scalar-count vs. detect-then-count vs. density regression, σ sweep - Lab:
academy/labs/saliency-and-explainability— Grad-CAM, Integrated Gradients, pointing-game scoring, rectangle-from-heatmap - Track: Detection and Segmentation Workflows
Text And Language¶
- Text Representations and Order
- Tokenization Mechanics
- Language Modeling Fundamentals
- Text Generation and Language Models
- Encoder-Decoder and Translation
- Clinic: Prompt Vs Retrieval Vs Fine-Tune — do it cold before the RAG lab
- Clinic: Published Result Trust — before anchoring your work on a benchmark number
- Black-Box LLM Optimization
- Lab:
academy/labs/encoder-decoder-translation— GRU+attention seq2seq, teacher-forced vs free-running loss, greedy vs beam decoding - Track: Text Workflows Beyond Classification
- Track: RAG And Prompting Workflows — two-loop evaluation with retrieval and rerank attribution
Generative Models¶
- Autoencoders and VAEs
- Generative Adversarial Networks
- Diffusion Models
- Steering Frozen Generative Models
- Lab:
academy/labs/autoencoder-and-vae— reconstruction vs. generation, β trade-off, posterior collapse - Lab:
academy/labs/gan-workflow— vanilla/LSGAN/WGAN-GP, mode coverage on a 4-mode mixture - Lab:
academy/labs/diffusion-models— linear vs cosine schedules, DDPM vs DDIM sampling step budgets - Lab:
academy/labs/generative-steering-and-search— textual inversion to a held-out class, embedding-direction transfer, UCB/beam search against a stochastic scorer, reward-hacking demo
Audio And Speech¶
Multi-Modal And Structured¶
Optimization And Efficiency¶
- Mixed Precision Training
- PyTorch Optimization Recipes
- Data Augmentation
- Track: Optimization, Regularization, and PEFT
Post-Model Algorithms And Theory¶
- Track: Structured Post-Model Algorithms
- Track: Problem Adaptation and Post-Processing
- Track: Theory Checkpoint Sheets
- Unsupervised Learning and Representation
When every phase is boring, move to Beyond The Academy.
Repair Map¶
If you already know your weakness, skip the phase ladder and go directly:
| Weakness | Go here |
|---|---|
| Splits and leakage are shaky | Phase 2 from the top |
| Cannot diagnose overfit vs. underfit | Clinic: Overfit Or Underfit? then Phase 3 |
| Training loops feel fragile | Phase 4 from the top |
| Transfer decisions are guesses | Clinic: Freeze Or Fine-Tune? then Phase 5 |
| Decisions collapse under time pressure | Phase 6, start with the mock task track |
| Threshold/operating-point choices are weak | Clinic: Threshold Under Asymmetric Cost then Phase 6 |
| Ensemble/complexity choices are weak | Clinic: Ensemble Temptation then Phase 3 |
| Need a specific domain (vision, text, audio) | Phase 7, pick the matching section |
| IOAI problem format is unfamiliar | IOAI Competition Surface, then Phase 6 |