Files
TinyTorch/modules/source/11_training/training_dev.ipynb
Vijay Janapa Reddi bfadc82ce6 Update generated notebooks and package exports
- Regenerate all .ipynb files from fixed .py modules
- Update tinytorch package exports with corrected implementations
- Sync package module index with current 16-module structure

These generated files reflect all the module fixes and ensure consistent
.py ↔ .ipynb conversion with the updated module implementations.
2025-09-18 16:42:57 -04:00

2175 lines
93 KiB
Plaintext

{
"cells": [
{
"cell_type": "markdown",
"id": "9722eef4",
"metadata": {
"cell_marker": "\"\"\""
},
"source": [
"# Training - Complete End-to-End ML Training Infrastructure\n",
"\n",
"Welcome to the Training module! You'll build the complete training infrastructure that orchestrates data loading, forward passes, loss computation, backpropagation, and optimization into a unified system.\n",
"\n",
"## Learning Goals\n",
"- Systems understanding: How training loops coordinate all ML system components and why training orchestration determines system reliability\n",
"- Core implementation skill: Build loss functions, evaluation metrics, and complete training loops with checkpointing and monitoring\n",
"- Pattern recognition: Understand how different loss functions affect learning dynamics and model behavior\n",
"- Framework connection: See how your training loop mirrors PyTorch's training patterns and state management\n",
"- Performance insight: Learn why training loop design affects convergence speed, memory usage, and debugging capability\n",
"\n",
"## Build → Use → Reflect\n",
"1. **Build**: Complete training infrastructure with loss functions, metrics, checkpointing, and progress monitoring\n",
"2. **Use**: Train real neural networks on CIFAR-10 and achieve meaningful accuracy on complex visual tasks\n",
"3. **Reflect**: Why does training loop design often determine the success or failure of ML projects?\n",
"\n",
"## What You'll Achieve\n",
"By the end of this module, you'll understand:\n",
"- Deep technical understanding of how training loops orchestrate complex ML systems into reliable, monitorable processes\n",
"- Practical capability to build production-ready training infrastructure with proper error handling and state management\n",
"- Systems insight into why training stability and reproducibility are critical for reliable ML systems\n",
"- Performance consideration of how training loop efficiency affects iteration speed and resource utilization\n",
"- Connection to production ML systems and how modern MLOps platforms build on these training patterns\n",
"\n",
"## Systems Reality Check\n",
"💡 **Production Context**: Modern ML training platforms like PyTorch Lightning and Hugging Face Transformers build sophisticated abstractions on top of basic training loops to handle distributed training, mixed precision, and fault tolerance\n",
"⚡ **Performance Note**: Training loop efficiency often matters more than model efficiency for development speed - good training infrastructure accelerates the entire ML development cycle"
]
},
{
"cell_type": "code",
"execution_count": null,
"id": "d79e429d",
"metadata": {
"nbgrader": {
"grade": false,
"grade_id": "training-imports",
"locked": false,
"schema_version": 3,
"solution": false,
"task": false
}
},
"outputs": [],
"source": [
"#| default_exp core.training\n",
"\n",
"#| export\n",
"import numpy as np\n",
"import sys\n",
"import os\n",
"from collections import defaultdict\n",
"import time\n",
"import pickle\n",
"\n",
"# Add module directories to Python path\n",
"sys.path.append(os.path.abspath('modules/source/02_tensor'))\n",
"sys.path.append(os.path.abspath('modules/source/03_activations'))\n",
"sys.path.append(os.path.abspath('modules/source/04_layers'))\n",
"sys.path.append(os.path.abspath('modules/source/05_dense'))\n",
"sys.path.append(os.path.abspath('modules/source/06_spatial'))\n",
"sys.path.append(os.path.abspath('modules/source/08_dataloader'))\n",
"sys.path.append(os.path.abspath('modules/source/09_autograd'))\n",
"sys.path.append(os.path.abspath('modules/source/10_optimizers'))\n",
"\n",
"# Helper function to set up import paths\n",
"# No longer needed, will use direct relative imports\n",
"\n",
"# Set up paths\n",
"# No longer needed\n",
"\n",
"# Import all the building blocks we need\n",
"from tensor_dev import Tensor\n",
"from activations_dev import ReLU, Sigmoid, Tanh, Softmax\n",
"from layers_dev import Dense\n",
"from dense_dev import Sequential, create_mlp\n",
"from spatial_dev import Conv2D, flatten\n",
"from dataloader_dev import Dataset, DataLoader\n",
"from autograd_dev import Variable\n",
"from optimizers_dev import SGD, Adam, StepLR"
]
},
{
"cell_type": "markdown",
"id": "2f3fe102",
"metadata": {
"cell_marker": "\"\"\""
},
"source": [
"## 🔧 DEVELOPMENT"
]
},
{
"cell_type": "markdown",
"id": "d29c83bd",
"metadata": {
"cell_marker": "\"\"\"",
"lines_to_next_cell": 1
},
"source": [
"## Step 1: Understanding Loss Functions\n",
"\n",
"### What are Loss Functions?\n",
"Loss functions measure how far our model's predictions are from the true values. They provide the \"signal\" that tells our optimizer which direction to update parameters.\n",
"\n",
"### The Mathematical Foundation\n",
"Training a neural network is an optimization problem:\n",
"```\n",
"θ* = argmin_θ L(f(x; θ), y)\n",
"```\n",
"Where:\n",
"- `θ` = model parameters (weights and biases)\n",
"- `f(x; θ)` = model predictions\n",
"- `y` = true labels\n",
"- `L` = loss function\n",
"- `θ*` = optimal parameters\n",
"\n",
"### Why Loss Functions Matter\n",
"- **Optimization target**: They define what \"good\" means for our model\n",
"- **Gradient source**: Provide gradients for backpropagation\n",
"- **Task-specific**: Different losses for different problems\n",
"- **Training dynamics**: Shape how the model learns\n",
"\n",
"### Common Loss Functions\n",
"\n",
"#### **Mean Squared Error (MSE)** - For Regression\n",
"```\n",
"MSE = (1/n) * Σ(y_pred - y_true)²\n",
"```\n",
"- **Use case**: Regression problems\n",
"- **Properties**: Penalizes large errors heavily\n",
"- **Gradient**: 2 * (y_pred - y_true)\n",
"\n",
"#### **Cross-Entropy Loss** - For Classification\n",
"```\n",
"CrossEntropy = -Σ y_true * log(y_pred)\n",
"```\n",
"- **Use case**: Multi-class classification\n",
"- **Properties**: Penalizes confident wrong predictions\n",
"- **Gradient**: y_pred - y_true (with softmax)\n",
"\n",
"#### **Binary Cross-Entropy** - For Binary Classification\n",
"```\n",
"BCE = -y_true * log(y_pred) - (1-y_true) * log(1-y_pred)\n",
"```\n",
"- **Use case**: Binary classification\n",
"- **Properties**: Symmetric around 0.5\n",
"- **Gradient**: (y_pred - y_true) / (y_pred * (1-y_pred))\n",
"\n",
"Let's implement these essential loss functions!"
]
},
{
"cell_type": "code",
"execution_count": null,
"id": "8efa2e22",
"metadata": {
"lines_to_next_cell": 1,
"nbgrader": {
"grade": false,
"grade_id": "mse-loss",
"locked": false,
"schema_version": 3,
"solution": true,
"task": false
}
},
"outputs": [],
"source": [
"#| export\n",
"class MeanSquaredError:\n",
" \"\"\"\n",
" Mean Squared Error Loss for Regression\n",
" \n",
" Measures the average squared difference between predictions and targets.\n",
" MSE = (1/n) * Σ(y_pred - y_true)²\n",
" \"\"\"\n",
" \n",
" def __init__(self):\n",
" \"\"\"Initialize MSE loss function.\"\"\"\n",
" pass\n",
" \n",
" def __call__(self, y_pred: Tensor, y_true: Tensor) -> Tensor:\n",
" \"\"\"\n",
" Compute MSE loss between predictions and targets.\n",
" \n",
" Args:\n",
" y_pred: Model predictions (shape: [batch_size, ...])\n",
" y_true: True targets (shape: [batch_size, ...])\n",
" \n",
" Returns:\n",
" Scalar loss value\n",
" \n",
" TODO: Implement Mean SquaredError loss computation.\n",
" \n",
" STEP-BY-STEP IMPLEMENTATION:\n",
" 1. Compute difference: diff = y_pred - y_true\n",
" 2. Square the differences: squared_diff = diff²\n",
" 3. Take mean over all elements: mean(squared_diff)\n",
" 4. Return as scalar Tensor\n",
" \n",
" EXAMPLE:\n",
" y_pred = Tensor([[1.0, 2.0], [3.0, 4.0]])\n",
" y_true = Tensor([[1.5, 2.5], [2.5, 3.5]])\n",
" loss = mse_loss(y_pred, y_true)\n",
" # Should return: mean([(1.0-1.5)², (2.0-2.5)², (3.0-2.5)², (4.0-3.5)²])\n",
" # = mean([0.25, 0.25, 0.25, 0.25]) = 0.25\n",
" \n",
" LEARNING CONNECTIONS:\n",
" - **Regression Optimization**: MSE loss guides models toward accurate numerical predictions\n",
" - **Gradient Properties**: MSE provides smooth gradients proportional to prediction error\n",
" - **Outlier Sensitivity**: Squared errors heavily penalize large mistakes\n",
" - **Production Usage**: Common in recommendation systems, time series, and financial modeling\n",
" \n",
" HINTS:\n",
" - Use tensor subtraction: y_pred - y_true\n",
" - Use tensor power: diff ** 2\n",
" - Use tensor mean: squared_diff.mean()\n",
" \"\"\"\n",
" ### BEGIN SOLUTION\n",
" diff = y_pred - y_true\n",
" squared_diff = diff * diff # Using multiplication for square\n",
" loss = np.mean(squared_diff.data)\n",
" return Tensor(loss)\n",
" ### END SOLUTION\n",
" \n",
" def forward(self, y_pred: Tensor, y_true: Tensor) -> Tensor:\n",
" \"\"\"Alternative interface for forward pass.\"\"\"\n",
" return self.__call__(y_pred, y_true)"
]
},
{
"cell_type": "markdown",
"id": "0a9c2f6b",
"metadata": {
"cell_marker": "\"\"\"",
"lines_to_next_cell": 1
},
"source": [
"### 🧪 Unit Test: MSE Loss\n",
"\n",
"Let's test our MSE loss implementation with known values."
]
},
{
"cell_type": "code",
"execution_count": null,
"id": "531d56c7",
"metadata": {
"lines_to_next_cell": 1,
"nbgrader": {
"grade": false,
"grade_id": "test-mse-loss",
"locked": false,
"schema_version": 3,
"solution": false,
"task": false
}
},
"outputs": [],
"source": [
"def test_unit_mse_loss():\n",
" \"\"\"Test MSE loss with comprehensive examples.\"\"\"\n",
" print(\"🔬 Unit Test: MSE Loss...\")\n",
" \n",
" mse = MeanSquaredError()\n",
" \n",
" # Test 1: Perfect predictions (loss should be 0)\n",
" y_pred = Tensor([[1.0, 2.0], [3.0, 4.0]])\n",
" y_true = Tensor([[1.0, 2.0], [3.0, 4.0]])\n",
" loss = mse(y_pred, y_true)\n",
" assert abs(loss.data) < 1e-6, f\"Perfect predictions should have loss ≈ 0, got {loss.data}\"\n",
" print(\"✅ Perfect predictions test passed\")\n",
" \n",
" # Test 2: Known loss computation\n",
" y_pred = Tensor([[1.0, 2.0]])\n",
" y_true = Tensor([[0.0, 1.0]])\n",
" loss = mse(y_pred, y_true)\n",
" expected = 1.0 # [(1-0)² + (2-1)²] / 2 = [1 + 1] / 2 = 1.0\n",
" assert abs(loss.data - expected) < 1e-6, f\"Expected loss {expected}, got {loss.data}\"\n",
" print(\"✅ Known loss computation test passed\")\n",
" \n",
" # Test 3: Batch processing\n",
" y_pred = Tensor([[1.0, 2.0], [3.0, 4.0]])\n",
" y_true = Tensor([[1.5, 2.5], [2.5, 3.5]])\n",
" loss = mse(y_pred, y_true)\n",
" expected = 0.25 # All squared differences are 0.25\n",
" assert abs(loss.data - expected) < 1e-6, f\"Expected batch loss {expected}, got {loss.data}\"\n",
" print(\"✅ Batch processing test passed\")\n",
" \n",
" # Test 4: Single value\n",
" y_pred = Tensor([5.0])\n",
" y_true = Tensor([3.0])\n",
" loss = mse(y_pred, y_true)\n",
" expected = 4.0 # (5-3)² = 4\n",
" assert abs(loss.data - expected) < 1e-6, f\"Expected single value loss {expected}, got {loss.data}\"\n",
" print(\"✅ Single value test passed\")\n",
" \n",
" print(\"🎯 MSE Loss: All tests passed!\")\n",
"\n",
"# Test function defined (called in main block) "
]
},
{
"cell_type": "code",
"execution_count": null,
"id": "14074504",
"metadata": {
"lines_to_next_cell": 1,
"nbgrader": {
"grade": false,
"grade_id": "crossentropy-loss",
"locked": false,
"schema_version": 3,
"solution": true,
"task": false
}
},
"outputs": [],
"source": [
"#| export\n",
"class CrossEntropyLoss:\n",
" \"\"\"\n",
" Cross-Entropy Loss for Multi-Class Classification\n",
" \n",
" Measures the difference between predicted probability distribution and true labels.\n",
" CrossEntropy = -Σ y_true * log(y_pred)\n",
" \"\"\"\n",
" \n",
" def __init__(self):\n",
" \"\"\"Initialize CrossEntropy loss function.\"\"\"\n",
" pass\n",
" \n",
" def __call__(self, y_pred: Tensor, y_true: Tensor) -> Tensor:\n",
" \"\"\"\n",
" Compute CrossEntropy loss between predictions and targets.\n",
" \n",
" Args:\n",
" y_pred: Model predictions (shape: [batch_size, num_classes])\n",
" y_true: True class indices (shape: [batch_size]) or one-hot (shape: [batch_size, num_classes])\n",
" \n",
" Returns:\n",
" Scalar loss value\n",
" \n",
" TODO: Implement Cross-Entropy loss computation.\n",
" \n",
" STEP-BY-STEP IMPLEMENTATION:\n",
" 1. Handle both class indices and one-hot encoded labels\n",
" 2. Apply softmax to predictions for probability distribution\n",
" 3. Compute log probabilities: log(softmax(y_pred))\n",
" 4. Calculate cross-entropy: -mean(y_true * log_probs)\n",
" 5. Return scalar loss\n",
" \n",
" EXAMPLE:\n",
" y_pred = Tensor([[2.0, 1.0, 0.1], [0.5, 2.1, 0.9]]) # Raw logits\n",
" y_true = Tensor([0, 1]) # Class indices\n",
" loss = crossentropy_loss(y_pred, y_true)\n",
" # Should apply softmax then compute -log(prob_of_correct_class)\n",
" \n",
" LEARNING CONNECTIONS:\n",
" - **Classification Foundation**: CrossEntropy is the standard loss for multi-class problems\n",
" - **Probability Interpretation**: Measures difference between predicted and true distributions\n",
" - **Information Theory**: Based on entropy and KL divergence concepts\n",
" - **Production Systems**: Used in image classification, NLP, and recommendation systems\n",
" \n",
" HINTS:\n",
" - Use softmax: exp(x) / sum(exp(x)) for probability distribution\n",
" - Add small epsilon (1e-15) to avoid log(0)\n",
" - Handle both class indices and one-hot encoding\n",
" - Use np.log for logarithm computation\n",
" \"\"\"\n",
" ### BEGIN SOLUTION\n",
" # Handle both 1D and 2D prediction arrays\n",
" if y_pred.data.ndim == 1:\n",
" # Reshape 1D to 2D for consistency (single sample)\n",
" y_pred_2d = y_pred.data.reshape(1, -1)\n",
" else:\n",
" y_pred_2d = y_pred.data\n",
" \n",
" # Apply softmax to get probability distribution\n",
" exp_pred = np.exp(y_pred_2d - np.max(y_pred_2d, axis=1, keepdims=True))\n",
" softmax_pred = exp_pred / np.sum(exp_pred, axis=1, keepdims=True)\n",
" \n",
" # Add small epsilon to avoid log(0)\n",
" epsilon = 1e-15\n",
" softmax_pred = np.clip(softmax_pred, epsilon, 1.0 - epsilon)\n",
" \n",
" # Handle class indices vs one-hot encoding\n",
" if len(y_true.data.shape) == 1:\n",
" # y_true contains class indices\n",
" batch_size = y_true.data.shape[0]\n",
" log_probs = np.log(softmax_pred[np.arange(batch_size), y_true.data.astype(int)])\n",
" loss = -np.mean(log_probs)\n",
" else:\n",
" # y_true is one-hot encoded\n",
" log_probs = np.log(softmax_pred)\n",
" loss = -np.mean(np.sum(y_true.data * log_probs, axis=1))\n",
" \n",
" return Tensor(loss)\n",
" ### END SOLUTION\n",
" \n",
" def forward(self, y_pred: Tensor, y_true: Tensor) -> Tensor:\n",
" \"\"\"Alternative interface for forward pass.\"\"\"\n",
" return self.__call__(y_pred, y_true)\n",
"\n",
"# Test function defined (called in main block)"
]
},
{
"cell_type": "markdown",
"id": "42426295",
"metadata": {
"cell_marker": "\"\"\"",
"lines_to_next_cell": 1
},
"source": [
"### 🧪 Unit Test: CrossEntropy Loss\n",
"\n",
"Let's test our CrossEntropy loss implementation."
]
},
{
"cell_type": "code",
"execution_count": null,
"id": "31e5f16a",
"metadata": {
"lines_to_next_cell": 1,
"nbgrader": {
"grade": false,
"grade_id": "test-crossentropy-loss",
"locked": false,
"schema_version": 3,
"solution": false,
"task": false
}
},
"outputs": [],
"source": [
"def test_unit_crossentropy_loss():\n",
" \"\"\"Test CrossEntropy loss with comprehensive examples.\"\"\"\n",
" print(\"🔬 Unit Test: CrossEntropy Loss...\")\n",
" \n",
" ce = CrossEntropyLoss()\n",
" \n",
" # Test 1: Perfect predictions\n",
" y_pred = Tensor([[10.0, 0.0, 0.0], [0.0, 10.0, 0.0]]) # Very confident correct predictions\n",
" y_true = Tensor([0, 1]) # Class indices\n",
" loss = ce(y_pred, y_true)\n",
" assert loss.data < 0.1, f\"Perfect predictions should have low loss, got {loss.data}\"\n",
" print(\"✅ Perfect predictions test passed\")\n",
" \n",
" # Test 2: Random predictions (should have higher loss)\n",
" y_pred = Tensor([[0.0, 0.0, 0.0], [0.0, 0.0, 0.0]]) # Uniform after softmax\n",
" y_true = Tensor([0, 1])\n",
" loss = ce(y_pred, y_true)\n",
" expected_random = -np.log(1.0/3.0) # log(1/num_classes) for uniform distribution\n",
" assert abs(loss.data - expected_random) < 0.1, f\"Random predictions should have loss ≈ {expected_random}, got {loss.data}\"\n",
" print(\"✅ Random predictions test passed\")\n",
" \n",
" # Test 3: Binary classification\n",
" y_pred = Tensor([[2.0, 1.0], [1.0, 2.0]])\n",
" y_true = Tensor([0, 1])\n",
" loss = ce(y_pred, y_true)\n",
" assert 0.0 < loss.data < 2.0, f\"Binary classification loss should be reasonable, got {loss.data}\"\n",
" print(\"✅ Binary classification test passed\")\n",
" \n",
" # Test 4: One-hot encoded labels\n",
" y_pred = Tensor([[2.0, 1.0, 0.0], [0.0, 2.0, 1.0]])\n",
" y_true = Tensor([[1.0, 0.0, 0.0], [0.0, 1.0, 0.0]]) # One-hot encoded\n",
" loss = ce(y_pred, y_true)\n",
" assert 0.0 < loss.data < 2.0, f\"One-hot encoded loss should be reasonable, got {loss.data}\"\n",
" print(\"✅ One-hot encoded labels test passed\")\n",
" \n",
" print(\"🎯 CrossEntropy Loss: All tests passed!\")\n",
"\n",
"# Test function defined (called in main block)"
]
},
{
"cell_type": "code",
"execution_count": null,
"id": "8b182b10",
"metadata": {
"lines_to_next_cell": 1,
"nbgrader": {
"grade": false,
"grade_id": "binary-crossentropy-loss",
"locked": false,
"schema_version": 3,
"solution": true,
"task": false
}
},
"outputs": [],
"source": [
"#| export\n",
"class BinaryCrossEntropyLoss:\n",
" \"\"\"\n",
" Binary Cross-Entropy Loss for Binary Classification\n",
" \n",
" Measures the difference between predicted probabilities and binary labels.\n",
" BCE = -y_true * log(y_pred) - (1-y_true) * log(1-y_pred)\n",
" \"\"\"\n",
" \n",
" def __init__(self):\n",
" \"\"\"Initialize Binary CrossEntropy loss function.\"\"\"\n",
" pass\n",
" \n",
" def __call__(self, y_pred: Tensor, y_true: Tensor) -> Tensor:\n",
" \"\"\"\n",
" Compute Binary CrossEntropy loss between predictions and targets.\n",
" \n",
" Args:\n",
" y_pred: Model predictions (shape: [batch_size, 1] or [batch_size])\n",
" y_true: True binary labels (shape: [batch_size, 1] or [batch_size])\n",
" \n",
" Returns:\n",
" Scalar loss value\n",
" \n",
" TODO: Implement Binary Cross-Entropy loss computation.\n",
" \n",
" STEP-BY-STEP IMPLEMENTATION:\n",
" 1. Apply sigmoid to predictions for probability values\n",
" 2. Clip probabilities to avoid log(0) and log(1)\n",
" 3. Compute: -y_true * log(y_pred) - (1-y_true) * log(1-y_pred)\n",
" 4. Take mean over batch\n",
" 5. Return scalar loss\n",
" \n",
" EXAMPLE:\n",
" y_pred = Tensor([[2.0], [0.0], [-1.0]]) # Raw logits\n",
" y_true = Tensor([[1.0], [1.0], [0.0]]) # Binary labels\n",
" loss = bce_loss(y_pred, y_true)\n",
" # Should apply sigmoid then compute binary cross-entropy\n",
" \n",
" LEARNING CONNECTIONS:\n",
" - **Binary Classification**: Standard loss for yes/no, spam/ham, fraud detection\n",
" - **Sigmoid Output**: Maps any real number to probability range [0,1]\n",
" - **Medical Diagnosis**: Common in disease detection and medical screening\n",
" - **A/B Testing**: Used for conversion prediction and user behavior modeling\n",
" \n",
" HINTS:\n",
" - Use sigmoid: 1 / (1 + exp(-x))\n",
" - Clip probabilities: np.clip(probs, epsilon, 1-epsilon)\n",
" - Handle both [batch_size] and [batch_size, 1] shapes\n",
" - Use np.log for logarithm computation\n",
" \"\"\"\n",
" ### BEGIN SOLUTION\n",
" # Use numerically stable implementation directly from logits\n",
" # This avoids computing sigmoid and log separately\n",
" logits = y_pred.data.flatten()\n",
" labels = y_true.data.flatten()\n",
" \n",
" # Numerically stable binary cross-entropy from logits\n",
" # Uses the identity: log(1 + exp(x)) = max(x, 0) + log(1 + exp(-abs(x)))\n",
" def stable_bce_with_logits(logits, labels):\n",
" # For each sample: -[y*log(sigmoid(x)) + (1-y)*log(1-sigmoid(x))]\n",
" # Which equals: -[y*log_sigmoid(x) + (1-y)*log_sigmoid(-x)]\n",
" # Where log_sigmoid(x) = x - log(1 + exp(x)) = x - softplus(x)\n",
" \n",
" # Compute log(sigmoid(x)) = x - log(1 + exp(x))\n",
" # Use numerical stability: log(1 + exp(x)) = max(0, x) + log(1 + exp(-abs(x)))\n",
" def log_sigmoid(x):\n",
" return x - np.maximum(0, x) - np.log(1 + np.exp(-np.abs(x)))\n",
" \n",
" # Compute log(1 - sigmoid(x)) = -x - log(1 + exp(-x))\n",
" def log_one_minus_sigmoid(x):\n",
" return -x - np.maximum(0, -x) - np.log(1 + np.exp(-np.abs(x)))\n",
" \n",
" # Binary cross-entropy: -[y*log_sigmoid(x) + (1-y)*log_sigmoid(-x)]\n",
" loss = -(labels * log_sigmoid(logits) + (1 - labels) * log_one_minus_sigmoid(logits))\n",
" return loss\n",
" \n",
" # Compute loss for each sample\n",
" losses = stable_bce_with_logits(logits, labels)\n",
" \n",
" # Take mean over batch\n",
" mean_loss = np.mean(losses)\n",
" \n",
" return Tensor(mean_loss)\n",
" ### END SOLUTION\n",
" \n",
" def forward(self, y_pred: Tensor, y_true: Tensor) -> Tensor:\n",
" \"\"\"Alternative interface for forward pass.\"\"\"\n",
" return self.__call__(y_pred, y_true)\n",
"\n",
"# Test function defined (called in main block)"
]
},
{
"cell_type": "markdown",
"id": "64b9a59a",
"metadata": {
"cell_marker": "\"\"\"",
"lines_to_next_cell": 1
},
"source": [
"### 🧪 Unit Test: Binary CrossEntropy Loss\n",
"\n",
"Let's test our Binary CrossEntropy loss implementation."
]
},
{
"cell_type": "code",
"execution_count": null,
"id": "9d3ddb43",
"metadata": {
"lines_to_next_cell": 1,
"nbgrader": {
"grade": false,
"grade_id": "test-binary-crossentropy-loss",
"locked": false,
"schema_version": 3,
"solution": false,
"task": false
}
},
"outputs": [],
"source": [
"def test_unit_binary_crossentropy_loss():\n",
" \"\"\"Test Binary CrossEntropy loss with comprehensive examples.\"\"\"\n",
" print(\"🔬 Unit Test: Binary CrossEntropy Loss...\")\n",
" \n",
" bce = BinaryCrossEntropyLoss()\n",
" \n",
" # Test 1: Perfect predictions\n",
" y_pred = Tensor([[10.0], [-10.0]]) # Very confident correct predictions\n",
" y_true = Tensor([[1.0], [0.0]])\n",
" loss = bce(y_pred, y_true)\n",
" assert loss.data < 0.1, f\"Perfect predictions should have low loss, got {loss.data}\"\n",
" print(\"✅ Perfect predictions test passed\")\n",
" \n",
" # Test 2: Random predictions (should have higher loss)\n",
" y_pred = Tensor([[0.0], [0.0]]) # 0.5 probability after sigmoid\n",
" y_true = Tensor([[1.0], [0.0]])\n",
" loss = bce(y_pred, y_true)\n",
" expected_random = -np.log(0.5) # log(0.5) for random guessing\n",
" assert abs(loss.data - expected_random) < 0.1, f\"Random predictions should have loss ≈ {expected_random}, got {loss.data}\"\n",
" print(\"✅ Random predictions test passed\")\n",
" \n",
" # Test 3: Batch processing\n",
" y_pred = Tensor([[1.0], [2.0], [-1.0]])\n",
" y_true = Tensor([[1.0], [1.0], [0.0]])\n",
" loss = bce(y_pred, y_true)\n",
" assert 0.0 < loss.data < 2.0, f\"Batch processing loss should be reasonable, got {loss.data}\"\n",
" print(\"✅ Batch processing test passed\")\n",
" \n",
" # Test 4: Edge cases\n",
" y_pred = Tensor([[100.0], [-100.0]]) # Extreme values\n",
" y_true = Tensor([[1.0], [0.0]])\n",
" loss = bce(y_pred, y_true)\n",
" assert loss.data < 0.1, f\"Extreme correct predictions should have low loss, got {loss.data}\"\n",
" print(\"✅ Edge cases test passed\")\n",
" \n",
" print(\"🎯 Binary CrossEntropy Loss: All tests passed!\")\n",
"\n",
"# Test function defined (called in main block) "
]
},
{
"cell_type": "markdown",
"id": "40ce7b15",
"metadata": {},
"source": [
"\"\"\"\n",
"# Step 2: Understanding Metrics\n",
"\n",
"## What are Metrics?\n",
"Metrics are measurements that help us understand how well our model is performing. Unlike loss functions, metrics are often more interpretable and align with business objectives.\n",
"\n",
"## Key Metrics for Classification\n",
"\n",
"### **Accuracy**\n",
"```\n",
"Accuracy = (Correct Predictions) / (Total Predictions)\n",
"```\n",
"- **Range**: [0, 1]\n",
"- **Interpretation**: Percentage of correct predictions\n",
"- **Good for**: Balanced datasets\n",
"\n",
"### **Precision**\n",
"```\n",
"Precision = True Positives / (True Positives + False Positives)\n",
"```\n",
"- **Range**: [0, 1]\n",
"- **Interpretation**: Of all positive predictions, how many were correct?\n",
"- **Good for**: When false positives are costly\n",
"\n",
"### **Recall (Sensitivity)**\n",
"```\n",
"Recall = True Positives / (True Positives + False Negatives)\n",
"```\n",
"- **Range**: [0, 1]\n",
"- **Interpretation**: Of all actual positives, how many did we find?\n",
"- **Good for**: When false negatives are costly\n",
"\n",
"## Key Metrics for Regression\n",
"\n",
"### **Mean Absolute Error (MAE)**\n",
"```\n",
"MAE = (1/n) * Σ|y_pred - y_true|\n",
"```\n",
"- **Range**: [0, ∞)\n",
"- **Interpretation**: Average absolute error\n",
"- **Good for**: Robust to outliers\n",
"\n",
"Let's implement these essential metrics!\n",
"\"\"\"\n",
"\n",
"Test function defined (called in main block)"
]
},
{
"cell_type": "code",
"execution_count": null,
"id": "ff9b65b9",
"metadata": {
"lines_to_next_cell": 1,
"nbgrader": {
"grade": false,
"grade_id": "accuracy-metric",
"locked": false,
"schema_version": 3,
"solution": true,
"task": false
}
},
"outputs": [],
"source": [
"#| export\n",
"class Accuracy:\n",
" \"\"\"\n",
" Accuracy Metric for Classification\n",
" \n",
" Computes the fraction of correct predictions.\n",
" Accuracy = (Correct Predictions) / (Total Predictions)\n",
" \"\"\"\n",
" \n",
" def __init__(self):\n",
" \"\"\"Initialize Accuracy metric.\"\"\"\n",
" pass\n",
" \n",
" def __call__(self, y_pred: Tensor, y_true: Tensor) -> float:\n",
" \"\"\"\n",
" Compute accuracy between predictions and targets.\n",
" \n",
" Args:\n",
" y_pred: Model predictions (shape: [batch_size, num_classes] or [batch_size])\n",
" y_true: True class labels (shape: [batch_size] or [batch_size])\n",
" \n",
" Returns:\n",
" Accuracy as a float value between 0 and 1\n",
" \n",
" TODO: Implement accuracy computation.\n",
" \n",
" STEP-BY-STEP IMPLEMENTATION:\n",
" 1. Convert predictions to class indices (argmax for multi-class)\n",
" 2. Convert true labels to class indices if needed\n",
" 3. Count correct predictions\n",
" 4. Divide by total predictions\n",
" 5. Return as float\n",
" \n",
" EXAMPLE:\n",
" y_pred = Tensor([[0.9, 0.1], [0.2, 0.8], [0.6, 0.4]]) # Probabilities\n",
" y_true = Tensor([0, 1, 0]) # True classes\n",
" accuracy = accuracy_metric(y_pred, y_true)\n",
" # Should return: 2/3 = 0.667 (first and second predictions correct)\n",
" \n",
" LEARNING CONNECTIONS:\n",
" - **Model Evaluation**: Primary metric for classification model performance\n",
" - **Business KPIs**: Often directly tied to business objectives and success metrics\n",
" - **Baseline Comparison**: Standard metric for comparing different models\n",
" - **Production Monitoring**: Real-time accuracy monitoring for model health\n",
" \n",
" HINTS:\n",
" - Use np.argmax(axis=1) for multi-class predictions\n",
" - Handle both probability and class index inputs\n",
" - Use np.mean() for averaging\n",
" - Return Python float, not Tensor\n",
" \"\"\"\n",
" ### BEGIN SOLUTION\n",
" # Convert predictions to class indices\n",
" if len(y_pred.data.shape) > 1 and y_pred.data.shape[1] > 1:\n",
" # Multi-class: use argmax\n",
" pred_classes = np.argmax(y_pred.data, axis=1)\n",
" else:\n",
" # Binary classification: threshold at 0.5\n",
" pred_classes = (y_pred.data.flatten() > 0.5).astype(int)\n",
" \n",
" # Convert true labels to class indices if needed\n",
" if len(y_true.data.shape) > 1 and y_true.data.shape[1] > 1:\n",
" # One-hot encoded\n",
" true_classes = np.argmax(y_true.data, axis=1)\n",
" else:\n",
" # Already class indices\n",
" true_classes = y_true.data.flatten().astype(int)\n",
" \n",
" # Compute accuracy\n",
" correct = np.sum(pred_classes == true_classes)\n",
" total = len(true_classes)\n",
" accuracy = correct / total\n",
" \n",
" return float(accuracy)\n",
" ### END SOLUTION\n",
" \n",
" def forward(self, y_pred: Tensor, y_true: Tensor) -> float:\n",
" \"\"\"Alternative interface for forward pass.\"\"\"\n",
" return self.__call__(y_pred, y_true)"
]
},
{
"cell_type": "markdown",
"id": "11d7f7a9",
"metadata": {
"cell_marker": "\"\"\"",
"lines_to_next_cell": 1
},
"source": [
"### 🧪 Unit Test: Accuracy Metric\n",
"\n",
"Let's test our Accuracy metric implementation."
]
},
{
"cell_type": "code",
"execution_count": null,
"id": "0fbb7dea",
"metadata": {
"lines_to_next_cell": 1,
"nbgrader": {
"grade": false,
"grade_id": "test-accuracy-metric",
"locked": false,
"schema_version": 3,
"solution": false,
"task": false
}
},
"outputs": [],
"source": [
"def test_unit_accuracy_metric():\n",
" \"\"\"Test Accuracy metric with comprehensive examples.\"\"\"\n",
" print(\"🔬 Unit Test: Accuracy Metric...\")\n",
" \n",
" accuracy = Accuracy()\n",
" \n",
" # Test 1: Perfect predictions\n",
" y_pred = Tensor([[0.9, 0.1], [0.1, 0.9], [0.8, 0.2]])\n",
" y_true = Tensor([0, 1, 0])\n",
" acc = accuracy(y_pred, y_true)\n",
" assert acc == 1.0, f\"Perfect predictions should have accuracy 1.0, got {acc}\"\n",
" print(\"✅ Perfect predictions test passed\")\n",
" \n",
" # Test 2: Half correct\n",
" y_pred = Tensor([[0.9, 0.1], [0.9, 0.1], [0.8, 0.2]]) # All predict class 0\n",
" y_true = Tensor([0, 1, 0]) # Classes: 0, 1, 0\n",
" acc = accuracy(y_pred, y_true)\n",
" expected = 2.0/3.0 # 2 out of 3 correct\n",
" assert abs(acc - expected) < 1e-6, f\"Half correct should have accuracy {expected}, got {acc}\"\n",
" print(\"✅ Half correct test passed\")\n",
" \n",
" # Test 3: Binary classification\n",
" y_pred = Tensor([[0.8], [0.3], [0.9], [0.1]]) # Predictions above/below 0.5\n",
" y_true = Tensor([1, 0, 1, 0])\n",
" acc = accuracy(y_pred, y_true)\n",
" assert acc == 1.0, f\"Binary classification should have accuracy 1.0, got {acc}\"\n",
" print(\"✅ Binary classification test passed\")\n",
" \n",
" # Test 4: Multi-class\n",
" y_pred = Tensor([[0.7, 0.2, 0.1], [0.1, 0.8, 0.1], [0.1, 0.1, 0.8]])\n",
" y_true = Tensor([0, 1, 2])\n",
" acc = accuracy(y_pred, y_true)\n",
" assert acc == 1.0, f\"Multi-class should have accuracy 1.0, got {acc}\"\n",
" print(\"✅ Multi-class test passed\")\n",
" \n",
" print(\"🎯 Accuracy Metric: All tests passed!\")\n",
"\n",
"# Test function defined (called in main block)"
]
},
{
"cell_type": "markdown",
"id": "89535c73",
"metadata": {
"cell_marker": "\"\"\"",
"lines_to_next_cell": 1
},
"source": [
"## Step 3: Building the Training Loop\n",
"\n",
"### What is a Training Loop?\n",
"A training loop is the orchestration logic that coordinates all components of neural network training:\n",
"\n",
"1. **Forward Pass**: Compute predictions\n",
"2. **Loss Computation**: Measure prediction quality\n",
"3. **Backward Pass**: Compute gradients\n",
"4. **Parameter Update**: Update model parameters\n",
"5. **Evaluation**: Compute metrics and validation performance\n",
"\n",
"### The Training Loop Architecture\n",
"```python\n",
"for epoch in range(num_epochs):\n",
" # Training phase\n",
" for batch in train_dataloader:\n",
" optimizer.zero_grad()\n",
" predictions = model(batch_x)\n",
" loss = loss_function(predictions, batch_y)\n",
" loss.backward()\n",
" optimizer.step()\n",
" \n",
" # Validation phase\n",
" for batch in val_dataloader:\n",
" predictions = model(batch_x)\n",
" val_loss = loss_function(predictions, batch_y)\n",
" accuracy = accuracy_metric(predictions, batch_y)\n",
"```\n",
"\n",
"### Why We Need a Trainer Class\n",
"- **Encapsulation**: Keeps training logic organized\n",
"- **Reusability**: Same trainer works with different models/datasets\n",
"- **Monitoring**: Built-in logging and progress tracking\n",
"- **Flexibility**: Easy to modify training behavior\n",
"\n",
"Let's build our Trainer class!"
]
},
{
"cell_type": "code",
"execution_count": null,
"id": "c8e5c58f",
"metadata": {
"lines_to_next_cell": 1,
"nbgrader": {
"grade": false,
"grade_id": "trainer-class",
"locked": false,
"schema_version": 3,
"solution": true,
"task": false
}
},
"outputs": [],
"source": [
"#| export\n",
"class Trainer:\n",
" \"\"\"\n",
" Training Loop Orchestrator\n",
" \n",
" Coordinates model training with loss functions, optimizers, and metrics.\n",
" \"\"\"\n",
" \n",
" def __init__(self, model, optimizer, loss_function, metrics=None):\n",
" \"\"\"\n",
" Initialize trainer with model and training components.\n",
" \n",
" Args:\n",
" model: Neural network model to train\n",
" optimizer: Optimizer for parameter updates\n",
" loss_function: Loss function for training\n",
" metrics: List of metrics to track (optional)\n",
" \n",
" TODO: Initialize the trainer with all necessary components.\n",
" \n",
" APPROACH:\n",
" 1. Store model, optimizer, loss function, and metrics\n",
" 2. Initialize history tracking for losses and metrics\n",
" 3. Set up training state (epoch, step counters)\n",
" 4. Prepare for training and validation loops\n",
" \n",
" EXAMPLE:\n",
" model = Sequential([Dense(10, 5), ReLU(), Dense(5, 2)])\n",
" optimizer = Adam(model.parameters, learning_rate=0.001)\n",
" loss_fn = CrossEntropyLoss()\n",
" metrics = [Accuracy()]\n",
" trainer = Trainer(model, optimizer, loss_fn, metrics)\n",
" \n",
" HINTS:\n",
" - Store all components as instance variables\n",
" - Initialize empty history dictionaries\n",
" - Set metrics to empty list if None provided\n",
" - Initialize epoch and step counters to 0\n",
" \"\"\"\n",
" ### BEGIN SOLUTION\n",
" self.model = model\n",
" self.optimizer = optimizer\n",
" self.loss_function = loss_function\n",
" self.metrics = metrics or []\n",
" \n",
" # Training history\n",
" self.history = {\n",
" 'train_loss': [],\n",
" 'val_loss': [],\n",
" 'epoch': []\n",
" }\n",
" \n",
" # Add metric history tracking\n",
" for metric in self.metrics:\n",
" metric_name = metric.__class__.__name__.lower()\n",
" self.history[f'train_{metric_name}'] = []\n",
" self.history[f'val_{metric_name}'] = []\n",
" \n",
" # Training state\n",
" self.current_epoch = 0\n",
" self.current_step = 0\n",
" ### END SOLUTION\n",
" \n",
" def train_epoch(self, dataloader):\n",
" \"\"\"\n",
" Train for one epoch on the given dataloader.\n",
" \n",
" Args:\n",
" dataloader: DataLoader containing training data\n",
" \n",
" Returns:\n",
" Dictionary with epoch training metrics\n",
" \n",
" TODO: Implement single epoch training logic.\n",
" \n",
" STEP-BY-STEP IMPLEMENTATION:\n",
" 1. Initialize epoch metrics tracking\n",
" 2. Iterate through batches in dataloader\n",
" 3. For each batch:\n",
" - Zero gradients\n",
" - Forward pass\n",
" - Compute loss\n",
" - Backward pass\n",
" - Update parameters\n",
" - Track metrics\n",
" 4. Return averaged metrics for the epoch\n",
" \n",
" LEARNING CONNECTIONS:\n",
" - **Training Loop Foundation**: Core pattern used in all deep learning frameworks\n",
" - **Gradient Accumulation**: Optimizer.zero_grad() prevents gradient accumulation bugs\n",
" - **Backpropagation**: loss.backward() computes gradients through entire network\n",
" - **Parameter Updates**: optimizer.step() applies computed gradients to model weights\n",
" \n",
" HINTS:\n",
" - Use optimizer.zero_grad() before each batch\n",
" - Call loss.backward() for gradient computation\n",
" - Use optimizer.step() for parameter updates\n",
" - Track running averages for metrics\n",
" \"\"\"\n",
" ### BEGIN SOLUTION\n",
" epoch_metrics = {'loss': 0.0}\n",
" \n",
" # Initialize metric tracking\n",
" for metric in self.metrics:\n",
" metric_name = metric.__class__.__name__.lower()\n",
" epoch_metrics[metric_name] = 0.0\n",
" \n",
" batch_count = 0\n",
" \n",
" for batch_x, batch_y in dataloader:\n",
" # Zero gradients\n",
" self.optimizer.zero_grad()\n",
" \n",
" # Forward pass\n",
" predictions = self.model(batch_x)\n",
" \n",
" # Compute loss\n",
" loss = self.loss_function(predictions, batch_y)\n",
" \n",
" # Backward pass (simplified - in real implementation would use autograd)\n",
" # loss.backward()\n",
" \n",
" # Update parameters\n",
" self.optimizer.step()\n",
" \n",
" # Track metrics\n",
" epoch_metrics['loss'] += loss.data\n",
" \n",
" for metric in self.metrics:\n",
" metric_name = metric.__class__.__name__.lower()\n",
" metric_value = metric(predictions, batch_y)\n",
" epoch_metrics[metric_name] += metric_value\n",
" \n",
" batch_count += 1\n",
" self.current_step += 1\n",
" \n",
" # Average metrics over all batches\n",
" for key in epoch_metrics:\n",
" epoch_metrics[key] /= batch_count\n",
" \n",
" return epoch_metrics\n",
" ### END SOLUTION\n",
" \n",
" def validate_epoch(self, dataloader):\n",
" \"\"\"\n",
" Validate for one epoch on the given dataloader.\n",
" \n",
" Args:\n",
" dataloader: DataLoader containing validation data\n",
" \n",
" Returns:\n",
" Dictionary with epoch validation metrics\n",
" \n",
" TODO: Implement single epoch validation logic.\n",
" \n",
" STEP-BY-STEP IMPLEMENTATION:\n",
" 1. Initialize epoch metrics tracking\n",
" 2. Iterate through batches in dataloader\n",
" 3. For each batch:\n",
" - Forward pass (no gradient computation)\n",
" - Compute loss\n",
" - Track metrics\n",
" 4. Return averaged metrics for the epoch\n",
" \n",
" LEARNING CONNECTIONS:\n",
" - **Model Evaluation**: Validation measures generalization to unseen data\n",
" - **Overfitting Detection**: Comparing train vs validation metrics reveals overfitting\n",
" - **Model Selection**: Validation metrics guide hyperparameter tuning and architecture choices\n",
" - **Early Stopping**: Validation loss plateaus indicate optimal training duration\n",
" \n",
" HINTS:\n",
" - No gradient computation needed for validation\n",
" - No parameter updates during validation\n",
" - Similar to train_epoch but simpler\n",
" \"\"\"\n",
" ### BEGIN SOLUTION\n",
" epoch_metrics = {'loss': 0.0}\n",
" \n",
" # Initialize metric tracking\n",
" for metric in self.metrics:\n",
" metric_name = metric.__class__.__name__.lower()\n",
" epoch_metrics[metric_name] = 0.0\n",
" \n",
" batch_count = 0\n",
" \n",
" for batch_x, batch_y in dataloader:\n",
" # Forward pass only (no gradients needed)\n",
" predictions = self.model(batch_x)\n",
" \n",
" # Compute loss\n",
" loss = self.loss_function(predictions, batch_y)\n",
" \n",
" # Track metrics\n",
" epoch_metrics['loss'] += loss.data\n",
" \n",
" for metric in self.metrics:\n",
" metric_name = metric.__class__.__name__.lower()\n",
" metric_value = metric(predictions, batch_y)\n",
" epoch_metrics[metric_name] += metric_value\n",
" \n",
" batch_count += 1\n",
" \n",
" # Average metrics over all batches\n",
" for key in epoch_metrics:\n",
" epoch_metrics[key] /= batch_count\n",
" \n",
" return epoch_metrics\n",
" ### END SOLUTION\n",
" \n",
" def fit(self, train_dataloader, val_dataloader=None, epochs=10, verbose=True, save_best=False, checkpoint_path=\"best_model.pkl\"):\n",
" \"\"\"\n",
" Train the model for specified number of epochs.\n",
" \n",
" Args:\n",
" train_dataloader: Training data\n",
" val_dataloader: Validation data (optional)\n",
" epochs: Number of training epochs\n",
" verbose: Whether to print training progress\n",
" \n",
" Returns:\n",
" Training history dictionary\n",
" \n",
" TODO: Implement complete training loop.\n",
" \n",
" STEP-BY-STEP IMPLEMENTATION:\n",
" 1. Loop through epochs\n",
" 2. For each epoch:\n",
" - Train on training data\n",
" - Validate on validation data (if provided)\n",
" - Update history\n",
" - Print progress (if verbose)\n",
" 3. Return complete training history\n",
" \n",
" LEARNING CONNECTIONS:\n",
" - **Epoch Management**: Organizing training into discrete passes through the dataset\n",
" - **Learning Curves**: History tracking enables visualization of training progress\n",
" - **Hyperparameter Tuning**: Training history guides learning rate and architecture decisions\n",
" - **Production Monitoring**: Training logs provide debugging and optimization insights\n",
" \n",
" HINTS:\n",
" - Use train_epoch() and validate_epoch() methods\n",
" - Update self.history with results\n",
" - Print epoch summary if verbose=True\n",
" \"\"\"\n",
" ### BEGIN SOLUTION\n",
" print(f\"Starting training for {epochs} epochs...\")\n",
" best_val_loss = float('inf')\n",
" \n",
" for epoch in range(epochs):\n",
" self.current_epoch = epoch\n",
" \n",
" # Training phase\n",
" train_metrics = self.train_epoch(train_dataloader)\n",
" \n",
" # Validation phase\n",
" val_metrics = {}\n",
" if val_dataloader is not None:\n",
" val_metrics = self.validate_epoch(val_dataloader)\n",
" \n",
" # Update history\n",
" self.history['epoch'].append(epoch)\n",
" self.history['train_loss'].append(train_metrics['loss'])\n",
" \n",
" if val_dataloader is not None:\n",
" self.history['val_loss'].append(val_metrics['loss'])\n",
" \n",
" # Update metric history\n",
" for metric in self.metrics:\n",
" metric_name = metric.__class__.__name__.lower()\n",
" self.history[f'train_{metric_name}'].append(train_metrics[metric_name])\n",
" if val_dataloader is not None:\n",
" self.history[f'val_{metric_name}'].append(val_metrics[metric_name])\n",
" \n",
" # Save best model checkpoint\n",
" if save_best and val_dataloader is not None:\n",
" if val_metrics['loss'] < best_val_loss:\n",
" best_val_loss = val_metrics['loss']\n",
" self.save_checkpoint(checkpoint_path)\n",
" if verbose:\n",
" print(f\" 💾 Saved best model (val_loss: {best_val_loss:.4f})\")\n",
" \n",
" # Print progress\n",
" if verbose:\n",
" train_loss = train_metrics['loss']\n",
" print(f\"Epoch {epoch+1}/{epochs} - train_loss: {train_loss:.4f}\", end=\"\")\n",
" \n",
" if val_dataloader is not None:\n",
" val_loss = val_metrics['loss']\n",
" print(f\" - val_loss: {val_loss:.4f}\", end=\"\")\n",
" \n",
" for metric in self.metrics:\n",
" metric_name = metric.__class__.__name__.lower()\n",
" train_metric = train_metrics[metric_name]\n",
" print(f\" - train_{metric_name}: {train_metric:.4f}\", end=\"\")\n",
" \n",
" if val_dataloader is not None:\n",
" val_metric = val_metrics[metric_name]\n",
" print(f\" - val_{metric_name}: {val_metric:.4f}\", end=\"\")\n",
" \n",
" print() # New line\n",
" \n",
" print(\"Training completed!\")\n",
" return self.history\n",
" ### END SOLUTION\n",
" \n",
" def save_checkpoint(self, filepath):\n",
" \"\"\"Save model checkpoint.\"\"\"\n",
" checkpoint = {\n",
" 'epoch': self.current_epoch,\n",
" 'model_state': self._get_model_state(),\n",
" 'history': self.history\n",
" }\n",
" \n",
" with open(filepath, 'wb') as f:\n",
" pickle.dump(checkpoint, f)\n",
" \n",
" def load_checkpoint(self, filepath):\n",
" \"\"\"Load model checkpoint.\"\"\"\n",
" with open(filepath, 'rb') as f:\n",
" checkpoint = pickle.load(f)\n",
" \n",
" self.current_epoch = checkpoint['epoch']\n",
" self.history = checkpoint['history']\n",
" self._set_model_state(checkpoint['model_state'])\n",
" \n",
" print(f\"✅ Loaded checkpoint from epoch {self.current_epoch}\")\n",
" \n",
" def _get_model_state(self):\n",
" \"\"\"Extract model parameters.\"\"\"\n",
" state = {}\n",
" for i, layer in enumerate(self.model.layers):\n",
" if hasattr(layer, 'weight'):\n",
" state[f'layer_{i}_weight'] = layer.weight.data.copy()\n",
" state[f'layer_{i}_bias'] = layer.bias.data.copy()\n",
" return state\n",
" \n",
" def _set_model_state(self, state):\n",
" \"\"\"Restore model parameters.\"\"\"\n",
" for i, layer in enumerate(self.model.layers):\n",
" if hasattr(layer, 'weight'):\n",
" layer.weight.data = state[f'layer_{i}_weight']\n",
" layer.bias.data = state[f'layer_{i}_bias']"
]
},
{
"cell_type": "markdown",
"id": "c3c15b00",
"metadata": {
"cell_marker": "\"\"\"",
"lines_to_next_cell": 1
},
"source": [
"### 🧪 Unit Test: Training Loop\n",
"\n",
"Let's test our Trainer class with a simple example."
]
},
{
"cell_type": "code",
"execution_count": null,
"id": "ba33e0d4",
"metadata": {
"lines_to_next_cell": 1,
"nbgrader": {
"grade": false,
"grade_id": "test-trainer",
"locked": false,
"schema_version": 3,
"solution": false,
"task": false
}
},
"outputs": [],
"source": [
"def test_unit_trainer():\n",
" \"\"\"Test Trainer class with comprehensive examples.\"\"\"\n",
" print(\"🔬 Unit Test: Trainer Class...\")\n",
" \n",
" # Create simple model and components\n",
" model = Sequential([Dense(2, 3), ReLU(), Dense(3, 2)]) # Simple model\n",
" optimizer = SGD([], learning_rate=0.01) # Empty parameters list for testing\n",
" loss_fn = MeanSquaredError()\n",
" metrics = [Accuracy()]\n",
" \n",
" # Create trainer\n",
" trainer = Trainer(model, optimizer, loss_fn, metrics)\n",
" \n",
" # Test 1: Trainer initialization\n",
" assert trainer.model is model, \"Model should be stored correctly\"\n",
" assert trainer.optimizer is optimizer, \"Optimizer should be stored correctly\"\n",
" assert trainer.loss_function is loss_fn, \"Loss function should be stored correctly\"\n",
" assert len(trainer.metrics) == 1, \"Metrics should be stored correctly\"\n",
" assert 'train_loss' in trainer.history, \"Training history should be initialized\"\n",
" print(\"✅ Trainer initialization test passed\")\n",
" \n",
" # Test 2: History structure\n",
" assert 'epoch' in trainer.history, \"History should track epochs\"\n",
" assert 'train_accuracy' in trainer.history, \"History should track training accuracy\"\n",
" assert 'val_accuracy' in trainer.history, \"History should track validation accuracy\"\n",
" print(\"✅ History structure test passed\")\n",
" \n",
" # Test 3: Training state\n",
" assert trainer.current_epoch == 0, \"Current epoch should start at 0\"\n",
" assert trainer.current_step == 0, \"Current step should start at 0\"\n",
" print(\"✅ Training state test passed\")\n",
" \n",
" print(\"🎯 Trainer Class: All tests passed!\")\n",
"\n",
"# Test function defined (called in main block)"
]
},
{
"cell_type": "markdown",
"id": "d3b578a7",
"metadata": {
"cell_marker": "\"\"\"",
"lines_to_next_cell": 1
},
"source": [
"### 🧪 Unit Test: Complete Training Comprehensive Test\n",
"\n",
"Let's test the complete training pipeline with all components working together.\n",
"\n",
"**This is a comprehensive test** - it tests all training components working together in a realistic scenario."
]
},
{
"cell_type": "code",
"execution_count": null,
"id": "f9db1638",
"metadata": {
"lines_to_next_cell": 1,
"nbgrader": {
"grade": true,
"grade_id": "test-training-comprehensive",
"locked": true,
"points": 25,
"schema_version": 3,
"solution": false,
"task": false
}
},
"outputs": [],
"source": [
"def test_module_training():\n",
" \"\"\"Test complete training pipeline with all components.\"\"\"\n",
" print(\"🔬 Integration Test: Complete Training Pipeline...\")\n",
" \n",
" try:\n",
" # Test 1: Loss functions work correctly\n",
" mse = MeanSquaredError()\n",
" ce = CrossEntropyLoss()\n",
" bce = BinaryCrossEntropyLoss()\n",
" \n",
" # MSE test\n",
" y_pred = Tensor([[1.0, 2.0]])\n",
" y_true = Tensor([[1.0, 2.0]])\n",
" loss = mse(y_pred, y_true)\n",
" assert abs(loss.data) < 1e-6, \"MSE should work for perfect predictions\"\n",
" \n",
" # CrossEntropy test\n",
" y_pred = Tensor([[10.0, 0.0], [0.0, 10.0]])\n",
" y_true = Tensor([0, 1])\n",
" loss = ce(y_pred, y_true)\n",
" assert loss.data < 1.0, \"CrossEntropy should work for good predictions\"\n",
" \n",
" # Binary CrossEntropy test\n",
" y_pred = Tensor([[10.0], [-10.0]])\n",
" y_true = Tensor([[1.0], [0.0]])\n",
" loss = bce(y_pred, y_true)\n",
" assert loss.data < 1.0, \"Binary CrossEntropy should work for good predictions\"\n",
" \n",
" print(\"✅ Loss functions work correctly\")\n",
" \n",
" # Test 2: Metrics work correctly\n",
" accuracy = Accuracy()\n",
" \n",
" y_pred = Tensor([[0.9, 0.1], [0.1, 0.9]])\n",
" y_true = Tensor([0, 1])\n",
" acc = accuracy(y_pred, y_true)\n",
" assert acc == 1.0, \"Accuracy should work for perfect predictions\"\n",
" \n",
" print(\"✅ Metrics work correctly\")\n",
" \n",
" # Test 3: Trainer integrates all components\n",
" model = Sequential([]) # Empty model for testing\n",
" optimizer = SGD([], learning_rate=0.01)\n",
" loss_fn = MeanSquaredError()\n",
" metrics = [Accuracy()]\n",
" \n",
" trainer = Trainer(model, optimizer, loss_fn, metrics)\n",
" \n",
" # Check trainer setup\n",
" assert trainer.model is model, \"Trainer should store model\"\n",
" assert trainer.optimizer is optimizer, \"Trainer should store optimizer\"\n",
" assert trainer.loss_function is loss_fn, \"Trainer should store loss function\"\n",
" assert len(trainer.metrics) == 1, \"Trainer should store metrics\"\n",
" \n",
" print(\"✅ Trainer integrates all components\")\n",
" \n",
" print(\"🎉 Complete training pipeline works correctly!\")\n",
" \n",
" # Test 4: Integration works end-to-end\n",
" print(\"✅ End-to-end integration successful\")\n",
" \n",
" except Exception as e:\n",
" print(f\"❌ Training pipeline test failed: {e}\")\n",
" raise\n",
" \n",
" print(\"🎯 Training Pipeline: All comprehensive tests passed!\")\n",
"\n",
"# Test function defined (called in main block)"
]
},
{
"cell_type": "markdown",
"id": "456150ec",
"metadata": {
"cell_marker": "\"\"\"",
"lines_to_next_cell": 1
},
"source": [
"## Step 4: ML Systems Thinking - Production Training Pipeline Analysis\n",
"\n",
"### 🏗️ Training Infrastructure at Scale\n",
"\n",
"Your training loop implementation provides the foundation for understanding how production ML systems orchestrate the entire training pipeline. Let's analyze the systems engineering challenges that arise when training models at scale.\n",
"\n",
"#### **Training Pipeline Architecture**\n",
"```python\n",
"class ProductionTrainingPipeline:\n",
" def __init__(self):\n",
" # Resource allocation and distributed coordination\n",
" self.gpu_memory_pool = GPUMemoryManager()\n",
" self.distributed_coordinator = DistributedTrainingCoordinator() \n",
" self.checkpoint_manager = CheckpointManager()\n",
" self.metrics_aggregator = MetricsAggregator()\n",
"```\n",
"\n",
"Real training systems must handle:\n",
"- **Multi-GPU coordination**: Synchronizing gradients across devices\n",
"- **Memory management**: Optimizing batch sizes for available GPU memory\n",
"- **Fault tolerance**: Recovering from hardware failures during long training runs\n",
"- **Resource scheduling**: Balancing compute, memory, and I/O across the cluster"
]
},
{
"cell_type": "code",
"execution_count": null,
"id": "604fbb39",
"metadata": {
"lines_to_next_cell": 1,
"nbgrader": {
"grade": false,
"grade_id": "training-pipeline-profiler",
"locked": false,
"schema_version": 3,
"solution": true,
"task": false
}
},
"outputs": [],
"source": [
"#| export\n",
"class TrainingPipelineProfiler:\n",
" \"\"\"\n",
" Production Training Pipeline Analysis and Optimization\n",
" \n",
" Monitors end-to-end training performance and identifies bottlenecks\n",
" across the complete training infrastructure.\n",
" \"\"\"\n",
" \n",
" def __init__(self, warning_threshold_seconds=5.0):\n",
" \"\"\"\n",
" Initialize training pipeline profiler.\n",
" \n",
" Args:\n",
" warning_threshold_seconds: Warn if any pipeline step exceeds this time\n",
" \"\"\"\n",
" self.warning_threshold = warning_threshold_seconds\n",
" self.profiling_data = defaultdict(list)\n",
" self.resource_usage = defaultdict(list)\n",
" \n",
" def profile_complete_training_step(self, model, dataloader, optimizer, loss_fn, batch_size=32):\n",
" \"\"\"\n",
" Profile complete training step including all pipeline components.\n",
" \n",
" TODO: Implement comprehensive training step profiling.\n",
" \n",
" STEP-BY-STEP IMPLEMENTATION:\n",
" 1. Time each component: data loading, forward pass, loss computation, backward pass, optimization\n",
" 2. Monitor memory usage throughout the pipeline\n",
" 3. Calculate throughput metrics (samples/second, batches/second)\n",
" 4. Identify pipeline bottlenecks and optimization opportunities\n",
" 5. Generate performance recommendations\n",
" \n",
" EXAMPLE:\n",
" profiler = TrainingPipelineProfiler()\n",
" step_metrics = profiler.profile_complete_training_step(model, dataloader, optimizer, loss_fn)\n",
" \n",
" LEARNING CONNECTIONS:\n",
" - **Performance Optimization**: Identifying bottlenecks in training pipeline\n",
" - **Resource Planning**: Understanding memory and compute requirements\n",
" - **Hardware Selection**: Data guides GPU vs CPU trade-offs\n",
" - **Production Scaling**: Optimizing training throughput for large models\n",
" print(f\"Training throughput: {step_metrics['samples_per_second']:.1f} samples/sec\")\n",
" \n",
" HINTS:\n",
" - Use time.time() for timing measurements\n",
" - Monitor before/after memory usage\n",
" - Calculate ratios: compute_time / total_time\n",
" - Identify which step is the bottleneck\n",
" \"\"\"\n",
" ### BEGIN SOLUTION\n",
" import time\n",
" \n",
" # Initialize timing and memory tracking\n",
" step_times = {}\n",
" memory_usage = {}\n",
" \n",
" # Get initial memory baseline (simplified - in production would use GPU monitoring)\n",
" baseline_memory = self._estimate_memory_usage()\n",
" \n",
" # 1. Data Loading Phase\n",
" data_start = time.time()\n",
" try:\n",
" batch_x, batch_y = next(iter(dataloader))\n",
" data_time = time.time() - data_start\n",
" step_times['data_loading'] = data_time\n",
" except:\n",
" # Handle case where dataloader is not iterable for testing\n",
" data_time = 0.001 # Minimal time for testing\n",
" step_times['data_loading'] = data_time\n",
" batch_x = Tensor(np.random.randn(batch_size, 10))\n",
" batch_y = Tensor(np.random.randint(0, 2, batch_size))\n",
" \n",
" memory_usage['after_data_loading'] = self._estimate_memory_usage()\n",
" \n",
" # 2. Forward Pass Phase\n",
" forward_start = time.time()\n",
" try:\n",
" predictions = model(batch_x)\n",
" forward_time = time.time() - forward_start\n",
" step_times['forward_pass'] = forward_time\n",
" except:\n",
" # Handle case for testing with simplified model\n",
" forward_time = 0.002\n",
" step_times['forward_pass'] = forward_time\n",
" predictions = Tensor(np.random.randn(batch_size, 2))\n",
" \n",
" memory_usage['after_forward_pass'] = self._estimate_memory_usage()\n",
" \n",
" # 3. Loss Computation Phase\n",
" loss_start = time.time()\n",
" loss = loss_fn(predictions, batch_y)\n",
" loss_time = time.time() - loss_start\n",
" step_times['loss_computation'] = loss_time\n",
" \n",
" memory_usage['after_loss_computation'] = self._estimate_memory_usage()\n",
" \n",
" # 4. Backward Pass Phase (simplified for testing)\n",
" backward_start = time.time()\n",
" # In real implementation: loss.backward()\n",
" backward_time = 0.003 # Simulated backward pass time\n",
" step_times['backward_pass'] = backward_time\n",
" \n",
" memory_usage['after_backward_pass'] = self._estimate_memory_usage()\n",
" \n",
" # 5. Optimization Phase\n",
" optimization_start = time.time()\n",
" try:\n",
" optimizer.step()\n",
" optimization_time = time.time() - optimization_start\n",
" step_times['optimization'] = optimization_time\n",
" except:\n",
" # Handle case for testing\n",
" optimization_time = 0.001\n",
" step_times['optimization'] = optimization_time\n",
" \n",
" memory_usage['after_optimization'] = self._estimate_memory_usage()\n",
" \n",
" # Calculate total time and throughput\n",
" total_time = sum(step_times.values())\n",
" samples_per_second = batch_size / total_time if total_time > 0 else 0\n",
" \n",
" # Identify bottleneck\n",
" bottleneck_step = max(step_times.items(), key=lambda x: x[1])\n",
" \n",
" # Calculate component percentages\n",
" component_percentages = {\n",
" step: (time_taken / total_time * 100) if total_time > 0 else 0\n",
" for step, time_taken in step_times.items()\n",
" }\n",
" \n",
" # Generate performance analysis\n",
" performance_analysis = self._analyze_pipeline_performance(step_times, memory_usage, component_percentages)\n",
" \n",
" # Store profiling data\n",
" self.profiling_data['total_time'].append(total_time)\n",
" self.profiling_data['samples_per_second'].append(samples_per_second)\n",
" self.profiling_data['bottleneck_step'].append(bottleneck_step[0])\n",
" \n",
" return {\n",
" 'step_times': step_times,\n",
" 'total_time': total_time,\n",
" 'samples_per_second': samples_per_second,\n",
" 'bottleneck_step': bottleneck_step[0],\n",
" 'bottleneck_time': bottleneck_step[1],\n",
" 'component_percentages': component_percentages,\n",
" 'memory_usage': memory_usage,\n",
" 'performance_analysis': performance_analysis\n",
" }\n",
" ### END SOLUTION\n",
" \n",
" def _estimate_memory_usage(self):\n",
" \"\"\"Estimate current memory usage (simplified implementation).\"\"\"\n",
" # In production: would use psutil.Process().memory_info().rss or GPU monitoring\n",
" import sys\n",
" return sys.getsizeof({}) * 1024 # Simplified estimate\n",
" \n",
" def _analyze_pipeline_performance(self, step_times, memory_usage, component_percentages):\n",
" \"\"\"Analyze training pipeline performance and generate recommendations.\"\"\"\n",
" analysis = []\n",
" \n",
" # Identify performance bottlenecks\n",
" max_step = max(step_times.items(), key=lambda x: x[1])\n",
" if max_step[1] > self.warning_threshold:\n",
" analysis.append(f\"⚠️ BOTTLENECK: {max_step[0]} taking {max_step[1]:.3f}s (>{self.warning_threshold}s threshold)\")\n",
" \n",
" # Analyze component balance\n",
" forward_pct = component_percentages.get('forward_pass', 0)\n",
" backward_pct = component_percentages.get('backward_pass', 0)\n",
" data_pct = component_percentages.get('data_loading', 0)\n",
" \n",
" if data_pct > 30:\n",
" analysis.append(\"📊 Data loading is >30% of total time - consider data pipeline optimization\")\n",
" \n",
" if forward_pct > 60:\n",
" analysis.append(\"🔄 Forward pass dominates (>60%) - consider model optimization or batch size tuning\")\n",
" \n",
" # Memory analysis\n",
" memory_keys = list(memory_usage.keys())\n",
" if len(memory_keys) > 1:\n",
" memory_growth = memory_usage[memory_keys[-1]] - memory_usage[memory_keys[0]]\n",
" if memory_growth > 1024 * 1024: # > 1MB growth\n",
" analysis.append(\"💾 Significant memory growth during training step - monitor for memory leaks\")\n",
" \n",
" return analysis"
]
},
{
"cell_type": "markdown",
"id": "8eb31853",
"metadata": {
"cell_marker": "\"\"\"",
"lines_to_next_cell": 1
},
"source": [
"### 🧪 Test: Training Pipeline Profiling\n",
"\n",
"Let's test our training pipeline profiler with a realistic training scenario."
]
},
{
"cell_type": "code",
"execution_count": null,
"id": "ec159c89",
"metadata": {
"lines_to_next_cell": 1,
"nbgrader": {
"grade": false,
"grade_id": "test-training-pipeline-profiler",
"locked": false,
"schema_version": 3,
"solution": false,
"task": false
}
},
"outputs": [],
"source": [
"def test_training_pipeline_profiler():\n",
" \"\"\"Test training pipeline profiler with comprehensive scenarios.\"\"\"\n",
" print(\"🔬 Unit Test: Training Pipeline Profiler...\")\n",
" \n",
" profiler = TrainingPipelineProfiler(warning_threshold_seconds=1.0)\n",
" \n",
" # Create test components\n",
" model = Sequential([Dense(10, 5), ReLU(), Dense(5, 2)])\n",
" optimizer = SGD([], learning_rate=0.01)\n",
" loss_fn = MeanSquaredError()\n",
" \n",
" # Create simple test dataloader\n",
" class TestDataLoader:\n",
" def __iter__(self):\n",
" return self\n",
" def __next__(self):\n",
" return Tensor(np.random.randn(32, 10)), Tensor(np.random.randint(0, 2, 32))\n",
" \n",
" dataloader = TestDataLoader()\n",
" \n",
" # Test training step profiling\n",
" metrics = profiler.profile_complete_training_step(model, dataloader, optimizer, loss_fn, batch_size=32)\n",
" \n",
" # Verify profiling results\n",
" assert 'step_times' in metrics, \"Should track step times\"\n",
" assert 'total_time' in metrics, \"Should track total time\"\n",
" assert 'samples_per_second' in metrics, \"Should calculate throughput\"\n",
" assert 'bottleneck_step' in metrics, \"Should identify bottleneck\"\n",
" assert 'performance_analysis' in metrics, \"Should provide performance analysis\"\n",
" \n",
" # Verify all pipeline steps are profiled\n",
" expected_steps = ['data_loading', 'forward_pass', 'loss_computation', 'backward_pass', 'optimization']\n",
" for step in expected_steps:\n",
" assert step in metrics['step_times'], f\"Should profile {step}\"\n",
" assert metrics['step_times'][step] >= 0, f\"Step time should be non-negative for {step}\"\n",
" \n",
" # Verify throughput calculation\n",
" assert metrics['samples_per_second'] >= 0, \"Throughput should be non-negative\"\n",
" \n",
" # Verify component percentages\n",
" total_percentage = sum(metrics['component_percentages'].values())\n",
" assert abs(total_percentage - 100.0) < 1.0, f\"Component percentages should sum to ~100%, got {total_percentage}\"\n",
" \n",
" print(\"✅ Training pipeline profiling test passed\")\n",
" \n",
" # Test performance analysis\n",
" assert isinstance(metrics['performance_analysis'], list), \"Performance analysis should be a list\"\n",
" print(\"✅ Performance analysis generation test passed\")\n",
" \n",
" print(\"🎯 Training Pipeline Profiler: All tests passed!\")\n",
"\n",
"# Test function defined (called in main block)"
]
},
{
"cell_type": "code",
"execution_count": null,
"id": "bba90077",
"metadata": {
"lines_to_next_cell": 1,
"nbgrader": {
"grade": false,
"grade_id": "production-training-optimizer",
"locked": false,
"schema_version": 3,
"solution": true,
"task": false
}
},
"outputs": [],
"source": [
"#| export\n",
"class ProductionTrainingOptimizer:\n",
" \"\"\"\n",
" Production Training Pipeline Optimization\n",
" \n",
" Optimizes training pipelines for production deployment with focus on\n",
" throughput, resource utilization, and system stability.\n",
" \"\"\"\n",
" \n",
" def __init__(self):\n",
" \"\"\"Initialize production training optimizer.\"\"\"\n",
" self.optimization_history = []\n",
" self.baseline_metrics = None\n",
" \n",
" def optimize_batch_size_for_throughput(self, model, loss_fn, optimizer, initial_batch_size=32, max_batch_size=512):\n",
" \"\"\"\n",
" Find optimal batch size for maximum training throughput.\n",
" \n",
" TODO: Implement batch size optimization for production throughput.\n",
" \n",
" STEP-BY-STEP IMPLEMENTATION:\n",
" 1. Test range of batch sizes from initial to maximum\n",
" 2. For each batch size, measure:\n",
" - Training throughput (samples/second)\n",
" - Memory usage\n",
" - Time per step\n",
" 3. Find optimal batch size balancing throughput and memory\n",
" 4. Handle memory limitations gracefully\n",
" 5. Return recommendations with trade-off analysis\n",
" \n",
" EXAMPLE:\n",
" optimizer = ProductionTrainingOptimizer()\n",
" optimal_config = optimizer.optimize_batch_size_for_throughput(model, loss_fn, optimizer)\n",
" print(f\"Optimal batch size: {optimal_config['batch_size']}\")\n",
" \n",
" LEARNING CONNECTIONS:\n",
" - **Memory vs Throughput**: Larger batches improve GPU utilization but use more memory\n",
" - **Hardware Optimization**: Optimal batch size depends on GPU memory and compute units\n",
" - **Training Dynamics**: Batch size affects gradient noise and convergence behavior\n",
" - **Production Cost**: Throughput optimization directly impacts cloud computing costs\n",
" print(f\"Expected throughput: {optimal_config['throughput']:.1f} samples/sec\")\n",
" \n",
" HINTS:\n",
" - Test powers of 2: 32, 64, 128, 256, 512\n",
" - Monitor memory usage to avoid OOM\n",
" - Calculate samples_per_second for each batch size\n",
" - Consider memory efficiency (throughput per MB)\n",
" \"\"\"\n",
" ### BEGIN SOLUTION\n",
" print(\"🔧 Optimizing batch size for production throughput...\")\n",
" \n",
" # Test batch sizes (powers of 2 for optimal GPU utilization)\n",
" test_batch_sizes = []\n",
" current_batch = initial_batch_size\n",
" while current_batch <= max_batch_size:\n",
" test_batch_sizes.append(current_batch)\n",
" current_batch *= 2\n",
" \n",
" optimization_results = []\n",
" profiler = TrainingPipelineProfiler()\n",
" \n",
" for batch_size in test_batch_sizes:\n",
" print(f\" Testing batch size: {batch_size}\")\n",
" \n",
" try:\n",
" # Create test data for this batch size\n",
" test_x = Tensor(np.random.randn(batch_size, 10))\n",
" test_y = Tensor(np.random.randint(0, 2, batch_size))\n",
" \n",
" # Create mock dataloader\n",
" class MockDataLoader:\n",
" def __init__(self, x, y):\n",
" self.x, self.y = x, y\n",
" def __iter__(self):\n",
" return self\n",
" def __next__(self):\n",
" return self.x, self.y\n",
" \n",
" dataloader = MockDataLoader(test_x, test_y)\n",
" \n",
" # Profile training step\n",
" metrics = profiler.profile_complete_training_step(\n",
" model, dataloader, optimizer, loss_fn, batch_size\n",
" )\n",
" \n",
" # Estimate memory usage (simplified)\n",
" estimated_memory_mb = batch_size * 10 * 4 / (1024 * 1024) # 4 bytes per float\n",
" memory_efficiency = metrics['samples_per_second'] / estimated_memory_mb if estimated_memory_mb > 0 else 0\n",
" \n",
" optimization_results.append({\n",
" 'batch_size': batch_size,\n",
" 'throughput': metrics['samples_per_second'],\n",
" 'total_time': metrics['total_time'],\n",
" 'estimated_memory_mb': estimated_memory_mb,\n",
" 'memory_efficiency': memory_efficiency,\n",
" 'bottleneck_step': metrics['bottleneck_step']\n",
" })\n",
" \n",
" except Exception as e:\n",
" print(f\" ⚠️ Batch size {batch_size} failed: {e}\")\n",
" # In production, this would typically be OOM\n",
" break\n",
" \n",
" # Find optimal configuration\n",
" if not optimization_results:\n",
" return {'error': 'No valid batch sizes found'}\n",
" \n",
" # Optimal = highest throughput that doesn't exceed memory limits\n",
" best_config = max(optimization_results, key=lambda x: x['throughput'])\n",
" \n",
" # Generate optimization analysis\n",
" analysis = self._generate_batch_size_analysis(optimization_results, best_config)\n",
" \n",
" # Store optimization history\n",
" self.optimization_history.append({\n",
" 'optimization_type': 'batch_size',\n",
" 'results': optimization_results,\n",
" 'best_config': best_config,\n",
" 'analysis': analysis\n",
" })\n",
" \n",
" return {\n",
" 'optimal_batch_size': best_config['batch_size'],\n",
" 'expected_throughput': best_config['throughput'],\n",
" 'estimated_memory_usage': best_config['estimated_memory_mb'],\n",
" 'all_results': optimization_results,\n",
" 'optimization_analysis': analysis\n",
" }\n",
" ### END SOLUTION\n",
" \n",
" def _generate_batch_size_analysis(self, results, best_config):\n",
" \"\"\"Generate analysis of batch size optimization results.\"\"\"\n",
" analysis = []\n",
" \n",
" # Throughput analysis\n",
" throughputs = [r['throughput'] for r in results]\n",
" max_throughput = max(throughputs)\n",
" min_throughput = min(throughputs)\n",
" \n",
" analysis.append(f\"📈 Throughput range: {min_throughput:.1f} - {max_throughput:.1f} samples/sec\")\n",
" analysis.append(f\"🎯 Optimal batch size: {best_config['batch_size']} ({max_throughput:.1f} samples/sec)\")\n",
" \n",
" # Memory efficiency analysis\n",
" memory_efficiencies = [r['memory_efficiency'] for r in results]\n",
" most_efficient = max(results, key=lambda x: x['memory_efficiency'])\n",
" \n",
" analysis.append(f\"💾 Most memory efficient: batch size {most_efficient['batch_size']} ({most_efficient['memory_efficiency']:.2f} samples/sec/MB)\")\n",
" \n",
" # Bottleneck analysis\n",
" bottleneck_counts = {}\n",
" for r in results:\n",
" step = r['bottleneck_step']\n",
" bottleneck_counts[step] = bottleneck_counts.get(step, 0) + 1\n",
" \n",
" common_bottleneck = max(bottleneck_counts.items(), key=lambda x: x[1])\n",
" analysis.append(f\"🔍 Common bottleneck: {common_bottleneck[0]} ({common_bottleneck[1]}/{len(results)} configurations)\")\n",
" \n",
" return analysis"
]
},
{
"cell_type": "markdown",
"id": "1281999e",
"metadata": {
"cell_marker": "\"\"\"",
"lines_to_next_cell": 1
},
"source": [
"### 🧪 Test: Production Training Optimization\n",
"\n",
"Let's test our production training optimizer."
]
},
{
"cell_type": "code",
"execution_count": null,
"id": "f82a0ee2",
"metadata": {
"nbgrader": {
"grade": false,
"grade_id": "test-production-optimizer",
"locked": false,
"schema_version": 3,
"solution": false,
"task": false
}
},
"outputs": [],
"source": [
"def test_production_training_optimizer():\n",
" \"\"\"Test production training optimizer with realistic scenarios.\"\"\"\n",
" print(\"🔬 Unit Test: Production Training Optimizer...\")\n",
" \n",
" optimizer_tool = ProductionTrainingOptimizer()\n",
" \n",
" # Create test components\n",
" model = Sequential([Dense(10, 5), ReLU(), Dense(5, 2)])\n",
" optimizer = SGD([], learning_rate=0.01)\n",
" loss_fn = MeanSquaredError()\n",
" \n",
" # Test batch size optimization\n",
" result = optimizer_tool.optimize_batch_size_for_throughput(\n",
" model, loss_fn, optimizer, \n",
" initial_batch_size=32, \n",
" max_batch_size=128\n",
" )\n",
" \n",
" # Verify optimization results\n",
" assert 'optimal_batch_size' in result, \"Should find optimal batch size\"\n",
" assert 'expected_throughput' in result, \"Should calculate expected throughput\"\n",
" assert 'estimated_memory_usage' in result, \"Should estimate memory usage\"\n",
" assert 'all_results' in result, \"Should provide all test results\"\n",
" assert 'optimization_analysis' in result, \"Should provide analysis\"\n",
" \n",
" # Verify optimal batch size is reasonable\n",
" assert result['optimal_batch_size'] >= 32, \"Optimal batch size should be at least initial size\"\n",
" assert result['optimal_batch_size'] <= 128, \"Optimal batch size should not exceed maximum\"\n",
" \n",
" # Verify throughput is positive\n",
" assert result['expected_throughput'] > 0, \"Expected throughput should be positive\"\n",
" \n",
" # Verify all results structure\n",
" all_results = result['all_results']\n",
" assert len(all_results) > 0, \"Should have tested at least one batch size\"\n",
" \n",
" for test_result in all_results:\n",
" assert 'batch_size' in test_result, \"Each result should have batch size\"\n",
" assert 'throughput' in test_result, \"Each result should have throughput\"\n",
" assert 'total_time' in test_result, \"Each result should have total time\"\n",
" assert test_result['throughput'] >= 0, \"Throughput should be non-negative\"\n",
" \n",
" print(\"✅ Batch size optimization test passed\")\n",
" \n",
" # Test optimization history tracking\n",
" assert len(optimizer_tool.optimization_history) == 1, \"Should track optimization history\"\n",
" history_entry = optimizer_tool.optimization_history[0]\n",
" assert history_entry['optimization_type'] == 'batch_size', \"Should track optimization type\"\n",
" assert 'results' in history_entry, \"Should store optimization results\"\n",
" assert 'best_config' in history_entry, \"Should store best configuration\"\n",
" \n",
" print(\"✅ Optimization history tracking test passed\")\n",
" \n",
" print(\"🎯 Production Training Optimizer: All tests passed!\")\n",
"\n",
"# Test function defined (called in main block)\n",
"\n",
"if __name__ == \"__main__\":\n",
" # Run all training tests\n",
" test_unit_simple_training_loop()\n",
" test_unit_batch_training()\n",
" test_unit_multiple_epochs()\n",
" test_unit_training_with_validation()\n",
" test_module_training_pipeline_integration()\n",
" test_training_pipeline_profiler()\n",
" \n",
" print(\"All tests passed!\")\n",
" print(\"Training module complete!\")"
]
},
{
"cell_type": "markdown",
"id": "b29aedd0",
"metadata": {
"cell_marker": "\"\"\""
},
"source": [
"## 🤔 ML Systems Thinking Questions\n",
"\n",
"*Take a moment to reflect on these questions. Consider how your training loop implementation connects to the broader challenges of production ML systems.*\n",
"\n",
"### 🏗️ Training Infrastructure Design\n",
"1. **Pipeline Architecture**: Your training loop orchestrates data loading, forward pass, loss computation, and optimization. How might this change when scaling to distributed training across multiple GPUs or machines?\n",
"\n",
"2. **Resource Management**: What happens to your training pipeline when GPU memory becomes the limiting factor? How do production systems handle out-of-memory errors during training?\n",
"\n",
"3. **Fault Tolerance**: If a training job crashes after 20 hours, how can production systems recover? What checkpointing strategies would you implement?\n",
"\n",
"### 📊 Production Training Operations\n",
"4. **Monitoring Strategy**: Beyond loss and accuracy, what metrics would you monitor in a production training system? How would you detect training instability or hardware failures?\n",
"\n",
"5. **Hyperparameter Optimization**: How would you systematically search for optimal batch sizes, learning rates, and model architectures at scale?\n",
"\n",
"6. **Data Pipeline Integration**: How does your training loop interact with data pipelines that might be processing terabytes of data? What happens when data arrives faster than the model can consume it?\n",
"\n",
"### ⚖️ Training at Scale\n",
"7. **Distributed Coordination**: When training on 1000 GPUs, how do you ensure all devices stay synchronized? What are the trade-offs between synchronous and asynchronous training?\n",
"\n",
"8. **Memory Optimization**: How would you implement gradient accumulation to simulate larger batch sizes? What other memory optimization techniques are critical for large models?\n",
"\n",
"9. **Training Efficiency**: What's the difference between training throughput (samples/second) and training efficiency (time to convergence)? How do you optimize for both?\n",
"\n",
"### 🔄 MLOps Integration\n",
"10. **Experiment Tracking**: How would you track thousands of training experiments with different configurations? What metadata is essential for reproducibility?\n",
"\n",
"11. **Model Lifecycle**: How does your training pipeline integrate with model versioning, A/B testing, and deployment systems?\n",
"\n",
"12. **Cost Optimization**: Training large models can cost thousands of dollars. How would you optimize training costs while maintaining model quality?\n",
"\n",
"*These questions connect your training implementation to the real challenges of production ML systems. Each question represents engineering decisions that impact the reliability, scalability, and cost-effectiveness of ML systems at scale.*"
]
},
{
"cell_type": "markdown",
"id": "a24eed33",
"metadata": {
"cell_marker": "\"\"\""
},
"source": [
"## 🎯 MODULE SUMMARY: Training Pipelines\n",
"\n",
"Congratulations! You've successfully implemented complete training pipelines:\n",
"\n",
"### What You've Accomplished\n",
"✅ **Training Loops**: End-to-end training with loss computation and optimization \n",
"✅ **Loss Functions**: Implementation and integration of loss calculations \n",
"✅ **Metrics Tracking**: Monitoring accuracy and loss during training \n",
"✅ **Integration**: Seamless compatibility with neural networks and optimizers \n",
"✅ **Real Applications**: Training real models on real data \n",
"✅ **Pipeline Profiling**: Production-grade performance analysis and optimization \n",
"✅ **Systems Thinking**: Understanding training infrastructure at scale \n",
"\n",
"### Key Concepts You've Learned\n",
"- **Training loops**: How to iterate over data, compute loss, and update parameters\n",
"- **Loss functions**: Quantifying model performance\n",
"- **Metrics tracking**: Monitoring progress and diagnosing issues\n",
"- **Integration patterns**: How training works with all components\n",
"- **Performance optimization**: Efficient training for large models\n",
"- **Pipeline profiling**: Identifying bottlenecks in training infrastructure\n",
"- **Production optimization**: Balancing throughput, memory, and resource utilization\n",
"\n",
"### Professional Skills Developed\n",
"- **Training orchestration**: Building robust training systems\n",
"- **Loss engineering**: Implementing and tuning loss functions\n",
"- **Metrics analysis**: Understanding and improving model performance\n",
"- **Integration testing**: Ensuring all components work together\n",
"- **Performance profiling**: Optimizing training pipelines for production\n",
"- **Systems design**: Understanding distributed training challenges\n",
"\n",
"### Ready for Advanced Applications\n",
"Your training pipeline implementations now enable:\n",
"- **Full model training**: End-to-end training of neural networks\n",
"- **Experimentation**: Testing different architectures and hyperparameters\n",
"- **Production systems**: Deploying trained models for real applications\n",
"- **Research**: Experimenting with new training strategies\n",
"- **Performance optimization**: Scaling training to production workloads\n",
"- **Infrastructure design**: Building reliable ML training systems\n",
"\n",
"### Connection to Real ML Systems\n",
"Your implementations mirror production systems:\n",
"- **PyTorch**: `torch.nn.Module`, `torch.optim`, and training loops\n",
"- **TensorFlow**: `tf.keras.Model`, `tf.keras.optimizers`, and fit methods\n",
"- **Industry Standard**: Every major ML framework uses these exact patterns\n",
"- **Production Tools**: Similar to Ray Train, Horovod, and distributed training frameworks\n",
"\n",
"### Next Steps\n",
"1. **Export your code**: `tito export 11_training`\n",
"2. **Test your implementation**: `tito test 11_training`\n",
"3. **Build evaluation pipelines**: Add benchmarking and validation\n",
"4. **Move to Module 12**: Add model compression and optimization!\n",
"\n",
"**Ready for compression?** Your training pipelines are now ready for real-world deployment!"
]
}
],
"metadata": {
"jupytext": {
"main_language": "python"
}
},
"nbformat": 4,
"nbformat_minor": 5
}