Files
TinyTorch/modules/02_tensor
Vijay Janapa Reddi 753ae52ae0 MAJOR: Implement beautiful module progression through strategic reordering
This commit implements the pedagogically optimal "inevitable discovery" module progression based on expert validation and educational design principles.

## Module Reordering Summary

**Previous Order (Problems)**:
- 05_losses → 06_autograd → 07_dataloader → 08_optimizers → 09_spatial → 10_training
- Issues: Autograd before optimizers, DataLoader before training, scattered dependencies

**New Order (Beautiful Progression)**:
- 05_losses → 06_optimizers → 07_autograd → 08_training → 09_spatial → 10_dataloader
- Benefits: Each module creates inevitable need for the next

## Pedagogical Flow Achieved

**05_losses** → "Need systematic weight updates" → **06_optimizers**
**06_optimizers** → "Need automatic gradients" → **07_autograd**
**07_autograd** → "Need systematic training" → **08_training**
**08_training** → "MLPs hit limits on images" → **09_spatial**
**09_spatial** → "Training is too slow" → **10_dataloader**

## Technical Changes

### Module Directory Renaming
- `06_autograd` → `07_autograd`
- `07_dataloader` → `10_dataloader`
- `08_optimizers` → `06_optimizers`
- `10_training` → `08_training`
- `09_spatial` → `09_spatial` (no change)

### System Integration Updates
- **MODULE_TO_CHECKPOINT mapping**: Updated in tito/commands/export.py
- **Test directories**: Renamed module_XX directories to match new numbers
- **Documentation**: Updated all references in MD files and agent configurations
- **CLI integration**: Updated next-steps suggestions for proper flow

### Agent Configuration Updates
- **Quality Assurance**: Updated module audit status with new numbers
- **Module Developer**: Updated work tracking with new sequence
- **Documentation**: Updated MASTER_PLAN_OF_RECORD.md with beautiful progression

## Educational Benefits

1. **Inevitable Discovery**: Each module naturally leads to the next
2. **Cognitive Load**: Concepts introduced exactly when needed
3. **Motivation**: Students understand WHY each tool is necessary
4. **Synthesis**: Everything flows toward complete ML systems understanding
5. **Professional Alignment**: Matches real ML engineering workflows

## Quality Assurance

-  All CLI commands still function
-  Checkpoint system mappings updated
-  Documentation consistency maintained
-  Test directory structure aligned
-  Agent configurations synchronized

**Impact**: This reordering transforms TinyTorch from a collection of modules into a coherent educational journey where each step naturally motivates the next, creating optimal conditions for deep learning systems understanding.
2025-09-24 15:56:47 -04:00
..

🔥 Module: Tensor

📊 Module Info

  • Difficulty: Intermediate
  • Time Estimate: 4-6 hours
  • Prerequisites: Setup module
  • Next Steps: Activations, Layers

Build the foundation of TinyTorch! This module implements the core Tensor class - the fundamental data structure that powers all neural networks and machine learning operations.

🎯 Learning Objectives

By the end of this module, you will:

  • Understand what tensors are and why they're essential for ML
  • Implement a complete Tensor class with core operations
  • Handle tensor shapes, data types, and memory management efficiently
  • Implement element-wise operations and reductions with proper broadcasting
  • Have a solid foundation for building neural networks

🧠 Build → Use → Understand

  1. Build: Complete Tensor class with arithmetic operations, shape management, and reductions
  2. Use: Create tensors, perform operations, and validate with real data
  3. Understand: How tensors serve as the foundation for all neural network computations

📚 What You'll Build

Core Tensor Class

# Creating tensors
x = Tensor([[1.0, 2.0], [3.0, 4.0]])
y = Tensor([[0.5, 1.5], [2.5, 3.5]])

# Properties
print(x.shape)    # (2, 2)
print(x.size)     # 4
print(x.dtype)    # float64

# Element-wise operations
z = x + y         # Addition
w = x * y         # Multiplication
p = x ** 2        # Exponentiation

# Shape manipulation
reshaped = x.reshape(4, 1)  # (4, 1)
transposed = x.T            # (2, 2) transposed

# Reductions
total = x.sum()             # Scalar sum
means = x.mean(axis=0)      # Mean along axis

Essential Operations

  • Arithmetic: Addition, subtraction, multiplication, division, powers
  • Shape management: Reshape, transpose, broadcasting rules
  • Reductions: Sum, mean, min, max along any axis
  • Memory handling: Efficient data storage and copying

🚀 Getting Started

Prerequisites Check

tito test --module setup  # Should pass ✅

Development Workflow

# Navigate to tensor module
cd modules/source/02_tensor

# Open development file
jupyter notebook tensor_dev.ipynb
# OR edit directly: code tensor_dev.py

Step-by-Step Implementation

  1. Basic Tensor class - Constructor and properties
  2. Shape management - Understanding tensor dimensions
  3. Arithmetic operations - Addition, multiplication, etc.
  4. Utility methods - Reshape, transpose, sum, mean
  5. Error handling - Robust edge case management

🧪 Testing Your Implementation

Inline Testing

# Test in the notebook or Python REPL
x = Tensor([[1.0, 2.0], [3.0, 4.0]])
print(f"Shape: {x.shape}")  # Should be (2, 2)
print(f"Sum: {x.sum()}")    # Should be 10.0

Module Tests

# Export your tensor implementation
tito export

# Test your implementation
tito test --module tensor

Manual Verification

# Create and test tensors
from tinytorch.core.tensor import Tensor

x = Tensor([1, 2, 3, 4, 5])
y = Tensor([2, 4, 6, 8, 10])

# Test operations
assert (x + y).data.tolist() == [3, 6, 9, 12, 15]
assert (x * 2).data.tolist() == [2, 4, 6, 8, 10]
print("✅ Basic operations working!")

🎯 Key Concepts

Tensors as Universal Data Structures

  • Scalars: 0-dimensional tensors (single numbers)
  • Vectors: 1-dimensional tensors (arrays)
  • Matrices: 2-dimensional tensors (common in ML)
  • Higher dimensions: Images (3D), video (4D), etc.

Why Tensors Matter in ML

  • Neural networks: All computations operate on tensors
  • GPU acceleration: operates on tensor primitives
  • Broadcasting: Efficient operations across different shapes
  • Vectorization: Process entire datasets simultaneously

Real-World Connections

  • PyTorch/TensorFlow: Your implementation mirrors production frameworks
  • NumPy: Foundation for scientific computing (we build similar abstractions)
  • Production systems: Understanding tensors is essential for ML engineering

Memory and Performance

  • Data layout: How tensors store data efficiently
  • Broadcasting: Smart operations without data copying
  • View vs Copy: Understanding memory management

🎉 Ready to Build?

The tensor module is where TinyTorch really begins. You're about to create the fundamental building block that will power neural networks, training loops, and production ML systems.

Take your time, test thoroughly, and enjoy building something that really works! 🔥