Files
TinyTorch/examples/README.md
Vijay Janapa Reddi 6491a7512e Clean up repository: remove temp files, organize modules, prepare for PyPI publication
- Removed temporary test files and audit reports
- Deleted backup and temp_holding directories
- Reorganized module structure (07->09 spatial, 09->07 dataloader)
- Added new modules: 11-14 (tokenization, embeddings, attention, transformers)
- Updated examples with historical ML milestones
- Cleaned up documentation structure
2025-09-24 10:13:37 -04:00

3.6 KiB

TinyTorch Examples: A Journey Through AI History

These examples tell the story of neural networks through historical breakthroughs. Each example represents a pivotal moment in AI history, and you'll build the same architectures that changed the field.

The Historical Journey

1957: The Perceptron - Where It All Began

perceptron_1957/rosenblatt_perceptron.py (Run after Module 4)

  • Frank Rosenblatt's first trainable neural network
  • Could learn linearly separable patterns
  • Sparked dreams of artificial intelligence
  • You'll build: Single-layer network for linear classification

1969: The XOR Problem - The First AI Winter

xor_1969/minsky_xor_problem.py (Run after Module 6)

  • Minsky & Papert proved perceptrons can't solve XOR
  • Led to decade-long "AI Winter" (1969-1980s)
  • Solution required hidden layers + nonlinearity + backpropagation
  • You'll build: Multi-layer perceptron that solves XOR

1998: LeNet - The Convolution Revolution

lenet_1998/train_mlp.py (Run after Module 9)

  • Yann LeCun's convolutional neural network
  • First practical system for reading handwritten digits
  • Deployed in banks for check processing
  • You'll build: Network for MNIST digit recognition

2012: AlexNet - The Deep Learning Explosion

alexnet_2012/train_cnn.py (Run after Module 10)

  • Alex Krizhevsky's ImageNet breakthrough
  • Proved deep networks could surpass traditional CV
  • Triggered the modern deep learning boom
  • You'll build: Deep CNN for CIFAR-10 classification

2018: GPT - The Transformer Era

gpt_2018/simple_tinygpt.py (Run after Module 14)

  • OpenAI's transformer architecture
  • Self-attention revolutionized NLP
  • Foundation for ChatGPT and modern AI
  • You'll build: Character-level language model

Running the Examples

Each example shows which modules are required:

# After Module 4: Can build architectures
python examples/perceptron_1957/rosenblatt_perceptron.py

# After Module 6: Can train with gradients  
python examples/xor_1969/minsky_xor_problem.py

# After Module 9: Can use convolutions
python examples/lenet_1998/train_mlp.py

# After Module 10: Full training pipeline
python examples/alexnet_2012/train_cnn.py

# After Module 14: Transformers work!
python examples/gpt_2018/simple_tinygpt.py

The Learning Flow

  1. Build modules → Core engine development
  2. Pass unit tests → Verify your implementation
  3. Complete moduletito module complete XX_modulename
  4. Pass integration tests → Automatic validation with other modules
  5. Unlock capability → New historical example available!
  6. Run example → See what you've enabled!

📚 See CAPABILITIES.md for the complete progression system

PyTorch-Style Code

All examples follow modern PyTorch conventions:

class HistoricNetwork:
    def __init__(self):
        # Define layers
        self.fc1 = Dense(input_size, hidden_size)
        self.activation = ReLU()
        self.fc2 = Dense(hidden_size, output_size)
    
    def forward(self, x):
        # Forward pass
        x = self.fc1(x)
        x = self.activation(x)
        x = self.fc2(x)
        return x

What You're Building

You're not just learning ML - you're rebuilding the breakthroughs that created modern AI:

  • 1957: Linear models that could learn
  • 1969: Multi-layer networks for complex patterns
  • 1998: Convolutional networks for vision
  • 2012: Deep networks that changed everything
  • 2018: Attention mechanisms powering ChatGPT

Each example runs on YOUR implementation. When GPT works, it's because YOU built every component from scratch!