mirror of
https://github.com/MLSysBook/TinyTorch.git
synced 2026-05-05 05:27:32 -05:00
- Removed temporary test files and audit reports - Deleted backup and temp_holding directories - Reorganized module structure (07->09 spatial, 09->07 dataloader) - Added new modules: 11-14 (tokenization, embeddings, attention, transformers) - Updated examples with historical ML milestones - Cleaned up documentation structure
104 lines
3.6 KiB
Markdown
104 lines
3.6 KiB
Markdown
# TinyTorch Examples: A Journey Through AI History
|
|
|
|
These examples tell the story of neural networks through historical breakthroughs. Each example represents a pivotal moment in AI history, and you'll build the same architectures that changed the field.
|
|
|
|
## The Historical Journey
|
|
|
|
### 1957: The Perceptron - Where It All Began
|
|
**`perceptron_1957/rosenblatt_perceptron.py`** (Run after Module 4)
|
|
- Frank Rosenblatt's first trainable neural network
|
|
- Could learn linearly separable patterns
|
|
- Sparked dreams of artificial intelligence
|
|
- **You'll build:** Single-layer network for linear classification
|
|
|
|
### 1969: The XOR Problem - The First AI Winter
|
|
**`xor_1969/minsky_xor_problem.py`** (Run after Module 6)
|
|
- Minsky & Papert proved perceptrons can't solve XOR
|
|
- Led to decade-long "AI Winter" (1969-1980s)
|
|
- Solution required hidden layers + nonlinearity + backpropagation
|
|
- **You'll build:** Multi-layer perceptron that solves XOR
|
|
|
|
### 1998: LeNet - The Convolution Revolution
|
|
**`lenet_1998/train_mlp.py`** (Run after Module 9)
|
|
- Yann LeCun's convolutional neural network
|
|
- First practical system for reading handwritten digits
|
|
- Deployed in banks for check processing
|
|
- **You'll build:** Network for MNIST digit recognition
|
|
|
|
### 2012: AlexNet - The Deep Learning Explosion
|
|
**`alexnet_2012/train_cnn.py`** (Run after Module 10)
|
|
- Alex Krizhevsky's ImageNet breakthrough
|
|
- Proved deep networks could surpass traditional CV
|
|
- Triggered the modern deep learning boom
|
|
- **You'll build:** Deep CNN for CIFAR-10 classification
|
|
|
|
### 2018: GPT - The Transformer Era
|
|
**`gpt_2018/simple_tinygpt.py`** (Run after Module 14)
|
|
- OpenAI's transformer architecture
|
|
- Self-attention revolutionized NLP
|
|
- Foundation for ChatGPT and modern AI
|
|
- **You'll build:** Character-level language model
|
|
|
|
## Running the Examples
|
|
|
|
Each example shows which modules are required:
|
|
|
|
```bash
|
|
# After Module 4: Can build architectures
|
|
python examples/perceptron_1957/rosenblatt_perceptron.py
|
|
|
|
# After Module 6: Can train with gradients
|
|
python examples/xor_1969/minsky_xor_problem.py
|
|
|
|
# After Module 9: Can use convolutions
|
|
python examples/lenet_1998/train_mlp.py
|
|
|
|
# After Module 10: Full training pipeline
|
|
python examples/alexnet_2012/train_cnn.py
|
|
|
|
# After Module 14: Transformers work!
|
|
python examples/gpt_2018/simple_tinygpt.py
|
|
```
|
|
|
|
## The Learning Flow
|
|
|
|
1. **Build modules** → Core engine development
|
|
2. **Pass unit tests** → Verify your implementation
|
|
3. **Complete module** → `tito module complete XX_modulename`
|
|
4. **Pass integration tests** → Automatic validation with other modules
|
|
5. **Unlock capability** → New historical example available!
|
|
6. **Run example** → See what you've enabled!
|
|
|
|
📚 **See [CAPABILITIES.md](CAPABILITIES.md) for the complete progression system**
|
|
|
|
## PyTorch-Style Code
|
|
|
|
All examples follow modern PyTorch conventions:
|
|
|
|
```python
|
|
class HistoricNetwork:
|
|
def __init__(self):
|
|
# Define layers
|
|
self.fc1 = Dense(input_size, hidden_size)
|
|
self.activation = ReLU()
|
|
self.fc2 = Dense(hidden_size, output_size)
|
|
|
|
def forward(self, x):
|
|
# Forward pass
|
|
x = self.fc1(x)
|
|
x = self.activation(x)
|
|
x = self.fc2(x)
|
|
return x
|
|
```
|
|
|
|
## What You're Building
|
|
|
|
You're not just learning ML - you're rebuilding the breakthroughs that created modern AI:
|
|
|
|
- **1957**: Linear models that could learn
|
|
- **1969**: Multi-layer networks for complex patterns
|
|
- **1998**: Convolutional networks for vision
|
|
- **2012**: Deep networks that changed everything
|
|
- **2018**: Attention mechanisms powering ChatGPT
|
|
|
|
Each example runs on YOUR implementation. When GPT works, it's because YOU built every component from scratch! |