Files
TinyTorch/site/quickstart-guide.md
Vijay Janapa Reddi 97e0563614 Add community and benchmark features with baseline validation
- Implement tito benchmark baseline and capstone commands
- Add SPEC-style normalization for baseline benchmarks
- Implement tito community join, update, leave, stats, profile commands
- Use project-local storage (.tinytorch/) for user data
- Add privacy-by-design with explicit consent prompts
- Update site documentation for community and benchmark features
- Add Marimo integration for online notebooks
- Clean up redundant milestone setup exploration docs
- Finalize baseline design: fast setup validation (~1 second) with normalized results
2025-11-20 00:17:21 -05:00

10 KiB

Quick Start Guide

From Zero to Building Neural Networks

Complete setup + first module in 15 minutes

Purpose: Get hands-on experience building ML systems in 15 minutes. Complete setup verification and build your first neural network component from scratch.

2-Minute Setup

Let's get you ready to build ML systems:

Step 1: One-Command Setup

# Clone repository
git clone https://github.com/mlsysbook/TinyTorch.git
cd TinyTorch

# Automated setup (handles everything!)
./setup-environment.sh

# Activate environment
source activate.sh

What this does:

  • Creates optimized virtual environment (arm64 on Apple Silicon)
  • Installs all dependencies (NumPy, Jupyter, Rich, PyTorch for validation)
  • Configures TinyTorch in development mode
  • Verifies installation

See TITO CLI Reference for detailed workflow and troubleshooting.

Step 2: Verify Setup

# Run system diagnostics
tito system doctor

You should see all green checkmarks. This confirms your environment is ready for hands-on ML systems building.

See Module Workflow for detailed commands and Troubleshooting if needed.

Step 3: Join the Community & Benchmark

After setup, join the global TinyTorch community and validate your setup:

# Join the community (optional)
tito community join

# Run baseline benchmark to validate setup
tito benchmark baseline

Community Features:

  • Join with optional information (country, institution, course type)
  • Track your progress automatically
  • See your cohort (Fall 2024, Spring 2025, etc.)
  • All data stored locally in .tinytorch/ directory

Baseline Benchmark:

  • Quick validation that everything works
  • Your "Hello World" moment!
  • Generates score and saves results locally

See Community Guide for complete features.

15-Minute First Module Walkthrough

Let's build your first neural network component following the TinyTorch workflow:

graph TD
    Start[Clone & Setup] --> Edit[Edit Module<br/>tensor_dev.ipynb]
    Edit --> Export[Export to Package<br/>tito module complete 01]
    Export --> Test[Test Import<br/>from tinytorch import Tensor]
    Test --> Next[Continue to Module 02]

    style Start fill:#e3f2fd
    style Edit fill:#fffbeb
    style Export fill:#f0fdf4
    style Test fill:#fef3c7
    style Next fill:#f3e5f5

See Student Workflow for the complete development cycle.

Module 01: Tensor Foundations

Learning Goal: Build N-dimensional arrays - the foundation of all neural networks

Time: 15 minutes

Action: Start with Module 01 to build tensor operations from scratch.

# Step 1: Edit the module source
cd modules/01_tensor
jupyter lab tensor_dev.ipynb

You'll implement core tensor operations:

  • N-dimensional array creation
  • Basic mathematical operations (add, multiply, matmul)
  • Shape manipulation (reshape, transpose)
  • Memory layout understanding

Key Implementation: Build the Tensor class that forms the foundation of all neural networks

# Step 2: Export to package when ready
tito module complete 01

This makes your implementation importable: from tinytorch import Tensor

See Student Workflow for the complete edit → export → validate cycle.

Achievement Unlocked: Foundation capability - "Can I create and manipulate the building blocks of ML?"

Next Step: Module 02 - Activations

Learning Goal: Add nonlinearity - the key to neural network intelligence

Time: 10 minutes

Action: Continue with Module 02 to add activation functions.

# Step 1: Edit the module
cd modules/02_activations
jupyter lab activations_dev.ipynb

You'll implement essential activation functions:

  • ReLU (Rectified Linear Unit) - the workhorse of deep learning
  • Softmax - for probability distributions
  • Understand gradient flow and numerical stability
  • Learn why nonlinearity enables learning

Key Implementation: Build activation functions that allow neural networks to learn complex patterns

# Step 2: Export when ready
tito module complete 02

See Student Workflow for the complete edit → export → validate cycle.

Achievement Unlocked: Intelligence capability - "Can I add nonlinearity to enable learning?"

Track Your Progress

After completing your first modules:

Check your new capabilities: Use the optional checkpoint system to track your progress:

tito checkpoint status  # View your completion tracking

This is helpful for self-assessment but not required for the core workflow.

See Student Workflow for the essential edit → export → validate cycle.

Validate with Historical Milestones

After exporting your modules, prove what you've built by running milestone scripts:

After Module 07: Build Rosenblatt's 1957 Perceptron - the first trainable neural network
After Module 07: Solve the 1969 XOR Crisis with multi-layer networks
After Module 08: Achieve 95%+ accuracy on MNIST with 1986 backpropagation
After Module 09: Hit 75%+ on CIFAR-10 with 1998 CNNs
After Module 13: Generate text with 2017 Transformers
After Module 18: Optimize for production with 2018 Torch Olympics

See Journey Through ML History for complete timeline, requirements, and expected results.

The Workflow: Edit modules → Export with tito module complete N → Run milestone scripts to validate

See Student Workflow for the complete cycle.

What You Just Accomplished

In 15 minutes, you've:

Setup Complete

Installed TinyTorch and verified your environment

Created Foundation

Implemented core tensor operations from scratch

First Capability

Earned your first ML systems capability checkpoint

Your Next Steps

Immediate Next Actions (Choose One):

Continue Building (Recommended): Begin Module 03 to add layers to your network.

Master the Workflow:

For Instructors:

Notebook Platforms:

  • Online (Viewing): Jupyter/MyBinder, Google Colab, Marimo - great for exploring notebooks
  • ⚠️ Important: Online notebooks are for viewing only. For full package experiments, milestone validation, and CLI tools, you need local installation (see Student Workflow)

Pro Tips for Continued Success

The TinyTorch Development Cycle:

  1. Edit module sources in modules/NN_name/ (e.g., modules/01_tensor/tensor_dev.ipynb)
  2. Export with tito module complete N
  3. Validate by running milestone scripts

See Student Workflow for detailed workflow guide and best practices.

You're Now a TinyTorch Builder

Ready to Build Production ML Systems

You've proven you can build ML components from scratch. Time to keep going!

Continue Building → TITO CLI Reference →

What makes TinyTorch different: You're not just learning about neural networks—you're building them from fundamental mathematical operations. Every line of code you write builds toward complete ML systems mastery.

Next milestone: After Module 08, you'll train real neural networks on actual datasets using 100% your own code!