Files
TinyTorch/site/credits.md
Vijay Janapa Reddi 0d2560c490 Update site documentation and development guides
- Improve site navigation and content structure
- Update development testing documentation
- Enhance site styling and visual consistency
- Update release notes and milestone templates
- Improve site rebuild script functionality
2025-11-13 10:42:51 -05:00

6.6 KiB

Credits & Acknowledgments

TinyTorch stands on the shoulders of giants.

This project draws inspiration from pioneering educational ML frameworks and owes its existence to the open source community's commitment to accessible ML education.


Core Inspirations

MiniTorch

minitorch.github.io by Sasha Rush (Cornell Tech)

TinyTorch's pedagogical DNA comes from MiniTorch's brilliant "build a framework from scratch" approach. MiniTorch pioneered teaching ML through implementation rather than usage, proving students gain deeper understanding by building systems themselves.

What MiniTorch teaches: Automatic differentiation through minimal, elegant implementations

How TinyTorch differs: Extends to full systems engineering including optimization, profiling, and production deployment across Foundation → Architecture → Optimization tiers

When to use MiniTorch: Excellent complement for deep mathematical understanding of autodifferentiation

Connection to TinyTorch: Modules 05-07 (Autograd, Optimizers, Training) share philosophical DNA with MiniTorch's core pedagogy


micrograd

github.com/karpathy/micrograd by Andrej Karpathy

Micrograd demonstrated that automatic differentiation—the heart of modern ML—can be taught in ~100 lines of elegant Python. Its clarity and simplicity inspired TinyTorch's emphasis on understandable implementations.

What micrograd teaches: Autograd engine in 100 beautiful lines of Python

How TinyTorch differs: Comprehensive framework covering vision, language, and production systems (20 modules vs. single-file implementation)

When to use micrograd: Perfect 2-hour introduction before starting TinyTorch

Connection to TinyTorch: Module 05 (Autograd) teaches the same core concepts with systems engineering focus


nanoGPT

github.com/karpathy/nanoGPT by Andrej Karpathy

nanoGPT's minimalist transformer implementation showed how to teach modern architectures without framework abstraction. TinyTorch's transformer modules (12, 13) follow this philosophy: clear, hackable implementations that reveal underlying mathematics.

What nanoGPT teaches: Clean transformer implementation for understanding GPT architecture

How TinyTorch differs: Build transformers from tensors up, understanding all dependencies from scratch

When to use nanoGPT: Complement to TinyTorch Modules 10-13 for transformer-specific deep-dive

Connection to TinyTorch: Module 13 (Transformers) culminates in similar architecture built from your own tensor operations


tinygrad

github.com/geohot/tinygrad by George Hotz

Tinygrad proves educational frameworks can achieve impressive performance. While TinyTorch optimizes for learning clarity over speed, tinygrad's emphasis on efficiency inspired our Optimization Tier's production-focused modules.

What tinygrad teaches: Performance-focused educational framework with actual GPU acceleration

How TinyTorch differs: Pedagogy-first with explicit systems thinking and scaffolding (educational over performant)

When to use tinygrad: After TinyTorch for performance optimization deep-dive and GPU programming

Connection to TinyTorch: Modules 14-19 (Optimization Tier) share production systems focus


Academic Foundation

Machine Learning Systems

mlsysbook.ai by Prof. Vijay Janapa Reddi (Harvard University)

TinyTorch began as hands-on exercises for the ML Systems textbook. The book's emphasis on systems engineering—memory hierarchies, performance optimization, production deployment—shaped TinyTorch's three-tier structure and systems-focused learning objectives.

What ML Systems Book teaches: Comprehensive textbook on production ML systems engineering

How TinyTorch differs: TinyTorch is the hands-on implementation companion

When to use: Parallel reading with TinyTorch for theoretical depth and production context

Connection to TinyTorch: TinyTorch modules directly implement concepts from ML Systems Book chapters


What Makes TinyTorch Unique

TinyTorch combines inspiration from these projects into a comprehensive ML systems course:

  • Comprehensive Scope: Only educational framework covering Foundation → Architecture → Optimization
  • Systems Thinking: Every module includes profiling, complexity analysis, production context
  • Historical Validation: Milestone system proving implementations through ML history (1957 → 2018)
  • Pedagogical Scaffolding: Progressive disclosure, Build → Use → Reflect methodology
  • Production Context: Direct connections to PyTorch, TensorFlow, and industry practices

For maximum educational value, we recommend:

  1. Start with micrograd (2 hours) — Understand autograd fundamentals
  2. Read Deep Learning Book Ch 6 — Mathematical foundations
  3. Begin TinyTorch (varies by path) — Build complete ML systems
  4. Reference ML Systems Book — Parallel reading for production context
  5. Explore tinygrad — Performance optimization and GPU programming

This progression takes you from mathematical foundations → systems implementation → production optimization.


Open Source Gratitude

TinyTorch exists because of open source software:

Core Dependencies:

  • NumPy: Numerical computing foundation
  • Jupyter: Interactive development environment
  • PyTorch: Reference implementation and validation framework
  • Rich: Beautiful CLI interfaces

Development Tools:

  • Jupyter Book: Documentation and course website
  • pytest: Testing infrastructure
  • GitHub: Version control and collaboration

Community Contributors

TinyTorch is built by students, educators, and ML engineers who believe in accessible systems education.

View all contributors on GitHub


How to Contribute

TinyTorch is open source and welcomes contributions:

See contribution guidelines


License

TinyTorch is released under the MIT License, ensuring it remains free and open for educational use.


Thank you to everyone building the future of accessible ML education.