Files
TinyTorch/tinytorch
Vijay Janapa Reddi 6f0c96c130 Fix gradient flow with PyTorch-style requires_grad tracking
- Updated Linear layer to use autograd operations (matmul, add) for proper gradient propagation
- Fixed Parameter class to wrap Variables with requires_grad=True
- Implemented proper MSELoss and CrossEntropyLoss with backward chaining
- Added broadcasting support in autograd operations for bias gradients
- Fixed memoryview errors in gradient data extraction
- All integration tests now pass - neural networks can learn via backpropagation
2025-09-29 10:46:58 -04:00
..