Files
TinyTorch/tinytorch
Vijay Janapa Reddi 87d5a7e381 fix(module-05): Add TransposeBackward and fix MatmulBackward for batched ops
TransposeBackward:
- New backward function for transpose operation
- Patch Tensor.transpose() to track gradients
- Critical for attention (Q @ K.T) gradient flow

MatmulBackward batched fix:
- Change np.dot to np.matmul for batched 3D+ tensors
- Use np.swapaxes instead of .T for proper batched transpose
- Fixes gradient shapes in attention mechanisms

Tests added:
- tests/05_autograd/test_batched_matmul_backward.py (3 tests)
- Updated tests/regression/test_gradient_flow_fixes.py (9 tests total)

All gradient flow issues for transformer training are now resolved!
2025-10-27 20:35:06 -04:00
..