Files
TinyTorch/modules/12_embeddings/module.yaml
Vijay Janapa Reddi 6491a7512e Clean up repository: remove temp files, organize modules, prepare for PyPI publication
- Removed temporary test files and audit reports
- Deleted backup and temp_holding directories
- Reorganized module structure (07->09 spatial, 09->07 dataloader)
- Added new modules: 11-14 (tokenization, embeddings, attention, transformers)
- Updated examples with historical ML milestones
- Cleaned up documentation structure
2025-09-24 10:13:37 -04:00

33 lines
1.1 KiB
YAML
Raw Blame History

This file contains ambiguous Unicode characters
This file contains Unicode characters that might be confused with other characters. If you think that this is intentional, you can safely ignore this warning. Use the Escape button to reveal them.
name: "Embeddings"
number: 12
description: "Dense vector representations that convert discrete tokens into continuous semantic spaces"
learning_objectives:
- "Implement embedding layers with efficient lookup operations"
- "Build sinusoidal and learned positional encoding systems"
- "Understand embedding memory scaling and optimization techniques"
- "Analyze how embedding choices affect model capacity and performance"
- "Design embedding systems for production language model deployment"
prerequisites:
- "02_tensor"
- "11_tokenization"
exports:
- "Embedding"
- "PositionalEncoding"
- "LearnedPositionalEmbedding"
- "EmbeddingProfiler"
systems_concepts:
- "Embedding table memory scaling O(vocab_size × embed_dim)"
- "Memory-bandwidth bound lookup operations"
- "Cache-friendly embedding access patterns"
- "Position encoding trade-offs and extrapolation"
- "Distributed embedding table management"
ml_systems_focus: "Memory-efficient embedding lookup, position encoding scalability, large-scale parameter management"
estimated_time: "4-5 hours"
next_modules:
- "13_attention"