AI Fundamentals — Hard
Key points
- Sparse attention focuses on specific token pairs
- Reduces computational and memory complexity for long sequences
- Alternative to full self-attention in transformers
Ready to go further?
Related questions
AI Fundamentals — Hard
Key points
Ready to go further?
Related questions