• Skip to primary navigation
  • Skip to content
  • Skip to footer
Probabilistic Nexus Probabilistic Nexus
  • Posts
  • Categories
  • Tags
  • About Me
    Posts by Tag
    Haotian Chen

    Haotian Chen

    Probabilistic model practitioner

    • LinkedIn
    • Generativ/e

    Posts by Tag

    • transformer 1
    • attention layer 1
    • low-rank approximation 1
    • cosine similarity 1
    • reconstruction error 1

    transformer

    Breaking Down Transformers: The Core of Attention as Low-Rank Approximation

    Why attention? How does it work?

    Back to Top ↑

    attention layer

    Breaking Down Transformers: The Core of Attention as Low-Rank Approximation

    Why attention? How does it work?

    Back to Top ↑

    low-rank approximation

    Breaking Down Transformers: The Core of Attention as Low-Rank Approximation

    Why attention? How does it work?

    Back to Top ↑

    cosine similarity

    Breaking Down Transformers: The Core of Attention as Low-Rank Approximation

    Why attention? How does it work?

    Back to Top ↑

    reconstruction error

    Breaking Down Transformers: The Core of Attention as Low-Rank Approximation

    Why attention? How does it work?

    Back to Top ↑
    • LinkedIn
    • Generativ/e
    • GitHub
    • Feed
    © 2024 Probabilistic Nexus. Powered by Jekyll & Minimal Mistakes.