Transformer

Tags: #machine learning #nlp #gpt

Equation

$$\text{Attention}(Q, K, V) = \text{softmax}(\frac{QK^T}{\sqrt{d_k}})V$$

Latex Code

                                 \text{Attention}(Q, K, V) = \text{softmax}(\frac{QK^T}{\sqrt{d_k}})V
                            

Have Fun

Let's Vote for the Most Difficult Equation!

Introduction

Equation



Latex Code

        \text{Attention}(Q, K, V) = \text{softmax}(\frac{QK^T}{\sqrt{d_k}})V
        

Explanation

Related Documents

Related Videos

Discussion

Comment to Make Wishes Come True

Leave your wishes (e.g. Passing Exams) in the comments and earn as many upvotes as possible to make your wishes come true


  • Les Finch
    I'm really stressing about this test; I hope I pass.
    2023-02-07 00:00

    Reply


    Kathryn Olson reply to Les Finch
    You can make it...
    2023-02-27 00:00:00.0

    Reply


  • Ernest White
    Getting a pass on this test would be a dream come true.
    2024-01-25 00:00

    Reply


    Joshua Moore reply to Ernest White
    Nice~
    2024-02-21 00:00:00.0

    Reply


  • Larry Mitchell
    I'm geared up to pass this exam.
    2023-06-02 00:00

    Reply


    Gregory Edwards reply to Larry Mitchell
    Nice~
    2023-06-27 00:00:00.0

    Reply


  • Ronald Walker
    I really wish that my latest paper can be accepted in this year’s NeurIPS and ICLR conferences. Hope that my wish will come true!!!
    2024-04-25 23:22

    Reply