FiveTech Support Forums

FiveWin / Harbour / xBase community
Board index latest AI news Transformers
Posts: 44158
Joined: Thu Oct 06, 2005 05:47 PM
Transformers
Posted: Tue Sep 23, 2025 07:08 AM
</s>flowchart TB
    A[Attention is All You Need] --&gt; B[Introduction]
    A --&gt; C[Background]
    A --&gt; D[Model Architecture]
    click D "https://forums.fivetechsupport.com/viewtopic.php?p=281677#p281677" "click"
    A --&gt; E[Why Self-Attention]
    click E "https://forums.fivetechsupport.com/viewtopic.php?p=281680#p281680" "click"
    A --&gt; F[Training]
    click F "https://forums.fivetechsupport.com/viewtopic.php?p=281681#p281681" "click"
    A --&gt; G[Results]
    click G "https://forums.fivetechsupport.com/viewtopic.php?p=281682#p281682" "click"
    A --&gt; H[Conclusions]
    click H "https://forums.fivetechsupport.com/viewtopic.php?p=281683#p281683" "click"
<e>
regards, saludos

Antonio Linares
www.fivetechsoft.com
Posts: 44158
Joined: Thu Oct 06, 2005 05:47 PM
Re: Transformers
Posted: Tue Sep 23, 2025 07:09 AM
</s>flowchart LR
    B[Introduction] --&gt; B1[Motivation]
    B --&gt; B2[RNN Sequential Nature]
    B --&gt; B3[Attention Mechanisms]
    B --&gt; B4[Transformer Overview]
    
    C[Background] --&gt; C1[Attention Function]
    C --&gt; C2[Self-Attention]
    C --&gt; C3[Multi-Head Attention]
<e>
regards, saludos

Antonio Linares
www.fivetechsoft.com
Posts: 44158
Joined: Thu Oct 06, 2005 05:47 PM
Re: Transformers
Posted: Tue Sep 23, 2025 07:10 AM
</s>flowchart TB
    D[Model Architecture] --&gt; D1[Encoder-Decoder]
    D --&gt; D2[Encoder Stack]
    D --&gt; D3[Decoder Stack]
    D --&gt; D4[Attention Mechanisms]
    D --&gt; D5[Position Encodings]
    D --&gt; D6[Embeddings]
<e>
regards, saludos

Antonio Linares
www.fivetechsoft.com
Posts: 44158
Joined: Thu Oct 06, 2005 05:47 PM
Re: Transformers
Posted: Tue Sep 23, 2025 07:11 AM
</s>flowchart LR
    D2[Encoder Stack] --&gt; D2a[6 Layers]
    D2 --&gt; D2b[Multi-Head Self-Attention]
    D2 --&gt; D2c[Feed Forward]
    D2 --&gt; D2d[Residual Connections]
    D2 --&gt; D2e[Layer Normalization]
    
    D3[Decoder Stack] --&gt; D3a[6 Layers]
    D3 --&gt; D3b[Masked Multi-Head Attention]
    D3 --&gt; D3c[Encoder-Decoder Attention]
    D3 --&gt; D3d[Feed Forward]
    D3 --&gt; D3e[Residual + LayerNorm]
<e>
regards, saludos

Antonio Linares
www.fivetechsoft.com
Posts: 44158
Joined: Thu Oct 06, 2005 05:47 PM
Re: Transformers
Posted: Tue Sep 23, 2025 07:11 AM
</s>flowchart TB
    D4[Attention Mechanisms] --&gt; D4a[Scaled Dot-Product]
    D4 --&gt; D4b[Multi-Head Attention]
    D4 --&gt; D4c[Self-Attention]
    D4 --&gt; D4d[Applications in Model]
    
    D4a --&gt; D4a1[Query, Key, Value]
    D4a --&gt; D4a2[Attention Weights]
    D4a --&gt; D4a3[Scaling Factor]
<e>
regards, saludos

Antonio Linares
www.fivetechsoft.com
Posts: 44158
Joined: Thu Oct 06, 2005 05:47 PM
Re: Transformers
Posted: Tue Sep 23, 2025 07:12 AM
</s>flowchart LR
    E[Why Self-Attention] --&gt; E1[Computational Complexity]
    E --&gt; E2[Parallelization]
    E --&gt; E3[Path Length]
    E --&gt; E4[Comparison with RNN]
    E --&gt; E5[Comparison with CNN]
    E --&gt; E6[Interpretability]
<e>
regards, saludos

Antonio Linares
www.fivetechsoft.com
Posts: 44158
Joined: Thu Oct 06, 2005 05:47 PM
Re: Transformers
Posted: Tue Sep 23, 2025 07:13 AM
</s>flowchart TB
    F[Training] --&gt; F1[Training Data]
    F --&gt; F2[Batching]
    F --&gt; F3[Hardware Setup]
    F --&gt; F4[Optimizer]
    F --&gt; F5[Regularization]
    F --&gt; F6[Learning Rate Schedule]
    
    F4 --&gt; F4a[Adam Optimizer]
    F4 --&gt; F4b[Beta Parameters]
    F4 --&gt; F4c[Learning Rate]
<e>
regards, saludos

Antonio Linares
www.fivetechsoft.com
Posts: 44158
Joined: Thu Oct 06, 2005 05:47 PM
Re: Transformers
Posted: Tue Sep 23, 2025 07:14 AM
</s>flowchart LR
    G[Results] --&gt; G1[Translation Quality]
    G --&gt; G2[Training Time]
    G --&gt; G3[BLEU Scores]
    G --&gt; G4[Comparison with Previous]
    G --&gt; G5[English-German]
    G --&gt; G6[English-French]
    
    G1 --&gt; G1a[WMT 2014]
    G1 --&gt; G1b[State-of-the-art]
<e>
regards, saludos

Antonio Linares
www.fivetechsoft.com
Posts: 44158
Joined: Thu Oct 06, 2005 05:47 PM
Re: Transformers
Posted: Tue Sep 23, 2025 07:15 AM
</s>flowchart TB
    H[Conclusions] --&gt; H1[Main Contributions]
    H --&gt; H2[Future Directions]
    H --&gt; H3[Limitations]
    H --&gt; H4[Impact]
    
    H1 --&gt; H1a[First Sequence Model]
    H1 --&gt; H1b[Based Entirely on Attention]
    H1 --&gt; H1c[Superior Performance]
<e>
regards, saludos

Antonio Linares
www.fivetechsoft.com

Continue the discussion