跳转至

文章研读

约 227 个字 预计阅读时间 1 分钟

记录一下一些论文和博客的研读笔记

Contents

  • 苏神围绕 Transformer 的一系列博客

    • https://spaces.ac.cn/content.html?tag=attention
  • CNN:

    • Neocognitron: A Self-organizing Neural Network Model for a Mechanism of Pattern Recognition Unaffected by Shift in Position

    • LeNet-5:

      • GradientBased Learning Applied to Document Recognition
    • AlexNet:

      • ImageNet Classification with Deep Convolutional Neural Networks
    • VGG:

      • Very Deep Convolutional Networks for Large-Scale Image Recognition
    • GoogleNet:

      • Going Deeper with Convolutions
    • ResNet:

      • Deep Residual Learning for Image Recognition
    • CNN Survey:

      • Convolutional Neural Networks: A Survey (Moez Krichen 2023)
  • RNN:

    • Neural networks and physical systems with emergent collective computational abilities

    • learning internal representations by error propagation

    • Finding Structure in Time

    • LSTM:

      • Long Short-Term Memory
    • Bi-RNN:

      • Bidirectional Recurrent Neural Networks
    • GRU:

      • Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation
    • RNN Survey:

      • Gate-Variants of Gated Recurrent Unit (GRU) Neural Networks
      • Recurrent Neural Networks and Long Short-Term Memory Networks: Tutorial and Survey (Benyamin Ghojogh)
  • Encoder-Decoder and Attention:

    • Learning Phrase Representations using RNN Encoder–Decoder for Statistical Machine Translation
    • Sequence to Sequence Learning with Neural Networks
    • Neural Machine Translation by Jointly Learning to Align and Translate
    • Effective Approaches to Attention-based Neural Machine Translation
  • Transformer:

    • Attention Is All You Need
    • BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding
    • T5
    • Longformer: The Long-Document Transformer
    • Performer: Efficient Attention with Faster Transformers
    • Swin Transformer