240 发简信
IP属地:北京
  • RoBERTa, 基于BERT的优化版预训练模型

    Roberta: A robustly optimized bert pretraining approachCitation: 1669 (2...

  • Resize,w 360,h 240
    Transformer-XL,处理超级长序列的Transformer

    Transformer-xl: Attentive language models beyond a fixed-length contextC...

  • Resize,w 360,h 240
    BERT, 18, 预训练模型

    Bert: Pre-training of deep bidirectional transformers for language under...

  • Resize,w 360,h 240
    Transformer, NeurIPS17, Attention is all you need

    Attention Is All You NeedCitation: 26532 (2021-09-04) 1. Motivation 重读经典...

  • Resize,w 360,h 240
    MIND, CIKM19, 天猫用户多兴趣表示的召回模型

    Multi-interest network with dynamic routing for recommendation at TmallC...

  • Resize,w 360,h 240
    ANIL, ICLR20, 重新审视MAML中的重要部分

    Rapid learning or feature reuse? towards understanding the effectiveness...

  • Resize,w 360,h 240
    元学习综述-分类

    Meta-Learning in Neural Networks: A SurveyCitation: 236 (2021-08-29) 1. ...

  • 元学习综述-introduction

    Meta-Learning in Neural Networks: A SurveyCitation: 236 (2021-08-29) 1. ...

  • Resize,w 360,h 240
    DRN, WWW18,深度强化学习推荐系统

    DRN: A Deep Reinforcement Learning Framework for News Recommendation Cit...