Roberta: A robustly optimized bert pretraining approachCitation: 1669 (2...
Transformer-xl: Attentive language models beyond a fixed-length contextC...
Bert: Pre-training of deep bidirectional transformers for language under...
Attention Is All You NeedCitation: 26532 (2021-09-04) 1. Motivation 重读经典...
Multi-interest network with dynamic routing for recommendation at TmallC...
Rapid learning or feature reuse? towards understanding the effectiveness...
Meta-Learning in Neural Networks: A SurveyCitation: 236 (2021-08-29) 1. ...
Meta-Learning in Neural Networks: A SurveyCitation: 236 (2021-08-29) 1. ...
DRN: A Deep Reinforcement Learning Framework for News Recommendation Cit...