Name
Affiliation
Papers
LI DONG
Microsoft Research
51
Collaborators
Citations 
PageRank 
102
582
31.86
Referers 
Referees 
References 
1842
851
508
Search Limit
1001000
Title
Citations
PageRank
Year
STABLEMOE: Stable Routing Strategy for Mixture of Experts00.342022
Swin Transformer V2: Scaling Up Capacity and Resolution00.342022
Controllable Natural Language Generation with Contrastive Prefixes00.342022
Kformer: Knowledge Injection in Transformer Feed-Forward Layers00.342022
Knowledge Neurons in Pretrained Transformers00.342022
XLM-E: Cross-lingual Language Model Pre-training via ELECTRA20.362022
CLIP Models are Few-shot Learners: Empirical Studies on VQA and Visual Entailment00.342022
THE-X: Privacy-Preserving Transformer Inference with Homomorphic Encryption00.342022
Multilingual Machine Translation Systems from Microsoft for WMT21 Shared Task.00.342021
Adapt-and-Distill - Developing Small, Fast and Effective Pretrained Language Models for Domains.00.342021
Self-Attention Attribution: Interpreting Information Interactions Inside Transformer00.342021
Learning to Sample Replacements for ELECTRA Pre-Training.00.342021
Allocating Large Vocabulary Capacity for Cross-Lingual Language Model Pre-Training.00.342021
mT6 - Multilingual Pretrained Text-to-Text Transformer with Translation Pairs.00.342021
Zero-Shot Cross-Lingual Transfer of Neural Machine Translation with Multilingual Pretrained Encoders.00.342021
Learning natural language interfaces with neural models00.342021
InfoXLM: An Information-Theoretic Framework for Cross-Lingual Language Model Pre-Training00.342021
MiniLMv2 - Multi-Head Self-Attention Relation Distillation for Compressing Pretrained Transformers.00.342021
Memory-Efficient Differentiable Transformer Architecture Search.00.342021
Can Monolingual Pretrained Models Help Cross-Lingual Classification?00.342020
UniLMv2: Pseudo-Masked Language Models for Unified Language Model Pre-Training00.342020
Harvesting and Refining Question-Answer Pairs for Unsupervised QA00.342020
MiniLM: Deep Self-Attention Distillation for Task-Agnostic Compression of Pre-Trained Transformers00.342020
Investigating Learning Dynamics of BERT Fine-Tuning.00.342020
Cross-Lingual Natural Language Generation Via Pre-Training00.342020
Unified Language Model Pre-training for Natural Language Understanding and Generation.80.442019
Visualizing and Understanding the Effectiveness of BERT70.492019
Multitask learning for biomedical named entity recognition with cross-sharing structure.00.342019
Inspecting Unification of Encoding and Matching with Transformer: A Case Study of Machine Reading Comprehension00.342019
Confidence Modeling For Neural Semantic Parsing10.352018
Coarse-To-Fine Decoding For Neural Semantic Parsing140.512018
Data-to-Text Generation with Content Selection and Planning30.392018
Proactive Resource Management for LTE in Unlicensed Spectrum: A Deep Learning Perspective.130.642018
Learning to Generate Product Reviews from Attributes.190.662017
Learning to Paraphrase for Question Answering.130.552017
Unsupervised Word and Dependency Path Embeddings for Aspect Term Extraction.190.702016
Long Short-Term Memory-Networks for Machine Reading.1063.452016
Solving and Generating Chinese Character Riddles.00.342016
Adaptive Multi-Compositionality for Recursive Neural Network Models.40.412016
Splusplus: A Feature-Rich Two-stage Classifier for Sentiment Analysis of Tweets40.452015
A joint segmentation and classification framework for sentence level sentiment classification190.592015
A hybrid neural model for type classification of entity mentions70.452015
A statistical parsing framework for sentiment classification121.512015
Question Answering Over Freebase With Multi-Column Convolutional Neural Networks781.942015
Ranking With Recursive Neural Networks And Its Application To Multi-Document Summarization401.252015
Adaptive Recursive Neural Network For Target-Dependent Twitter Sentiment Classification561.532014
A Joint Segmentation and Classification Framework for Sentiment Analysis.30.382014
The Automated Acquisition of Suggestions from Tweets.70.582013
Unraveling The Origin Of Exponential Law In Intra-Urban Human Mobility311.432013
MoodLens: an emoticon-based sentiment analysis system for chinese tweets1123.272012
  • 1
  • 2