Name
Affiliation
Papers
OLIVIER PIETQUIN
Supélec, IMS Research Group, Metz, France
147
Collaborators
Citations 
PageRank 
253
664
68.60
Referers 
Referees 
References 
1335
1303
1240
Search Limit
1001000
Title
Citations
PageRank
Year
Lazy-MDPs: Towards Interpretable RL by Learning When to Act.00.342022
Continuous Control with Action Quantization from Demonstrations.00.342022
Learning Natural Language Generation with Truncated Reinforcement Learning00.342022
On the role of population heterogeneity in emergent communication00.342022
Solving N-Player Dynamic Routing Games with Congestion: A Mean-Field Approach.00.342022
Offline Reinforcement Learning With Pseudometric Learning00.342021
Show me the Way: Intrinsic Motivation from Demonstrations00.342021
Mean Field Games Flock! The Reinforcement Learning Way.00.342021
Adversarially Guided Actor-Critic00.342021
Self-Imitation Advantage Learning00.342021
What Matters for On-Policy Deep Actor-Critic Methods? A Large-Scale Study00.342021
Hyperparameter Selection for Imitation Learning00.342021
Fictitious Play for Mean Field Games: Continuous Time Analysis and Applications00.342020
On The Convergence Of Model Free Learning In Mean Field Games00.342020
HIGhER: Improving instruction following with Hindsight Generation for Experience Replay00.342020
Supervised Seeded Iterated Learning for Interactive Language Learning.00.342020
Countering Language Drift with Seeded Iterated Learning00.342020
Self-Attentional Credit Assignment for Transfer in Reinforcement Learning10.362020
Foolproof Cooperative Learning.00.342020
CopyCAT:: Taking Control of Neural Policies with Constant Attacks00.342020
Leverage the Average: an Analysis of KL Regularization in Reinforcement Learning00.342020
Scaling up budgeted reinforcement learning.00.342019
Deep Conservative Policy Iteration00.342019
Observational Learning by Reinforcement Learning00.342019
Self-Educated Language Agent with Hindsight Experience Replay for Instruction Following.00.342019
Targeted Attacks on Deep Reinforcement Learning Agents through Adversarial Observations.00.342019
A Theory of Regularized Markov Decision Processes.00.342019
Learning from a Learner00.342019
Foolproof Cooperative Learning.00.342019
Playing the Game of Universal Adversarial Perturbations.10.342018
Observe and Look Further: Achieving Consistent Performance on Atari.100.462018
End-to-end optimization of goal-driven and visually grounded dialogue systems.210.812017
Noisy Networks for Exploration.461.442017
Modulating early visual processing by language.290.922017
LIG-CRIStAL System for the WMT17 Automatic Post-Editing Task.00.342017
Leveraging Demonstrations for Deep Reinforcement Learning on Robotics Problems with Sparse Rewards.260.812017
Is the Bellman residual a bad proxy?00.342017
Observational Learning by Reinforcement Learning.00.342017
On the Use of Non-Stationary Strategies for Solving Two-Player Zero-Sum Markov Games.50.422016
Listen and Translate: A Proof of Concept for End-to-End Speech-to-Text Translation.10.352016
Should one minimize the expected Bellman residual or maximize the mean value?00.342016
Difference of Convex Functions Programming Applied to Control with Expert Data.00.342016
Score-based Inverse Reinforcement Learning.00.342016
MultiVec: a Multilingual and Multilevel Representation Learning Toolkit for NLP.100.562016
Softened Approximate Policy Iteration for Markov Games.40.422016
Compact and Interpretable Dialogue State Representation with Genetic Sparse Distributed Memory.00.342016
Imitation Learning Applied to Embodied Conversational Agents20.382015
Bayesian Credible Intervals for Online and Active Learning of Classification Trees00.342015
Inverse reinforcement learning in relational domains40.412015
Optimism in Active Learning.20.392015
  • 1
  • 2