Name
Affiliation
Papers
DEVI PARIKH
Virginia Tech, Blacksburg, VA USA
149
Collaborators
Citations 
PageRank 
278
2929
132.01
Referers 
Referees 
References 
6354
2365
2416
Search Limit
1001000
Title
Citations
PageRank
Year
Long Video Generation with Time-Agnostic VQGAN and Time-Sensitive Transformer.00.342022
Episodic Memory Question Answering.00.342022
VX2TEXT: End-to-End Learning of Video-Based Text Generation From Multimodal Inputs00.342021
Contrast and Classify - Training Robust VQA Models.00.342021
Feel The Music: Automatically Generating A Dance For An Input Song00.342020
Lemotif - An Affective Visual Journal Using Deep Neural Networks.00.342020
SQuINTing at VQA Models - Introspecting VQA Models With Sub-Questions.00.342020
Exploring Crowd Co-creation Scenarios for Sketches00.342020
Predicting A Creator's Preferences In, and From, Interactive Generative Art00.342020
Neuro-Symbolic Generative Art: A Preliminary Study00.342020
Sim-to-Real Transfer for Vision-and-Language Navigation.00.342020
12-in-1: Multi-Task Vision and Language Representation Learning90.542020
Dialog without Dialog Data: Learning Visual Dialog Agents from VQA Data00.342020
Grad-CAM: Visual Explanations from Deep Networks via Gradient-Based Localization.3007.452020
Where Are You? Localization from Embodied Dialog.00.342020
Embodied Multimodal Multitask Learning.10.352020
Audio Visual Scene-Aware Dialog20.422019
Probabilistic Neural-symbolic Models for Interpretable Visual Question Answering.10.342019
Embodied Question Answering In Photorealistic Environments With Point Cloud Perception40.402019
Align2ground: Weakly Supervised Phrase Grounding Guided By Image-Caption Alignment30.412019
Lemotif: Abstract Visual Depictions of your Emotional States in Life.00.342019
Counterfactual Visual Explanations.00.342019
CLEVR-Dialog: A Diagnostic Dataset for Multi-Round Reasoning in Visual Dialog.10.352019
Taking A Hint: Leveraging Explanations To Make Vision And Language Models More Grounded91.172019
Modeling the Long Term Future in Model-Based Reinforcement Learning20.352019
RUBi: Reducing Unimodal Biases in Visual Question Answering.20.352019
Embodied Visual Recognition.00.342019
Emergence of Compositional Language with Deep Generational Transmission.20.372019
Trick or TReAT: Thematic Reinforcement for Artistic Typography.00.342019
Learning Dynamics Model in Reinforcement Learning by Incorporating the Long Term Future.20.362019
Improving Generative Visual Dialog by Answering Diverse Questions10.352019
Response to "Visual Dialogue without Vision or Dialogue" (Massiceti et al., 2018).00.342019
Cross-channel Communication Networks00.342019
Neural Modular Control for Embodied Question Answering.90.442018
Punny Captions: Witty Wordplay in Image Descriptions.10.362018
Do explanations make VQA models more predictable to a human?40.382018
Pythia v0.1: the Winning Entry to the VQA Challenge 2018.90.502018
Embodied Question Answering20.402018
Audio Visual Scene-Aware Dialog (AVSD) Challenge at DSTC7.30.422018
End-To-End Audio Visual Scene-Aware Dialog Using Multimodal Attention-Based Video Features50.462018
Deal or No Deal? End-to-End Learning of Negotiation Dialogues.361.212017
Active Learning for Visual Question Answering: An Empirical Study.50.432017
C-VQA: A Compositional Split of the Visual Question Answering (VQA) v1.0 Dataset.50.432017
ParlAI: A Dialog Research Software Platform.150.732017
Evaluating Visual Conversational Agents via Cooperative Human-AI Games.40.432017
It Takes Two to Tango: Towards Theory of AI's Mind.40.452017
LR-GAN: Layered Recursive Generative Adversarial Networks for Image Generation.170.652017
Best of Both Worlds: Transferring Knowledge from Discriminative Learning to a Generative Visual Dialog Model.130.592017
Cooperative Learning with Visual Attributes.10.352017
Visual Dialog.00.342017
  • 1
  • 2