The llama 3 herd of models A Dubey, A Jauhri, A Pandey, A Kadian, A Al-Dahle, A Letman, A Mathur, ... arXiv preprint arXiv:2407.21783, 2024 | 1109 | 2024 |
Transform-retrieve-generate: Natural language-centric outside-knowledge visual question answering F Gao, Q Ping, G Thattai, A Reganti, YN Wu, P Natarajan Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern …, 2022 | 90 | 2022 |
Embodied bert: A transformer model for embodied, language-guided visual task completion A Suglia, Q Gao, J Thomason, G Thattai, G Sukhatme arXiv preprint arXiv:2108.04927, 2021 | 77 | 2021 |
Dialfred: Dialogue-enabled agents for embodied instruction following X Gao, Q Gao, R Gong, K Lin, G Thattai, GS Sukhatme IEEE Robotics and Automation Letters 7 (4), 10049-10056, 2022 | 71 | 2022 |
Alexa arena: A user-centric interactive platform for embodied ai Q Gao, G Thattai, S Shakiah, X Gao, S Pansare, V Sharma, G Sukhatme, ... Advances in Neural Information Processing Systems 36, 2024 | 25 | 2024 |
Learning better visual dialog agents with pretrained visual-linguistic representation T Tu, Q Ping, G Thattai, G Tur, P Natarajan Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern …, 2021 | 24 | 2021 |
LRTA: A transparent neural-symbolic reasoning framework with modular supervision for visual question answering W Liang, F Niu, A Reganti, G Thattai, G Tur arXiv preprint arXiv:2011.10731, 2020 | 22 | 2020 |
A thousand words are worth more than a picture: Natural language-centric outside-knowledge visual question answering F Gao, Q Ping, G Thattai, A Reganti, YN Wu, P Natarajan arXiv preprint arXiv:2201.05299, 2022 | 19 | 2022 |
Neural architecture search for parameter-efficient fine-tuning of large pre-trained language models N Lawton, A Kumar, G Thattai, A Galstyan, GV Steeg arXiv preprint arXiv:2305.16597, 2023 | 16 | 2023 |
Givl: Improving geographical inclusivity of vision-language models with pre-training methods D Yin, F Gao, G Thattai, M Johnston, KW Chang Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern …, 2023 | 16 | 2023 |
Luminous: Indoor scene generation for embodied ai challenges Y Zhao, K Lin, Z Jia, Q Gao, G Thattai, J Thomason, GS Sukhatme arXiv preprint arXiv:2111.05527, 2021 | 16 | 2021 |
Learning to act with affordance-aware multimodal neural slam Z Jia, K Lin, Y Zhao, Q Gao, G Thattai, GS Sukhatme 2022 IEEE/RSJ International Conference on Intelligent Robots and Systems …, 2022 | 15 | 2022 |
" polyaural" array processing for automatic speech recognition in degraded environments. RM Stern, EB Gouvêa, G Thattai Interspeech, 926-929, 2007 | 14 | 2007 |
Are we there yet? learning to localize in embodied instruction following S Storks, Q Gao, G Thattai, G Tur arXiv preprint arXiv:2101.03431, 2021 | 10 | 2021 |
Interactive teaching for conversational ai Q Ping, F Niu, G Thattai, J Chengottusseriyil, Q Gao, A Reganti, ... arXiv preprint arXiv:2012.00958, 2020 | 10 | 2020 |
Lemma: Learning language-conditioned multi-robot manipulation R Gong, X Gao, Q Gao, S Shakiah, G Thattai, GS Sukhatme IEEE Robotics and Automation Letters, 2023 | 9 | 2023 |
The 2004 BBN/LIMSI 20xRT english conversational telephone speech system R Prasad, S Matsoukas, CL Kao, J Ma, DX Xu, T Colthurst, G Thattai, ... Proc. Rich Transcription Workshop, 2004 | 8 | 2004 |
Towards reasoning-aware explainable vqa R Vaideeswaran, F Gao, A Mathur, G Thattai arXiv preprint arXiv:2211.05190, 2022 | 6 | 2022 |
Alexa, play with robot: Introducing the first alexa prize simbot challenge on embodied ai H Shi, L Ball, G Thattai, D Zhang, L Hu, Q Gao, S Shakiah, X Gao, ... arXiv preprint arXiv:2308.05221, 2023 | 5 | 2023 |
Opend: A benchmark for language-driven door and drawer opening Y Zhao, Q Gao, L Qiu, G Thattai, GS Sukhatme arXiv preprint arXiv:2212.05211, 2022 | 5 | 2022 |