Seguir
William Merrill
Título
Citado por
Citado por
Ano
CORD-19: The COVID-19 open research dataset
LL Wang, K Lo, Y Chandrasekhar, R Reas, J Yang, D Eide, K Funk, ...
Workshop on NLP for COVID-19, 2020
10422020
How language model hallucinations can snowball
M Zhang, O Press, W Merrill, A Liu, NA Smith
arXiv preprint arXiv:2305.13534, 2023
2702023
Olmo: Accelerating the science of language models
D Groeneveld, I Beltagy, P Walsh, A Bhagia, R Kinney, O Tafjord, AH Jha, ...
ACL 🏆 Best Theme Paper, 2024
1642024
ReCLIP: A Strong Zero-Shot Baseline for Referring Expression Comprehension
S Subramanian, W Merrill, T Darrell, M Gardner, S Singh, A Rohrbach
Empirical Methods in Natural Language Processing, 2022
1252022
Saturated transformers are constant-depth threshold circuits
W Merrill, A Sabharwal, NA Smith
Transactions of the Association for Computational Linguistics 10, 843-856, 2022
1092022
Competency problems: On finding and removing artifacts in language data
M Gardner, W Merrill, J Dodge, ME Peters, A Ross, S Singh, N Smith
Empirical Methods in Natural Language Processing, 2021
1032021
The Expressive Power of Transformers with Chain of Thought
W Merrill, A Sabharwal
ICLR 2024, 2023
982023
The Parallelism Tradeoff: Limitations of Log-Precision Transformers
W Merrill, A Sabharwal
TACL, 2022
942022
A formal hierarchy of RNN architectures
W Merrill, G Weiss, Y Goldberg, R Schwartz, NA Smith, E Yahav
Association of Computational Linguistics, 2020
892020
Sequential neural networks as automata
W Merrill
Deep Learning and Formal Languages (ACL workshop), 2019
832019
Provable limitations of acquiring meaning from ungrounded form: What will future language models understand?
W Merrill, Y Goldberg, R Schwartz, NA Smith
Transactions of the Association for Computational Linguistics 9, 1047-1060, 2021
752021
What formal languages can transformers express? a survey
L Strobl, W Merrill, G Weiss, D Chiang, D Angluin
Transactions of the Association for Computational Linguistics 12, 543-561, 2024
62*2024
The illusion of state in state-space models
W Merrill, J Petty, A Sabharwal
ICML, 2024
492024
A Logic for Expressing Log-Precision Transformers
W Merrill, A Sabharwal
NeurIPS 2023, 2022
482022
Let's Think Dot by Dot: Hidden Computation in Transformer Language Models
J Pfau, W Merrill, SR Bowman
Conference on Language Models, 2024
422024
Context-free transductions with neural stacks
Y Hao, W Merrill, D Angluin, R Frank, N Amsel, A Benz, S Mendelsohn
BlackboxNLP, 2018
422018
A tale of two circuits: Grokking as competition of sparse and dense subnetworks
W Merrill, N Tsilivis, A Shukla
arXiv preprint arXiv:2303.11873, 2023
392023
Effects of parameter norm growth during transformer training: Inductive bias from gradient descent
W Merrill, V Ramanujan, Y Goldberg, R Schwartz, N Smith
Empirical Methods in Natural Language Processing, 2021
372021
End-to-end graph-based TAG parsing with neural networks
J Kasai, R Frank, P Xu, W Merrill, O Rambow
NAACL, 2018
162018
Entailment Semantics Can Be Extracted from an Ideal Language Model
W Merrill, A Warstadt, T Linzen
CoNLL 2022, 2022
152022
O sistema não pode efectuar a operação agora. Tente mais tarde.
Artigos 1–20