Forrest Davis

Assistant Professor of Computer Science | Colgate University.

prof_pic.jpg

322 Bernstein Hall

Colgate University

Hamilton, NY 13346

I am an Assistant Professor of Computer Science at Colgate University. Before joining Colgate, I was a Postdoctoral Associate in the Linguistics and Philosophy Department at the Massachusetts Institute of Technology. I received my Ph.D. from the Department of Linguistics at Cornell University, where I was primarily advised by Marten van Schijndel, and my B.A. in Computer Science and Mathematics from Columbia University.

I am broadly interested in mismatches between our experiences with language and our knowledge of language. That is, I try to find in linguistic data systematic deviations from what we might expect given our knowledge of grammar. Findings like this excite me because they suggest cases where our minds extend beyond mere correspondence with experience. My primary tool at the moment is neural language models (e.g., large language models) trained on text data. Drawing on insights from psycholinguistics, linguistic theory, and cross-linguistic variation, I expose limitations in current AI models and tie these limitations to properties of training data. My dissertation titled “On the Limitations of Data: Mismatches between Neural Models of Language and Humans” sketches out my perspective with case studies.

Besides research and teaching, I love coffee, watching movies, swimming, hiking, and skiing. I cohabitate with a cat named Figaro, who spends his time sleeping throughout our home, sitting on my keyboard, and meowing for food.

news

Jan 10, 2026 Presented “What is It? Language Models and Expletive Forms” with Meg Gotowski as a talk at the 2026 Linguistic Society of America Annual Meeting. January 8-11, 2026.
Nov 8, 2025 Presented “Being (an LM) and Nothingness: How Models Interpret Expletive Forms” with Meg Gotowski as a poster at the 50th Annual Boston University Conference on Language Development. November 6-9, 2025.
Apr 11, 2025 “Discourse Sensitivity in Attraction Effects: The Interplay Between Language Model Size and Training Data” with Sanghee J. Kim accepted as a talk at the 8th Annual Meeting of the Society for Computation in Linguistics. July 18-20, 2025.
Sep 4, 2024 Humans vs. Machines: Comparing Adjective Learning Performance” accepted as a poster at the 2025 Linguistic Society of America Annual Meeting (with Megan Gotowski). January 9-12, 2025.
Jul 31, 2024 “Training an NLP Scholar at a Small Liberal Arts College: A Backwards Designed Course Proposal” to be presented as a talk at the Sixth Workshop on Teaching NLP (with Grusha Prasad). August 15, 2024.

selected publications

2025

  1. Discourse Sensitivity in Attraction Effects: The Interplay Between Language Model Size and Training Data
    Sanghee Kim, and Forrest Davis
    In Society for Computation in Lingusitics, 2025

2023

  1. Can Language Models Be Tricked by Language Illusions? Easier with Syntax, Harder with Semantics
    Yuhan ZhangEdward Gibson, and Forrest Davis
    In Proceedings of the 27th Conference on Computational Natural Language Learning (CoNLL), Dec 2023

2022

  1. Incremental Processing of Principle B: Mismatches Between Neural Models and Humans
    Forrest Davis
    In Proceedings of the 26th Conference on Computational Natural Language Learning (CoNLL), Dec 2022

2020

  1. Recurrent Neural Network Language Models Always Learn English-Like Relative Clause Attachment
    Forrest Davis, and Marten van Schijndel
    In Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics, Jul 2020