Skip to main content

Showing 1–50 of 66 results for author: Eisner, J

Searching in archive cs. Search in all archives.
.
  1. arXiv:2602.07812  [pdf, ps, other

    cs.CL

    LLMs Know More About Numbers than They Can Say

    Authors: Fengting Yuchi, Li Du, Jason Eisner

    Abstract: Although state-of-the-art LLMs can solve math problems, we find that they make errors on numerical comparisons with mixed notation: "Which is larger, $5.7 \times 10^2$ or $580$?" This raises a fundamental question: Do LLMs even know how big these numbers are? We probe the hidden states of several smaller open-source LLMs. A single linear projection of an appropriate hidden layer encodes the log-ma… ▽ More

    Submitted 17 February, 2026; v1 submitted 7 February, 2026; originally announced February 2026.

    Comments: EACL 2026 (Oral), camera-ready version with GitHub link

  2. arXiv:2512.23693  [pdf, ps, other

    cs.CL

    Fine-Tuning LLMs with Fine-Grained Human Feedback on Text Spans

    Authors: Sky CH-Wang, Justin Svegliato, Helen Appel, Jason Eisner

    Abstract: We present a method and dataset for fine-tuning language models with preference supervision using feedback-driven improvement chains. Given a model response, an annotator provides fine-grained feedback by marking ``liked'' and ``disliked'' spans and specifying what they liked or disliked about them. The base model then rewrites the disliked spans accordingly, proceeding from left to right, forming… ▽ More

    Submitted 29 December, 2025; originally announced December 2025.

    ACM Class: I.2.7

  3. arXiv:2512.23665  [pdf, ps, other

    cs.PL

    Automating the Analysis of Parsing Algorithms (and other Dynamic Programs)

    Authors: Tim Vieira, Ryan Cotterell, Jason Eisner

    Abstract: Much algorithmic research in NLP aims to efficiently manipulate rich formal structures. An algorithm designer typically seeks to provide guarantees about their proposed algorithm -- for example, that its running time or space complexity is upper-bounded as a certain function of its input size. They may also wish to determine the necessary properties of the quantities derived by the algorithm to sy… ▽ More

    Submitted 29 December, 2025; originally announced December 2025.

    Comments: 2021 manuscript; accepted by TACL but not revised for publication

    ACM Class: F.3.1; I.2.7

  4. arXiv:2512.23049  [pdf, ps, other

    cs.CL

    Accelerating Language Model Workflows with Prompt Choreography

    Authors: TJ Bai, Jason Eisner

    Abstract: Large language models are increasingly deployed in multi-agent workflows. We introduce Prompt Choreography, a framework that efficiently executes LLM workflows by maintaining a dynamic, global KV cache. Each LLM call can attend to an arbitrary, reordered subset of previously encoded messages. Parallel calls are supported. Though caching messages' encodings sometimes gives different results from re… ▽ More

    Submitted 28 December, 2025; originally announced December 2025.

    Comments: to appear in TACL (final preprint of 2025-10-12); 10 pages + appendices

    ACM Class: I.2.7; I.5.1; I.5.5; C.1.4

  5. arXiv:2504.20168  [pdf, other

    cs.CL cs.AI cs.LG

    MICE for CATs: Model-Internal Confidence Estimation for Calibrating Agents with Tools

    Authors: Nishant Subramani, Jason Eisner, Justin Svegliato, Benjamin Van Durme, Yu Su, Sam Thomson

    Abstract: Tool-using agents that act in the world need to be both useful and safe. Well-calibrated model confidences can be used to weigh the risk versus reward of potential actions, but prior work shows that many models are poorly calibrated. Inspired by interpretability literature exploring the internals of models, we propose a novel class of model-internal confidence estimators (MICE) to better assess co… ▽ More

    Submitted 28 April, 2025; originally announced April 2025.

    Comments: Accepted at NAACL 2025. Code: https://github.com/microsoft/mice_for_cats

  6. arXiv:2504.13139  [pdf, other

    cs.CL cs.AI cs.LG

    Syntactic and Semantic Control of Large Language Models via Sequential Monte Carlo

    Authors: João Loula, Benjamin LeBrun, Li Du, Ben Lipkin, Clemente Pasti, Gabriel Grand, Tianyu Liu, Yahya Emara, Marjorie Freedman, Jason Eisner, Ryan Cotterell, Vikash Mansinghka, Alexander K. Lew, Tim Vieira, Timothy J. O'Donnell

    Abstract: A wide range of LM applications require generating text that conforms to syntactic or semantic constraints. Imposing such constraints can be naturally framed as probabilistic conditioning, but exact generation from the resulting distribution -- which can differ substantially from the LM's base distribution -- is generally intractable. In this work, we develop an architecture for controlled LM gene… ▽ More

    Submitted 18 April, 2025; v1 submitted 17 April, 2025; originally announced April 2025.

    Comments: 34 pages, 4 figures

  7. arXiv:2504.05410  [pdf, ps, other

    cs.CL cs.AI cs.LG

    Fast Controlled Generation from Language Models with Adaptive Weighted Rejection Sampling

    Authors: Benjamin Lipkin, Benjamin LeBrun, Jacob Hoover Vigly, João Loula, David R. MacIver, Li Du, Jason Eisner, Ryan Cotterell, Vikash Mansinghka, Timothy J. O'Donnell, Alexander K. Lew, Tim Vieira

    Abstract: The dominant approach to generating from language models subject to some constraint is locally constrained decoding (LCD), incrementally sampling tokens at each time step such that the constraint is never violated. Typically, this is achieved through token masking: looping over the vocabulary and excluding non-conforming tokens. There are two important problems with this approach. (i) Evaluating t… ▽ More

    Submitted 18 August, 2025; v1 submitted 7 April, 2025; originally announced April 2025.

    Comments: COLM 2025

  8. LLM-Rubric: A Multidimensional, Calibrated Approach to Automated Evaluation of Natural Language Texts

    Authors: Helia Hashemi, Jason Eisner, Corby Rosset, Benjamin Van Durme, Chris Kedzie

    Abstract: This paper introduces a framework for the automated evaluation of natural language texts. A manually constructed rubric describes how to assess multiple dimensions of interest. To evaluate a text, a large language model (LLM) is prompted with each rubric question and produces a distribution over potential responses. The LLM predictions often fail to agree well with human judges -- indeed, the huma… ▽ More

    Submitted 30 December, 2024; originally announced January 2025.

    Comments: Updated version of 17 June 2024

    ACM Class: I.2.1; I.2.6; I.2.7

    Journal ref: Proceedings of ACL 2024 (Volume 1: Long Papers), pp. 13806-13834

  9. arXiv:2412.02081  [pdf, other

    cs.CL

    Let's Think Var-by-Var: Large Language Models Enable Ad Hoc Probabilistic Reasoning

    Authors: Shepard Xia, Brian Lu, Jason Eisner

    Abstract: A hallmark of intelligence is the ability to flesh out underspecified situations using "common sense." We propose to extract that common sense from large language models (LLMs), in a form that can feed into probabilistic inference. We focus our investigation on $\textit{guesstimation}$ questions such as "How much are Airbnb listings in Newark, NJ?" Formulating a sensible answer without access to d… ▽ More

    Submitted 2 December, 2024; originally announced December 2024.

  10. arXiv:2410.07063  [pdf, ps, other

    cs.LG

    InAttention: Linear Context Scaling for Transformers

    Authors: Joseph Eisner

    Abstract: VRAM requirements for transformer models scale quadratically with context length due to the self-attention mechanism. In this paper we modify the decoder-only transformer, replacing self-attention with InAttention, which scales linearly with context length during inference by having tokens attend only to initial states. Benchmarking shows that InAttention significantly reduces VRAM usage during in… ▽ More

    Submitted 9 October, 2024; originally announced October 2024.

  11. arXiv:2406.14739  [pdf, other

    cs.CL

    Learning to Retrieve Iteratively for In-Context Learning

    Authors: Yunmo Chen, Tongfei Chen, Harsh Jhamtani, Patrick Xia, Richard Shin, Jason Eisner, Benjamin Van Durme

    Abstract: We introduce iterative retrieval, a novel framework that empowers retrievers to make iterative decisions through policy optimization. Finding an optimal portfolio of retrieved items is a combinatorial optimization problem, generally considered NP-hard. This approach provides a learned approximation to such a solution, meeting specific task requirements under a given family of large language models… ▽ More

    Submitted 20 June, 2024; originally announced June 2024.

  12. arXiv:2403.04746  [pdf, other

    cs.CL cs.AI cs.LG

    LLMs in the Imaginarium: Tool Learning through Simulated Trial and Error

    Authors: Boshi Wang, Hao Fang, Jason Eisner, Benjamin Van Durme, Yu Su

    Abstract: Tools are essential for large language models (LLMs) to acquire up-to-date information and take consequential actions in external environments. Existing work on tool-augmented LLMs primarily focuses on the broad coverage of tools and the flexibility of adding new tools. However, a critical aspect that has surprisingly been understudied is simply how accurately an LLM uses tools for which it has be… ▽ More

    Submitted 7 March, 2024; originally announced March 2024.

    Comments: Code and data available at https://github.com/microsoft/simulated-trial-and-error

  13. arXiv:2312.17710  [pdf, other

    cs.CL cs.LG

    Principled Gradient-based Markov Chain Monte Carlo for Text Generation

    Authors: Li Du, Afra Amini, Lucas Torroba Hennigen, Xinyan Velocity Yu, Jason Eisner, Holden Lee, Ryan Cotterell

    Abstract: Recent papers have demonstrated the possibility of energy-based text generation by adapting gradient-based sampling algorithms, a paradigm of MCMC algorithms that promises fast convergence. However, as we show in this paper, previous attempts on this approach to text generation all fail to sample correctly from the target language model distributions. To address this limitation, we consider the pr… ▽ More

    Submitted 29 December, 2023; originally announced December 2023.

    Comments: Preprint

  14. arXiv:2312.17249  [pdf, other

    cs.CL cs.AI cs.LG

    Do Androids Know They're Only Dreaming of Electric Sheep?

    Authors: Sky CH-Wang, Benjamin Van Durme, Jason Eisner, Chris Kedzie

    Abstract: We design probes trained on the internal representations of a transformer language model to predict its hallucinatory behavior on three grounded generation tasks. To train the probes, we annotate for span-level hallucination on both sampled (organic) and manually edited (synthetic) reference outputs. Our probes are narrowly trained and we find that they are sensitive to their training domain: they… ▽ More

    Submitted 8 June, 2024; v1 submitted 28 December, 2023; originally announced December 2023.

    Comments: ACL 2024 (Findings) Camera-Ready

  15. arXiv:2312.13614  [pdf, other

    cs.LG cs.CL

    Structure-Aware Path Inference for Neural Finite State Transducers

    Authors: Weiting Tan, Chu-cheng Lin, Jason Eisner

    Abstract: Neural finite-state transducers (NFSTs) form an expressive family of neurosymbolic sequence transduction models. An NFST models each string pair as having been generated by a latent path in a finite-state transducer. As they are deep generative models, both training and inference of NFSTs require inference networks that approximate posterior distributions over such latent variables. In this paper,… ▽ More

    Submitted 21 December, 2023; originally announced December 2023.

    Comments: In Proceedings of ICBINB Workshop at NeurIPS 2023

  16. arXiv:2312.02073  [pdf, other

    cs.CL cs.AI cs.LG

    A Glitch in the Matrix? Locating and Detecting Language Model Grounding with Fakepedia

    Authors: Giovanni Monea, Maxime Peyrard, Martin Josifoski, Vishrav Chaudhary, Jason Eisner, Emre Kıcıman, Hamid Palangi, Barun Patra, Robert West

    Abstract: Large language models (LLMs) have an impressive ability to draw on novel information supplied in their context. Yet the mechanisms underlying this contextual grounding remain unknown, especially in situations where contextual information contradicts factual knowledge stored in the parameters, which LLMs also excel at recalling. Favoring the contextual information is critical for retrieval-augmente… ▽ More

    Submitted 10 June, 2024; v1 submitted 4 December, 2023; originally announced December 2023.

    Comments: Accepted at ACL 2024 (main conference)

  17. arXiv:2311.09796  [pdf, other

    cs.CL cs.AI

    Interpreting User Requests in the Context of Natural Language Standing Instructions

    Authors: Nikita Moghe, Patrick Xia, Jacob Andreas, Jason Eisner, Benjamin Van Durme, Harsh Jhamtani

    Abstract: Users of natural language interfaces, generally powered by Large Language Models (LLMs),often must repeat their preferences each time they make a similar request. We describe an approach to LLM-based dialogue modeling in which persistent user constraints and preferences -- collectively termed standing instructions -- as additional context for such interfaces. For example, when a user states "I'm h… ▽ More

    Submitted 7 March, 2024; v1 submitted 16 November, 2023; originally announced November 2023.

    Comments: Updated with results from LLaMA-2

  18. arXiv:2309.13075  [pdf, other

    cs.AI cs.CL cs.LG

    SCREWS: A Modular Framework for Reasoning with Revisions

    Authors: Kumar Shridhar, Harsh Jhamtani, Hao Fang, Benjamin Van Durme, Jason Eisner, Patrick Xia

    Abstract: Large language models (LLMs) can improve their accuracy on various tasks through iteratively refining and revising their output based on feedback. We observe that these revisions can introduce errors, in which case it is better to roll back to a previous result. Further, revisions are typically homogeneous: they use the same reasoning method that produced the initial answer, which may not correct… ▽ More

    Submitted 20 September, 2023; originally announced September 2023.

  19. arXiv:2307.04008  [pdf, other

    cs.CL

    Toward Interactive Dictation

    Authors: Belinda Z. Li, Jason Eisner, Adam Pauls, Sam Thomson

    Abstract: Voice dictation is an increasingly important text input modality. Existing systems that allow both dictation and editing-by-voice restrict their command language to flat templates invoked by trigger words. In this work, we study the feasibility of allowing users to interrupt their dictation with spoken editing commands in open-ended natural language. We introduce a new task and dataset, TERTiUS, t… ▽ More

    Submitted 8 July, 2023; originally announced July 2023.

    Comments: 17 pages, 5 tables, 4 figures; ACL

  20. arXiv:2307.02982  [pdf, other

    cs.CL cs.DS cs.FL

    Efficient Semiring-Weighted Earley Parsing

    Authors: Andreas Opedal, Ran Zmigrod, Tim Vieira, Ryan Cotterell, Jason Eisner

    Abstract: This paper provides a reference description, in the form of a deduction system, of Earley's (1970) context-free parsing algorithm with various speed-ups. Our presentation includes a known worst-case runtime improvement from Earley's $O (N^3|G||R|)$, which is unworkable for the large grammars that arise in natural language processing, to $O (N^3|G|)$, which matches the runtime of CKY on a binarized… ▽ More

    Submitted 6 July, 2023; originally announced July 2023.

    Comments: Main conference long paper at ACL 2023

  21. arXiv:2305.20076  [pdf, other

    cs.CL cs.AI

    Decision-Oriented Dialogue for Human-AI Collaboration

    Authors: Jessy Lin, Nicholas Tomlin, Jacob Andreas, Jason Eisner

    Abstract: We describe a class of tasks called decision-oriented dialogues, in which AI assistants such as large language models (LMs) must collaborate with one or more humans via natural language to help them make complex decisions. We formalize three domains in which users face everyday decisions: (1) choosing an assignment of reviewers to conference papers, (2) planning a multi-step itinerary in a city, a… ▽ More

    Submitted 5 May, 2024; v1 submitted 31 May, 2023; originally announced May 2023.

    Comments: TACL 2024, pre-MIT Press publication version

  22. arXiv:2305.12272  [pdf, other

    cs.CL cs.AI cs.LG

    Autoregressive Modeling with Lookahead Attention

    Authors: Li Du, Hongyuan Mei, Jason Eisner

    Abstract: To predict the next token, autoregressive models ordinarily examine the past. Could they also benefit from also examining hypothetical futures? We consider a novel Transformer-based autoregressive architecture that estimates the next-token distribution by extrapolating multiple continuations of the past, according to some proposal distribution, and attending to these extended strings. This archite… ▽ More

    Submitted 20 May, 2023; originally announced May 2023.

  23. arXiv:2301.06862  [pdf, other

    cs.DS cs.CL

    Algorithms for Acyclic Weighted Finite-State Automata with Failure Arcs

    Authors: Anej Svete, Benjamin Dayan, Tim Vieira, Ryan Cotterell, Jason Eisner

    Abstract: Weighted finite-state automata (WSFAs) are commonly used in NLP. Failure transitions are a useful extension for compactly representing backoffs or interpolation in $n$-gram models and CRFs, which are special cases of WFSAs. The pathsum in ordinary acyclic WFSAs is efficiently computed by the backward algorithm in time $O(|E|)$, where $E$ is the set of transitions. However, this does not allow fail… ▽ More

    Submitted 11 July, 2023; v1 submitted 17 January, 2023; originally announced January 2023.

    Comments: 9 pages, Proceedings of EMNLP 2022

  24. arXiv:2212.10520  [pdf, other

    cs.CL

    Privacy-Preserving Domain Adaptation of Semantic Parsers

    Authors: Fatemehsadat Mireshghallah, Yu Su, Tatsunori Hashimoto, Jason Eisner, Richard Shin

    Abstract: Task-oriented dialogue systems often assist users with personal or confidential matters. For this reason, the developers of such a system are generally prohibited from observing actual usage. So how can they know where the system is failing and needs more training data or new functionality? In this work, we study ways in which realistic user utterances can be generated synthetically, to help incre… ▽ More

    Submitted 8 June, 2023; v1 submitted 20 December, 2022; originally announced December 2022.

    Comments: ACL 2023

  25. A Measure-Theoretic Characterization of Tight Language Models

    Authors: Li Du, Lucas Torroba Hennigen, Tiago Pimentel, Clara Meister, Jason Eisner, Ryan Cotterell

    Abstract: Language modeling, a central task in natural language processing, involves estimating a probability distribution over strings. In most cases, the estimated distribution sums to 1 over all finite strings. However, in some pathological cases, probability mass can ``leak'' onto the set of infinite sequences. In order to characterize the notion of leakage more precisely, this paper offers a measure-th… ▽ More

    Submitted 21 August, 2023; v1 submitted 20 December, 2022; originally announced December 2022.

    Comments: 25 pages; ACL 2023 camera ready

  26. arXiv:2210.15097  [pdf, other

    cs.CL cs.AI cs.LG

    Contrastive Decoding: Open-ended Text Generation as Optimization

    Authors: Xiang Lisa Li, Ari Holtzman, Daniel Fried, Percy Liang, Jason Eisner, Tatsunori Hashimoto, Luke Zettlemoyer, Mike Lewis

    Abstract: Given a language model (LM), maximum probability is a poor decoding objective for open-ended generation, because it produces short and repetitive text. On the other hand, sampling can often produce incoherent text that drifts from the original topics. We propose contrastive decoding (CD), a reliable decoding approach that optimizes a contrastive objective subject to a plausibility constraint. The… ▽ More

    Submitted 10 July, 2023; v1 submitted 26 October, 2022; originally announced October 2022.

    Comments: Main conference long paper at ACL 2023

  27. arXiv:2209.07800  [pdf, other

    cs.CL

    The Whole Truth and Nothing But the Truth: Faithful and Controllable Dialogue Response Generation with Dataflow Transduction and Constrained Decoding

    Authors: Hao Fang, Anusha Balakrishnan, Harsh Jhamtani, John Bufe, Jean Crawford, Jayant Krishnamurthy, Adam Pauls, Jason Eisner, Jacob Andreas, Dan Klein

    Abstract: In a real-world dialogue system, generated text must be truthful and informative while remaining fluent and adhering to a prescribed style. Satisfying these constraints simultaneously is difficult for the two predominant paradigms in language generation: neural language modeling and rule-based generation. We describe a hybrid architecture for dialogue response generation that combines the strength… ▽ More

    Submitted 26 May, 2023; v1 submitted 16 September, 2022; originally announced September 2022.

    Comments: Findings of ACL 2023

  28. arXiv:2209.06809  [pdf, other

    cs.FL cs.CL

    On the Intersection of Context-Free and Regular Languages

    Authors: Clemente Pasti, Andreas Opedal, Tiago Pimentel, Tim Vieira, Jason Eisner, Ryan Cotterell

    Abstract: The Bar-Hillel construction is a classic result in formal language theory. It shows, by a simple construction, that the intersection of a context-free language and a regular language is itself context-free. In the construction, the regular language is specified by a finite-state automaton. However, neither the original construction (Bar-Hillel et al., 1961) nor its weighted extension (Nederhof and… ▽ More

    Submitted 18 May, 2023; v1 submitted 14 September, 2022; originally announced September 2022.

    Comments: EACL 2023 camera ready version. Our code is available in https://github.com/rycolab/bar-hillel

  29. arXiv:2206.10668  [pdf, ps, other

    cs.CL

    BenchCLAMP: A Benchmark for Evaluating Language Models on Syntactic and Semantic Parsing

    Authors: Subhro Roy, Sam Thomson, Tongfei Chen, Richard Shin, Adam Pauls, Jason Eisner, Benjamin Van Durme

    Abstract: Recent work has shown that generation from a prompted or fine-tuned language model can perform well at semantic parsing when the output is constrained to be a valid semantic representation. We introduce BenchCLAMP, a Benchmark to evaluate Constrained LAnguage Model Parsing, that includes context-free grammars for seven semantic parsing datasets and two syntactic parsing datasets with varied output… ▽ More

    Submitted 10 January, 2024; v1 submitted 21 June, 2022; originally announced June 2022.

    Comments: Neural Information Processing Systems (NeurIPS 2023) Track on Datasets and Benchmarks

  30. arXiv:2205.12422  [pdf, other

    cs.CL cs.AI cs.PL

    Non-Programmers Can Label Programs Indirectly via Active Examples: A Case Study with Text-to-SQL

    Authors: Ruiqi Zhong, Charlie Snell, Dan Klein, Jason Eisner

    Abstract: Can non-programmers annotate natural language utterances with complex programs that represent their meaning? We introduce APEL, a framework in which non-programmers select among candidate programs generated by a seed semantic parser (e.g., Codex). Since they cannot understand the candidate programs, we ask them to select indirectly by examining the programs' input-ouput examples. For each utteranc… ▽ More

    Submitted 23 October, 2023; v1 submitted 24 May, 2022; originally announced May 2022.

  31. arXiv:2205.12228  [pdf, other

    cs.CL

    When More Data Hurts: A Troubling Quirk in Developing Broad-Coverage Natural Language Understanding Systems

    Authors: Elias Stengel-Eskin, Emmanouil Antonios Platanios, Adam Pauls, Sam Thomson, Hao Fang, Benjamin Van Durme, Jason Eisner, Yu Su

    Abstract: In natural language understanding (NLU) production systems, users' evolving needs necessitate the addition of new features over time, indexed by new symbols added to the meaning representation space. This requires additional training data and results in ever-growing datasets. We present the first systematic investigation of this incremental symbol learning scenario. Our analysis reveals a troublin… ▽ More

    Submitted 8 November, 2022; v1 submitted 24 May, 2022; originally announced May 2022.

    Comments: EMNLP 2022

  32. arXiv:2201.00044  [pdf, other

    cs.LG cs.AI cs.LO

    Transformer Embeddings of Irregularly Spaced Events and Their Participants

    Authors: Chenghao Yang, Hongyuan Mei, Jason Eisner

    Abstract: The neural Hawkes process (Mei & Eisner, 2017) is a generative model of irregularly spaced sequences of discrete events. To handle complex domains with many event types, Mei et al. (2020a) further consider a setting in which each event in the sequence updates a deductive database of facts (via domain-specific pattern-matching rules); future events are then conditioned on the database contents. The… ▽ More

    Submitted 6 May, 2022; v1 submitted 31 December, 2021; originally announced January 2022.

    Comments: ICLR 2022 Final

  33. arXiv:2109.06966  [pdf, other

    cs.CL

    Searching for More Efficient Dynamic Programs

    Authors: Tim Vieira, Ryan Cotterell, Jason Eisner

    Abstract: Computational models of human language often involve combinatorial problems. For instance, a probabilistic parser may marginalize over exponentially many trees to make predictions. Algorithms for such problems often employ dynamic programming and are not always unique. Finding one with optimal asymptotic runtime can be unintuitive, time-consuming, and error-prone. Our work aims to automate this la… ▽ More

    Submitted 14 September, 2021; originally announced September 2021.

  34. arXiv:2104.08768  [pdf, other

    cs.CL

    Constrained Language Models Yield Few-Shot Semantic Parsers

    Authors: Richard Shin, Christopher H. Lin, Sam Thomson, Charles Chen, Subhro Roy, Emmanouil Antonios Platanios, Adam Pauls, Dan Klein, Jason Eisner, Benjamin Van Durme

    Abstract: We explore the use of large pretrained language models as few-shot semantic parsers. The goal in semantic parsing is to generate a structured meaning representation given a natural language input. However, language models are trained to generate natural language. To bridge the gap, we use language models to paraphrase inputs into a controlled sublanguage resembling English that can be automaticall… ▽ More

    Submitted 16 November, 2021; v1 submitted 18 April, 2021; originally announced April 2021.

    Comments: EMNLP 2021. Code is available at https://github.com/microsoft/semantic_parsing_with_constrained_lm

  35. Learning How to Ask: Querying LMs with Mixtures of Soft Prompts

    Authors: Guanghui Qin, Jason Eisner

    Abstract: Natural-language prompts have recently been used to coax pretrained language models into performing other AI tasks, using a fill-in-the-blank paradigm (Petroni et al., 2019) or a few-shot extrapolation paradigm (Brown et al., 2020). For example, language models retain factual knowledge from their training corpora that can be extracted by asking them to "fill in the blank" in a sentential prompt. H… ▽ More

    Submitted 13 April, 2021; originally announced April 2021.

    Comments: NAACL-HLT 2021 camera-ready

    Journal ref: NAACL-HLT 2021

  36. arXiv:2011.00717  [pdf, other

    cs.LG stat.ML

    Noise-Contrastive Estimation for Multivariate Point Processes

    Authors: Hongyuan Mei, Tom Wan, Jason Eisner

    Abstract: The log-likelihood of a generative model often involves both positive and negative terms. For a temporal multivariate point process, the negative term sums over all the possible event types at each time and also integrates over all the possible times. As a result, maximum likelihood estimation is expensive. We show how to instead apply a version of noise-contrastive estimation---a general paramete… ▽ More

    Submitted 1 November, 2020; originally announced November 2020.

    Comments: NeurIPS 2020 camera-ready

  37. arXiv:2010.11939  [pdf, other

    cs.LG cs.CL stat.ML

    Limitations of Autoregressive Models and Their Alternatives

    Authors: Chu-Cheng Lin, Aaron Jaech, Xin Li, Matthew R. Gormley, Jason Eisner

    Abstract: Standard autoregressive language models perform only polynomial-time computation to compute the probability of the next symbol. While this is attractive, it means they cannot model distributions whose next-symbol probability is hard to compute. Indeed, they cannot even model them well enough to solve associated easy decision problems for which an engineer might want to consult a language model. Th… ▽ More

    Submitted 30 May, 2021; v1 submitted 22 October, 2020; originally announced October 2020.

    Comments: NAACL 2021 (same content, more relaxed layout)

  38. arXiv:2010.10503  [pdf, ps, other

    cs.PL cs.SC

    Evaluation of Logic Programs with Built-Ins and Aggregation: A Calculus for Bag Relations

    Authors: Matthew Francis-Landau, Tim Vieira, Jason Eisner

    Abstract: We present a scheme for translating logic programs, which may use aggregation and arithmetic, into algebraic expressions that denote bag relations over ground terms of the Herbrand universe. To evaluate queries against these relations, we develop an operational semantics based on term rewriting of the algebraic expressions. This approach can exploit arithmetic identities and recovers a range of us… ▽ More

    Submitted 20 October, 2020; originally announced October 2020.

    Comments: An earlier version of this paper appeared at WRLA 2020

  39. Task-Oriented Dialogue as Dataflow Synthesis

    Authors: Semantic Machines, Jacob Andreas, John Bufe, David Burkett, Charles Chen, Josh Clausman, Jean Crawford, Kate Crim, Jordan DeLoach, Leah Dorner, Jason Eisner, Hao Fang, Alan Guo, David Hall, Kristin Hayes, Kellie Hill, Diana Ho, Wendy Iwaszuk, Smriti Jha, Dan Klein, Jayant Krishnamurthy, Theo Lanman, Percy Liang, Christopher H Lin, Ilya Lintsbakh , et al. (21 additional authors not shown)

    Abstract: We describe an approach to task-oriented dialogue in which dialogue state is represented as a dataflow graph. A dialogue agent maps each user utterance to a program that extends this graph. Programs include metacomputation operators for reference and revision that reuse dataflow fragments from previous turns. Our graph-based state enables the expression and manipulation of complex user intents, an… ▽ More

    Submitted 10 February, 2021; v1 submitted 23 September, 2020; originally announced September 2020.

    Journal ref: Transactions of the Association for Computational Linguistics 2020 Vol. 8, 556-571

  40. arXiv:2006.16723  [pdf, other

    cs.LG cs.AI cs.DB cs.LO stat.ML

    Neural Datalog Through Time: Informed Temporal Modeling via Logical Specification

    Authors: Hongyuan Mei, Guanghui Qin, Minjie Xu, Jason Eisner

    Abstract: Learning how to predict future events from patterns of past events is difficult when the set of possible event types is large. Training an unrestricted neural model might overfit to spurious patterns. To exploit domain-specific knowledge of how past events might affect an event's present probability, we propose using a temporal deductive database to track structured facts over time. Rules serve to… ▽ More

    Submitted 16 August, 2020; v1 submitted 30 June, 2020; originally announced June 2020.

    Comments: ICML 2020 camera-ready (new Appendix A.3, rewritten Appendix F)

  41. arXiv:2005.13962  [pdf, other

    cs.CL

    A Corpus for Large-Scale Phonetic Typology

    Authors: Elizabeth Salesky, Eleanor Chodroff, Tiago Pimentel, Matthew Wiesner, Ryan Cotterell, Alan W Black, Jason Eisner

    Abstract: A major hurdle in data-driven research on typology is having sufficient data in many languages to draw meaningful conclusions. We present VoxClamantis v1.0, the first large-scale corpus for phonetic typology, with aligned segments and estimated phoneme-level labels in 690 readings spanning 635 languages, along with acoustic-phonetic measures of vowels and sibilants. Access to such data can greatly… ▽ More

    Submitted 28 May, 2020; originally announced May 2020.

    Comments: Accepted to ACL2020

  42. arXiv:1910.00163  [pdf, other

    cs.CL cs.LG

    Specializing Word Embeddings (for Parsing) by Information Bottleneck

    Authors: Xiang Lisa Li, Jason Eisner

    Abstract: Pre-trained word embeddings like ELMo and BERT contain rich syntactic and semantic information, resulting in state-of-the-art performance on various tasks. We propose a very fast variational information bottleneck (VIB) method to nonlinearly compress these embeddings, keeping only the information that helps a discriminative parser. We compress each word embedding to either a discrete tag or a cont… ▽ More

    Submitted 30 September, 2019; originally announced October 2019.

    Comments: Accepted for publication at EMNLP 2019

  43. arXiv:1906.11298  [pdf, other

    cs.CL cs.LG

    A Generative Model for Punctuation in Dependency Trees

    Authors: Xiang Lisa Li, Dingquan Wang, Jason Eisner

    Abstract: Treebanks traditionally treat punctuation marks as ordinary words, but linguists have suggested that a tree's "true" punctuation marks are not observed (Nunberg, 1990). These latent "underlying" marks serve to delimit or separate constituents in the syntax tree. When the tree's yield is rendered as a written sentence, a string rewriting mechanism transduces the underlying marks into "surface" mark… ▽ More

    Submitted 26 June, 2019; originally announced June 2019.

  44. arXiv:1906.04726  [pdf, other

    cs.CL

    What Kind of Language Is Hard to Language-Model?

    Authors: Sabrina J. Mielke, Ryan Cotterell, Kyle Gorman, Brian Roark, Jason Eisner

    Abstract: How language-agnostic are current state-of-the-art NLP tools? Are there some types of language that are easier to model with current methods? In prior work (Cotterell et al., 2018) we attempted to address this question for language modeling, and observed that recurrent neural network language models do not perform equally well over all the high-resource European languages found in the Europarl cor… ▽ More

    Submitted 25 February, 2020; v1 submitted 11 June, 2019; originally announced June 2019.

    Comments: Published at ACL 2019

  45. arXiv:1905.05570  [pdf, other

    cs.LG cs.AI stat.ML

    Imputing Missing Events in Continuous-Time Event Streams

    Authors: Hongyuan Mei, Guanghui Qin, Jason Eisner

    Abstract: Events in the world may be caused by other, unobserved events. We consider sequences of events in continuous time. Given a probability model of complete sequences, we propose particle smoothing---a form of sequential importance sampling---to impute the missing events in an incomplete sequence. We develop a trainable family of proposal distributions based on a type of bidirectional continuous-time… ▽ More

    Submitted 14 May, 2019; originally announced May 2019.

    Comments: ICML 2019 camera-ready. The first version of this work appeared on OpenReview in September 2018

  46. arXiv:1905.01420  [pdf, ps, other

    cs.CL

    Contextualization of Morphological Inflection

    Authors: Ekaterina Vylomova, Ryan Cotterell, Timothy Baldwin, Trevor Cohn, Jason Eisner

    Abstract: Critical to natural language generation is the production of correctly inflected text. In this paper, we isolate the task of predicting a fully inflected sentence from its partially lemmatized version. Unlike traditional morphological inflection or surface realization, our task input does not provide ``gold'' tags that specify what morphological features to realize on each lemmatized word; rather,… ▽ More

    Submitted 3 May, 2019; originally announced May 2019.

    Comments: NAACL 2019

  47. arXiv:1810.11101  [pdf, other

    cs.CL

    UniMorph 2.0: Universal Morphology

    Authors: Christo Kirov, Ryan Cotterell, John Sylak-Glassman, Géraldine Walther, Ekaterina Vylomova, Patrick Xia, Manaal Faruqui, Sabrina J. Mielke, Arya D. McCarthy, Sandra Kübler, David Yarowsky, Jason Eisner, Mans Hulden

    Abstract: The Universal Morphology UniMorph project is a collaborative effort to improve how NLP handles complex morphology across the world's languages. The project releases annotated morphological data using a universal tagset, the UniMorph schema. Each inflected form is associated with a lemma, which typically carries its underlying lexical meaning, and a bundle of morphological features from our schema.… ▽ More

    Submitted 25 February, 2020; v1 submitted 25 October, 2018; originally announced October 2018.

    Comments: LREC 2018

  48. arXiv:1810.07125  [pdf, other

    cs.CL

    The CoNLL--SIGMORPHON 2018 Shared Task: Universal Morphological Reinflection

    Authors: Ryan Cotterell, Christo Kirov, John Sylak-Glassman, Géraldine Walther, Ekaterina Vylomova, Arya D. McCarthy, Katharina Kann, Sabrina J. Mielke, Garrett Nicolai, Miikka Silfverberg, David Yarowsky, Jason Eisner, Mans Hulden

    Abstract: The CoNLL--SIGMORPHON 2018 shared task on supervised learning of morphological generation featured data sets from 103 typologically diverse languages. Apart from extending the number of languages involved in earlier supervised tasks of generating inflected forms, this year the shared task also featured a new second task which asked participants to inflect words in sentential context, similar to a… ▽ More

    Submitted 25 February, 2020; v1 submitted 16 October, 2018; originally announced October 2018.

    Comments: CoNLL 2018. arXiv admin note: text overlap with arXiv:1706.09031

  49. arXiv:1807.02747  [pdf, other

    cs.CL

    On the Complexity and Typology of Inflectional Morphological Systems

    Authors: Ryan Cotterell, Christo Kirov, Mans Hulden, Jason Eisner

    Abstract: We quantify the linguistic complexity of different languages' morphological systems. We verify that there is an empirical trade-off between paradigm size and irregularity: a language's inflectional paradigms may be either large in size or highly irregular, but never both. Our methodology measures paradigm irregularity as the entropy of the surface realization of a paradigm -- how hard it is to joi… ▽ More

    Submitted 7 July, 2018; originally announced July 2018.

    Comments: TACL 2018

  50. arXiv:1807.02745  [pdf, other

    cs.CL

    A Deep Generative Model of Vowel Formant Typology

    Authors: Ryan Cotterell, Jason Eisner

    Abstract: What makes some types of languages more probable than others? For instance, we know that almost all spoken languages contain the vowel phoneme /i/; why should that be? The field of linguistic typology seeks to answer these questions and, thereby, divine the mechanisms that underlie human language. In our work, we tackle the problem of vowel system typology, i.e., we propose a generative probabilit… ▽ More

    Submitted 7 July, 2018; originally announced July 2018.

    Comments: NAACL 2018