Skip to main content

Showing 1–4 of 4 results for author: Broestl, N

Searching in archive cs. Search in all archives.
.
  1. arXiv:2510.00001  [pdf, ps, other

    cs.LG cs.AI cs.SE

    Methodological Framework for Quantifying Semantic Test Coverage in RAG Systems

    Authors: Noah Broestl, Adel Nasser Abdalla, Rajprakash Bale, Hersh Gupta, Max Struever

    Abstract: Reliably determining the performance of Retrieval-Augmented Generation (RAG) systems depends on comprehensive test questions. While a proliferation of evaluation frameworks for LLM-powered applications exists, current practices lack a systematic method to ensure these test sets adequately cover the underlying knowledge base, leaving developers with significant blind spots. To address this, we pres… ▽ More

    Submitted 13 August, 2025; originally announced October 2025.

    Comments: 7 pages, 3 figures, 1 table, 1 algo

  2. arXiv:2508.14119  [pdf, ps, other

    cs.CY cs.AI cs.HC

    Documenting Deployment with Fabric: A Repository of Real-World AI Governance

    Authors: Mackenzie Jorgensen, Kendall Brogle, Katherine M. Collins, Lujain Ibrahim, Arina Shah, Petra Ivanovic, Noah Broestl, Gabriel Piles, Paul Dongha, Hatim Abdulhussein, Adrian Weller, Jillian Powers, Umang Bhatt

    Abstract: Artificial intelligence (AI) is increasingly integrated into society, from financial services and traffic management to creative writing. Academic literature on the deployment of AI has mostly focused on the risks and harms that result from the use of AI. We introduce Fabric, a publicly available repository of deployed AI use cases to outline their governance mechanisms. Through semi-structured in… ▽ More

    Submitted 29 August, 2025; v1 submitted 18 August, 2025; originally announced August 2025.

    Comments: AIES 2025

  3. arXiv:2505.18779  [pdf, ps, other

    cs.CY

    Evaluating Intra-firm LLM Alignment Strategies in Business Contexts

    Authors: Noah Broestl, Benjamin Lange, Cristina Voinea, Geoff Keeling, Rachael Lam

    Abstract: Instruction-tuned Large Language Models (LLMs) are increasingly deployed as AI Assistants in firms for support in cognitive tasks. These AI assistants carry embedded perspectives which influence factors across the firm including decision-making, collaboration, and organizational culture. This paper argues that firms must align the perspectives of these AI Assistants intentionally with their object… ▽ More

    Submitted 24 May, 2025; originally announced May 2025.

    Comments: 9 pages, 0 figures

  4. arXiv:2107.04313  [pdf, other

    cs.CV

    Memes in the Wild: Assessing the Generalizability of the Hateful Memes Challenge Dataset

    Authors: Hannah Rose Kirk, Yennie Jun, Paulius Rauba, Gal Wachtel, Ruining Li, Xingjian Bai, Noah Broestl, Martin Doff-Sotta, Aleksandar Shtedritski, Yuki M. Asano

    Abstract: Hateful memes pose a unique challenge for current machine learning systems because their message is derived from both text- and visual-modalities. To this effect, Facebook released the Hateful Memes Challenge, a dataset of memes with pre-extracted text captions, but it is unclear whether these synthetic examples generalize to `memes in the wild'. In this paper, we collect hateful and non-hateful m… ▽ More

    Submitted 9 July, 2021; originally announced July 2021.

    Comments: Accepted paper at ACL WOAH 2021