Skip to main content

Showing 1–50 of 85 results for author: Smith, V

Searching in archive cs. Search in all archives.
.
  1. arXiv:2604.11061  [pdf, ps, other

    cs.LG cs.AI

    Pando: Do Interpretability Methods Work When Models Won't Explain Themselves?

    Authors: Ziqian Zhong, Aashiq Muhamed, Mona T. Diab, Virginia Smith, Aditi Raghunathan

    Abstract: Mechanistic interpretability is often motivated for alignment auditing, where a model's verbal explanations can be absent, incomplete, or misleading. Yet many evaluations do not control whether black-box prompting alone can recover the target behavior, so apparent gains from white-box tools may reflect elicitation rather than internal signal; we call this the elicitation confounder. We introduce P… ▽ More

    Submitted 13 April, 2026; originally announced April 2026.

  2. arXiv:2603.21461  [pdf, ps, other

    cs.LG cs.AI cs.CL

    DSPA: Dynamic SAE Steering for Data-Efficient Preference Alignment

    Authors: James Wedgwood, Aashiq Muhamed, Mona T. Diab, Virginia Smith

    Abstract: Preference alignment is usually achieved by weight-updating training on preference data, which adds substantial alignment-stage compute and provides limited mechanistic visibility. We propose Dynamic SAE Steering for Preference Alignment (DSPA), an inference-time method that makes sparse autoencoder (SAE) steering prompt-conditional. From preference triples, DSPA computes a conditional-difference… ▽ More

    Submitted 22 March, 2026; originally announced March 2026.

  3. arXiv:2603.12151  [pdf, ps, other

    cs.LG cs.AI

    IsoCompute Playbook: Optimally Scaling Sampling Compute for LLM RL

    Authors: Zhoujun Cheng, Yutao Xie, Yuxiao Qu, Amrith Setlur, Shibo Hao, Varad Pimpalkhute, Tongtong Liang, Feng Yao, Zhengzhong Liu, Eric Xing, Virginia Smith, Ruslan Salakhutdinov, Zhiting Hu, Taylor Killian, Aviral Kumar

    Abstract: While scaling laws guide compute allocation for LLM pre-training, analogous prescriptions for reinforcement learning (RL) post-training of large language models (LLMs) remain poorly understood. We study the compute-optimal allocation of sampling compute for on-policy RL methods in LLMs, framing scaling as a compute-constrained optimization over three resources: parallel rollouts per problem, numbe… ▽ More

    Submitted 12 March, 2026; originally announced March 2026.

    Comments: 29 pages, 27 figures. Under review

  4. arXiv:2603.03319  [pdf, ps, other

    cs.CL cs.AI

    Automated Concept Discovery for LLM-as-a-Judge Preference Analysis

    Authors: James Wedgwood, Chhavi Yadav, Virginia Smith

    Abstract: Large Language Models (LLMs) are increasingly used as scalable evaluators of model outputs, but their preference judgments exhibit systematic biases and can diverge from human evaluations. Prior work on LLM-as-a-judge has largely focused on a small, predefined set of hypothesized biases, leaving open the problem of automatically discovering unknown drivers of LLM preferences. We address this gap b… ▽ More

    Submitted 9 February, 2026; originally announced March 2026.

  5. arXiv:2601.18779  [pdf, ps, other

    cs.LG cs.AI cs.CL

    POPE: Learning to Reason on Hard Problems via Privileged On-Policy Exploration

    Authors: Yuxiao Qu, Amrith Setlur, Virginia Smith, Ruslan Salakhutdinov, Aviral Kumar

    Abstract: Reinforcement learning (RL) has improved the reasoning abilities of large language models (LLMs), yet state-of-the-art methods still fail to learn on many training problems. On hard problems, on-policy RL rarely explores even a single correct rollout, yielding zero reward and no learning signal for driving improvement. We find that natural solutions to remedy this exploration problem from classica… ▽ More

    Submitted 26 January, 2026; originally announced January 2026.

  6. arXiv:2510.20214  [pdf, ps, other

    cs.CV

    Towards Objective Obstetric Ultrasound Assessment: Contrastive Representation Learning for Fetal Movement Detection

    Authors: Talha Ilyas, Duong Nhu, Allison Thomas, Arie Levin, Lim Wei Yap, Shu Gong, David Vera Anaya, Yiwen Jiang, Deval Mehta, Ritesh Warty, Vinayak Smith, Maya Reddy, Euan Wallace, Wenlong Cheng, Zongyuan Ge, Faezeh Marzbanrad

    Abstract: Accurate fetal movement (FM) detection is essential for assessing prenatal health, as abnormal movement patterns can indicate underlying complications such as placental dysfunction or fetal distress. Traditional methods, including maternal perception and cardiotocography (CTG), suffer from subjectivity and limited accuracy. To address these challenges, we propose Contrastive Ultrasound Video Repre… ▽ More

    Submitted 23 October, 2025; originally announced October 2025.

    Comments: This is the preprint version of the manuscript submitted to IEEE Journal of Biomedical and Health Informatics (JBHI) for review

  7. arXiv:2510.12595  [pdf, ps, other

    cs.LG

    Research in Collaborative Learning Does Not Serve Cross-Silo Federated Learning in Practice

    Authors: Kevin Kuo, Chhavi Yadav, Virginia Smith

    Abstract: Cross-silo federated learning (FL) is a promising approach to enable cross-organization collaboration in machine learning model development without directly sharing private data. Despite growing organizational interest driven by data protection regulations such as GDPR and HIPAA, the adoption of cross-silo FL remains limited in practice. In this paper, we conduct an interview study to understand t… ▽ More

    Submitted 14 October, 2025; originally announced October 2025.

    Comments: Main text: 23 pages, 2 tables, 2 figures

  8. arXiv:2510.10390  [pdf, ps, other

    cs.CL cs.AI cs.LG

    RefusalBench: Generative Evaluation of Selective Refusal in Grounded Language Models

    Authors: Aashiq Muhamed, Leonardo F. R. Ribeiro, Markus Dreyer, Virginia Smith, Mona T. Diab

    Abstract: The ability of language models in RAG systems to selectively refuse to answer based on flawed context is critical for safety, yet remains a significant failure point. Our large-scale study reveals that even frontier models struggle in this setting, with refusal accuracy dropping below 50% on multi-document tasks, while exhibiting either dangerous overconfidence or overcaution. Static benchmarks fa… ▽ More

    Submitted 11 October, 2025; originally announced October 2025.

  9. arXiv:2509.21837  [pdf, ps, other

    cs.CL

    Semantic Agreement Enables Efficient Open-Ended LLM Cascades

    Authors: Duncan Soiffer, Steven Kolawole, Virginia Smith

    Abstract: Cascade systems route computational requests to smaller models when possible and defer to larger models only when necessary, offering a promising approach to balance cost and quality in LLM deployment. However, they face a fundamental challenge in open-ended text generation: determining output reliability when generation quality lies on a continuous spectrum, often with multiple valid responses. T… ▽ More

    Submitted 27 October, 2025; v1 submitted 25 September, 2025; originally announced September 2025.

    Comments: 2025 Conference on Empirical Methods in Natural Language Processing (EMNLP) Industry Track

  10. arXiv:2506.18728  [pdf, ps, other

    cs.LG

    PARALLELPROMPT: Extracting Parallelism from Large Language Model Queries

    Authors: Steven Kolawole, Keshav Santhanam, Virginia Smith, Pratiksha Thaker

    Abstract: LLM serving systems typically treat user prompts as monolithic inputs, optimizing inference through decoding tricks or inter-query batching. However, many real-world prompts contain latent semantic parallelism--decomposable structures where subtasks can be executed independently to reduce latency while preserving meaning. We introduce PARALLELPROMPT, the first benchmark for measuring intra-query p… ▽ More

    Submitted 20 October, 2025; v1 submitted 23 June, 2025; originally announced June 2025.

    Comments: 39th Conference on Neural Information Processing Systems (NeurIPS 2025) Datasets and Benchmarks Track

  11. arXiv:2506.15699  [pdf, ps, other

    cs.LG cs.AI

    BLUR: A Benchmark for LLM Unlearning Robust to Forget-Retain Overlap

    Authors: Shengyuan Hu, Neil Kale, Pratiksha Thaker, Yiwei Fu, Steven Wu, Virginia Smith

    Abstract: Machine unlearning has the potential to improve the safety of large language models (LLMs) by removing sensitive or harmful information post hoc. A key challenge in unlearning involves balancing between forget quality (effectively unlearning undesirable information) and retain quality (maintaining good performance on other, general tasks). Unfortunately, as we show, current LLM unlearning benchmar… ▽ More

    Submitted 28 May, 2025; originally announced June 2025.

  12. arXiv:2506.09026  [pdf, ps, other

    cs.LG cs.CL

    e3: Learning to Explore Enables Extrapolation of Test-Time Compute for LLMs

    Authors: Amrith Setlur, Matthew Y. R. Yang, Charlie Snell, Jeremy Greer, Ian Wu, Virginia Smith, Max Simchowitz, Aviral Kumar

    Abstract: Test-time scaling offers a promising path to improve LLM reasoning by utilizing more compute at inference time; however, the true promise of this paradigm lies in extrapolation (i.e., improvement in performance on hard problems as LLMs keep "thinking" for longer, beyond the maximum token budget they were trained on). Surprisingly, we find that most existing reasoning models do not extrapolate well… ▽ More

    Submitted 13 June, 2025; v1 submitted 10 June, 2025; originally announced June 2025.

  13. arXiv:2506.06488  [pdf, ps, other

    cs.LG cs.CR stat.ML

    Membership Inference Attacks for Unseen Classes

    Authors: Pratiksha Thaker, Neil Kale, Zhiwei Steven Wu, Virginia Smith

    Abstract: The state-of-the-art for membership inference attacks on machine learning models is a class of attacks based on shadow models that mimic the behavior of the target model on subsets of held-out nonmember data. However, we find that this class of attacks is fundamentally limited because of a key assumption -- that the shadow models can replicate the target model's behavior on the distribution of int… ▽ More

    Submitted 25 October, 2025; v1 submitted 6 June, 2025; originally announced June 2025.

    Comments: Preprint

  14. arXiv:2505.20254  [pdf, ps, other

    cs.LG cs.AI cs.CL stat.ML

    Position: Mechanistic Interpretability Should Prioritize Feature Consistency in SAEs

    Authors: Xiangchen Song, Aashiq Muhamed, Yujia Zheng, Lingjing Kong, Zeyu Tang, Mona T. Diab, Virginia Smith, Kun Zhang

    Abstract: Sparse Autoencoders (SAEs) are a prominent tool in mechanistic interpretability (MI) for decomposing neural network activations into interpretable features. However, the aspiration to identify a canonical set of features is challenged by the observed inconsistency of learned SAE features across different training runs, undermining the reliability and efficiency of MI research. This position paper… ▽ More

    Submitted 26 May, 2025; originally announced May 2025.

  15. arXiv:2504.08192  [pdf, other

    cs.LG cs.AI cs.CL cs.CR

    SAEs $\textit{Can}$ Improve Unlearning: Dynamic Sparse Autoencoder Guardrails for Precision Unlearning in LLMs

    Authors: Aashiq Muhamed, Jacopo Bonato, Mona Diab, Virginia Smith

    Abstract: Machine unlearning is a promising approach to improve LLM safety by removing unwanted knowledge from the model. However, prevailing gradient-based unlearning methods suffer from issues such as high computational costs, hyperparameter instability, poor sequential unlearning capability, vulnerability to relearning attacks, low data efficiency, and lack of interpretability. While Sparse Autoencoders… ▽ More

    Submitted 10 April, 2025; originally announced April 2025.

  16. arXiv:2504.04626  [pdf, other

    cs.LG

    Exact Unlearning of Finetuning Data via Model Merging at Scale

    Authors: Kevin Kuo, Amrith Setlur, Kartik Srinivas, Aditi Raghunathan, Virginia Smith

    Abstract: Approximate unlearning has gained popularity as an approach to efficiently update an LLM so that it behaves (roughly) as if it was not trained on a subset of data to begin with. However, existing methods are brittle in practice and can easily be attacked to reveal supposedly unlearned information. To alleviate issues with approximate unlearning, we instead propose SIFT-Masks (SIgn-Fixed Tuning-Mas… ▽ More

    Submitted 6 April, 2025; originally announced April 2025.

    Comments: 9 pages, 10 figures

  17. arXiv:2504.01883  [pdf, other

    cs.AI cs.CL cs.IR cs.LG

    CoRAG: Collaborative Retrieval-Augmented Generation

    Authors: Aashiq Muhamed, Mona Diab, Virginia Smith

    Abstract: Retrieval-Augmented Generation (RAG) models excel in knowledge-intensive tasks, especially under few-shot learning constraints. We introduce CoRAG, a framework extending RAG to collaborative settings, where clients jointly train a shared model using a collaborative passage store. To evaluate CoRAG, we introduce CRAB, a benchmark for collaborative homogeneous open-domain question answering. Our exp… ▽ More

    Submitted 2 April, 2025; originally announced April 2025.

    Comments: NAACL 2024

  18. arXiv:2411.03730  [pdf, ps, other

    cs.LG cs.CR cs.CV

    NeurIPS 2023 Competition: Privacy Preserving Federated Learning Document VQA

    Authors: Marlon Tobaben, Mohamed Ali Souibgui, Rubèn Tito, Khanh Nguyen, Raouf Kerkouche, Kangsoo Jung, Joonas Jälkö, Lei Kang, Andrey Barsky, Vincent Poulain d'Andecy, Aurélie Joseph, Aashiq Muhamed, Kevin Kuo, Virginia Smith, Yusuke Yamasaki, Takumi Fukami, Kenta Niwa, Iifan Tyou, Hiro Ishii, Rio Yokota, Ragul N, Rintu Kutum, Josep Llados, Ernest Valveny, Antti Honkela , et al. (2 additional authors not shown)

    Abstract: The Privacy Preserving Federated Learning Document VQA (PFL-DocVQA) competition challenged the community to develop provably private and communication-efficient solutions in a federated setting for a real-life use case: invoice processing. The competition introduced a dataset of real invoice documents, along with associated questions and answers requiring information extraction and reasoning over… ▽ More

    Submitted 3 June, 2025; v1 submitted 6 November, 2024; originally announced November 2024.

    Comments: 33 pages, 7 figures; published in TMLR 06/2025 https://openreview.net/forum?id=3HKNwejEEq

    Journal ref: Transactions on Machine Learning Research, ISSN 2835-8856, 2025

  19. arXiv:2411.00743  [pdf, other

    cs.LG cs.AI cs.CL

    Decoding Dark Matter: Specialized Sparse Autoencoders for Interpreting Rare Concepts in Foundation Models

    Authors: Aashiq Muhamed, Mona Diab, Virginia Smith

    Abstract: Understanding and mitigating the potential risks associated with foundation models (FMs) hinges on developing effective interpretability methods. Sparse Autoencoders (SAEs) have emerged as a promising tool for disentangling FM representations, but they struggle to capture rare, yet crucial concepts in the data. We introduce Specialized Sparse Autoencoders (SSAEs), designed to illuminate these elus… ▽ More

    Submitted 1 November, 2024; originally announced November 2024.

  20. arXiv:2410.02879  [pdf, other

    cs.CL

    Position: LLM Unlearning Benchmarks are Weak Measures of Progress

    Authors: Pratiksha Thaker, Shengyuan Hu, Neil Kale, Yash Maurya, Zhiwei Steven Wu, Virginia Smith

    Abstract: Unlearning methods have the potential to improve the privacy and safety of large language models (LLMs) by removing sensitive or harmful information post hoc. The LLM unlearning research community has increasingly turned toward empirical benchmarks to assess the effectiveness of such methods. In this paper, we find that existing benchmarks provide an overly optimistic and potentially misleading vi… ▽ More

    Submitted 8 April, 2025; v1 submitted 3 October, 2024; originally announced October 2024.

    Comments: Appears in IEEE Secure and Trustworthy Machine Learning (SaTML) '25

  21. arXiv:2407.02348  [pdf, ps, other

    cs.LG

    Agreement-Based Cascading for Efficient Inference

    Authors: Steven Kolawole, Don Dennis, Ameet Talwalkar, Virginia Smith

    Abstract: Adaptive inference schemes reduce the cost of machine learning inference by assigning smaller models to easier examples, attempting to avoid invocation of larger models when possible. In this work we explore a simple, effective adaptive inference technique we term Agreement-Based Cascading (ABC). ABC builds a cascade of models of increasing size/complexity, and uses agreement between ensembles of… ▽ More

    Submitted 24 September, 2025; v1 submitted 2 July, 2024; originally announced July 2024.

    Comments: Published at TMLR (July 2025)

    Journal ref: TMLR 2025

  22. arXiv:2406.17660  [pdf, other

    cs.LG

    Grass: Compute Efficient Low-Memory LLM Training with Structured Sparse Gradients

    Authors: Aashiq Muhamed, Oscar Li, David Woodruff, Mona Diab, Virginia Smith

    Abstract: Large language model (LLM) training and finetuning are often bottlenecked by limited GPU memory. While existing projection-based optimization methods address this by projecting gradients into a lower-dimensional subspace to reduce optimizer state memory, they typically rely on dense projection matrices, which can introduce computational and memory overheads. In this work, we propose Grass (GRAdien… ▽ More

    Submitted 25 June, 2024; originally announced June 2024.

  23. arXiv:2406.14532  [pdf, other

    cs.LG cs.CL

    RL on Incorrect Synthetic Data Scales the Efficiency of LLM Math Reasoning by Eight-Fold

    Authors: Amrith Setlur, Saurabh Garg, Xinyang Geng, Naman Garg, Virginia Smith, Aviral Kumar

    Abstract: Training on model-generated synthetic data is a promising approach for finetuning LLMs, but it remains unclear when it helps or hurts. In this paper, we investigate this question for math reasoning via an empirical study, followed by building a conceptual understanding of our observations. First, we find that while the typical approach of finetuning a model on synthetic correct or positive problem… ▽ More

    Submitted 20 June, 2024; originally announced June 2024.

  24. arXiv:2406.13356  [pdf, other

    cs.LG

    Unlearning or Obfuscating? Jogging the Memory of Unlearned LLMs via Benign Relearning

    Authors: Shengyuan Hu, Yiwei Fu, Zhiwei Steven Wu, Virginia Smith

    Abstract: Machine unlearning is a promising approach to mitigate undesirable memorization of training data in ML models. However, in this work we show that existing approaches for unlearning in LLMs are surprisingly susceptible to a simple set of $\textit{benign relearning attacks}$. With access to only a small and potentially loosely related set of data, we find that we can ''jog'' the memory of unlearned… ▽ More

    Submitted 17 March, 2025; v1 submitted 19 June, 2024; originally announced June 2024.

    Comments: ICLR 2025, 32 pages, 8 figures, 9 tables

  25. arXiv:2406.05233  [pdf, other

    cs.LG cs.DC

    Federated LoRA with Sparse Communication

    Authors: Kevin Kuo, Arian Raje, Kousik Rajesh, Virginia Smith

    Abstract: Low-rank adaptation (LoRA) is a natural method for finetuning in communication-constrained machine learning settings such as cross-device federated learning. Prior work that has studied LoRA in the context of federated learning has focused on improving LoRA's robustness to heterogeneity and privacy. In this work, we instead consider techniques for further improving communication-efficiency in fede… ▽ More

    Submitted 7 June, 2024; originally announced June 2024.

    Comments: 12 pages (excluding references), 8 figures

  26. arXiv:2403.05598  [pdf, other

    cs.CR cs.LG

    Privacy Amplification for the Gaussian Mechanism via Bounded Support

    Authors: Shengyuan Hu, Saeed Mahloujifar, Virginia Smith, Kamalika Chaudhuri, Chuan Guo

    Abstract: Data-dependent privacy accounting frameworks such as per-instance differential privacy (pDP) and Fisher information loss (FIL) confer fine-grained privacy guarantees for individuals in a fixed training dataset. These guarantees can be desirable compared to vanilla DP in real world settings as they tightly upper-bound the privacy leakage for a $\textit{specific}$ individual in an $\textit{actual}$… ▽ More

    Submitted 7 March, 2024; originally announced March 2024.

    Comments: 23 pages, 4 figures

  27. arXiv:2403.04099  [pdf, other

    cs.LG

    Many-Objective Multi-Solution Transport

    Authors: Ziyue Li, Tian Li, Virginia Smith, Jeff Bilmes, Tianyi Zhou

    Abstract: Optimizing the performance of many objectives (instantiated by tasks or clients) jointly with a few Pareto stationary solutions (models) is critical in machine learning. However, previous multi-objective optimization methods often focus on a few number of objectives and cannot scale to many objectives that outnumber the solutions, leading to either subpar performance or ignored objectives. We intr… ▽ More

    Submitted 6 March, 2024; originally announced March 2024.

  28. arXiv:2403.03329  [pdf, other

    cs.CL

    Guardrail Baselines for Unlearning in LLMs

    Authors: Pratiksha Thaker, Yash Maurya, Shengyuan Hu, Zhiwei Steven Wu, Virginia Smith

    Abstract: Recent work has demonstrated that finetuning is a promising approach to 'unlearn' concepts from large language models. However, finetuning can be expensive, as it requires both generating a set of examples and running iterations of finetuning to update the model. In this work, we show that simple guardrail-based approaches such as prompting and filtering can achieve unlearning results comparable t… ▽ More

    Submitted 11 June, 2024; v1 submitted 5 March, 2024; originally announced March 2024.

    Comments: Preliminary work, accepted to ICLR workshop SeT-LLM 2024

  29. arXiv:2402.16187  [pdf, other

    cs.CR cs.CL cs.LG

    No Free Lunch in LLM Watermarking: Trade-offs in Watermarking Design Choices

    Authors: Qi Pang, Shengyuan Hu, Wenting Zheng, Virginia Smith

    Abstract: Advances in generative models have made it possible for AI-generated text, code, and images to mirror human-generated content in many applications. Watermarking, a technique that aims to embed information in the output of a model to verify its source, is useful for mitigating the misuse of such AI-generated content. However, we show that common design choices in LLM watermarking schemes make the r… ▽ More

    Submitted 13 November, 2024; v1 submitted 25 February, 2024; originally announced February 2024.

  30. arXiv:2402.05406  [pdf, ps, other

    cs.LG cs.CL

    Everybody Prune Now: Structured Pruning of LLMs with only Forward Passes

    Authors: Steven Kolawole, Lucio Dery, Jean-François Kagy, Virginia Smith, Graham Neubig, Ameet Talwalkar

    Abstract: Structured pruning is a promising approach to create smaller, faster large language models. However, existing methods typically rely on computing the gradient via backward passes, which can inflate memory requirements and compute costs. In this work we introduce Bonsai, a gradient-free structured pruning method that eliminates the need for backpropagation, significantly reducing memory requirement… ▽ More

    Submitted 22 January, 2026; v1 submitted 7 February, 2024; originally announced February 2024.

    Comments: 19 pages, 6 fiigures, 16 tables

  31. arXiv:2312.15551  [pdf, ps, other

    cs.LG cs.CR stat.ML

    On the Benefits of Public Representations for Private Transfer Learning under Distribution Shift

    Authors: Pratiksha Thaker, Amrith Setlur, Zhiwei Steven Wu, Virginia Smith

    Abstract: Public pretraining is a promising approach to improve differentially private model training. However, recent work has noted that many positive research results studying this paradigm only consider in-distribution tasks, and may not apply to settings where there is distribution shift between the pretraining and finetuning data -- a scenario that is likely when finetuning private tasks due to the se… ▽ More

    Submitted 8 September, 2025; v1 submitted 24 December, 2023; originally announced December 2023.

    Comments: Published in NeurIPS 2024

  32. arXiv:2312.03318  [pdf, other

    cs.LG cs.CV stat.ML

    Complementary Benefits of Contrastive Learning and Self-Training Under Distribution Shift

    Authors: Saurabh Garg, Amrith Setlur, Zachary Chase Lipton, Sivaraman Balakrishnan, Virginia Smith, Aditi Raghunathan

    Abstract: Self-training and contrastive learning have emerged as leading techniques for incorporating unlabeled data, both under distribution shift (unsupervised domain adaptation) and when it is absent (semi-supervised learning). However, despite the popularity and compatibility of these techniques, their efficacy in combination remains unexplored. In this paper, we undertake a systematic empirical investi… ▽ More

    Submitted 6 December, 2023; originally announced December 2023.

    Comments: NeurIPS 2023

  33. arXiv:2310.01424  [pdf, ps, other

    cs.CL cs.AI

    Identifying and Mitigating Privacy Risks Stemming from Language Models: A Survey

    Authors: Victoria Smith, Ali Shahin Shamsabadi, Carolyn Ashurst, Adrian Weller

    Abstract: Large Language Models (LLMs) have shown greatly enhanced performance in recent years, attributed to increased size and extensive training data. This advancement has led to widespread interest and adoption across industries and the public. However, training data memorization in Machine Learning models scales with model size, particularly concerning for LLMs. Memorized text sequences have the potent… ▽ More

    Submitted 18 June, 2024; v1 submitted 27 September, 2023; originally announced October 2023.

    Comments: 15 pages

  34. arXiv:2304.12180  [pdf, other

    cs.NE cs.AI cs.LG

    Variance-Reduced Gradient Estimation via Noise-Reuse in Online Evolution Strategies

    Authors: Oscar Li, James Harrison, Jascha Sohl-Dickstein, Virginia Smith, Luke Metz

    Abstract: Unrolled computation graphs are prevalent throughout machine learning but present challenges to automatic differentiation (AD) gradient estimation methods when their loss functions exhibit extreme local sensitivtiy, discontinuity, or blackbox characteristics. In such scenarios, online evolution strategies methods are a more capable alternative, while being more parallelizable than vanilla evolutio… ▽ More

    Submitted 9 December, 2023; v1 submitted 21 April, 2023; originally announced April 2023.

    Comments: NeurIPS 2023. 41 pages. Code available at https://github.com/OscarcarLi/Noise-Reuse-Evolution-Strategies

  35. arXiv:2302.10093  [pdf, other

    cs.LG

    Progressive Ensemble Distillation: Building Ensembles for Efficient Inference

    Authors: Don Kurian Dennis, Abhishek Shetty, Anish Sevekari, Kazuhito Koishida, Virginia Smith

    Abstract: We study the problem of progressive ensemble distillation: Given a large, pretrained teacher model $g$, we seek to decompose the model into smaller, low-inference cost student models $f_i$, such that progressively evaluating additional models in this ensemble leads to improved predictions. The resulting ensemble allows for flexibly tuning accuracy vs. inference cost at runtime, which is useful for… ▽ More

    Submitted 9 November, 2023; v1 submitted 20 February, 2023; originally announced February 2023.

  36. arXiv:2302.08533  [pdf, other

    cs.LG cs.DC

    Federated Learning as a Network Effects Game

    Authors: Shengyuan Hu, Dung Daniel Ngo, Shuran Zheng, Virginia Smith, Zhiwei Steven Wu

    Abstract: Federated Learning (FL) aims to foster collaboration among a population of clients to improve the accuracy of machine learning without directly sharing local data. Although there has been rich literature on designing federated learning algorithms, most prior works implicitly assume that all clients are willing to participate in a FL scheme. In practice, clients may not benefit from joining in FL,… ▽ More

    Submitted 16 February, 2023; originally announced February 2023.

    Comments: 14 pages of main text, 26 pages in total

  37. arXiv:2302.02931  [pdf, other

    cs.LG

    Bitrate-Constrained DRO: Beyond Worst Case Robustness To Unknown Group Shifts

    Authors: Amrith Setlur, Don Dennis, Benjamin Eysenbach, Aditi Raghunathan, Chelsea Finn, Virginia Smith, Sergey Levine

    Abstract: Training machine learning models robust to distribution shifts is critical for real-world applications. Some robust training algorithms (e.g., Group DRO) specialize to group shifts and require group information on all training points. Other methods (e.g., CVaR DRO) that do not need group annotations can be overly conservative, since they naively upweight high loss points which may form a contrived… ▽ More

    Submitted 11 October, 2023; v1 submitted 6 February, 2023; originally announced February 2023.

    Journal ref: ICLR 2023

  38. arXiv:2212.08930  [pdf, other

    cs.LG

    On Noisy Evaluation in Federated Hyperparameter Tuning

    Authors: Kevin Kuo, Pratiksha Thaker, Mikhail Khodak, John Nguyen, Daniel Jiang, Ameet Talwalkar, Virginia Smith

    Abstract: Hyperparameter tuning is critical to the success of federated learning applications. Unfortunately, appropriately selecting hyperparameters is challenging in federated networks. Issues of scale, privacy, and heterogeneity introduce noise in the tuning process and make it difficult to evaluate the performance of various hyperparameters. In this work, we perform the first systematic study on the eff… ▽ More

    Submitted 15 May, 2023; v1 submitted 17 December, 2022; originally announced December 2022.

    Comments: v1: 19 pages, 15 figures, submitted to MLSys2023; v2: Fixed citation formatting; v3: Fixed typo, update acks v4: MLSys2023 camera-ready

  39. arXiv:2212.00309  [pdf, other

    cs.LG cs.CR

    Differentially Private Adaptive Optimization with Delayed Preconditioners

    Authors: Tian Li, Manzil Zaheer, Ken Ziyu Liu, Sashank J. Reddi, H. Brendan McMahan, Virginia Smith

    Abstract: Privacy noise may negate the benefits of using adaptive optimizers in differentially private model training. Prior works typically address this issue by using auxiliary information (e.g., public data) to boost the effectiveness of adaptive optimization. In this work, we explore techniques to estimate and efficiently adapt to gradient geometry in private adaptive optimization without auxiliary data… ▽ More

    Submitted 7 June, 2023; v1 submitted 1 December, 2022; originally announced December 2022.

    Comments: Accepted by ICLR 2023

  40. arXiv:2211.15458  [pdf, other

    cs.LG cs.CL

    Validating Large Language Models with ReLM

    Authors: Michael Kuchnik, Virginia Smith, George Amvrosiadis

    Abstract: Although large language models (LLMs) have been touted for their ability to generate natural-sounding text, there are growing concerns around possible negative effects of LLMs such as data memorization, bias, and inappropriate language. Unfortunately, the complexity and generation capacities of LLMs make validating (and correcting) such concerns difficult. In this work, we introduce ReLM, a system… ▽ More

    Submitted 8 May, 2023; v1 submitted 21 November, 2022; originally announced November 2022.

  41. arXiv:2208.00467  [pdf, other

    cs.CV cs.LG

    COCOA: Cross Modality Contrastive Learning for Sensor Data

    Authors: Shohreh Deldari, Hao Xue, Aaqib Saeed, Daniel V. Smith, Flora D. Salim

    Abstract: Self-Supervised Learning (SSL) is a new paradigm for learning discriminative representations without labelled data and has reached comparable or even state-of-the-art results in comparison to supervised counterparts. Contrastive Learning (CL) is one of the most well-known approaches in SSL that attempts to learn general, informative representations of data. CL methods have been mostly developed fo… ▽ More

    Submitted 3 August, 2022; v1 submitted 31 July, 2022; originally announced August 2022.

    Comments: 27 pages, 10 figures, 6 tables, Accepted with minor revision at IMWUT Vol. 6 No. 3

  42. arXiv:2206.09262  [pdf, other

    cs.LG cs.DC

    Motley: Benchmarking Heterogeneity and Personalization in Federated Learning

    Authors: Shanshan Wu, Tian Li, Zachary Charles, Yu Xiao, Ziyu Liu, Zheng Xu, Virginia Smith

    Abstract: Personalized federated learning considers learning models unique to each client in a heterogeneous network. The resulting client-specific models have been purported to improve metrics such as accuracy, fairness, and robustness in federated networks. However, despite a plethora of work in this area, it remains unclear: (1) which personalization techniques are most effective in various settings, and… ▽ More

    Submitted 26 September, 2022; v1 submitted 18 June, 2022; originally announced June 2022.

    Comments: 40 pages, 10 figures, 7 tables. EMNIST and Landmarks fine-tuning results are corrected in (and after) v5. Code: https://github.com/google-research/federated/tree/master/personalization_benchmark

  43. arXiv:2206.07902  [pdf, other

    cs.LG cs.CR stat.ML

    On Privacy and Personalization in Cross-Silo Federated Learning

    Authors: Ziyu Liu, Shengyuan Hu, Zhiwei Steven Wu, Virginia Smith

    Abstract: While the application of differential privacy (DP) has been well-studied in cross-device federated learning (FL), there is a lack of work considering DP and its implications for cross-silo FL, a setting characterized by a limited number of clients each containing many data subjects. In cross-silo FL, usual notions of client-level DP are less suitable as real-world privacy regulations typically con… ▽ More

    Submitted 17 October, 2022; v1 submitted 15 June, 2022; originally announced June 2022.

    Comments: NeurIPS 2022, 37 pages

  44. arXiv:2206.02353  [pdf, other

    cs.LG cs.CV

    Beyond Just Vision: A Review on Self-Supervised Representation Learning on Multimodal and Temporal Data

    Authors: Shohreh Deldari, Hao Xue, Aaqib Saeed, Jiayuan He, Daniel V. Smith, Flora D. Salim

    Abstract: Recently, Self-Supervised Representation Learning (SSRL) has attracted much attention in the field of computer vision, speech, natural language processing (NLP), and recently, with other types of modalities, including time series from sensors. The popularity of self-supervised learning is driven by the fact that traditional models typically require a huge amount of well-annotated data for training… ▽ More

    Submitted 7 June, 2022; v1 submitted 6 June, 2022; originally announced June 2022.

    Comments: 36 pages, 5 figures, 9 tables, Survey paper

  45. arXiv:2206.01367  [pdf, other

    cs.LG cs.CR

    Adversarial Unlearning: Reducing Confidence Along Adversarial Directions

    Authors: Amrith Setlur, Benjamin Eysenbach, Virginia Smith, Sergey Levine

    Abstract: Supervised learning methods trained with maximum likelihood objectives often overfit on training data. Most regularizers that prevent overfitting look to increase confidence on additional examples (e.g., data augmentation, adversarial training), or reduce it on training data (e.g., label smoothing). In this work we propose a complementary regularization strategy that reduces confidence on self-gen… ▽ More

    Submitted 2 June, 2022; originally announced June 2022.

  46. arXiv:2205.14840  [pdf, other

    cs.LG

    Maximizing Global Model Appeal in Federated Learning

    Authors: Yae Jee Cho, Divyansh Jhunjhunwala, Tian Li, Virginia Smith, Gauri Joshi

    Abstract: Federated learning typically considers collaboratively training a global model using local data at edge clients. Clients may have their own individual requirements, such as having a minimal training loss threshold, which they expect to be met by the global model. However, due to client heterogeneity, the global model may not meet each client's requirements, and only a small subset may find the glo… ▽ More

    Submitted 4 February, 2023; v1 submitted 30 May, 2022; originally announced May 2022.

  47. arXiv:2203.10190  [pdf, other

    cs.LG cs.CY

    Fair Federated Learning via Bounded Group Loss

    Authors: Shengyuan Hu, Zhiwei Steven Wu, Virginia Smith

    Abstract: Fair prediction across protected groups is an important constraint for many federated learning applications. However, prior work studying group fair federated learning lacks formal convergence or fairness guarantees. In this work we propose a general framework for provably fair federated learning. In particular, we explore and extend the notion of Bounded Group Loss as a theoretically-grounded app… ▽ More

    Submitted 12 October, 2022; v1 submitted 18 March, 2022; originally announced March 2022.

    Comments: 19 pages

  48. arXiv:2202.05963  [pdf, other

    cs.LG cs.CR stat.ML

    Private Adaptive Optimization with Side Information

    Authors: Tian Li, Manzil Zaheer, Sashank J. Reddi, Virginia Smith

    Abstract: Adaptive optimization methods have become the default solvers for many machine learning tasks. Unfortunately, the benefits of adaptivity may degrade when training with differential privacy, as the noise added to ensure privacy reduces the effectiveness of the adaptive preconditioner. To this end, we propose AdaDPS, a general framework that uses non-sensitive side information to precondition the gr… ▽ More

    Submitted 24 June, 2022; v1 submitted 11 February, 2022; originally announced February 2022.

    Comments: ICML 2022

  49. arXiv:2111.04131  [pdf, other

    cs.LG cs.PF

    Plumber: Diagnosing and Removing Performance Bottlenecks in Machine Learning Data Pipelines

    Authors: Michael Kuchnik, Ana Klimovic, Jiri Simsa, Virginia Smith, George Amvrosiadis

    Abstract: Input pipelines, which ingest and transform input data, are an essential part of training Machine Learning (ML) models. However, it is challenging to implement efficient input pipelines, as it requires reasoning about parallelism, asynchrony, and variability in fine-grained profiling information. Our analysis of over two million ML jobs in Google datacenters reveals that a significant fraction of… ▽ More

    Submitted 21 March, 2022; v1 submitted 7 November, 2021; originally announced November 2021.

  50. arXiv:2109.06141  [pdf, other

    cs.LG cs.IT math.OC stat.ML

    On Tilted Losses in Machine Learning: Theory and Applications

    Authors: Tian Li, Ahmad Beirami, Maziar Sanjabi, Virginia Smith

    Abstract: Exponential tilting is a technique commonly used in fields such as statistics, probability, information theory, and optimization to create parametric distribution shifts. Despite its prevalence in related fields, tilting has not seen widespread use in machine learning. In this work, we aim to bridge this gap by exploring the use of tilting in risk minimization. We study a simple extension to ERM -… ▽ More

    Submitted 1 June, 2023; v1 submitted 13 September, 2021; originally announced September 2021.

    Comments: arXiv admin note: substantial text overlap with arXiv:2007.01162