Skip to main content

Showing 1–50 of 61 results for author: Bhatia, P

Searching in archive cs. Search in all archives.
.
  1. arXiv:2511.20718  [pdf, ps, other

    cs.LG cs.AI cs.CL

    ST-PPO: Stabilized Off-Policy Proximal Policy Optimization for Multi-Turn Agents Training

    Authors: Chenliang Li, Adel Elmahdy, Alex Boyd, Zhongruo Wang, Alfredo Garcia, Parminder Bhatia, Taha Kass-Hout, Cao Xiao, Mingyi Hong

    Abstract: PPO has been widely adopted for training large language models (LLMs) at the token level in multi-turn dialogue and reasoning tasks. However, its performance is often unstable and prone to collapse. Through empirical analysis, we identify two main sources of instability in this setting: (1)~token-level importance sampling, which is misaligned with the natural granularity of multi-turn environments… ▽ More

    Submitted 25 November, 2025; originally announced November 2025.

  2. arXiv:2511.01096  [pdf, ps, other

    stat.ML cs.LG

    Hyper Hawkes Processes: Interpretable Models of Marked Temporal Point Processes

    Authors: Alex Boyd, Andrew Warrington, Taha Kass-Hout, Parminder Bhatia, Danica Xiao

    Abstract: Foundational marked temporal point process (MTPP) models, such as the Hawkes process, often use inexpressive model families in order to offer interpretable parameterizations of event data. On the other hand, neural MTPPs models forego this interpretability in favor of absolute predictive performance. In this work, we present a new family MTPP models: the hyper Hawkes process (HHP), which aims to b… ▽ More

    Submitted 2 November, 2025; originally announced November 2025.

  3. arXiv:2510.11883  [pdf

    cs.CV cs.AI

    MammoDINO: Anatomically Aware Self-Supervision for Mammographic Images

    Authors: Sicheng Zhou, Lei Wu, Cao Xiao, Parminder Bhatia, Taha Kass-Hout

    Abstract: Self-supervised learning (SSL) has transformed vision encoder training in general domains but remains underutilized in medical imaging due to limited data and domain specific biases. We present MammoDINO, a novel SSL framework for mammography, pretrained on 1.4 million mammographic images. To capture clinically meaningful features, we introduce a breast tissue aware data augmentation sampler for b… ▽ More

    Submitted 13 October, 2025; originally announced October 2025.

    Comments: 5 pages

    MSC Class: 1.2

  4. arXiv:2509.21249  [pdf, ps, other

    cs.CV cs.AI cs.LG

    Decipher-MR: A Vision-Language Foundation Model for 3D MRI Representations

    Authors: Zhijian Yang, Noel DSouza, Istvan Megyeri, Xiaojian Xu, Amin Honarmandi Shandiz, Farzin Haddadpour, Krisztian Koos, Laszlo Rusko, Emanuele Valeriano, Bharadwaj Swaninathan, Lei Wu, Parminder Bhatia, Taha Kass-Hout, Erhan Bas

    Abstract: Magnetic Resonance Imaging (MRI) is a critical medical imaging modality in clinical diagnosis and research, yet its complexity and heterogeneity pose challenges for automated analysis, particularly in scalable and generalizable machine learning applications. While foundation models have revolutionized natural language and vision tasks, their application to MRI remains limited due to data scarcity… ▽ More

    Submitted 25 September, 2025; originally announced September 2025.

  5. arXiv:2509.17281  [pdf

    cs.LG cs.AI cs.CY

    Training the next generation of physicians for artificial intelligence-assisted clinical neuroradiology: ASNR MICCAI Brain Tumor Segmentation (BraTS) 2025 Lighthouse Challenge education platform

    Authors: Raisa Amiruddin, Nikolay Y. Yordanov, Nazanin Maleki, Pascal Fehringer, Athanasios Gkampenis, Anastasia Janas, Kiril Krantchev, Ahmed Moawad, Fabian Umeh, Salma Abosabie, Sara Abosabie, Albara Alotaibi, Mohamed Ghonim, Mohanad Ghonim, Sedra Abou Ali Mhana, Nathan Page, Marko Jakovljevic, Yasaman Sharifi, Prisha Bhatia, Amirreza Manteghinejad, Melisa Guelen, Michael Veronesi, Virginia Hill, Tiffany So, Mark Krycia , et al. (23 additional authors not shown)

    Abstract: High-quality reference standard image data creation by neuroradiology experts for automated clinical tools can be a powerful tool for neuroradiology & artificial intelligence education. We developed a multimodal educational approach for students and trainees during the MICCAI Brain Tumor Segmentation Lighthouse Challenge 2025, a landmark initiative to develop accurate brain tumor segmentation algo… ▽ More

    Submitted 21 September, 2025; originally announced September 2025.

    Comments: 23 pages, 9 figures, 1 table, 3 supplementary tables

  6. arXiv:2507.06261  [pdf, ps, other

    cs.CL cs.AI

    Gemini 2.5: Pushing the Frontier with Advanced Reasoning, Multimodality, Long Context, and Next Generation Agentic Capabilities

    Authors: Gheorghe Comanici, Eric Bieber, Mike Schaekermann, Ice Pasupat, Noveen Sachdeva, Inderjit Dhillon, Marcel Blistein, Ori Ram, Dan Zhang, Evan Rosen, Luke Marris, Sam Petulla, Colin Gaffney, Asaf Aharoni, Nathan Lintz, Tiago Cardal Pais, Henrik Jacobsson, Idan Szpektor, Nan-Jiang Jiang, Krishna Haridasan, Ahmed Omran, Nikunj Saunshi, Dara Bahri, Gaurav Mishra, Eric Chu , et al. (3410 additional authors not shown)

    Abstract: In this report, we introduce the Gemini 2.X model family: Gemini 2.5 Pro and Gemini 2.5 Flash, as well as our earlier Gemini 2.0 Flash and Flash-Lite models. Gemini 2.5 Pro is our most capable model yet, achieving SoTA performance on frontier coding and reasoning benchmarks. In addition to its incredible coding and reasoning skills, Gemini 2.5 Pro is a thinking model that excels at multimodal unde… ▽ More

    Submitted 19 December, 2025; v1 submitted 7 July, 2025; originally announced July 2025.

    Comments: 72 pages, 17 figures

  7. arXiv:2505.18503  [pdf, ps, other

    cs.CV

    Focus on What Matters: Enhancing Medical Vision-Language Models with Automatic Attention Alignment Tuning

    Authors: Aofei Chang, Le Huang, Alex James Boyd, Parminder Bhatia, Taha Kass-Hout, Cao Xiao, Fenglong Ma

    Abstract: Medical Large Vision-Language Models (Med-LVLMs) often exhibit suboptimal attention distribution on visual inputs, leading to hallucinated or inaccurate outputs. Existing mitigation methods primarily rely on inference-time interventions, which are limited in attention adaptation or require additional supervision. To address this, we propose A$^3$Tune, a novel fine-tuning framework for Automatic At… ▽ More

    Submitted 24 May, 2025; originally announced May 2025.

    Comments: Accepted to ACL2025 (main)

  8. arXiv:2505.17100  [pdf, ps, other

    cs.CL

    Any Large Language Model Can Be a Reliable Judge: Debiasing with a Reasoning-based Bias Detector

    Authors: Haoyan Yang, Runxue Bao, Cao Xiao, Jun Ma, Parminder Bhatia, Shangqian Gao, Taha Kass-Hout

    Abstract: LLM-as-a-Judge has emerged as a promising tool for automatically evaluating generated outputs, but its reliability is often undermined by potential biases in judgment. Existing efforts to mitigate these biases face key limitations: in-context learning-based methods fail to address rooted biases due to the evaluator's limited capacity for self-reflection, whereas fine-tuning is not applicable to al… ▽ More

    Submitted 27 October, 2025; v1 submitted 21 May, 2025; originally announced May 2025.

    Comments: Accepted at NeurIPS 2025 (Camera-Ready Version)

  9. arXiv:2505.00059  [pdf, ps, other

    cs.CL cs.SD eess.AS

    BERSting at the Screams: A Benchmark for Distanced, Emotional and Shouted Speech Recognition

    Authors: Paige Tuttösí, Mantaj Dhillon, Luna Sang, Shane Eastwood, Poorvi Bhatia, Quang Minh Dinh, Avni Kapoor, Yewon Jin, Angelica Lim

    Abstract: Some speech recognition tasks, such as automatic speech recognition (ASR), are approaching or have reached human performance in many reported metrics. Yet, they continue to struggle in complex, real-world, situations, such as with distanced speech. Previous challenges have released datasets to address the issue of distanced ASR, however, the focus remains primarily on distance, specifically relyin… ▽ More

    Submitted 30 July, 2025; v1 submitted 30 April, 2025; originally announced May 2025.

    Comments: Accepted to Computer Speech and Language, Special issue: Multi-Speaker, Multi-Microphone, and Multi-Modal Distant Speech Recognition. Project Webpage and Data access : https://huggingface.co/datasets/Rosie-Lab/BERSt

  10. arXiv:2503.04639  [pdf, other

    cs.CV cs.LG

    Enhancing SAM with Efficient Prompting and Preference Optimization for Semi-supervised Medical Image Segmentation

    Authors: Aishik Konwer, Zhijian Yang, Erhan Bas, Cao Xiao, Prateek Prasanna, Parminder Bhatia, Taha Kass-Hout

    Abstract: Foundational models such as the Segment Anything Model (SAM) are gaining traction in medical imaging segmentation, supporting multiple downstream tasks. However, such models are supervised in nature, still relying on large annotated datasets or prompts supplied by experts. Conventional techniques such as active learning to alleviate such limitations are limited in scope and still necessitate conti… ▽ More

    Submitted 6 March, 2025; originally announced March 2025.

    Comments: Accepted to CVPR 2025

  11. arXiv:2503.02157  [pdf, other

    cs.CV cs.AI

    MedHEval: Benchmarking Hallucinations and Mitigation Strategies in Medical Large Vision-Language Models

    Authors: Aofei Chang, Le Huang, Parminder Bhatia, Taha Kass-Hout, Fenglong Ma, Cao Xiao

    Abstract: Large Vision Language Models (LVLMs) are becoming increasingly important in the medical domain, yet Medical LVLMs (Med-LVLMs) frequently generate hallucinations due to limited expertise and the complexity of medical applications. Existing benchmarks fail to effectively evaluate hallucinations based on their underlying causes and lack assessments of mitigation strategies. To address this gap, we in… ▽ More

    Submitted 3 March, 2025; originally announced March 2025.

    Comments: Preprint, under review

  12. arXiv:2412.19634  [pdf, ps, other

    stat.ML cs.LG

    Deep Continuous-Time State-Space Models for Marked Event Sequences

    Authors: Yuxin Chang, Alex Boyd, Cao Xiao, Taha Kass-Hout, Parminder Bhatia, Padhraic Smyth, Andrew Warrington

    Abstract: Marked temporal point processes (MTPPs) model sequences of events occurring at irregular time intervals, with wide-ranging applications in fields such as healthcare, finance and social networks. We propose the state-space point process (S2P2) model, a novel and performant model that leverages techniques derived for modern deep state-space models (SSMs) to overcome limitations of existing MTPP mode… ▽ More

    Submitted 22 October, 2025; v1 submitted 27 December, 2024; originally announced December 2024.

    Comments: NeurIPS 2025 Spotlight

  13. arXiv:2410.23605  [pdf, other

    cs.CL

    Dynamic Uncertainty Ranking: Enhancing Retrieval-Augmented In-Context Learning for Long-Tail Knowledge in LLMs

    Authors: Shuyang Yu, Runxue Bao, Parminder Bhatia, Taha Kass-Hout, Jiayu Zhou, Cao Xiao

    Abstract: Large language models (LLMs) can learn vast amounts of knowledge from diverse domains during pre-training. However, long-tail knowledge from specialized domains is often scarce and underrepresented, rarely appearing in the models' memorization. Prior work has shown that in-context learning (ICL) with retriever augmentation can help LLMs better capture long-tail knowledge, reducing their reliance o… ▽ More

    Submitted 7 February, 2025; v1 submitted 30 October, 2024; originally announced October 2024.

    Comments: Accepted by NAACL 2025

  14. arXiv:2410.12831  [pdf, ps, other

    eess.IV cs.AI cs.CV

    Segment as You Wish -- Free-Form Language-Based Segmentation for Medical Images

    Authors: Longchao Da, Rui Wang, Xiaojian Xu, Parminder Bhatia, Taha Kass-Hout, Hua Wei, Cao Xiao

    Abstract: Medical imaging is crucial for diagnosing a patient's health condition, and accurate segmentation of these images is essential for isolating regions of interest to ensure precise diagnosis and treatment planning. Existing methods primarily rely on bounding boxes or point-based prompts, while few have explored text-related prompts, despite clinicians often describing their observations and instruct… ▽ More

    Submitted 29 June, 2025; v1 submitted 2 October, 2024; originally announced October 2024.

    Comments: 19 pages, 9 as main content. The paper was accepted to KDD2025

    MSC Class: 68T45; 68U10; 92C55 ACM Class: I.2.7; I.4.9; H.3.3; I.2.6

  15. arXiv:2410.09079  [pdf, other

    cs.CL cs.AI cs.LG

    BIPEFT: Budget-Guided Iterative Search for Parameter Efficient Fine-Tuning of Large Pretrained Language Models

    Authors: Aofei Chang, Jiaqi Wang, Han Liu, Parminder Bhatia, Cao Xiao, Ting Wang, Fenglong Ma

    Abstract: Parameter Efficient Fine-Tuning (PEFT) offers an efficient solution for fine-tuning large pretrained language models for downstream tasks. However, most PEFT strategies are manually designed, often resulting in suboptimal performance. Recent automatic PEFT approaches aim to address this but face challenges such as search space entanglement, inefficiency, and lack of integration between parameter b… ▽ More

    Submitted 4 October, 2024; originally announced October 2024.

    Comments: Accepted to EMNLP 2024 (Findings)

  16. arXiv:2410.04585  [pdf, other

    cs.CL

    Reasoning-Enhanced Healthcare Predictions with Knowledge Graph Community Retrieval

    Authors: Pengcheng Jiang, Cao Xiao, Minhao Jiang, Parminder Bhatia, Taha Kass-Hout, Jimeng Sun, Jiawei Han

    Abstract: Large language models (LLMs) have demonstrated significant potential in clinical decision support. Yet LLMs still suffer from hallucinations and lack fine-grained contextual medical knowledge, limiting their high-stake healthcare applications such as clinical diagnosis. Traditional retrieval-augmented generation (RAG) methods attempt to address these limitations but frequently retrieve sparse or i… ▽ More

    Submitted 20 April, 2025; v1 submitted 6 October, 2024; originally announced October 2024.

    Comments: ICLR 2025 Camera-Ready

  17. arXiv:2405.16412  [pdf, other

    cs.CL cs.LG

    KG-FIT: Knowledge Graph Fine-Tuning Upon Open-World Knowledge

    Authors: Pengcheng Jiang, Lang Cao, Cao Xiao, Parminder Bhatia, Jimeng Sun, Jiawei Han

    Abstract: Knowledge Graph Embedding (KGE) techniques are crucial in learning compact representations of entities and relations within a knowledge graph, facilitating efficient reasoning and knowledge discovery. While existing methods typically focus either on training KGE models solely based on graph structure or fine-tuning pre-trained language models with classification data in KG, KG-FIT leverages LLM-gu… ▽ More

    Submitted 27 October, 2024; v1 submitted 25 May, 2024; originally announced May 2024.

    Comments: NeurIPS 2024

  18. arXiv:2405.15973  [pdf, other

    cs.CV cs.AI cs.CL cs.LG

    Enhancing Visual-Language Modality Alignment in Large Vision Language Models via Self-Improvement

    Authors: Xiyao Wang, Jiuhai Chen, Zhaoyang Wang, Yuhang Zhou, Yiyang Zhou, Huaxiu Yao, Tianyi Zhou, Tom Goldstein, Parminder Bhatia, Furong Huang, Cao Xiao

    Abstract: Large vision-language models (LVLMs) have achieved impressive results in visual question-answering and reasoning tasks through vision instruction tuning on specific datasets. However, there remains significant room for improvement in aligning visual and language modalities. Existing methods often depend on external models or data, leading to uncontrollable and unstable alignment results. In this p… ▽ More

    Submitted 8 February, 2025; v1 submitted 24 May, 2024; originally announced May 2024.

    Comments: NAACL 2025 Findings

  19. arXiv:2403.10351  [pdf, other

    cs.CL

    TriSum: Learning Summarization Ability from Large Language Models with Structured Rationale

    Authors: Pengcheng Jiang, Cao Xiao, Zifeng Wang, Parminder Bhatia, Jimeng Sun, Jiawei Han

    Abstract: The advent of large language models (LLMs) has significantly advanced natural language processing tasks like text summarization. However, their large size and computational demands, coupled with privacy concerns in data transmission, limit their use in resource-constrained and privacy-centric settings. To overcome this, we introduce TriSum, a framework for distilling LLMs' text summarization abili… ▽ More

    Submitted 15 March, 2024; originally announced March 2024.

    Comments: NAACL'24

  20. arXiv:2403.08845  [pdf, other

    cs.LG cs.AI

    Bifurcated Attention: Accelerating Massively Parallel Decoding with Shared Prefixes in LLMs

    Authors: Ben Athiwaratkun, Sujan Kumar Gonugondla, Sanjay Krishna Gouda, Haifeng Qian, Hantian Ding, Qing Sun, Jun Wang, Jiacheng Guo, Liangfu Chen, Parminder Bhatia, Ramesh Nallapati, Sudipta Sengupta, Bing Xiang

    Abstract: This study introduces bifurcated attention, a method designed to enhance language model inference in shared-context batch decoding scenarios. Our approach addresses the challenge of redundant memory IO costs, a critical factor contributing to latency in high batch sizes and extended context lengths. Bifurcated attention achieves this by strategically dividing the attention mechanism during increme… ▽ More

    Submitted 11 July, 2024; v1 submitted 13 March, 2024; originally announced March 2024.

  21. arXiv:2310.18642  [pdf

    cs.CV cs.AI

    One-shot Localization and Segmentation of Medical Images with Foundation Models

    Authors: Deepa Anand, Gurunath Reddy M, Vanika Singhal, Dattesh D. Shanbhag, Shriram KS, Uday Patil, Chitresh Bhushan, Kavitha Manickam, Dawei Gui, Rakesh Mullick, Avinash Gopal, Parminder Bhatia, Taha Kass-Hout

    Abstract: Recent advances in Vision Transformers (ViT) and Stable Diffusion (SD) models with their ability to capture rich semantic features of the image have been used for image correspondence tasks on natural images. In this paper, we examine the ability of a variety of pre-trained ViT (DINO, DINOv2, SAM, CLIP) and SD models, trained exclusively on natural images, for solving the correspondence problems o… ▽ More

    Submitted 28 October, 2023; originally announced October 2023.

    Comments: Accepted at NeurIPS 2023 R0-FoMo Workshop

  22. arXiv:2310.16872  [pdf, other

    eess.IV cs.CV

    SonoSAMTrack -- Segment and Track Anything on Ultrasound Images

    Authors: Hariharan Ravishankar, Rohan Patil, Vikram Melapudi, Harsh Suthar, Stephan Anzengruber, Parminder Bhatia, Kass-Hout Taha, Pavan Annangi

    Abstract: In this paper, we present SonoSAMTrack - that combines a promptable foundational model for segmenting objects of interest on ultrasound images called SonoSAM, with a state-of-the art contour tracking model to propagate segmentations on 2D+t and 3D ultrasound datasets. Fine-tuned and tested exclusively on a rich, diverse set of objects from $\approx200$k ultrasound image-mask pairs, SonoSAM demonst… ▽ More

    Submitted 16 November, 2023; v1 submitted 25 October, 2023; originally announced October 2023.

  23. arXiv:2310.11248  [pdf, other

    cs.LG cs.CL cs.SE

    CrossCodeEval: A Diverse and Multilingual Benchmark for Cross-File Code Completion

    Authors: Yangruibo Ding, Zijian Wang, Wasi Uddin Ahmad, Hantian Ding, Ming Tan, Nihal Jain, Murali Krishna Ramanathan, Ramesh Nallapati, Parminder Bhatia, Dan Roth, Bing Xiang

    Abstract: Code completion models have made significant progress in recent years, yet current popular evaluation datasets, such as HumanEval and MBPP, predominantly focus on code completion tasks within a single file. This over-simplified setting falls short of representing the real-world software development scenario where repositories span multiple files with numerous cross-file dependencies, and accessing… ▽ More

    Submitted 16 November, 2023; v1 submitted 17 October, 2023; originally announced October 2023.

    Comments: To appear at NeurIPS 2023 (Datasets and Benchmarks Track)

  24. arXiv:2307.02435  [pdf, other

    cs.LG cs.CL cs.SE

    Exploring Continual Learning for Code Generation Models

    Authors: Prateek Yadav, Qing Sun, Hantian Ding, Xiaopeng Li, Dejiao Zhang, Ming Tan, Xiaofei Ma, Parminder Bhatia, Ramesh Nallapati, Murali Krishna Ramanathan, Mohit Bansal, Bing Xiang

    Abstract: Large-scale code generation models such as Codex and CodeT5 have achieved impressive performance. However, libraries are upgraded or deprecated very frequently and re-training large-scale language models is computationally expensive. Therefore, Continual Learning (CL) is an important aspect that remains underexplored in the code domain. In this paper, we introduce a benchmark called CodeTask-CL th… ▽ More

    Submitted 5 July, 2023; originally announced July 2023.

    Comments: ACL 2023

  25. arXiv:2306.03203  [pdf, other

    cs.CL cs.SE

    A Static Evaluation of Code Completion by Large Language Models

    Authors: Hantian Ding, Varun Kumar, Yuchen Tian, Zijian Wang, Rob Kwiatkowski, Xiaopeng Li, Murali Krishna Ramanathan, Baishakhi Ray, Parminder Bhatia, Sudipta Sengupta, Dan Roth, Bing Xiang

    Abstract: Large language models trained on code have shown great potential to increase productivity of software developers. Several execution-based benchmarks have been proposed to evaluate functional correctness of model-generated code on simple programming problems. Nevertheless, it is expensive to perform the same evaluation on complex real-world projects considering the execution cost. On the contrary,… ▽ More

    Submitted 5 June, 2023; originally announced June 2023.

    Comments: Accepted by ACL 2023 industry track

  26. arXiv:2306.01631  [pdf, other

    cs.LG cs.AI q-bio.QM

    Bi-level Contrastive Learning for Knowledge-Enhanced Molecule Representations

    Authors: Pengcheng Jiang, Cao Xiao, Tianfan Fu, Parminder Bhatia, Taha Kass-Hout, Jimeng Sun, Jiawei Han

    Abstract: Molecular representation learning is vital for various downstream applications, including the analysis and prediction of molecular properties and side effects. While Graph Neural Networks (GNNs) have been a popular framework for modeling molecular data, they often struggle to capture the full complexity of molecular representations. In this paper, we introduce a novel method called GODE, which acc… ▽ More

    Submitted 16 February, 2025; v1 submitted 2 June, 2023; originally announced June 2023.

    Comments: AAAI 2025

  27. arXiv:2303.05378  [pdf, other

    cs.LG cs.SE

    Greener yet Powerful: Taming Large Code Generation Models with Quantization

    Authors: Xiaokai Wei, Sujan Gonugondla, Wasi Ahmad, Shiqi Wang, Baishakhi Ray, Haifeng Qian, Xiaopeng Li, Varun Kumar, Zijian Wang, Yuchen Tian, Qing Sun, Ben Athiwaratkun, Mingyue Shang, Murali Krishna Ramanathan, Parminder Bhatia, Bing Xiang

    Abstract: ML-powered code generation aims to assist developers to write code in a more productive manner, by intelligently generating code blocks based on natural language prompts. Recently, large pretrained deep learning models have substantially pushed the boundary of code generation and achieved impressive performance. Despite their great power, the huge number of model parameters poses a significant thr… ▽ More

    Submitted 9 March, 2023; originally announced March 2023.

    Comments: 10 pages, 7 figures, 10 tables

  28. arXiv:2302.14383  [pdf, other

    cs.LG cs.AI cs.CL cs.CV

    Linear Spaces of Meanings: Compositional Structures in Vision-Language Models

    Authors: Matthew Trager, Pramuditha Perera, Luca Zancato, Alessandro Achille, Parminder Bhatia, Stefano Soatto

    Abstract: We investigate compositional structures in data embeddings from pre-trained vision-language models (VLMs). Traditionally, compositionality has been associated with algebraic operations on embeddings of words from a pre-existing vocabulary. In contrast, we seek to approximate representations from an encoder as combinations of a smaller set of vectors in the embedding space. These vectors can be see… ▽ More

    Submitted 11 January, 2024; v1 submitted 28 February, 2023; originally announced February 2023.

    Comments: 18 pages, 9 figures, 7 tables

    Journal ref: Proceedings of the IEEE/CVF International Conference on Computer Vision 2023 (pp. 15395-15404)

  29. arXiv:2212.10264  [pdf, other

    cs.LG cs.CL cs.SE

    ReCode: Robustness Evaluation of Code Generation Models

    Authors: Shiqi Wang, Zheng Li, Haifeng Qian, Chenghao Yang, Zijian Wang, Mingyue Shang, Varun Kumar, Samson Tan, Baishakhi Ray, Parminder Bhatia, Ramesh Nallapati, Murali Krishna Ramanathan, Dan Roth, Bing Xiang

    Abstract: Code generation models have achieved impressive performance. However, they tend to be brittle as slight edits to a prompt could lead to very different generations; these robustness properties, critical for user experience when deployed in real-life applications, are not well understood. Most existing works on robustness in text or code tasks have focused on classification, while robustness in gene… ▽ More

    Submitted 20 December, 2022; originally announced December 2022.

    Comments: Code and data available at https://github.com/amazon-science/recode

  30. arXiv:2212.10007  [pdf, other

    cs.CL cs.SE

    CoCoMIC: Code Completion By Jointly Modeling In-file and Cross-file Context

    Authors: Yangruibo Ding, Zijian Wang, Wasi Uddin Ahmad, Murali Krishna Ramanathan, Ramesh Nallapati, Parminder Bhatia, Dan Roth, Bing Xiang

    Abstract: While pre-trained language models (LM) for code have achieved great success in code completion, they generate code conditioned only on the contents within the file, i.e., in-file context, but ignore the rich semantics in other files within the same project, i.e., cross-file context, a critical source of information that is especially useful in modern modular software development. Such overlooking… ▽ More

    Submitted 24 May, 2023; v1 submitted 20 December, 2022; originally announced December 2022.

  31. arXiv:2210.14868  [pdf, other

    cs.LG cs.CL

    Multi-lingual Evaluation of Code Generation Models

    Authors: Ben Athiwaratkun, Sanjay Krishna Gouda, Zijian Wang, Xiaopeng Li, Yuchen Tian, Ming Tan, Wasi Uddin Ahmad, Shiqi Wang, Qing Sun, Mingyue Shang, Sujan Kumar Gonugondla, Hantian Ding, Varun Kumar, Nathan Fulton, Arash Farahani, Siddhartha Jain, Robert Giaquinto, Haifeng Qian, Murali Krishna Ramanathan, Ramesh Nallapati, Baishakhi Ray, Parminder Bhatia, Sudipta Sengupta, Dan Roth, Bing Xiang

    Abstract: We present new benchmarks on evaluation code generation models: MBXP and Multilingual HumanEval, and MathQA-X. These datasets cover over 10 programming languages and are generated using a scalable conversion framework that transpiles prompts and test cases from the original Python datasets into the corresponding data in the target language. Using these benchmarks, we are able to assess the perform… ▽ More

    Submitted 28 March, 2023; v1 submitted 26 October, 2022; originally announced October 2022.

    Comments: Code and data release: https://github.com/amazon-research/mxeval

  32. arXiv:2210.01185  [pdf, other

    cs.CL

    ContraCLM: Contrastive Learning For Causal Language Model

    Authors: Nihal Jain, Dejiao Zhang, Wasi Uddin Ahmad, Zijian Wang, Feng Nan, Xiaopeng Li, Ming Tan, Ramesh Nallapati, Baishakhi Ray, Parminder Bhatia, Xiaofei Ma, Bing Xiang

    Abstract: Despite exciting progress in causal language models, the expressiveness of the representations is largely limited due to poor discrimination ability. To remedy this issue, we present ContraCLM, a novel contrastive learning framework at both token-level and sequence-level. We assess ContraCLM on a variety of downstream tasks. We show that ContraCLM enhances discrimination of the representations and… ▽ More

    Submitted 2 May, 2023; v1 submitted 3 October, 2022; originally announced October 2022.

    Comments: 10 pages

    Journal ref: ACL 2023

  33. arXiv:2205.09240  [pdf, other

    cs.IR cs.AI cs.CY

    Debiasing Neural Retrieval via In-batch Balancing Regularization

    Authors: Yuantong Li, Xiaokai Wei, Zijian Wang, Shen Wang, Parminder Bhatia, Xiaofei Ma, Andrew Arnold

    Abstract: People frequently interact with information retrieval (IR) systems, however, IR models exhibit biases and discrimination towards various demographics. The in-processing fair ranking methods provide a trade-offs between accuracy and fairness through adding a fairness-related regularization term in the loss function. However, there haven't been intuitive objective functions that depend on the click… ▽ More

    Submitted 18 May, 2022; originally announced May 2022.

    Comments: 9 pages, 1 figure, and 3 tables. A version appears in the Proceedings of the 4th Workshop on Gender Bias in Natural Language Processing (GeBNLP), 2022

  34. arXiv:2203.11239  [pdf, other

    cs.CL

    DQ-BART: Efficient Sequence-to-Sequence Model via Joint Distillation and Quantization

    Authors: Zheng Li, Zijian Wang, Ming Tan, Ramesh Nallapati, Parminder Bhatia, Andrew Arnold, Bing Xiang, Dan Roth

    Abstract: Large-scale pre-trained sequence-to-sequence models like BART and T5 achieve state-of-the-art performance on many generative NLP tasks. However, such models pose a great challenge in resource-constrained scenarios owing to their large memory requirements and high latency. To alleviate this issue, we propose to jointly distill and quantize the model, where knowledge is transferred from the full-pre… ▽ More

    Submitted 21 March, 2022; originally announced March 2022.

    Comments: ACL 2022

  35. arXiv:2202.01340  [pdf, other

    cs.LG

    An Artificial Intelligence Dataset for Solar Energy Locations in India

    Authors: Anthony Ortiz, Dhaval Negandhi, Sagar R Mysorekar, Joseph Kiesecker, Shivaprakash K Nagaraju, Caleb Robinson, Priyal Bhatia, Aditi Khurana, Jane Wang, Felipe Oviedo, Juan Lavista Ferres

    Abstract: Rapid development of renewable energy sources, particularly solar photovoltaics (PV), is critical to mitigate climate change. As a result, India has set ambitious goals to install 500 gigawatts of solar energy capacity by 2030. Given the large footprint projected to meet renewables energy targets, the potential for land use conflicts over environmental values is high. To expedite development of so… ▽ More

    Submitted 30 June, 2022; v1 submitted 31 January, 2022; originally announced February 2022.

    Comments: Accepted for publication in Nature Scientific Data

  36. arXiv:2111.06012  [pdf, other

    cs.CL cs.LG

    Kronecker Factorization for Preventing Catastrophic Forgetting in Large-scale Medical Entity Linking

    Authors: Denis Jered McInerney, Luyang Kong, Kristjan Arumae, Byron Wallace, Parminder Bhatia

    Abstract: Multi-task learning is useful in NLP because it is often practically desirable to have a single model that works across a range of tasks. In the medical domain, sequential training on tasks may sometimes be the only way to train models, either because access to the original (potentially sensitive) data is no longer available, or simply owing to the computational costs inherent to joint retraining.… ▽ More

    Submitted 10 November, 2021; originally announced November 2021.

  37. arXiv:2110.08455  [pdf, other

    cs.CL

    Knowledge Enhanced Pretrained Language Models: A Compreshensive Survey

    Authors: Xiaokai Wei, Shen Wang, Dejiao Zhang, Parminder Bhatia, Andrew Arnold

    Abstract: Pretrained Language Models (PLM) have established a new paradigm through learning informative contextualized representations on large-scale text corpus. This new paradigm has revolutionized the entire field of natural language processing, and set the new state-of-the-art performance for a wide variety of NLP tasks. However, though PLMs could store certain knowledge/facts from training corpus, thei… ▽ More

    Submitted 15 October, 2021; originally announced October 2021.

  38. arXiv:2107.11094  [pdf, other

    cs.CL

    Improving Early Sepsis Prediction with Multi Modal Learning

    Authors: Fred Qin, Vivek Madan, Ujjwal Ratan, Zohar Karnin, Vishaal Kapoor, Parminder Bhatia, Taha Kass-Hout

    Abstract: Sepsis is a life-threatening disease with high morbidity, mortality and healthcare costs. The early prediction and administration of antibiotics and intravenous fluids is considered crucial for the treatment of sepsis and can save potentially millions of lives and billions in health care costs. Professional clinical care practitioners have proposed clinical criterion which aid in early detection o… ▽ More

    Submitted 23 July, 2021; originally announced July 2021.

  39. arXiv:2105.13225  [pdf, other

    cs.CL cs.AI

    Neural Entity Recognition with Gazetteer based Fusion

    Authors: Qing Sun, Parminder Bhatia

    Abstract: Incorporating external knowledge into Named Entity Recognition (NER) systems has been widely studied in the generic domain. In this paper, we focus on clinical domain where only limited data is accessible and interpretability is important. Recent advancement in technology and the acceleration of clinical trials has resulted in the discovery of new drugs, procedures as well as medical conditions. T… ▽ More

    Submitted 27 May, 2021; originally announced May 2021.

    Journal ref: the Association for Computational Linguistics (ACL) 2021

  40. arXiv:2105.12682  [pdf, other

    cs.CL cs.AI

    Zero-shot Medical Entity Retrieval without Annotation: Learning From Rich Knowledge Graph Semantics

    Authors: Luyang Kong, Christopher Winestock, Parminder Bhatia

    Abstract: Medical entity retrieval is an integral component for understanding and communicating information across various health systems. Current approaches tend to work well on specific medical domains but generalize poorly to unseen sub-specialties. This is of increasing concern under a public health crisis as new medical conditions and drug treatments come to light frequently. Zero-shot retrieval is cha… ▽ More

    Submitted 26 May, 2021; originally announced May 2021.

  41. arXiv:2104.13498  [pdf, other

    cs.CL cs.LG

    Towards Clinical Encounter Summarization: Learning to Compose Discharge Summaries from Prior Notes

    Authors: Han-Chin Shing, Chaitanya Shivade, Nima Pourdamghani, Feng Nan, Philip Resnik, Douglas Oard, Parminder Bhatia

    Abstract: The records of a clinical encounter can be extensive and complex, thus placing a premium on tools that can extract and summarize relevant information. This paper introduces the task of generating discharge summaries for a clinical encounter. Summaries in this setting need to be faithful, traceable, and scale to multiple long documents, motivating the use of extract-then-abstract summarization casc… ▽ More

    Submitted 27 April, 2021; originally announced April 2021.

  42. arXiv:2010.00784  [pdf, other

    cs.CL

    An Empirical Investigation Towards Efficient Multi-Domain Language Model Pre-training

    Authors: Kristjan Arumae, Qing Sun, Parminder Bhatia

    Abstract: Pre-training large language models has become a standard in the natural language processing community. Such models are pre-trained on generic data (e.g. BookCorpus and English Wikipedia) and often fine-tuned on tasks in the same domain. However, in order to achieve state-of-the-art performance on out of domain tasks such as clinical named entity recognition and relation extraction, additional in d… ▽ More

    Submitted 1 October, 2020; originally announced October 2020.

    Comments: arXiv admin note: text overlap with arXiv:2004.03794

  43. arXiv:2009.07241  [pdf, other

    stat.ML cs.LG

    Improve black-box sequential anomaly detector relevancy with limited user feedback

    Authors: Luyang Kong, Lifan Chen, Ming Chen, Parminder Bhatia, Laurent Callot

    Abstract: Anomaly detectors are often designed to catch statistical anomalies. End-users typically do not have interest in all of the detected outliers, but only those relevant to their application. Given an existing black-box sequential anomaly detector, this paper proposes a method to improve its user relevancy using a small number of human feedback. As our first contribution, the method is agnostic to th… ▽ More

    Submitted 15 September, 2020; originally announced September 2020.

  44. arXiv:2007.12731  [pdf, other

    cs.IR cs.AI cs.CL

    COVID-19 Knowledge Graph: Accelerating Information Retrieval and Discovery for Scientific Literature

    Authors: Colby Wise, Vassilis N. Ioannidis, Miguel Romero Calvo, Xiang Song, George Price, Ninad Kulkarni, Ryan Brand, Parminder Bhatia, George Karypis

    Abstract: The coronavirus disease (COVID-19) has claimed the lives of over 350,000 people and infected more than 6 million people worldwide. Several search engines have surfaced to provide researchers with additional tools to find and retrieve information from the rapidly growing corpora on COVID-19. These engines lack extraction and visualization tools necessary to retrieve and interpret complex relations… ▽ More

    Submitted 24 July, 2020; originally announced July 2020.

  45. arXiv:2007.09186  [pdf, other

    cs.IR

    AWS CORD-19 Search: A Neural Search Engine for COVID-19 Literature

    Authors: Parminder Bhatia, Lan Liu, Kristjan Arumae, Nima Pourdamghani, Suyog Deshpande, Ben Snively, Mona Mona, Colby Wise, George Price, Shyam Ramaswamy, Xiaofei Ma, Ramesh Nallapati, Zhiheng Huang, Bing Xiang, Taha Kass-Hout

    Abstract: Coronavirus disease (COVID-19) has been declared as a pandemic by WHO with thousands of cases being reported each day. Numerous scientific articles are being published on the disease raising the need for a service which can organize, and query them in a reliable fashion. To support this cause we present AWS CORD-19 Search (ACS), a public, COVID-19 specific, neural search engine that is powered by… ▽ More

    Submitted 7 October, 2020; v1 submitted 17 July, 2020; originally announced July 2020.

  46. arXiv:2007.00492  [pdf, other

    cs.CL cs.CY cs.LG stat.ML

    Towards User Friendly Medication Mapping Using Entity-Boosted Two-Tower Neural Network

    Authors: Shaoqing Yuan, Parminder Bhatia, Busra Celikkaya, Haiyang Liu, Kyunghwan Choi

    Abstract: Recent advancements in medical entity linking have been applied in the area of scientific literature and social media data. However, with the adoption of telemedicine and conversational agents such as Alexa in healthcare settings, medical name inference has become an important task. Medication name inference is the task of mapping user friendly medication names from a free-form text to a concept i… ▽ More

    Submitted 9 October, 2020; v1 submitted 17 June, 2020; originally announced July 2020.

  47. arXiv:2006.13299  [pdf, other

    cs.CL cs.LG

    Supervised Understanding of Word Embeddings

    Authors: Halid Ziya Yerebakan, Parmeet Bhatia, Yoshihisa Shinagawa

    Abstract: Pre-trained word embeddings are widely used for transfer learning in natural language processing. The embeddings are continuous and distributed representations of the words that preserve their similarities in compact Euclidean spaces. However, the dimensions of these spaces do not provide any clear interpretation. In this study, we have obtained supervised projections in the form of the linear key… ▽ More

    Submitted 23 June, 2020; originally announced June 2020.

  48. arXiv:2004.04295  [pdf, ps, other

    cs.CL

    Severing the Edge Between Before and After: Neural Architectures for Temporal Ordering of Events

    Authors: Miguel Ballesteros, Rishita Anubhai, Shuai Wang, Nima Pourdamghani, Yogarshi Vyas, Jie Ma, Parminder Bhatia, Kathleen McKeown, Yaser Al-Onaizan

    Abstract: In this paper, we propose a neural architecture and a set of training methods for ordering events by predicting temporal relations. Our proposed models receive a pair of events within a span of text as input and they identify temporal relations (Before, After, Equal, Vague) between them. Given that a key challenge with this task is the scarcity of annotated data, our models rely on either pretrain… ▽ More

    Submitted 8 April, 2020; originally announced April 2020.

  49. arXiv:2004.03794  [pdf, other

    cs.CL

    CALM: Continuous Adaptive Learning for Language Modeling

    Authors: Kristjan Arumae, Parminder Bhatia

    Abstract: Training large language representation models has become a standard in the natural language processing community. This allows for fine tuning on any number of specific tasks, however, these large high capacity models can continue to train on domain specific unlabeled data to make initialization even more robust for supervised tasks. We demonstrate that in practice these pre-trained models present… ▽ More

    Submitted 7 April, 2020; originally announced April 2020.

  50. arXiv:1911.09787  [pdf, other

    cs.CL cs.IR cs.LG stat.ML

    LATTE: Latent Type Modeling for Biomedical Entity Linking

    Authors: Ming Zhu, Busra Celikkaya, Parminder Bhatia, Chandan K. Reddy

    Abstract: Entity linking is the task of linking mentions of named entities in natural language text, to entities in a curated knowledge-base. This is of significant importance in the biomedical domain, where it could be used to semantically annotate a large volume of clinical records and biomedical literature, to standardized concepts described in an ontology such as Unified Medical Language System (UMLS).… ▽ More

    Submitted 20 January, 2020; v1 submitted 21 November, 2019; originally announced November 2019.

    Comments: AAAI 2020 Conference