Skip to main content

Showing 1–5 of 5 results for author: Agro, M T

Searching in archive cs. Search in all archives.
.
  1. arXiv:2601.08645  [pdf, ps, other

    cs.CL

    A Parallel Cross-Lingual Benchmark for Multimodal Idiomaticity Understanding

    Authors: Dilara Torunoğlu-Selamet, Dogukan Arslan, Rodrigo Wilkens, Wei He, Doruk Eryiğit, Thomas Pickard, Adriana S. Pagano, Aline Villavicencio, Gülşen Eryiğit, Ágnes Abuczki, Aida Cardoso, Alesia Lazarenka, Dina Almassova, Amalia Mendes, Anna Kanellopoulou, Antoni Brosa-Rodríguez, Baiba Saulite, Beata Wojtowicz, Bolette Pedersen, Carlos Manuel Hidalgo-Ternero, Chaya Liebeskind, Danka Jokić, Diego Alves, Eleni Triantafyllidi, Erik Velldal , et al. (53 additional authors not shown)

    Abstract: Potentially idiomatic expressions (PIEs) construe meanings inherently tied to the everyday experience of a given language community. As such, they constitute an interesting challenge for assessing the linguistic (and to some extent cultural) capabilities of NLP systems. In this paper, we present XMPIE, a parallel multilingual and multimodal dataset of potentially idiomatic expressions. The dataset… ▽ More

    Submitted 24 February, 2026; v1 submitted 13 January, 2026; originally announced January 2026.

  2. arXiv:2507.07741  [pdf, ps, other

    cs.CL cs.SD eess.AS

    Code-Switching in End-to-End Automatic Speech Recognition: A Systematic Literature Review

    Authors: Maha Tufail Agro, Atharva Kulkarni, Karima Kadaoui, Zeerak Talat, Hanan Aldarmaki

    Abstract: Motivated by a growing research interest into automatic speech recognition (ASR), and the growing body of work for languages in which code-switching (CS) often occurs, we present a systematic literature review of code-switching in end-to-end ASR models. We collect and manually annotate papers published in peer reviewed venues. We document the languages considered, datasets, metrics, model choices,… ▽ More

    Submitted 10 July, 2025; originally announced July 2025.

  3. arXiv:2506.12552  [pdf, ps, other

    cs.CL cs.AI cs.LG

    Profiling News Media for Factuality and Bias Using LLMs and the Fact-Checking Methodology of Human Experts

    Authors: Zain Muhammad Mujahid, Dilshod Azizov, Maha Tufail Agro, Preslav Nakov

    Abstract: In an age characterized by the proliferation of mis- and disinformation online, it is critical to empower readers to understand the content they are reading. Important efforts in this direction rely on manual or automatic fact-checking, which can be challenging for emerging claims with limited information. Such scenarios can be handled by assessing the reliability and the political bias of the sou… ▽ More

    Submitted 14 June, 2025; originally announced June 2025.

    Comments: Accepted to Findings of the Association for Computational Linguistics (ACL) 2025

  4. arXiv:2409.19806  [pdf, other

    cs.SD cs.AI eess.AS

    PALM: Few-Shot Prompt Learning for Audio Language Models

    Authors: Asif Hanif, Maha Tufail Agro, Mohammad Areeb Qazi, Hanan Aldarmaki

    Abstract: Audio-Language Models (ALMs) have recently achieved remarkable success in zero-shot audio recognition tasks, which match features of audio waveforms with class-specific text prompt features, inspired by advancements in Vision-Language Models (VLMs). Given the sensitivity of zero-shot performance to the choice of hand-crafted text prompts, many prompt learning techniques have been developed for VLM… ▽ More

    Submitted 29 September, 2024; originally announced September 2024.

    Comments: EMNLP 2024 (Main)

  5. arXiv:2305.16337  [pdf, other

    cs.CL cs.AI

    Handling Realistic Label Noise in BERT Text Classification

    Authors: Maha Tufail Agro, Hanan Aldarmaki

    Abstract: Labels noise refers to errors in training labels caused by cheap data annotation methods, such as web scraping or crowd-sourcing, which can be detrimental to the performance of supervised classifiers. Several methods have been proposed to counteract the effect of random label noise in supervised classification, and some studies have shown that BERT is already robust against high rates of randomly… ▽ More

    Submitted 20 October, 2023; v1 submitted 23 May, 2023; originally announced May 2023.