Skip to main content

Showing 1–50 of 205 results for author: Cha, S

Searching in archive cs. Search in all archives.
.
  1. arXiv:2604.05349  [pdf, ps, other

    cs.HC cs.SE

    Symetra: Visual Analytics for the Parameter Tuning Process of Symbolic Execution Engines

    Authors: Donghee Hong, Minjong Kim, Sooyoung Cha, Jaemin Jo

    Abstract: Symbolic execution engines such as KLEE automatically generate test cases to maximize branch coverage, but their numerous parameters make it difficult to understand the parameters' impact, leading the user to rely on suboptimal default configurations. While automated tuners have shown promising results, they provide limited insights into why certain configurations work well, motivating the need fo… ▽ More

    Submitted 6 April, 2026; originally announced April 2026.

  2. arXiv:2603.28708  [pdf

    cs.LG cs.DC

    GPU-Accelerated Optimization of Transformer-Based Neural Networks for Real-Time Inference

    Authors: Soutrik Mukherjee, Sangwhan Cha

    Abstract: This paper presents the design and evaluation of a GPU-accelerated inference pipeline for transformer models using NVIDIA TensorRT with mixed-precision optimization. We evaluate BERT-base (110M parameters) and GPT-2 (124M parameters) across batch sizes from 1 to 32 and sequence lengths from 32 to 512. The system achieves up to 64.4x speedup over CPU baselines, sub-10 ms latency for single-sample i… ▽ More

    Submitted 30 March, 2026; originally announced March 2026.

    Comments: 10 pages, 8 figures, 15 tables

  3. arXiv:2603.27950  [pdf, ps, other

    cs.LG

    Scaling Atomistic Protein Binder Design with Generative Pretraining and Test-Time Compute

    Authors: Kieran Didi, Zuobai Zhang, Guoqing Zhou, Danny Reidenbach, Zhonglin Cao, Sooyoung Cha, Tomas Geffner, Christian Dallago, Jian Tang, Michael M. Bronstein, Martin Steinegger, Emine Kucukbenli, Arash Vahdat, Karsten Kreis

    Abstract: Protein interaction modeling is central to protein design, which has been transformed by machine learning with applications in drug discovery and beyond. In this landscape, structure-based de novo binder design is cast as either conditional generative modeling or sequence optimization via structure predictors ("hallucination"). We argue that this is a false dichotomy and propose Proteina-Complexa,… ▽ More

    Submitted 29 March, 2026; originally announced March 2026.

    Comments: ICLR 2026 Oral Presentation. Project page: https://research.nvidia.com/labs/genair/proteina-complexa/

  4. arXiv:2603.25551  [pdf, ps, other

    cs.AI

    Voxtral TTS

    Authors: Mistral-AI, :, Alexander H. Liu, Alexis Tacnet, Andy Ehrenberg, Andy Lo, Chen-Yo Sun, Guillaume Lample, Henry Lagarde, Jean-Malo Delignon, Jaeyoung Kim, John Harvill, Khyathi Raghavi Chandu, Lorenzo Signoretti, Margaret Jennings, Patrick von Platen, Pavankumar Reddy Muddireddy, Rohin Arora, Sanchit Gandhi, Samuel Humeau, Soham Ghosh, Srijan Mishra, Van Phung, Abdelaziz Bounhar, Abhinav Rastogi , et al. (164 additional authors not shown)

    Abstract: We introduce Voxtral TTS, an expressive multilingual text-to-speech model that generates natural speech from as little as 3 seconds of reference audio. Voxtral TTS adopts a hybrid architecture that combines auto-regressive generation of semantic speech tokens with flow-matching for acoustic tokens. These tokens are encoded and decoded with Voxtral Codec, a speech tokenizer trained from scratch wit… ▽ More

    Submitted 6 April, 2026; v1 submitted 26 March, 2026; originally announced March 2026.

  5. arXiv:2603.20851  [pdf, ps, other

    cs.CL cs.AI

    Can ChatGPT Really Understand Modern Chinese Poetry?

    Authors: Shanshan Wang, Derek F. Wong, Jingming Yao, Lidia S. Chao

    Abstract: ChatGPT has demonstrated remarkable capabilities on both poetry generation and translation, yet its ability to truly understand poetry remains unexplored. Previous poetry-related work merely analyzed experimental outcomes without addressing fundamental issues of comprehension. This paper introduces a comprehensive framework for evaluating ChatGPT's understanding of modern poetry. We collaborated w… ▽ More

    Submitted 21 March, 2026; originally announced March 2026.

    Comments: Accepted by EACL 2026

  6. arXiv:2603.19158  [pdf, ps, other

    cs.CV

    Adaptive Auxiliary Prompt Blending for Target-Faithful Diffusion Generation

    Authors: Kwanyoung Lee, SeungJu Cha, Yebin Ahn, Hyunwoo Oh, Sungho Koh, Dong-Jin Kim

    Abstract: Diffusion-based text-to-image (T2I) models have made remarkable progress in generating photorealistic and semantically rich images. However, when the target concepts lie in low-density regions of the training distribution, these models often produce semantically misaligned or structurally inconsistent results. This limitation arises from the long-tailed nature of text-image datasets, where rare co… ▽ More

    Submitted 19 March, 2026; originally announced March 2026.

    Comments: Accepted in CVPR 2026 (main track). 10 pages, 6 figures; supplementary material included (14 pages, 11 figures)

  7. arXiv:2603.19157  [pdf, ps, other

    cs.CV

    ADAPT: Attention Driven Adaptive Prompt Scheduling and InTerpolating Orthogonal Complements for Rare Concepts Generation

    Authors: Kwanyoung Lee, Hyunwoo Oh, SeungJu Cha, Sungho Koh, Dong-Jin Kim

    Abstract: Generating rare compositional concepts in text-to-image synthesis remains a challenge for diffusion models, particularly for attributes that are uncommon in the training data. While recent approaches, such as R2F, address this challenge by utilizing LLM for prompt scheduling, they suffer from inherent variance due to the randomness of language models and suboptimal guidance from iterative text emb… ▽ More

    Submitted 19 March, 2026; originally announced March 2026.

    Comments: Accepted in CVPR 2026 (findings). 10 pages, 4 figures; supplementary material included (8 pages, 10 figures)

  8. arXiv:2603.13201  [pdf, ps, other

    cs.CL

    Neuron-Aware Data Selection In Instruction Tuning For Large Language Models

    Authors: Xin Chen, Junchao Wu, Shu Yang, Runzhe Zhan, Zeyu Wu, Min Yang, Shujian Huang, Lidia S. Chao, Derek F. Wong

    Abstract: Instruction Tuning (IT) has been proven to be an effective approach to unlock the powerful capabilities of large language models (LLMs). Recent studies indicate that excessive IT data can degrade LLMs performance, while carefully selecting a small subset of high-quality IT data can significantly enhance their capabilities. Therefore, identifying the most efficient subset data from the IT dataset t… ▽ More

    Submitted 13 March, 2026; originally announced March 2026.

  9. arXiv:2603.09216  [pdf, ps, other

    cs.DC

    PIM-SHERPA: Software Method for On-device LLM Inference by Resolving PIM Memory Attribute and Layout Inconsistencies

    Authors: Sunjung Lee, Sanghoon Cha, Hyeonsu Kim, Seungwoo Seo, Yuhwan Ro, Sukhan Lee, Byeongho Kim, Yongjun Park, Kyomin Sohn, Seungwon Lee, Jaehoon Yu

    Abstract: On-device deployments of large language models (LLMs) are rapidly proliferating across mobile and edge platforms. LLM inference comprises a compute-intensive prefill phase and a memory bandwidth-intensive decode phase, and the decode phase has been widely recognized as well-suited to processing-in-memory (PIM) in both academia and industry. However, practical PIM-enabled systems face two obstacles… ▽ More

    Submitted 10 March, 2026; originally announced March 2026.

    Comments: 13 pages, 13 figures

  10. arXiv:2603.09062  [pdf, ps, other

    cs.LG

    Dynamic Multi-period Experts for Online Time Series Forecasting

    Authors: Seungha Hong, Sukang Chae, Suyeon Kim, Sanghwan Jang, Hwanjo Yu

    Abstract: Online Time Series Forecasting (OTSF) requires models to continuously adapt to concept drift. However, existing methods often treat concept drift as a monolithic phenomenon. To address this limitation, we first redefine concept drift by categorizing it into two distinct types: Recurring Drift, where previously seen patterns reappear, and Emergent Drift, where entirely new patterns emerge. We then… ▽ More

    Submitted 9 March, 2026; originally announced March 2026.

    Comments: WWW 2026

  11. arXiv:2603.08483  [pdf, ps, other

    cs.CV cs.AI cs.LG

    X-AVDT: Audio-Visual Cross-Attention for Robust Deepfake Detection

    Authors: Youngseo Kim, Kwan Yun, Seokhyeon Hong, Sihun Cha, Colette Suhjung Koo, Junyong Noh

    Abstract: The surge of highly realistic synthetic videos produced by contemporary generative systems has significantly increased the risk of malicious use, challenging both humans and existing detectors. Against this backdrop, we take a generator-side view and observe that internal cross-attention mechanisms in these models encode fine-grained speech-motion alignment, offering useful correspondence cues for… ▽ More

    Submitted 9 March, 2026; originally announced March 2026.

    Journal ref: CVPR 2026

  12. arXiv:2603.07392  [pdf, ps, other

    cs.CL

    Can Large Language Models Keep Up? Benchmarking Online Adaptation to Continual Knowledge Streams

    Authors: Jiyeon Kim, Hyunji Lee, Dylan Zhou, Sue Hyun Park, Seunghyun Yoon, Trung Bui, Franck Dernoncourt, Sungmin Cha, Minjoon Seo

    Abstract: LLMs operating in dynamic real-world contexts often encounter knowledge that evolves continuously or emerges incrementally. To remain accurate and effective, models must adapt to newly arriving information on the fly. We introduce Online Adaptation to Continual Knowledge Streams(OAKS) to evaluate this capability, establishing a benchmark for online adaptation over streaming, continually updating k… ▽ More

    Submitted 7 March, 2026; originally announced March 2026.

  13. arXiv:2603.05437  [pdf, ps, other

    cs.CV cs.AI

    SAIL: Similarity-Aware Guidance and Inter-Caption Augmentation-based Learning for Weakly-Supervised Dense Video Captioning

    Authors: Ye-Chan Kim, SeungJu Cha, Si-Woo Kim, Minju Jeon, Hyungee Kim, Dong-Jin Kim

    Abstract: Weakly-Supervised Dense Video Captioning aims to localize and describe events in videos trained only on caption annotations, without temporal boundaries. Prior work introduced an implicit supervision paradigm based on Gaussian masking and complementary captioning. However, existing method focuses merely on generating non-overlapping masks without considering their semantic relationship to correspo… ▽ More

    Submitted 9 March, 2026; v1 submitted 5 March, 2026; originally announced March 2026.

    Comments: Accepted to CVPR 2026

  14. arXiv:2603.04247  [pdf, ps, other

    cs.LG cs.AI

    Online Learning for Multi-Layer Hierarchical Inference under Partial and Policy-Dependent Feedback

    Authors: Haoran Zhang, Seohyeon Cha, Hasan Burhan Beytur, Kevin S Chan, Gustavo de Veciana, Haris Vikalo

    Abstract: Hierarchical inference systems route tasks across multiple computational layers, where each node may either finalize a prediction locally or offload the task to a node in the next layer for further processing. Learning optimal routing policies in such systems is challenging: inference loss is defined recursively across layers, while feedback on prediction error is revealed only at a terminal oracl… ▽ More

    Submitted 4 March, 2026; originally announced March 2026.

    Comments: preprint

  15. arXiv:2602.23638  [pdf, ps, other

    cs.LG cs.AI

    FedRot-LoRA: Mitigating Rotational Misalignment in Federated LoRA

    Authors: Haoran Zhang, Dongjun Kim, Seohyeon Cha, Haris Vikalo

    Abstract: Federated LoRA provides a communication-efficient mechanism for fine-tuning large language models on decentralized data. In practice, however, a discrepancy between the factor-wise averaging used to preserve low rank and the mathematically correct aggregation of local updates can cause significant aggregation error and unstable training. We argue that a major source of this problem is rotational m… ▽ More

    Submitted 26 February, 2026; originally announced February 2026.

    Comments: preprint

    MSC Class: 68T05 (Primary); 68T07 (Secondary) ACM Class: I.2.11; I.2.7

  16. arXiv:2602.20217  [pdf, ps, other

    cs.LG cs.AI

    KnapSpec: Self-Speculative Decoding via Adaptive Layer Selection as a Knapsack Problem

    Authors: Seongjin Cha, Gyuwan Kim, Dongsu Han, Tao Yang, Insu Han

    Abstract: Self-speculative decoding (SSD) accelerates LLM inference by skipping layers to create an efficient draft model, yet existing methods often rely on static heuristics that ignore the dynamic computational overhead of attention in long-context scenarios. We propose KnapSpec, a training-free framework that reformulates draft model selection as a knapsack problem to maximize tokens-per-time throughput… ▽ More

    Submitted 23 February, 2026; originally announced February 2026.

  17. arXiv:2602.11298  [pdf, ps, other

    cs.AI

    Voxtral Realtime

    Authors: Mistral-AI, :, Alexander H. Liu, Andy Ehrenberg, Andy Lo, Chen-Yo Sun, Guillaume Lample, Jean-Malo Delignon, Khyathi Raghavi Chandu, Patrick von Platen, Pavankumar Reddy Muddireddy, Rohin Arora, Sanchit Gandhi, Sandeep Subramanian, Soham Ghosh, Srijan Mishra, Abhinav Rastogi, Adrien Sadé, Alan Jeffares, Albert Jiang, Alexandre Cahill, Alexandre Gavaudan, Alexandre Sablayrolles, Amélie Héliou, Amos You , et al. (144 additional authors not shown)

    Abstract: We introduce Voxtral Realtime, a natively streaming automatic speech recognition model that matches offline transcription quality at sub-second latency. Unlike approaches that adapt offline models through chunking or sliding windows, Voxtral Realtime is trained end-to-end for streaming, with explicit alignment between audio and text streams. Our architecture builds on the Delayed Streams Modeling… ▽ More

    Submitted 6 April, 2026; v1 submitted 11 February, 2026; originally announced February 2026.

  18. arXiv:2602.05902  [pdf, ps, other

    cs.LG cs.AI

    Regularized Calibration with Successive Rounding for Post-Training Quantization

    Authors: Seohyeon Cha, Huancheng Chen, Dongjun Kim, Haoran Zhang, Kevin Chan, Gustavo de Veciana, Haris Vikalo

    Abstract: Large language models (LLMs) deliver robust performance across diverse applications, yet their deployment often faces challenges due to the memory and latency costs of storing and accessing billions of parameters. Post-training quantization (PTQ) enables efficient inference by mapping pretrained weights to low-bit formats without retraining, but its effectiveness depends critically on both the qua… ▽ More

    Submitted 5 February, 2026; originally announced February 2026.

  19. arXiv:2602.05339  [pdf, ps, other

    cs.CV cs.LG

    Consistency-Preserving Concept Erasure via Unsafe-Safe Pairing and Directional Fisher-weighted Adaptation

    Authors: Yongwoo Kim, Sungmin Cha, Hyunsoo Kim, Jaewon Lee, Donghyun Kim

    Abstract: With the increasing versatility of text-to-image diffusion models, the ability to selectively erase undesirable concepts (e.g., harmful content) has become indispensable. However, existing concept erasure approaches primarily focus on removing unsafe concepts without providing guidance toward corresponding safe alternatives, which often leads to failure in preserving the structural and semantic co… ▽ More

    Submitted 5 February, 2026; originally announced February 2026.

  20. arXiv:2601.21794  [pdf, ps, other

    cs.LG

    Knowledge Vector Weakening: Efficient Training-free Unlearning for Large Vision-Language Models

    Authors: Yejin Kim, Dongjun Hwang, Sungmin Cha, Junsuk Choe

    Abstract: Large Vision-Language Models (LVLMs) are widely adopted for their strong multimodal capabilities, yet they raise serious concerns such as privacy leakage and harmful content generation. Machine unlearning has emerged as a promising solution for removing the influence of specific data from trained models. However, existing approaches largely rely on gradient-based optimization, incurring substantia… ▽ More

    Submitted 29 January, 2026; originally announced January 2026.

  21. Deep Learning Based Facial Retargeting Using Local Patches

    Authors: Yeonsoo Choi, Inyup Lee, Sihun Cha, Seonghyeon Kim, Sunjin Jung, Junyong Noh

    Abstract: In the era of digital animation, the quest to produce lifelike facial animations for virtual characters has led to the development of various retargeting methods. While the retargeting facial motion between models of similar shapes has been very successful, challenges arise when the retargeting is performed on stylized or exaggerated 3D characters that deviate significantly from human facial struc… ▽ More

    Submitted 13 January, 2026; originally announced January 2026.

    Comments: Eurographics 25

    Journal ref: Computer Graphics Forum 2024

  22. arXiv:2601.07125  [pdf, ps, other

    cs.IR cs.CL cs.CV

    ReinPool: Reinforcement Learning Pooling Multi-Vector Embeddings for Retrieval System

    Authors: Sungguk Cha, DongWook Kim, Mintae Kim, Youngsub Han, Byoung-Ki Jeon, Sangyeob Lee

    Abstract: Multi-vector embedding models have emerged as a powerful paradigm for document retrieval, preserving fine-grained visual and textual details through token-level representations. However, this expressiveness comes at a staggering cost: storing embeddings for every token inflates index sizes by over $1000\times$ compared to single-vector approaches, severely limiting scalability. We introduce \textb… ▽ More

    Submitted 11 January, 2026; originally announced January 2026.

    Comments: 5 pages

  23. arXiv:2511.03929  [pdf, ps, other

    cs.LG cs.AI cs.CV

    NVIDIA Nemotron Nano V2 VL

    Authors: NVIDIA, :, Amala Sanjay Deshmukh, Kateryna Chumachenko, Tuomas Rintamaki, Matthieu Le, Tyler Poon, Danial Mohseni Taheri, Ilia Karmanov, Guilin Liu, Jarno Seppanen, Guo Chen, Karan Sapra, Zhiding Yu, Adi Renduchintala, Charles Wang, Peter Jin, Arushi Goel, Mike Ranzinger, Lukas Voegtle, Philipp Fischer, Timo Roman, Wei Ping, Boxin Wang, Zhuolin Yang , et al. (99 additional authors not shown)

    Abstract: We introduce Nemotron Nano V2 VL, the latest model of the Nemotron vision-language series designed for strong real-world document understanding, long video comprehension, and reasoning tasks. Nemotron Nano V2 VL delivers significant improvements over our previous model, Llama-3.1-Nemotron-Nano-VL-8B, across all vision and text domains through major enhancements in model architecture, datasets, and… ▽ More

    Submitted 6 November, 2025; v1 submitted 5 November, 2025; originally announced November 2025.

  24. arXiv:2510.25818  [pdf, ps, other

    cs.LG cs.AI

    ScaleDiff: Higher-Resolution Image Synthesis via Efficient and Model-Agnostic Diffusion

    Authors: Sungho Koh, SeungJu Cha, Hyunwoo Oh, Kwanyoung Lee, Dong-Jin Kim

    Abstract: Text-to-image diffusion models often exhibit degraded performance when generating images beyond their training resolution. Recent training-free methods can mitigate this limitation, but they often require substantial computation or are incompatible with recent Diffusion Transformer models. In this paper, we propose ScaleDiff, a model-agnostic and highly efficient framework for extending the resolu… ▽ More

    Submitted 29 October, 2025; originally announced October 2025.

    Comments: NeurIPS 2025. Code: https://github.com/KSH00906/ScaleDiff

  25. arXiv:2510.24541  [pdf, ps, other

    cs.CL

    Open Korean Historical Corpus: A Millennia-Scale Diachronic Collection of Public Domain Texts

    Authors: Seyoung Song, Nawon Kim, Songeun Chae, Kiwoong Park, Jiho Jin, Haneul Yoo, Kyunghyun Cho, Alice Oh

    Abstract: The history of the Korean language is characterized by a discrepancy between its spoken and written forms and a pivotal shift from Chinese characters to the Hangul alphabet. However, this linguistic evolution has remained largely unexplored in NLP due to a lack of accessible historical corpora. To address this gap, we introduce the Open Korean Historical Corpus, a large-scale, openly licensed data… ▽ More

    Submitted 5 March, 2026; v1 submitted 28 October, 2025; originally announced October 2025.

    Comments: LREC 2026

  26. arXiv:2510.23371  [pdf, ps, other

    cs.LG cs.CE

    Towards a Generalizable AI for Materials Discovery: Validation through Immersion Coolant Screening

    Authors: Hyunseung Kim, Dae-Woong Jeong, Changyoung Park, Won-Ji Lee, Ha-Eun Lee, Ji-Hye Lee, Rodrigo Hormazabal, Sung Moon Ko, Sumin Lee, Soorin Yim, Chanhui Lee, Sehui Han, Sang-Ho Cha, Woohyung Lim

    Abstract: Artificial intelligence (AI) has emerged as a powerful accelerator of materials discovery, yet most existing models remain problem-specific, requiring additional data collection and retraining for each new property. Here we introduce and validate GATE (Geometrically Aligned Transfer Encoder) -- a generalizable AI framework that jointly learns 34 physicochemical properties spanning thermal, electri… ▽ More

    Submitted 31 October, 2025; v1 submitted 27 October, 2025; originally announced October 2025.

    Comments: 16 pages, 4 figures

  27. arXiv:2510.20780  [pdf, ps, other

    cs.CL cs.AI

    Are Large Reasoning Models Good Translation Evaluators? Analysis and Performance Boost

    Authors: Runzhe Zhan, Zhihong Huang, Xinyi Yang, Lidia S. Chao, Min Yang, Derek F. Wong

    Abstract: Recent advancements in large reasoning models (LRMs) have introduced an intermediate "thinking" process prior to generating final answers, improving their reasoning capabilities on complex downstream tasks. However, the potential of LRMs as evaluators for machine translation (MT) quality remains underexplored. We provides the first systematic analysis of LRM-as-a-judge in MT evaluation. We identif… ▽ More

    Submitted 23 October, 2025; originally announced October 2025.

    Comments: NeurIPS 2025

  28. arXiv:2510.14557  [pdf, ps, other

    cs.LG cs.AR

    MX+: Pushing the Limits of Microscaling Formats for Efficient Large Language Model Serving

    Authors: Jungi Lee, Junyong Park, Soohyun Cha, Jaehoon Cho, Jaewoong Sim

    Abstract: Reduced-precision data formats are crucial for cost-effective serving of large language models (LLMs). While numerous reduced-precision formats have been introduced thus far, they often require intrusive modifications to the software frameworks or are rather unconventional for widespread adoption across hardware vendors. In this paper, we instead focus on recent industry-driven variants of block f… ▽ More

    Submitted 16 October, 2025; originally announced October 2025.

    Comments: To appear at the 58th International Symposium on Microarchitecture (MICRO 2025)

  29. arXiv:2510.12981  [pdf, ps, other

    cs.LG

    Reference-Specific Unlearning Metrics Can Hide the Truth: A Reality Check

    Authors: Sungjun Cho, Dasol Hwang, Frederic Sala, Sangheum Hwang, Kyunghyun Cho, Sungmin Cha

    Abstract: Current unlearning metrics for generative models evaluate success based on reference responses or classifier outputs rather than assessing the core objective: whether the unlearned model behaves indistinguishably from a model that never saw the unwanted data. This reference-specific approach creates systematic blind spots, allowing models to appear successful while retaining unwanted knowledge acc… ▽ More

    Submitted 14 October, 2025; originally announced October 2025.

    Comments: 20 pages, 11 figures

  30. arXiv:2510.10517  [pdf, ps, other

    cs.PL cs.AI cs.SE

    ECO: Enhanced Code Optimization via Performance-Aware Prompting for Code-LLMs

    Authors: Su-Hyeon Kim, Joonghyuk Hahn, Sooyoung Cha, Yo-Sub Han

    Abstract: Code runtime optimization-the task of rewriting a given code to a faster one-remains challenging, as it requires reasoning about performance trade-offs involving algorithmic and structural choices. Recent approaches employ code-LLMs with slow-fast code pairs provided as optimization guidance, but such pair-based methods obscure the causal factors of performance gains and often lead to superficial… ▽ More

    Submitted 12 October, 2025; originally announced October 2025.

  31. arXiv:2510.10013  [pdf, ps, other

    cs.CL

    Path Drift in Large Reasoning Models:How First-Person Commitments Override Safety

    Authors: Yuyi Huang, Runzhe Zhan, Lidia S. Chao, Ailin Tao, Derek F. Wong

    Abstract: As large language models (LLMs) are increasingly deployed for complex reasoning tasks, Long Chain-of-Thought (Long-CoT) prompting has emerged as a key paradigm for structured inference. Despite early-stage safeguards enabled by alignment techniques such as RLHF, we identify a previously underexplored vulnerability: reasoning trajectories in Long-CoT models can drift from aligned paths, resulting i… ▽ More

    Submitted 11 October, 2025; originally announced October 2025.

  32. arXiv:2510.07243  [pdf, ps, other

    cs.CL cs.AI

    LeMAJ (Legal LLM-as-a-Judge): Bridging Legal Reasoning and LLM Evaluation

    Authors: Joseph Enguehard, Morgane Van Ermengem, Kate Atkinson, Sujeong Cha, Arijit Ghosh Chowdhury, Prashanth Kallur Ramaswamy, Jeremy Roghair, Hannah R Marlowe, Carina Suzana Negreanu, Kitty Boxall, Diana Mincu

    Abstract: Evaluating large language model (LLM) outputs in the legal domain presents unique challenges due to the complex and nuanced nature of legal analysis. Current evaluation approaches either depend on reference data, which is costly to produce, or use standardized assessment methods, both of which have significant limitations for legal applications. Although LLM-as-a-Judge has emerged as a promising… ▽ More

    Submitted 8 October, 2025; originally announced October 2025.

    Comments: Published in Natural Legal Language Processing - EMNLP Workshop 2025

  33. arXiv:2510.04835  [pdf, ps, other

    cs.SE

    InsightQL: Advancing Human-Assisted Fuzzing with a Unified Code Database and Parameterized Query Interface

    Authors: Wentao Gao, Renata Borovica-Gajic, Sang Kil Cha, Tian Qiu, Van-Thuan Pham

    Abstract: Fuzzing is a highly effective automated testing method for uncovering software vulnerabilities. Despite advances in fuzzing techniques, such as coverage-guided greybox fuzzing, many fuzzers struggle with coverage plateaus caused by fuzz blockers, limiting their ability to find deeper vulnerabilities. Human expertise can address these challenges, but analyzing fuzzing results to guide this support… ▽ More

    Submitted 6 October, 2025; originally announced October 2025.

  34. arXiv:2510.04120  [pdf, ps, other

    cs.CL cs.AI

    Unveiling LLMs' Metaphorical Understanding: Exploring Conceptual Irrelevance, Context Leveraging and Syntactic Influence

    Authors: Fengying Ye, Shanshan Wang, Lidia S. Chao, Derek F. Wong

    Abstract: Metaphor analysis is a complex linguistic phenomenon shaped by context and external factors. While Large Language Models (LLMs) demonstrate advanced capabilities in knowledge integration, contextual reasoning, and creative generation, their mechanisms for metaphor comprehension remain insufficiently explored. This study examines LLMs' metaphor-processing abilities from three perspectives: (1) Conc… ▽ More

    Submitted 5 October, 2025; originally announced October 2025.

  35. arXiv:2510.02025  [pdf, ps, other

    cs.CL

    Style over Story: Measuring LLM Narrative Preferences via Structured Selection

    Authors: Donghoon Jung, Jiwoo Choi, Songeun Chae, Seohyon Jung

    Abstract: We introduce a constraint-selection-based experiment design for measuring narrative preferences of Large Language Models (LLMs). This design offers an interpretable lens on LLMs' narrative behavior. We developed a library of 200 narratology-grounded constraints and prompted selections from six LLMs under three different instruction types: basic, quality-focused, and creativity-focused. Findings de… ▽ More

    Submitted 6 January, 2026; v1 submitted 2 October, 2025; originally announced October 2025.

  36. arXiv:2510.00829  [pdf, ps, other

    cs.CL

    Exposing the Cracks: Vulnerabilities of Retrieval-Augmented LLM-based Machine Translation

    Authors: Yanming Sun, Runzhe Zhan, Chi Seng Cheang, Han Wu, Xuebo Liu, Yuyao Niu, Fengying Ye, Kaixin Lan, Lidia S. Chao, Derek F. Wong

    Abstract: \textbf{RE}trieval-\textbf{A}ugmented \textbf{L}LM-based \textbf{M}achine \textbf{T}ranslation (REAL-MT) shows promise for knowledge-intensive tasks like idiomatic translation, but its reliability under noisy retrieval contexts remains poorly understood despite this being a common challenge in real-world deployment. To address this gap, we propose a noise synthesis framework and new metrics to eva… ▽ More

    Submitted 17 November, 2025; v1 submitted 1 October, 2025; originally announced October 2025.

    Comments: Accepted by AAAI 2026

  37. arXiv:2509.23667  [pdf, ps, other

    cs.LG

    Why Alignment Must Precede Distillation: A Minimal Working Explanation

    Authors: Sungmin Cha, Kyunghyun Cho

    Abstract: For efficiency, preference alignment is often performed on compact, knowledge-distilled (KD) models. We argue this common practice introduces a significant limitation by overlooking a key property of the alignment's reference model: its distributional recall. We show that the standard KD -> Align workflow diminishes the model's capacity to align rare yet desirable behaviors, even under strong pref… ▽ More

    Submitted 28 September, 2025; originally announced September 2025.

    Comments: Preprint

  38. arXiv:2509.23592  [pdf, ps, other

    cs.LG cs.AI

    Toward a Holistic Approach to Continual Model Merging

    Authors: Hoang Phan, Sungmin Cha, Tung Lam Tran, Qi Lei

    Abstract: We present a holistic framework for Continual Model Merging (CMM) that intervenes at three critical stages: pre-merging, during merging, and post-merging-to address two fundamental challenges in continual learning. In particular, conventional approaches either maintain a growing list of per-domain task vectors, leading to scalability issues or rely solely on weight-space merging when old data is i… ▽ More

    Submitted 19 February, 2026; v1 submitted 27 September, 2025; originally announced September 2025.

    Comments: Accepted to Workshop on Continual Learning in Computer Vision, ICCV 2025

  39. arXiv:2509.21606  [pdf, ps, other

    cs.LG

    Task-Agnostic Federated Continual Learning via Replay-Free Gradient Projection

    Authors: Seohyeon Cha, Huancheng Chen, Haris Vikalo

    Abstract: Federated continual learning (FCL) enables distributed client devices to learn from streaming data across diverse and evolving tasks. A major challenge to continual learning, catastrophic forgetting, is exacerbated in decentralized settings by the data heterogeneity, constrained communication and privacy concerns. We propose Federated gradient Projection-based Continual Learning with Task Identity… ▽ More

    Submitted 10 November, 2025; v1 submitted 25 September, 2025; originally announced September 2025.

  40. arXiv:2509.11511  [pdf, ps, other

    stat.ML cs.LG

    Learning Majority-to-Minority Transformations with MMD and Triplet Loss for Imbalanced Classification

    Authors: Suman Cha, Hyunjoong Kim

    Abstract: Class imbalance in supervised classification often degrades model performance by biasing predictions toward the majority class, particularly in critical applications such as medical diagnosis and fraud detection. Traditional oversampling techniques, including SMOTE and its variants, generate synthetic minority samples via local interpolation but fail to capture global data distributions in high-di… ▽ More

    Submitted 14 September, 2025; originally announced September 2025.

    Comments: .19 pages, 6 figures

  41. arXiv:2509.01620  [pdf, ps, other

    cs.CL cs.AI

    Benchmarking the Detection of LLMs-Generated Modern Chinese Poetry

    Authors: Shanshan Wang, Junchao Wu, Fengying Ye, Jingming Yao, Lidia S. Chao, Derek F. Wong

    Abstract: The rapid development of advanced large language models (LLMs) has made AI-generated text indistinguishable from human-written text. Previous work on detecting AI-generated text has made effective progress, but has not involved modern Chinese poetry. Due to the distinctive characteristics of modern Chinese poetry, it is difficult to identify whether a poem originated from humans or AI. The prolife… ▽ More

    Submitted 1 September, 2025; originally announced September 2025.

    Comments: Accepted by EMNLP 2025

  42. arXiv:2509.00768  [pdf, ps, other

    cs.AI cond-mat.mtrl-sci cs.CL

    Aligning Reasoning LLMs for Materials Discovery with Physics-aware Rejection Sampling

    Authors: Lee Hyun, Sohee Yoon, Jinwoo Park, Sue In Chae, Seongeon Park, Jooyeon Ahn, Yebin Jung, Youjung Chung, Hogeun Chang, Sujin Park, Myeonginn Kang, Jina Kim, Ho-Gyeong Kim, Myeonghun Jeong

    Abstract: AI-driven materials discovery that couples automated experimentation with algorithmic decision-making requires process aware recipe to property predictors that are accurate, calibrated, and physically admissible. We approach this as a reasoning problem with large reasoning models (LRMs). To instill reasoning capability into language models, we curate reasoning traces from a teacher model to train… ▽ More

    Submitted 2 October, 2025; v1 submitted 31 August, 2025; originally announced September 2025.

    Comments: 16 pages, 6 figures

  43. arXiv:2508.13380  [pdf, ps, other

    cs.LG

    Batching-Aware Joint Model Onloading and Offloading for Hierarchical Multi-Task Inference

    Authors: Seohyeon Cha, Kevin Chan, Gustavo de Veciana, Haris Vikalo

    Abstract: The growing demand for intelligent services on resource-constrained edge devices has spurred the development of collaborative inference systems that distribute workloads across end devices, edge servers, and the cloud. While most existing frameworks focus on single-task, single-model scenarios, many real-world applications (e.g., autonomous driving and augmented reality) require concurrent executi… ▽ More

    Submitted 18 August, 2025; originally announced August 2025.

  44. arXiv:2508.13152  [pdf, ps, other

    cs.CL cs.AI

    RepreGuard: Detecting LLM-Generated Text by Revealing Hidden Representation Patterns

    Authors: Xin Chen, Junchao Wu, Shu Yang, Runzhe Zhan, Zeyu Wu, Ziyang Luo, Di Wang, Min Yang, Lidia S. Chao, Derek F. Wong

    Abstract: Detecting content generated by large language models (LLMs) is crucial for preventing misuse and building trustworthy AI systems. Although existing detection methods perform well, their robustness in out-of-distribution (OOD) scenarios is still lacking. In this paper, we hypothesize that, compared to features used by existing detection methods, the internal representations of LLMs contain more com… ▽ More

    Submitted 18 August, 2025; originally announced August 2025.

    Comments: Accepted to TACL 2025. This version is a pre-MIT Press publication version

  45. arXiv:2507.23242  [pdf, ps, other

    cs.CV cs.CL cs.LG

    Annotation-Free Reinforcement Learning Query Rewriting via Verifiable Search Reward

    Authors: Sungguk Cha, DongWook Kim, Taeseung Hahn, Mintae Kim, Youngsub Han, Byoung-Ki Jeon

    Abstract: Optimizing queries for Retrieval-Augmented Generation (RAG) systems poses a significant challenge, particularly across diverse modal indices. We introduce RL-QR, a novel annotation-free reinforcement learning framework for query rewriting that eliminates the need for costly human-annotated data. By leveraging verifiable search rewards derived from index-aligned synthetic queries, RL-QR overcomes h… ▽ More

    Submitted 12 December, 2025; v1 submitted 31 July, 2025; originally announced July 2025.

  46. arXiv:2507.18750  [pdf, ps, other

    cs.MM cs.SD eess.AS

    CatchPhrase: EXPrompt-Guided Encoder Adaptation for Audio-to-Image Generation

    Authors: Hyunwoo Oh, SeungJu Cha, Kwanyoung Lee, Si-Woo Kim, Dong-Jin Kim

    Abstract: We propose CatchPhrase, a novel audio-to-image generation framework designed to mitigate semantic misalignment between audio inputs and generated images. While recent advances in multi-modal encoders have enabled progress in cross-modal generation, ambiguity stemming from homographs and auditory illusions continues to hinder accurate alignment. To address this issue, CatchPhrase generates enriched… ▽ More

    Submitted 24 July, 2025; originally announced July 2025.

  47. arXiv:2507.18632  [pdf, ps, other

    cs.CV cs.AI cs.LG cs.MM

    SIDA: Synthetic Image Driven Zero-shot Domain Adaptation

    Authors: Ye-Chan Kim, SeungJu Cha, Si-Woo Kim, Taewhan Kim, Dong-Jin Kim

    Abstract: Zero-shot domain adaptation is a method for adapting a model to a target domain without utilizing target domain image data. To enable adaptation without target images, existing studies utilize CLIP's embedding space and text description to simulate target-like style features. Despite the previous achievements in zero-shot domain adaptation, we observe that these text-driven methods struggle to cap… ▽ More

    Submitted 24 July, 2025; originally announced July 2025.

    Comments: Accepted to ACM MM 2025

  48. arXiv:2507.12723  [pdf, ps, other

    cs.SD cs.MM eess.AS

    Cross-Modal Watermarking for Authentic Audio Recovery and Tamper Localization in Synthesized Audiovisual Forgeries

    Authors: Minyoung Kim, Sehwan Park, Sungmin Cha, Paul Hongsuck Seo

    Abstract: Recent advances in voice cloning and lip synchronization models have enabled Synthesized Audiovisual Forgeries (SAVFs), where both audio and visuals are manipulated to mimic a target speaker. This significantly increases the risk of misinformation by making fake content seem real. To address this issue, existing methods detect or localize manipulations but cannot recover the authentic audio that c… ▽ More

    Submitted 16 July, 2025; originally announced July 2025.

    Comments: 5 pages, 2 figures, Interspeech 2025

  49. arXiv:2506.16617  [pdf, ps, other

    cs.AI cs.HC

    The Role of Explanation Styles and Perceived Accuracy on Decision Making in Predictive Process Monitoring

    Authors: Soobin Chae, Suhwan Lee, Hanna Hauptmann, Hajo A. Reijers, Xixi Lu

    Abstract: Predictive Process Monitoring (PPM) often uses deep learning models to predict the future behavior of ongoing processes, such as predicting process outcomes. While these models achieve high accuracy, their lack of interpretability undermines user trust and adoption. Explainable AI (XAI) aims to address this challenge by providing the reasoning behind the predictions. However, current evaluations o… ▽ More

    Submitted 19 June, 2025; originally announced June 2025.

    Comments: Accepted at CAiSE'25

  50. arXiv:2506.16172  [pdf, ps, other

    cs.CL

    SGIC: A Self-Guided Iterative Calibration Framework for RAG

    Authors: Guanhua Chen, Yutong Yao, Lidia S. Chao, Xuebo Liu, Derek F. Wong

    Abstract: Recent research in retrieval-augmented generation (RAG) has concentrated on retrieving useful information from candidate documents. However, numerous methodologies frequently neglect the calibration capabilities of large language models (LLMs), which capitalize on their robust in-context reasoning prowess. This work illustrates that providing LLMs with specific cues substantially improves their ca… ▽ More

    Submitted 19 June, 2025; originally announced June 2025.